\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b22190de3162d3e03481cae8904ce563912756eea7ac7e7feec0acbbdae5e6f6
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:c2f39be75173b4cc57498c5845b14c464bf9af92b49716df9661c8135ac2cf7a
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b2d58f4dc20eb7e17e09af0e8c5de55fa04f9f4f697e10c6be81a7a6a08381b2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:27efa0ab1d9c69047a924bc1dcd2b56108f5f390a6da56afb6ab7a37cb4578e2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/volume-expander-operator@sha256:b775df99f891e2cee9c2169e9c948a8d61d65b4cc90f36752b46a56e2d4be602
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/apimatic-kubernetes-operator@sha256:a5f68205291caa9bb7903180afc974fe86252ba87e671978c8b49e0206976d06
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: sonar-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        - quay.io/openshift-community-operators/sonar-operator@sha256:5e7a87f52955162f090c1d7fc69f6d0917444b553355a11c061ab60f289a17d7
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    defaultChannel: alpha
    packageName: sonar-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-documentdb-controller@sha256:8a991c76f281ec30258b6317bd623894b78a4f39d9790a1cbb32e6f1b7ff63f5
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        - quay.io/community-operator-pipeline-prod/dynatrace-operator@sha256:c475f9cc0f6caa2bb316ff42a35a40b1d4bf8ef7c8a3384d9c78eb205bda7076
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        - quay.io/community-operator-pipeline-prod/kaoto-operator@sha256:4b6cc43eba8efbe17ffec746431d6cf6b045886ff865673a08a0820abecebdcd
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:5ea2dd102074798d1ee005dc03fbbd24f4324df831f279355dc2dfc262c9d880
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:e4dec1f39f8e24043cf16703efcac2cc3b953d3f4c37b3956fce0c65ef829764
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:ed8230e90f0ab0586a743d5e760a7469345b1136971aad17ef675c1beccf1982
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:08995782514e496c24d888352e9d1e6b788dd87adb761e0c2e745be2975ffcaf
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:7b324e897fbcc881ececa529822d7b738f82cd4c3adf4cc54b66080b888f5d9c
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:e2a329fc74ed5f17e395367126c7e96197178e8f886d475b657a7910a4a14152
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:0218d68f97e89b48bbdd9c7cd750d4a666cf33af1cb8a76507c067a332a7b6bd
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:2bc56e149c14920662b549a5367f98daf9afad0df29b4746b195f6afc0c9c597
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
          createdAt: "2025-09-19T22:01:50Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
        - quay.io/community-operator-pipeline-prod/ack-eks-controller@sha256:120566a02e487aaa8823323423fccfaa377135037e2e1cb9f3ff14313cf3bfeb
        version: 1.9.2
      entries:
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        - quay.io/community-operator-pipeline-prod/pelorus-operator@sha256:f54c42a30675897a971cac278d5d8ea24edf3121570264c07cde04f11cc4e7a8
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:57af9e32c31dad769bda449ebe35e09ea135e8a7d06cddb42be4d1a09005f518
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ab3cb65b88c34fd58524a2a51cd510c611924e5860834965bad0b4418a8b6dd3
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
        - registry.redhat.io/openshift4/ose-local-storage-operator-bundle@sha256:d303fc63f5b849f073e34a7791ea92cdc3cff269cd1b593e27480ae96cac1a15
        version: 4.18.0-202509100344
      entries:
      - name: local-storage-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: local-storage-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: local-storage-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: local-storage-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: local-storage-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: local-storage-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: local-storage-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: local-storage-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: local-storage-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: local-storage-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: local-storage-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: local-storage-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: local-storage-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: local-storage-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: local-storage-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: local-storage-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
          createdAt: "2024-09-04T17:53:42Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:07cafb83ba90378c1fa60338cf951a9d7145ec245710637e90fdb508875dcc45
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:1d86075469312e84f3d16379efcd6130ec9e8b23a1343e3945f071d82beff6e3
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:dc10544e1bb24883f38373a0445944745282e5a69ed1744f428f98871a7a39c9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:31b3658dd79c18111c21e99aea52cc198f8e95b458a1a4f8d2bfc4d21e9bdacb
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:1b8008a16a0f53c09ecc5a219964194689a1aeb01acee8313aa49a9c26259429
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:a539064234088a4ee13b8a9666ca9507208fe38189ee9b8b673a38a58aa75599
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:b137e1345038c9e566767f24cce60c03ec278f5a2efc28fa08e0bd8bfa016486
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:c001dc96a1478a58f30461ab018e40bede13a78f3cdce869a70dcb8b9bba22b1
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:92a3162c705537147005a065cdddbb410e1863d80c08d8b0f27b7ff5588500cd
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:8fbfffb0bc3adfb4d72961dad381e1daba414bf4a7084f03b5208f6be4191ceb
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:035855c3d4b32a29232846d73dc398e81f941fd90a03e33241ca06f6e80b2204
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:f86e96edb0418811df328debdee93b4a51cf7f651134ee35e9c50b9dbfc29fd8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:084f1e54ef0d7f457d5a7cfc57b567ad8d85237c12f707c82f7a30a54e3652ba
        version: 1.7.17
      entries:
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:9f5479ac795eb50fa5ea597828b2d50ddaa4878faa1a07b3de0e611cb4af5fc6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/operator-bundle@sha256:bc8a47d5f8ffe3a2fe48819f08c82525677131cfb4f3c0facd82470c0f10acee
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        - registry.connect.redhat.com/loadcore/keysight-nimbusmosaic-dusim-bundle@sha256:97d5f906ef83272a8d7a2a249f38f8b9b186f8e9b64729a049a7eac0dbda8b1a
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        - registry.connect.redhat.com/loadcore/keysight-wap-appsec-bundle@sha256:e5049593a52fb26c124de7005b45919f26fa0430293f18283a20c8cde981b025
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cli-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager/cli-manager-rhel9-operator@sha256:be0fc230fef111cff36a8b846bbac9aaa98aae94ad9200416e841aa487121de4
        - registry.redhat.io/cli-manager/cli-manager-rhel9@sha256:a6b7857c1acd5173c770e6ba1d03947eb13804d2512f7f7cd6bb96e5e5cdaa17
        - registry.redhat.io/cli-manager/cli-manager-operator-bundle@sha256:74c88f1cfef4cc31b794fc5ed8dcc10d3b0e6869c228d808f1d98819e69554b6
        version: 0.1.1
      entries:
      - name: cli-manager-operator.v0.1.1
        version: 0.1.1
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:a0e4c432590e65818e7ebaab127d483d68fdc5ccae9a13820665a8aef122fde7
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:242d754d5fe1f2a07091ae93da284b8912e0f4b4c973200cf6d1241e6abb88a9
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:437a786a301ea720cdb5651dff878a9f51470bab46ba28daae4928c7074c9157
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
          createdAt: "2024-11-21T23:48:37Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVSNetwork is the Schema for the ovsnetworks API
            displayName: OVSNetwork
            kind: OVSNetwork
            name: ovsnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-operator-bundle@sha256:4c8271d0bf6d94061fe8df8110b041cb0e1a840fb6263315f552166fc5053e73
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:2e71d4125a35f6e7bd2a04cdd8cb6aae058d745eda372ae3d3fd37783778f486
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:57a57f4f3a0adb8b7234683f2968352988ec9a414824f4f444cca096825a5df5
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:45835aba95af72aa8e73b630d1d2b4c1079217531f025f4715a2eb8e8856205f
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:8b412541a797aaff1b49aa229e310957a8d7da03a3d38e6bbfd65f2ca36c2bdd
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:6eb244a301ae5395b69b1bfcd0baac37c094f4c38a1a330c0a70fb81e71f6157
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:3b87d49b200ac4fdeac97f0e97a46001e91c8f4680e2975e90bd94ec37864403
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:0f18db6a5bbc952bee8f3e4ac9e93e807d1e85cd132affe94b1f4005c1e128ce
        - registry.redhat.io/openshift4/ose-sriov-rdma-cni-rhel9@sha256:65864bc235689ff38b63d2a7b2c72874ee155241a493fa9429fbe87d5100cfcf
        version: 4.18.0-202509100344
      entries:
      - name: sriov-network-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: sriov-network-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: sriov-network-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: sriov-network-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: sriov-network-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: sriov-network-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: sriov-network-operator.v4.18.0-202505260733
        version: 4.18.0-202505260733
      - name: sriov-network-operator.v4.18.0-202505121333
        version: 4.18.0-202505121333
      - name: sriov-network-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: sriov-network-operator.v4.18.0-202504221235
        version: 4.18.0-202504221235
      - name: sriov-network-operator.v4.18.0-202504160002
        version: 4.18.0-202504160002
      - name: sriov-network-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: sriov-network-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: sriov-network-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: sriov-network-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: sriov-network-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: sriov-network-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle-marketplace@sha256:4145f7a04d7dbdda226b75af4e4cabca63a803f290b1581ea4ae0dc223eca7df
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        - registry.connect.redhat.com/questlabs/alloy-operator@sha256:fe4e83fc4a50e1aa04e6a9ba2a57acc737077b86681c1d7c597646fad337f221
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - registry.connect.redhat.com/kove/kove-operator@sha256:04bf970760c4dfe910465b0ef6a18e9c17fa70f6388131f12ae0aa0cd646f663
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vertical-pod-autoscaler.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "autoscaling.k8s.io/v1",
                "kind": "VerticalPodAutoscaler",
                "metadata": {
                  "name": "myapp-vpa"
                },
                "spec": {
                  "resourcePolicy": {
                    "containerPolicies": [
                      {
                        "containerName": "*",
                        "controlledResources": [
                          "cpu",
                          "memory"
                        ],
                        "maxAllowed": {
                          "cpu": 1,
                          "memory": "500Mi"
                        },
                        "minAllowed": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    ]
                  },
                  "targetRef": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "myapp-deployment"
                  }
                }
              },
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "minReplicas": 2,
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "safetyMarginFraction": 0.15
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
          createdAt: "2024-12-05T23:41:17Z"
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1beta2
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1beta2
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: Vertical Pod Autoscaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - verticalpodautoscaler
        - vpa
        - autoscaling
        - scaling
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-operator-bundle@sha256:f5df5d878ef0a0b494c9bafe99c35f3152d30f6c426be789fcd636dd13eff6e6
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:5da084c4dc81251a707db1bd9347c13b39a1e01660120aa2e6a47f9d1d981fe9
        version: 4.18.0-202509100344
      entries:
      - name: vertical-pod-autoscaler.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: vertical-pod-autoscaler.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: vertical-pod-autoscaler.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: vertical-pod-autoscaler.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: vertical-pod-autoscaler.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: vertical-pod-autoscaler.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: vertical-pod-autoscaler.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: vertical-pod-autoscaler.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: vertical-pod-autoscaler.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: vertical-pod-autoscaler.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: vertical-pod-autoscaler.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: vertical-pod-autoscaler.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: vertical-pod-autoscaler.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: vertical-pod-autoscaler.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: vertical-pod-autoscaler.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator-metadata@sha256:d6d0e71f5451e8255edfff87ef28ba7743fe7057efd53b25bbbd01bf9e1d0307
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/lws-operator
    name: leader-worker-set
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: leader-worker-set.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "LeaderWorkerSetOperator",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-04-14T08:12:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lws-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/lws-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LeaderWorkerSetOperator
            name: leaderworkersetoperators.operator.openshift.io
            version: v1
        description: "Leader Worker Set Operator is based on the [LWS](https://lws.sigs.k8s.io/docs/)
          open source project.\nLeader Worker Set: An API for deploying a group of
          pods as a unit of replication. \nIt aims to address common deployment patterns
          of AI/ML inference workloads, \nespecially multi-host inference workloads
          where the LLM will be sharded and run across multiple devices on multiple
          nodes.\n"
        displayName: Leader Worker Set Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - leader-worker-set
        links:
        - name: Leader Worker Set
          url: https://lws.sigs.k8s.io/docs/overview/
        - name: Operand Source Code
          url: https://github.com/openshift/kubernetes-sigs-lws
        - name: Operator Source Code
          url: https://github.com/openshift/lws-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/lws-operator
        relatedImages:
        - registry.redhat.io/leader-worker-set/lws-rhel9@sha256:affb303b1173c273231bb50fef07310b0e220d2f08bfc0aa5912d0825e3e0d4f
        - registry.redhat.io/leader-worker-set/lws-operator-bundle@sha256:f641cf3b2d6659e26eb0ae49c67a3416152a2fbc2f26f490889a2bd169b9ea06
        - registry.redhat.io/leader-worker-set/lws-rhel9-operator@sha256:c202bfa15626262ff22682b64ac57539d28dd35f5960c490f5afea75cef34309
        version: 1.0.0
      entries:
      - name: leader-worker-set.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: leader-worker-set
    provider:
      name: Red Hat
      url: https://github.com/openshift/lws-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:72a54c9457e311001b25c1b414b7001c6d7f9a34b89efb1b4198084586386f31
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:7a04dfd687dfe1646bd59a39c9e88781ffb815085bbd15f6305f6eac5f59bc49
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:d1a9246def81537e3417a6d48d1e818459583db658b54a87ca5af6911c8a3f53
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:e88676c0346c245b315f1a3bf7f85b10515bae3620c786fedda0533b39d2a400
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:66bd3eba0b68c317fe128c6713488cb6374f9647095768d3f7eed8120bf02e77
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        - quay.io/openshift-community-operators/authorino-operator@sha256:8fe34f94f87c567401cc2bbbbcc3a3e719364b80a2b17851496f86c85af5c09a
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino:v0.20.0
        - quay.io/community-operator-pipeline-prod/authorino-operator@sha256:92ad93f27c34f1406fc05a248d04709b0109adf37f2f9424f48fc6a73d9d359e
        - quay.io/kuadrant/authorino-operator:v0.16.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/logging-operator@sha256:d3ce3322f7f25c7b8bfabdd6e9fd32b833239fc39592240e0ccff3a13757fb17
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:f252d6f89a7cdcc520c3af461ca57d0dfabf736b9c38277d77477c4c880e182c
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:97d03647567ddb6f04edea39244bc0814e8f7e557251060d9b241a5fabe04dfc
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:f3497acc3216d820014cce7332e9d0683f4ca6438782ace02210192f787ca097
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:23f2f66e9f6d79d05e2be5a10c394834f093f4677ab06d8935ed33980408e8d3
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:c6d6c3e0408b851cdc9589cc0ee642d77397cdfaa40f50cf42557c5aaf9c0690
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:ff314ab983699455c32ab0ef27bb552d00825974a2d7088df2000f09afd8a2e2
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:5f10cfc1bc8571f44a99d88755adade0743106219673bbdd2e4d3a04ee714778
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:9f90cd7ab8c0ab5da6eb12f7f0dbd177df62f07c38fb759a97989899cfcbe75d
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:71ff93b137b6569f8fdefbd446cf12cd95477e4294bdea98fb5318e5d7d1cdd6
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle@sha256:2a28b64f7321df23928c8226b3eb3161cf520d38cc57d875b3cbfc828c164edc
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        - quay.io/community-operator-pipeline-prod/hazelcast-platform-operator@sha256:1209876fc9103eb2d69c34e97e4e2e0b384a9a778ae7e6a53222b8552fbf2f26
        version: 5.13.0
      entries:
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kms-controller@sha256:a2c60484a6556949318aa1d730c601b5bb05b1795b945a607209ef3434cce72f
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - k8s.gcr.io/redis:v1
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - quay.io/openshift-community-operators/ripsaw@sha256:ff8d796d5bd6c8cb0e155b3ced631122397758eb472470223e30352eeddba480
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:68cdc37990bc7da8a6bbe3a4908378c8026e5ff980bf8c44c0277f689006fc51
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:97762fcd8bfeec64fe4d3d43a6e60bc705ff2c8773354357e545adf31c8f85c5
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:4fa7a30eb3f679a495aeb96794056ffd003237686b2b60e693c8cabf6bc15bae
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6dce519730a6eb73bc18bef8372cdde65c29af3cb52eb78afc808d8582526fd3
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:9f2c01e88b29154ba7a0f70a1c69a49c86a567808cf057cb5d115159278cf2cf
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:edec37e045212f74d160724ffc7b2c8ccf4e60eb95629aae80360494d78d5433
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:465f0bef1f167b1d16ba9ae5f2d9a1133d11786b6809698a84c5c962387dbf6f
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8be098242f971aefa61a4a24d2c6630ecaa7856ece842166c38887a66bce1bda
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6afbc7ffc98918e3a557b0b3d9b2c3aa80caa47b9a73205fac9e6572c66f4df7
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:eefdd05485167d490f0011d9e5d2bb6112db1daea25940c79e1f55b3af254548
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:05e66583757da895fd23c3af9830ae1ac6a84d107107e5b6bbaece57e4d79c9a
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:02f53d3395b2b98f92a96b466c83ca15c79a75d82451e2ed8ba921375e727c7f
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8abe0994052e2a328e54c9d30c920dc331e9b816a4606234a3023b9ef65ec4ed
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:ef3f85e1227e886a131d3ddec3fb431aef6d3daed46115b288246b9ace0e69e2
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8b9b06ead5e014b920c82034ded615413dca53cf36f804005408726c9bbda66f
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:51db956f1472ae0498c57b9488448c34cf7e24afab6b9997eb3689b001d73277
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:392f9299c75a3ca4f593eaebdfe515145696f12db9f990bb004cfa58dd6bf0ef
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:df4e4dce45f87e684e7e1044a7edb2456160141afb6ba57b9af28088ffbac407
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:fecad1a330421341e6946252f95735ed177a690eab6250be097413ea29ce9473
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:9dcf24fecf3ee0d999601c276ac15e8634c275590f2c42b8a6bcb83984bf1445
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a9950264bfefc536b4e0a0adc2348f92aa41da55dd50a2324ddad0253291b67c
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:4d59e60cc75de82bd3f3a063237cef8a5598d315f6d53e803d37589a94929b0e
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:40d17d4d0632c272779da152f710fb7c2de50acc3791eb0b9fe445a13c0cb102
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:8e321dc1bf74ba331b62d21bff4d8993956ab935cf5be32f01e7f7aa6897bf22
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:16fbdaaf303ab40202d49438737f3307926594a5259081eeeb39595b7ea41f8e
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-zero-trust-workload-identity-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zero-trust-workload-identity-manager.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "ZeroTrustWorkloadIdentityManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/part-of": "zero-trust-workload-identity-manager"
                  },
                  "name": "zerotrustworkloadidentitymanager-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
          createdAt: 2025-06-16T11:15:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: zero-trust-workload-identity-manager
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/zero-trust-workload-identity-manager
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterFederatedTrustDomain
            name: clusterfederatedtrustdomains.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterSPIFFEID
            name: clusterspiffeids.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterStaticEntry
            name: clusterstaticentries.spire.spiffe.io
            version: v1alpha1
          - kind: SpiffeCSIDriver
            name: spiffecsidrivers.operator.openshift.io
            version: v1alpha1
          - kind: SpireAgent
            name: spireagents.operator.openshift.io
            version: v1alpha1
          - kind: SpireOIDCDiscoveryProvider
            name: spireoidcdiscoveryproviders.operator.openshift.io
            version: v1alpha1
          - kind: SpireServer
            name: spireservers.operator.openshift.io
            version: v1alpha1
          - kind: ZeroTrustWorkloadIdentityManager
            name: zerotrustworkloadidentitymanagers.operator.openshift.io
            version: v1alpha1
        description: The Zero Trust Workload Identity Manager is an Openshift Day-2
          Operator designed to automate the setup and management of SPIFFE/SPIRE components
          (like spire-server, spire-agent, spiffe-csi-driver, and oidc-discovery-provider)
          within OpenShift clusters. It enables zero-trust security by dynamically
          issuing and rotating workload identities, enforcing strict identity-based
          authentication across workloads. This manager abstracts complex SPIRE configurations,
          streamlining workload onboarding with secure identities and improving the
          cluster's overall security posture.
        displayName: Zero Trust Workload Identity Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zero-trust-workload-identity-manager
        - ztwim
        - spire
        - spiffe-spire
        - security
        links:
        - name: Documentation
          url: https://github.com/openshift/zero-trust-workload-identity-manager/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-agent-rhel9@sha256:0281d6ab233a685ba0f739629584aacd9254e0b13a697fe8708efa41289e746e
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-operator-bundle@sha256:1974903a3e2b692eedab2602395fdc4a6e2ffbc79ed4e633705c05ce7771e1b7
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-helper-rhel9@sha256:a39c80c6e1495ea166a858c8ed03304c6455420c077d937c99bf8732a6829790
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-controller-manager-rhel9@sha256:67f0e6670195c84515e7d78139a2951e463db11445740fe1d9a449a7f338556c
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-oidc-discovery-provider-rhel9@sha256:5ba7e94b7ea06852d32002b055df556ff2a2fdc9a2e8f793b837c4856d773782
        - registry.access.redhat.com/ubi9@sha256:861e833044a903f689ecfa404424494a7e387ab39cf7949c54843285d13a9774
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:7d170ba9b3de2e88dc0d0a6b0c13630ca3b23eed96ea5e3c725516046d7c7fc7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-csi-driver-rhel9@sha256:a16ac0a4363f7cffada39a2f17aa42f5de3faf78b3688ae5ebf2ecf3a4759cb7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-server-rhel9@sha256:70f97c613feeaadfed99d37aabd62e1cb2247825292ad9be806bc2fd065c9166
        version: 0.1.0
      entries:
      - name: zero-trust-workload-identity-manager.v0.1.0
        version: 0.1.0
      name: tech-preview-v0.1
    defaultChannel: tech-preview-v0.1
    packageName: openshift-zero-trust-workload-identity-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-operator-bundle-rhel8@sha256:0fbbb0b32adadc5746251103d25bba6bd365b2ff60e289e4694061fc92c3ce6e
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        - registry.connect.redhat.com/paloalto/pcc-operator@sha256:c7200a8c652ee544b0e0bb4d9427eab3a3b2dee6ca61ee88abb8052316382673
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/sap-hana-express-operator@sha256:0fa4f29bd7669690be87a08405bed95d3ead2293a125b28149de2b10b02d2cd6
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/community-operator-pipeline-prod/falcon-operator@sha256:3916adf5083c83838564b1ed039decbadb2efa4ec288699ef0137a66983a3cc4
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.connect.redhat.com/hitachi/hnas-csi-operator-bundle@sha256:79e6ba3db78aa55a9f3839a172843d284d91e9cdaae9cb2c8e7e705f3d03115d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:60b293c23587072b3376f83afee95b0207228c5c7b80483b025adda572c3551d
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        - quay.io/community-operator-pipeline-prod/ack-rds-controller@sha256:80b0d6e33fc94471d05e05f6d370be3a84fcb5c2044486e007d57360eb69d9d7
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-messaging-topology-operator@sha256:4c62fb6d686a20a28437905d4e1a9b93ac05853b4dbf29fcce34421a43268724
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:b10c121395f5bc46d61fdeddd35a0c8d9ae4e67a9a77d8ada90e0ea53666323e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle-marketplace@sha256:9d84d46dd1eb74afa18466c73766402b46c2885ef483600aff83065f8065d8d1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        - registry.redhat.io/jboss-eap/eap-operator-bundle@sha256:0db70367917668964e2f4f5efdcef864ccc3f065848023e52caf4abdec728d9d
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:096946dd896b6d9d937eb6529c3a23a84ccb129649c5a0389d0ab83989c28b53
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/jhipster-online-operator@sha256:5f2436297c026832e5c15241d60090da581daa5502e352ea33d77f22ef3e7a52
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        - quay.io/community-operator-pipeline-prod/argocd-operator@sha256:483854621b95dcefa03d18a970600b271726422ae3362479f73dcfca9085c92d
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/mtr/mtr-operator-bundle@sha256:190a6fa2a6a9b6b182cdf9cab18814d05c203b54ca18dc533e553fa5dcca779e
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:2808a0397495982b4ea0001ede078803a043d5c9b0285662b08044fe4c11f243
        version: 1.0.2
      deprecation:
        message: |
          The preview-1.0 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:a91cee14f47824ce49759628d06bf4e48276e67dae00b50123d3233d78531720
        version: 1.1.0
      deprecation:
        message: |
          The preview-1.1 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:4477810692429ac498b6ec1b295e6ead0f8fea25e91f6ddf17255d7e349b82e1
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:ff2bb666c2696fed365df55de78141a02e372044647b8031e6d06e7583478af4
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        version: 1.2.2
      deprecation:
        message: |
          The stable-1.2 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:8da62ba1c19c905bc1b87a6233ead475b047a766dc2acb7569149ac5cfe7f0f1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        version: 1.3.3
      deprecation:
        message: |
          The stable-1.3 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-load-core-bundle@sha256:f62a7f8eb74d0d5309fede6c9c0201d1f0c3736f93f1aeec7446089d64fce28f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/openshift-community-operators/sigstore-helm-operator@sha256:ff4e548a8a003bd23e5076362bafb2e22f77b8445c5e3585d99cd5ce26fcaf56
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:5a5e2f775760aa68c85d0f69b700d23ef2a6be7bb415bd96401d03eb6077f90c
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:79a9191f0cd5d11af004ead87ba9eb5359c67fd658c346db4c3cea3b16b23987
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:54ea11add0a1d988e0c965968f9a65e011978190f76fa2f826f1378474bac1d4
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
          createdAt: "2024-10-11T14:05:47Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8337924489a1747d4bbca6dd6ac42210731cf721b549b5830846b271621e2ac1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:6d0aeffbccaf9df83b42c530bf82707aa7c1cc16f72a5ef30b66fbd562c01ab3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8c05a303be5c536f4051ef291475b578d4d38ca0946fb4b4dd8bca376ef432aa
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:e91c2e5473f3e24c185dd25e9843f333e6d7419d9c5f4d4396669717b5969683
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:c543dbbb66fcac02345e1d3c12333780721bf049dca0f0a27d4c55b15a5d3e14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:a65780c5680a6fbbd0e3d0eed423b14a9c9b3d7d17f801c1eb6eb6b777384ca5
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fd6fb83bbeb9010f7e91f0afc91beac0bf38cb79c2de8f533a2bd07b1b238262
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:936fd68d7659683de7789fcf34d2e5873bd4bd1534c01c2d80501647f7c3c442
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1dd8f0f08ff1323d910fea8adfa13cef64c2fd2f01968e4d9d55f71cf42835d3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:365115407964d182de7e7564438c6fd089e2c0bcc0f94ec48621ca0cc2bf0fe4
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4bea07e101740afc36a1617363a79b1c474efcacd584f6e475a53b3f2395a079
        version: 2.7.2
      entries:
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:3d0409c1a0b17c51d3dbd5aa0371edd597e3c5ba194aad987fe867dd804d1923
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    - currentCSV: mtv-operator.v2.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
          createdAt: "2025-09-09T18:06:51Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:41cc528bf93000095b4eb56aaff9a6dfb124d1711a2f31f0745163ff934ac96e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:0c8ebfdd52053dbcc221a3df7d603163a821dd9e0ac6f54977e081adc6cb4504
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a299acb18821691cddba52904140cef8bf12632fbf2e752b33118a574a5ff394
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:ad3b548405e8c4c075ef7d4c0eb0d0430684066ed7eab4563a1f8378eaf3d939
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4b37b1aecbbf087257be5ba2bb559f2de005aacb4cd371192cdde1a9891739cf
        - registry.redhat.io/migration-toolkit-virtualization/mtv-vsphere-xcopy-volume-populator-rhel9@sha256:c0a4a232ff8adfb7d57e0761251ad0d79a0f02d27c4bd813f62f6b943deb6585
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f90f4eaaaa02b289704b19dd171282f3d11f370ed6cea992e55964729ad5116e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:258efbf073f78f8b5b48fdda22ef055e2315046cfecc9644da51959289c836d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d57b39674260f3bb1ae4c164022c2c7be3cc9645f1c46a83e43f40965633451a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:aeaa9ffcb0ffbb0f69b2712a3031f7e2dfc0824273b5c9c0f6349bc701a0e89e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:e438c5fccc957ad26ecfc3633561342d3b6f2a6d5a4c7839109564008de62222
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:83a523392a9c715972bba7044c7c3d3944594cc89778449e1f59fbd6a2659baf
        version: 2.9.4
      entries:
      - name: mtv-operator.v2.9.4
        version: 2.9.4
      - name: mtv-operator.v2.9.3
        version: 2.9.3
      - name: mtv-operator.v2.9.2
        version: 2.9.2
      - name: mtv-operator.v2.9.1
        version: 2.9.1
      - name: mtv-operator.v2.9.0
        version: 2.9.0
      name: release-v2.9
    defaultChannel: release-v2.9
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:0ee0bd2449f183277af845e312257b4a1ced4de865831b15e1758a397d8d29eb
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:2cf058c819b40512094f84f4d374457455370ec4994d34cee6e39b9d87841f05
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        - quay.io/openshift-community-operators/camel-karavan-operator@sha256:6dc78ecdf6e1936872626c02d41986d6553c1500002d85692c97e86fa77c129d
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:f182078b0077b6df6d47ddcab0204e963c1475486bd61b8eccb73b46eb8b2995
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c8b3505ed491c1bf6a913f161715c8ac483d8d9adf1fd29916ea4edad987cd67
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:af5a13351f960c0938c68604b772f24743c0e997701838a7784c53406471e72a
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:0d9f5293d69eebd91fffd83ae371141e640cab8f9e2d559a83714ce0f94fee08
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:d6675617c19b4ac7c1913078eaf9ffbe3d03761b570dffa002d9de11f337f19f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:2a35a0ef852cb4d76e963df50449712a06ff13391c30675aaf931e21cb561b8a
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:00f7ce0148988e93679d4b53b45ae8d2ad54f308b908940e7b28504318f718c6
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:c352f5633beef2eaaa324dda66b9b484e5ea79efbbc3b781079f2e9d10c880a9
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-bundle@sha256:47a7bafff2715fbef6130ca0b84560b6c2358275821b72ae54ad46020e02bee5
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hive-operator:1.2.4953-c248a51
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:be43412a2e4bad2cd9d15e4fbabe0292585c5bbbd65da12d519f30f17d2fe594
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:822fd8e65c241cebedd835bfebf67645199561cfdd6148d5afffb8b35e499ae3
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        - quay.io/openshift-community-operators/hive-operator@sha256:9a3cb682afc66ad662045b3d8ff214203cc00be60f66f74564820f6dcf99cd69
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        - registry.connect.redhat.com/fortanix/ccm-node-agent-dcap-operator-bundle@sha256:545724c273dfd6a1fa923d468b2d054f39ec3186d01586a8b3fe494c0f22f8dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/ndmspc-operator@sha256:b1f47ce0cd6930d902dc3538a1a87140dbe9257dd50773a518fc2350185f90cc
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        - quay.io/community-operator-pipeline-prod/klusterlet@sha256:9efef9a034c9f12505a0d81bc9697c4a3ab7e3d05d1e7aaa4180490d70ddf6a2
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:30264b9f49912cc0e6bed59001869e3168caace20b292c9dad4cc2d4947d289d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:d7cc6d2575eb252fb3e2468106e801f192da7844b0b153dc3c6a12cf7cc0a8fe
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:60e99bb91309e448c652898dac070a6d920fda164f27879d94632c88a9e61ffa
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:11edce56874ff1ee666752ec5fa521f3f6ae78c52cf1f289ceb97a86e6c083aa
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:5e4c319ce871a0dcbab7fff23325e6233c9dbcb2592fdeb1a9ba7c57ac2a63b2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:d86bb21fdeec35b1fc9789023a558750da76e036e0790dcfce97dcd137dcb6fb
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:7d94521a5cbdd34f8307653a3e1b596ec86415b2165fabdbabc8ed8bed35473d
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:3d5ad015a953708cc53ac2ea92b1b1d99c870a801b26e790d41458adc0a958e5
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:fea5b977249013181616f60b95c5e6d224cd0f248717171d946ff0e4d6eed6c0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-memorydb-controller@sha256:7f2e65d817479fcc5658cac745cef325b2bfd6256eeed8b819f1713afd59fe68
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patch-operator@sha256:437f5b7c661d3e9bc2c8ac13b0c8f3565f685b9e5ef52791312bc64d55e2e476
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:21f4648e0e71b33ef24b4921ab34b9e2b5e90eac25ade735d51331ac4f7c4cf3
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:2ed6bfdf37e36c63cde3633443553dee410201200ad351e6d91a5651f8e20ecc
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - quay.io/helmoperators/cockroachdb:v5.0.4
        - quay.io/openshift-community-operators/cockroachdb@sha256:f42337e7b85a46d83c94694638e2312e10ca16a03542399a65ba783c94a32b63
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        - quay.io/openshift-community-operators/cockroachdb@sha256:d3016b1507515fc7712f9c47fd9082baf9ccb070aaab58ed0ef6e5abdedde8ba
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/netobserv-operator@sha256:73af4040707ce79baa3115deed1e9148436b248ef3570f1175a6d31fdf9596c4
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:394d122efcf0da16bb68bfef1292395cda1bdb21909ae55f8fcc54a16814f1cf
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:62f93c5989d1d8217a3bac076addac05c0906a1a1f1eeaacb87bcf256812fdf9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:ea8c060444aa8e89dfbd90d61d6ec3900ab4d6634f6aa8067face1e6974b9065
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        - registry.connect.redhat.com/loadcore/keysight-wap-notifications-bundle@sha256:9c2ebb98476bf3374b71a838c394017e7d31aaa045bd52b3a66924b6f6b7b067
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.connect.redhat.com/cpopen/ibm-licensing-operator@sha256:a9f68d204b6acb50ac5023e5f2314ad0d06e06483f87290381d1e92476a79579
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator-certified.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-operator-bundle@sha256:4512f786fed0d9b2285b5a2827ec3b597452f90c217e27398f2f0f552377d844
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:48b37460dd7245cf661009267c3353962f0f195efd6a3007c9cc70b51b91a9e3
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:9949a581df3960a770117b6cf6b9fee1c63ea12ca09ac2d677f1cbe0a83fbd93
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - quay.io/akaris/sosreport-operator:0.0.2
        - quay.io/openshift-community-operators/sosreport-operator@sha256:f0d28a3e2ac2e1cbd0c307797b728ad89ac33375135d72054a7b765e274e4bea
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/node-maintenance-operator@sha256:80dfaab6f7a5c34cebb45c874ce134bf60e3697e14a95b69f1972dd98ecc172a
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=4.11.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:75eaf7b8b71e1a15e002f6cc5430e5e50039699d7b52c3150f5f71b1595f00c5
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-operator-bundle@sha256:41ece1ea773fd610cf8a57483a4d2846e0bb3ce81db701021a3e0dd91e13b7c8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:f7159e880ec55da49ac774db0c454800055e7cd3f37ba94fbcb629d08aac1146
        version: 4.18.0-202509100344
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-operator-bundle@sha256:4f8908ba97517826fda14cab3624ed42a6348fb96f0be4abf9cd2a7b68f7d3e9
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-wap-migration-bundle@sha256:3deff8b48ab1a6f0e5fd0f8315bc1e191da55525cb3cf0f315f2f7afdb47a6c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - registry.connect.redhat.com/noiro/aci-operator-bundle@sha256:80e25268ab5aa6ea418deae5ccc197fbffe2b9cdf4f27894f5beeb9315f6b123
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:20a4f2c56cb186fff4e8d31021d84b690512cc65c3c9629ffe623a0213ca9b0a
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        - quay.io/community-operator-pipeline-prod/mariadb-operator@sha256:3cbbff7ae75fb4f4629c7b8b58ca85ce67b23d5da558241301266d010e8b1188
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.connect.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:99e845965f7eca4b5a3361c494a385292d655e7c1906185197826c773cf9aaf6
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        - registry.connect.redhat.com/qc-qranium/aic-operator-bundle@sha256:581b21f44d6b993888681ed702ed10b9d1bb8e52dcca53a91e856febed2613a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patterns-operator@sha256:99ac398b06fa9d9a7626b99cb47fb380d0cf4a4307fa11da8e486df35966debb
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        - quay.io/openshift-community-operators/varnish-operator@sha256:46de72c5abad80cef3d6426be967131c51c2c15029a6a38744975578c805a44e
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:3f7fc52c79f3e4ef7ddd6f5efc8eec1ee4836bbf2076e7dd5f68bb3b7c6e9d8f
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:34d4f950bce7863f15d5f1e7d6555a3ccf97ec31dc3ee140f68442890c4f0f96
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:96b77663961daea3a336b97e34e9412281bfb2727854dc970a69edd029b9af85
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:5170255ecd45b9bb7a70abec0f6eeb0518bbc2871de759e468dd61a80261c0ad
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:9a3f5b6c403fb84c9cb93c9d7cf902c80990196983607c8ca568d4b558938e17
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/kasten-rhmp-bundle@sha256:c6b58e5f6c9b2a1f5c4bb770d1fa1db6eaaee8ae7d9751a48338623d0d812438
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        - quay.io/openshift-community-operators/loki-helm-operator@sha256:fa3662ec16e6cb3f866ad46982a50d8fdefb43c68f3caa13291239013e03af13
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/openshift-community-operators/infinispan@sha256:3ba9d20e83825f3c041a7dfd195b0800f74fa0f5cd1e6218fbd3547fedd85d64
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:f165693aa550903d7a26b429ed4d093f60194a7a04148e65e61c52d67b527fad
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:5e1a1fdb5a78a34155abe77754c774c65335ea8cbfdec2a222ee22a062c2db46
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:e43b0581415af648c0b82830377e354a951814d3eb7825bbe31b9f3e633917c5
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/devspaces-operator-bundle@sha256:a48dd94d5616183843ba939146b7e4d985210cdd902038ff80ff6df3df3d74ca
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:18452ba29f4e0f52b765e8f9359e4143ebfe1203160e6f64ce46ec9d4f095a62
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:9716402247ed874f19574a0d5125d1035fd198f36081d9129c1d4d4ad2f714a6
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:397feef6fc1dffb7e2c3798bbe3bfebb86d6173c225eb61473029cc4500b0cbd
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a70fceaad3b6a23e6f247b4dff41fa64349c4bbf841b66adbd0f8309c6acce74
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a7a4d20b5acc59dbb8c99a37c19b2749f15ed0be7f841490a829796865c5a956
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:01ca6a65a041a47efd611673a720e69464f02424d21feadc496c356065ff2f65
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        - registry.connect.redhat.com/appian/appian-operator-bundle@sha256:2cfb2d2e43fa207c8c06df6de8d3e90123c4cd52cdb6a4c6d618ec576b79c2eb
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:4ec16d88828e580fd9223125043ca49569324c4f06b7a65bbaa7d76923de2bad
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:7f959a4076c36d76f8f1489a92d6984d1cfd2ae5c0533a78e8b16010eb6d4d3c
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:a3132caf9cddf6d6186b3e982a6f6232a208fc727b433ee6d200a3b72f50ca88
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:d89f4d647534569f401790d6e10bc42433ae791be86558b100a4027a74b92cd3
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
          createdAt: "2025-09-30T14:21:42Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/flux-operator@sha256:557ae37fa6a6942e3a9d06966df67721011ac7905e2241867c975948092f2936
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
        version: 0.30.0
      entries:
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-operator-bundle@sha256:84d65eed99300b0ba87369cbf9a9fc17a37781f8b6f9de2819d0a159f9071a6c
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:e140415e0361cabeab192ba6d22634283cd47749dd25e9cee4a775d3c409cd97
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:4b7796235a0a545db83570add7a675fdd9d82b17f931eb81d0cafbedb1fbf22c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:9f9dadba5a35c8ab98ddacee22593a247a31e96a4b6e1b4f1a946c2e34006bb9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:ac244e06ecd2f8a5e4de7b121efb8e82e3e7d23194878cb8327ff0b2affb5f04
        - registry.redhat.io/openshift4/ose-smb-csi-driver-operator-bundle@sha256:91d04cbe0eee6c4fe0230d0d179366942e6c54afb65bb77a43a681e1a2411caf
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        version: 4.18.0-202509100344
      entries:
      - name: smb-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: smb-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: smb-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: smb-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: smb-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: smb-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: smb-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: smb-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: smb-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: smb-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: smb-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: smb-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: smb-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: smb-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: smb-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        - registry.connect.redhat.com/enterprisedb/pg4k-pgd@sha256:fb206b95690e9604c41715369c69125a906b866dc35cfc5c63638ce7a4337fa2
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:26af4e950a33755114298386a70229672a09ce45b6203025856f460a55c4c8a0
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        - registry.connect.redhat.com/amd/amd-t2-january-bundle@sha256:7c87354aa07e783a2cbf4ab29877a8ba56ea4269bc1f41bfbe08c558b3666841
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.connect.redhat.com/loc-devops/operator-bundle@sha256:b9f7a678ceec2d367d55499a48a15a9dd729bc34b0e707989be23c97fc3c5f4e
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "aes_key_rotation_interval": "2160h",
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_location_id": "devspatcher",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "enable_server_session": "true",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "session_duration_seconds": "86400",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          },
                          {
                            "email": "admin@mycompany.com",
                            "hash": "$2y$10/yYwtXRrOagYO2jNtwVzGQPPEfDvX0QllV8NXbku",
                            "userID": "b7e2c2e2-2e7a-4c7b-9e2a-7e6b8c1f2a3d",
                            "username": "Admin MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
          createdAt: "2025-09-24T08:25:29Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/edb-hcp-operator@sha256:87833c7edfa714b433fb066533bd415049b4ae96588ea8457a85a62326801d5c
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
        version: 1.3.0
      entries:
      - name: edb-hcp-operator.v1.3.0
        version: 1.3.0
      - name: edb-hcp-operator.v1.2.3
        version: 1.2.3
      - name: edb-hcp-operator.v1.2.2
        version: 1.2.2
      - name: edb-hcp-operator.v1.2.1
        version: 1.2.1
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator-bundle@sha256:a26586b5f2d748f56dc29fbe3871d84de422e4103ce13c93dbe477718c582487
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        - quay.io/openshift-community-operators/yaks@sha256:0031de9a68502cb65923c4e17453195663f59e46b1cda37619a857add503c016
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:91632b89b4001e2ffa0f1aef56207eb48f0275e20c635e7f1ff7c6de57a59ec2
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:eed8048d085eef7eefaa643a09627521fc2596ebd26280e05c79dddaa0b01afe
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        version: 4.17.13-rhodf
      entries:
      - name: mcg-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: mcg-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: mcg-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: mcg-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: mcg-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: mcg-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: mcg-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: mcg-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: mcg-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: mcg-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: mcg-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: mcg-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: mcg-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: mcg-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: mcg-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{},"bucketNotifications":{"enabled":false}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:ed5c7c8cbbe455b6d31b1c55871585a0679a91800f53743d81d4cc36c6aac7cc
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:88387ba7a9e85c1bcf57966558db57c099fd301ea80a68e43afc2bca3dfe94a7
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.18.11-rhodf
      entries:
      - name: mcg-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: mcg-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: mcg-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: mcg-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: mcg-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: mcg-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: mcg-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: mcg-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: mcg-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: mcg-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: mcg-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: mcg-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-operator-bundle@sha256:b637a02d23bd8364cc8914421049eb60169c163ac70bff2f33591df1a1193002
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e9612ec99daa171c403b5ddf37788526e5375b83987e5de9b1ebe519198607b4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:34d177abb56b948eaf9ce6fa62e37427ddee9fcecca48e752d09f72c69847a46
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:932e83fe9c9b54934767a7c5af683b2ab08bd2b4f0524c191f1c766dc64d9c3d
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e1c19bbcd4c010ef2bd4f760eb24b2c96cdb82b4e025361a433e6a241eb23cf5
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:a6e02e1876fd5382ba4921dd65650b2d747666feb96c839f34e99d2972dfc323
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cb711fcb32b5ca0f34c4e3c67624bc4e035de4247617d988a9e841fd58c1c911
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        - quay.io/openshift-community-operators/limitador-operator@sha256:7aa0a4fe55e5e099e2e6e4c2cd44b9bfeba69d8cfade10500ccba3092fb60dd2
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        - quay.io/community-operator-pipeline-prod/limitador-operator@sha256:450f9bf43b57b83e9975a1865981af22e28c725193573354d1180f406887664a
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/flux@sha256:cd94037ff3055ab6ce40b0670f900248a5fd233cc67a2a4f92c1559cf9d2f3ad
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/deployment-validation-operator@sha256:2803298c7cf3eeaffe9da3153142058d94d65ff1dae90efd8819a4cd23e0518f
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        - quay.io/community-operator-pipeline-prod/eginnovations-operator@sha256:19b12c75f7a2dfab027881ed9f0b09d25e3baa936d21f7be3ec1e0177ed9fdf2
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/keycloak-permissions-operator@sha256:4de19e52e84cdba0963bdd852a63912f097280800d65add8d474c0953da478a0
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-framework-bundle@sha256:53d056ac37177263d79098cda775deff86a327c65106ea5a40109b08a856cf84
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        - registry.connect.redhat.com/loadcore/keysight-wap-msg-broker-bundle@sha256:a7dfca98b4e09c20317cbb486e7893de8d88faf5db2ee26549ba81332ec526cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle@sha256:a90a748aa70ee5f1b8fd375a7877db1292792ba46f7b49f629533c82a25c43a0
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.connect.redhat.com/cpopen/ibm-zen-operator@sha256:5a9120d5977e934ca3801eab44ed48ecd4a9fee10c0b4980594d33cb02e64855
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        - registry.connect.redhat.com/airlock/microgateway@sha256:80d49a299a97bed33c28941f93312e86b949ce945e843c93da6bca49865c81d0
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        - registry.connect.redhat.com/airlock/microgateway@sha256:087d4d726f8080111d2b18cd0f9ab290f4560333fef308e553991484b6210c61
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/openshift-community-operators/nlp-server@sha256:6f80f44822301252540c6b808a20cea327f0eedeadc7643d31fc743de5c5e696
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:fdc0c228ca38db30f89812e41c09ec96d9276ed844b0801013375bbffd6c1a16
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:ea606dcfa4cbe4df7b27c10d3fc45f464845518c4896845079ca9194600ceb6f
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - registry.connect.redhat.com/loadcore/keysight-kcos-ingress-bundle@sha256:f04b8f15b069c06eac5d9a97f81a2f069704237eea998d35727b89854c793746
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - quay.io/community-operator-pipeline-prod/dell-csm-operator@sha256:6e2e813b7ca426a793da717d42f3620edbc3eaa84be710600f2a5262924378b4
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:0739d5035590cece6d2a573aa1e799b0b0cab8b548a41cc2488cf465486cf4a5
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:3028579f4909be6b8dcd970e778bc88d0576c76e55b98c886738b8f601cdf4c1
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:e50343879b4c3c4002c5e7ff67f4664df441cf5f239f59fa98b7fddd75516d41
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-rhel9-operator-bundle@sha256:896a615487c0ffd5335b869f0dc0c0ec1c9200dbcafa52f2493003fecbc030bc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        - quay.io/openshift-community-operators/alvearie-imaging-ingestion@sha256:f65cc3cb83b1c75385970ce70c6a84b15b143e682ad0d1784148f1643df0991e
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:add5d14653f57ac1e3254c2d1c70fc9232959f76ee6286fe5cb572696a5a4fd3
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:7fc4d97b5ee63085672fc9a562ebdad3d4c00d1b555e06679d1dd9ac945f1bb9
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:05e12cc4caaebb7f185f651f8ca876c442c02d79794e561b743beba0d189fa04
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:84796005b6e71efa5d783c55b1f72bcc3762944bf32740cb75498bf8b908bffb
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:424e138a17e2af91603e834ff65d529400d6793771ddaab92d3988f4e6d2f2e3
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:9d505bf561650eb9785811ba50c920480d85679eac6a611a1c5794544e175934
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3abd8c3f97c9bba127b6e8c7efcc078c90c9d7cc35d74cfcd9bc0d303f62d2ed
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:21c17448932f053f1d4845111288a317786b190154bae9da7947c315ba12f796
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:023b9cc1e4ec2542b8d68b636b21a2e8c4eb143424581c62d16598fddb47da6a
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3843fd48c57be102576db93e8483a30f8efe8cd46b5e09c704ac0bc5f18553a3
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:55bb4133abc945ce7b75e6c01c6beaf63c19c129fc8fbfb48f1b53b8e28aae97
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:43b959fcdcf107b70e4224d80b63d1c123560c3af055109bbcc7d68476546584
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:2845843f32b895dee0242e31380b44625e3f8381caee7ecb3258648311c302f7
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:c7cd3ffa88eae188e94abcc0af56d328aff43463b98a3bc3c0b574f29373944b
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: openshift
      provider-url: ""
    name: pf-status-relay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pf-status-relay-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pfstatusrelay.openshift.io/v1alpha1",
                "kind": "PFLACPMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "pf-status-relay-operator"
                  },
                  "name": "pflacpmonitor-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.19.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PFLACPMonitor is the Schema for the pflacpmonitors API
            displayName: PFLACPMonitor
            kind: PFLACPMonitor
            name: pflacpmonitors.pfstatusrelay.openshift.io
            version: v1alpha1
        description: It deploys the pf-status-relay application.
        displayName: pf-status-relay-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pf-status-relay
        - sriov
        - lacp
        links:
        - name: Pf Status Relay Operator
          url: https://pf-status-relay-operator.domain
        maintainers:
        - email: marguerr@redhat.com
          name: Marcelo Guerrero
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: openshift
        relatedImages:
        - registry.redhat.io/openshift4/pf-status-relay-rhel9@sha256:8cdc0714d02144f5b32e26c97f7678d0377e0c3826c475a784e76a42f781d3e3
        - registry.redhat.io/openshift4/pf-status-relay-operator-bundle@sha256:fb9635a614b6676fbe6a532d07b3206f12ee525b7947378785e79951d43940b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/pf-status-relay-rhel9-operator@sha256:e6719af6fdc99c652238f3c31a014782811eb46da91e65f93c083a7a001e2496
        version: 4.18.0-202509100344
      entries:
      - name: pf-status-relay-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: pf-status-relay-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: pf-status-relay-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: pf-status-relay-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: pf-status-relay-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: pf-status-relay-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: pf-status-relay-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: pf-status-relay-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: pf-status-relay-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: pf-status-relay-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: pf-status-relay-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: pf-status-relay-operator
    provider:
      name: openshift
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/rhpam-7/rhpam-operator-bundle@sha256:46e7d899d64ea8cbfc35221d05535db1023c41c1d3a7c580cb03e7e7d769144b
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:4ae986761a69d25697b9b6a3a99de88d8587beba136c3de29b90624e2ea6738f
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:adbfebc73bc5a9cd12d17e9e1fbdf87c2176aae7a7217bcefcec3068e51bc904
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    defaultChannel: stable-v0.2
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-operator-bundle@sha256:f596dc9323da89f5e45e71192f0fa698b7bb29b8205eaa0fc1659d68fa7401f4
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:8e223a537f487bc3cef10769adc6fb619085dafe3ea797cfd52c20e3066f694f
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle-rhmp@sha256:81476f897726c85293e64e6188ed314e3e1e637c568c6b0ff8f475a876bee29c
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-rhmp@sha256:1502ccda3c067d74689d0341008f05362bc23904c8e35b20e020fe0dca82aef7
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.connect.redhat.com/nec/sanstoragecsi-operator-bundle@sha256:1957586c13e3bf80bdc688f519be440ec1dcc4ffe63343f526018dcb93e21cea
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/gitlab/gitlab-runner-operator-bundle@sha256:68b22c412a44d5fa09b45e204d7c8e7b6ca6afa4ed6879de608dbcb0b1929676
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/openshift-community-operators/pcc-operator@sha256:e99a466ddb52ad44c435ccf952754d685744fb5b7b7e4cf77dd542e530e4f968
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53-controller@sha256:227a306cc90998489bfb681cb29b416238b404bc2db28a09b8949d4cca2d87a4
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-operator-bundle@sha256:e83707d0b396c38cf55046d3159f1f427188404aa11dc47d7b1c1f25605182de
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.3-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:64a623d5d536c790dee04c1604ec2f4429f8b8f1973b4232b543e8cbe23d6590
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:7c9b4327d8677f48c48888230fc38eec8c20c7f646dde568dbd96b5e4c8f7967
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:8797a3be94ce6f4e4da15e4efefe41ea5b6ffb4cdc6d698638c80976bb22491c
        version: 4.17.6
      entries:
      - name: lvms-operator.v4.17.6
        version: 4.17.6
      - name: lvms-operator.v4.17.5
        version: 4.17.5
      - name: lvms-operator.v4.17.4
        version: 4.17.4
      - name: lvms-operator.v4.17.3
        version: 4.17.3
      - name: lvms-operator.v4.17.2
        version: 4.17.2
      - name: lvms-operator.v4.17.1
        version: 4.17.1
      - name: lvms-operator.v4.17.0
        version: 4.17.0
      name: stable-4.17
    - currentCSV: lvms-operator.v4.18.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:5d0534e08790e91d5d355eac2e5c187a217c1ec6eb59464ea9a771490ac040e9
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:6e809f8393e9c3004a9b8d80eb6ea708c0ab1e124083c481b48c01a359684588
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:b3b1c4836a82a7ec6c9b02e4a86c352f4ae8c45a3fe027ae7200869afe3ab028
        version: 4.18.3
      entries:
      - name: lvms-operator.v4.18.3
        version: 4.18.3
      - name: lvms-operator.v4.18.2
        version: 4.18.2
      - name: lvms-operator.v4.18.1
        version: 4.18.1
      - name: lvms-operator.v4.18.0
        version: 4.18.0
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:ca70ebce43d451d040aedcec036413f986e620234fd9a56d45983805396db0b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:4aec6aa2d2b9a24167bb6bc92db7c2c323847b7a2f3ad7f998c96bf1d2903366
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        - registry.connect.redhat.com/netapp/netapp-supported-apache-spark-on-openshift-operator@sha256:4f15a0fc375156fafdf2d504f9980a8eb4451808b5554357877345cacf8f3594
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-key-manager-operator@sha256:7a998ea17ef537ec5e998e1627b065ae04ab01b7988196da1cdd1763e05a9924
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:6ef2956b297795a9d0c83119e71309549975df3ae5652e6a98cef035ace5d51a
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:8831faee6979acfb5bd031b64cbc09fb90e9ea4f48bbff15e9d39ccdf6bf3d95
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        - quay.io/openshift-community-operators/prometheus-exporter-operator@sha256:e62276f4c7e583d1fb24997fa970662f1a8d510de0baea159ba796b48cff709b
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        - quay.io/community-operator-pipeline-prod/ack-dynamodb-controller@sha256:547bdd3d1037dbc3b189f4c2f8e1d49379e1d547cd9063391d6c745e1a103b80
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:59ff4966f4f653e49d42d97bfed6351d3730128db74f666ee4ecacff11790baf
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:01f2ca529d2486f113bcefc9fedce6a6fd07bcb48aaf534394b5b04c353f8853
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.connect.redhat.com/cpopen/ds-operator@sha256:6ec2387e45d9557cea2df749831e136120945794773ed0d48e0b04f0b3d9e115
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        - quay.io/community-operator-pipeline-prod/activemq-artemis-operator@sha256:1e3c6ff2d2612c3e85d40b1974b5a3c4f603b77353752389fa588728bd4263c4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-athena-controller@sha256:8cff9258583f8acbdd26d263a61ffa11792cf9f4c1d65410bb4c71f9f7aaf178
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - quay.io/openshift-community-operators/devopsinabox@sha256:1c2ad76102151a56ecd753d9b86fd9f155d64b912067738d1858d33a02557928
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        - registry.connect.redhat.com/smilecdr/smilecdr@sha256:c44f5661a598aa18c4c10372918a45edb0f26d7d2772c47a55c66e4ff874cb5f
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - registry.connect.redhat.com/mariadb/mariadb-enterprise-operator-bundle@sha256:73e7f243a6676d559068bf8ab80cc0ffe0db58152675603ca8bd360334439f26
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator-bundle@sha256:df0dcd31867bb1c3c4321db4412a744fd1b3ba8faddb509d191fab92407b6677
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        - quay.io/community-operator-pipeline-prod/apicurio-registry@sha256:7087ec1377c2612edf3b8a76f6e115d66c3ec2b4b20f5d8374638d6bf0037e93
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-pipes-controller@sha256:e46b540219ca42c3b6c08ff99afabf09daea2dca5a335cda6013343df2f35878
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        - registry.redhat.io/openshift-serverless-1/logic-operator-bundle@sha256:5fff2717f7b08df2c90a2be7bfb36c27e13be188d23546497ed9ce266f1c03f4
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:53d8cf8bf5c1103860ffc4110e2a422371761e76ee79edfef8680e947407c5ce
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        version: 1.3.2
      deprecation:
        message: '''builds-1.3'' has reached the end of maintenance support. Please
          upgrade to ''builds-1.5''.'
      entries:
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:aadba00eadb3ac46d321a14498fc3dffd5be9afbf698d80adddde0a7a1681fc4
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:50b7f0a5f3fdd98cfb0294d664cc01ffd1e5359be66a3da14f5552e9cb3bae2c
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:29d46db99d97fe28758c0efd2acf27203ce9958734b82cf37c3a9063b5159506
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
          createdAt: "2025-09-09T14:14:54Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:ca10a842900d6515f714a17916a32dfd5a2909543e4d339acefc7e3e41b8d7ac
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:31d145c7074ad6cd85684b5191a3033533bf2149124c7ead5367871da5ed0dd8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        version: 4.17.13-rhodf
      entries:
      - name: ocs-client-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-client-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-client-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-client-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-client-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-client-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-client-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-client-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-client-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-client-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-client-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-client-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-client-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-client-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-client-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
          createdAt: "2025-09-11T15:23:25Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:60a91d6e4fe7e0ab999eab2224ef09fd0681e74920b23b306c7a9893bbb29cb8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:725eabb38876b93ab7693d75013d0292756a2efc00b8608f5716f31f8c324a53
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        version: 4.18.11-rhodf
      entries:
      - name: ocs-client-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-client-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-client-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-client-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-client-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-client-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-client-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-client-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-client-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-client-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-client-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-client-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-sso-bundle@sha256:f2c5a8abd1a9ec7665edf1e6206a6212cb752082b09532e4c218bc0990e48287
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:3dbb71004362a70683a5ee9f0916bca75c0d1c351d144677fd9d85bcb2ca8bf7
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:89d480ae76480829660d4cccfede74717faeb7e62a4a48b932750e41a945783d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a71b6fe33997db924669253992436e1e02d88769bb97376c0871848c2d330fed
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:1074f45a625df6ea412a9ffef3d4f4e53f3107ca45bb1de9bb030c47ce6518d9
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:0543e615d8a48ba6aaf90a8de0b0a79d37689c64bfe71f91c340455d9a359426
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a93fdee9e8d05ebe65e9dd8682fa7bc99c54f9d9f26d2cc8519f529255805da8
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:9d825d79fffd39a1149b262d48d08616795e5cfdb4c4bc40f5f192242f71042d
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: nexus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        - quay.io/openshift-community-operators/nexus-operator@sha256:277b94985c0dc982abb76b5dd8c93b37305ccca32c000d67552d540d4918cbba
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:b83c3263e77904ce1c0f95b7bfab969605c97297a5598fd794d64daf508c52c0
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:ea88a3090291f3742ccd4641785c19a80f4d3bd485a4f9c362a7645f73896db3
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/jaeger@sha256:907dfe1d01f73c8a385da3661ddbe03db31e0b9ad6f5ab2336460402d0d915a0
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kinesis-controller@sha256:de760dbe730e38cd9fadebe5fdb69299c04512e6a70f567ce678fc0bedc53e9c
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:a5645cb55c3e6c22655c6c8606b9bd51a49db44aecc3fe52e0eb80cca83d17c4
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:19e133c41b8f947aa83afba0417d8db3570cc2e1368813996fa5d135fdca4e04
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        - registry.connect.redhat.com/ibm/ibm-application-gateway-operator-bundle@sha256:e4ece7113fa7056980507408ac29d60676cf51c20d71a058a1accf765a6ea7e3
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchservice-controller@sha256:699f483cf96822576e3b79b690f256e2dbf4ca8d4426daf4298c980aefdeba4f
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/percona-xtradb-cluster-operator@sha256:d0c9d03130774d961f62032f684c494313787354186d028e9709e2463f69431c
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        - registry.connect.redhat.com/loadcore/keysight-wap-logging-bundle@sha256:2aeeefa7493eed8960004904567894244418982209e1b3a40488b759463edcdb
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.102
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-09-03T09:23:19Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.102
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator-bundle@sha256:fdef8f42c5425b9607915439daf3e5e2b77705cd220a083a00b42ee2e122d894
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:80c6d45a2ca70f55809620a82c309abb1e8272a0515f9c1a2905b3302c1467a1
        version: 0.1.102
      entries:
      - name: hibernation-operator.v0.1.102
        version: 0.1.102
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-commonui-operator@sha256:87788afbc24851deb2c9c3f14b773bab174b2a2f02ac2fa0dd656a97433ffca2
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dns-operator@sha256:d5eadbf006aade69ac18156c891614da899d0779ba9e8fbaf755a256c46f055c
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/openshift-community-operators/gitops-primer@sha256:97af48469275f1b3a669f8e24a97a80ce9897fc5b12643a4f298b82dbe26e20b
        - quay.io/openshift/origin-oauth-proxy:4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso-operators/watcher-operator-bundle@sha256:01ab69b9c33a91cb194ad644bb64884b358af78467e2bb766ef373e4c36e795d
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4478b8a78bdde23e0d40476f8ad9e4bf0628fba7603a390e8de7e6a319abc1ec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:6c2ab763239f6e5a2a72890120bf1f20f6277850cec59a7e11cea2b6a06f12e8
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:8f30d4c7c491e9a804a8abf80570d1e1924cd0d787c6abb727331541c435d5ff
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:2b6b1a0b56cfb348ed4035b13d31501ff5757c101ed21027b330a7852576b891
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4c4c98b6cecd51c6644eb79b85789b7ebbcdc662edbb66ce435ea3b5e58d9273
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:b8aa74c5b2635a6d3413eeac6fc0f6eb9a2e459089aaabbf626ea3ef92a67ce0
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:73d91324ccc995e9a6db525f194bee340c17dd624ce520c767ea8fcf2346f786
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:d29abe14abc67d8836cc6f7656556a0a7fa8f4a8eea3f78e40265b339e7430f5
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:f7c4ed28a549fe7414312bd22c2ae140f1f28d30656fb9fb0c8ade05a8f82088
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c59956d70416fab0c6d7b3ecd8c9cfa3cf9415f75cfdfe2214f44b0b4aced594
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:7cace73b03bb02202a04b6026eb9f7323610fd0ccd5ba18a6d00fd66577a96b4
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c173a99026499700c08092f8418c79f39527d9b914690b199b41ed781c279de1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:042623e1faacc9fc22222a1c3e546a561abf7fc05d41ae9c915a69795f6bec7e
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:67b2b8783a04460019057cf826a74c18747c8c64ee7fe8ce83c8d0b7cd87cb36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/resource-locker-operator@sha256:9cc08639be2af4b72a9b816420d87dfa290a168fb83edc947a04b3e06cdb675b
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/lib-bucket-provisioner@sha256:132ecc88db79fb8b117a7a7974ca8f704600bac489923d918b78fd00378d5a5f
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-operator-bundle@sha256:aeae57b2fc706b782a3e3905e5c1825fe831f96f5b1c9478e1f2b217307a892b
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataDirect Networks
      provider-url: https://www.ddn.com
    name: vmstore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vmstore-csi-driver-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.tintri.com/v1alpha1",
                "kind": "VmstoreCsi",
                "metadata": {
                  "name": "vmstorecsi-sample"
                },
                "spec": {
                  "config": {
                    "debug": true,
                    "vmstore_map": {
                      "vmstore2": {
                        "dataIp": "\u003cdataIp\u003e",
                        "managementAddress": "\u003cmanagementAddress\u003e",
                        "mountPoint": "/tintri-csi",
                        "password": "tintri99",
                        "username": "admin",
                        "vmstoreFS": "/tintri/csi"
                      }
                    }
                  },
                  "csi_snapshotter": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                    "resources": null,
                    "tag": "v6.3.2",
                    "timeout": "120m"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "tintri/vmstore-openshift-file-driver",
                    "tag": "v1.1.0"
                  },
                  "priorityClassName": "system-cluster-critical",
                  "provisioner": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                    "resources": null,
                    "tag": "v4.0.1",
                    "timeout": "120m"
                  },
                  "registrar": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                    "resources": null,
                    "tag": "v2.10.1"
                  },
                  "resizer": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-resizer",
                    "resources": null,
                    "tag": "v1.12.0"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "512m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "64Mi"
                    }
                  },
                  "secretName": "vmstore-csi-file-driver-config",
                  "volumeNamePrefix": "pvc-vmstore"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-13T21:26:54Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: VmstoreCsi
            name: vmstorecsis.charts.tintri.com
            version: v1alpha1
        description: "A Container Storage Interface (CSI) Driver for Tintri Vmstore
          filesystems. This operator eases the installation\nand management of the
          Vmstore CSI Driver in an OpenShift or other Operator Lifecycle Manager enabled
          cluster.\n\n\n## Prerequisites\nMake sure that the top level folder where
          the CSI volumes will be created exists on VMStore and is shared over NFS.\nFor
          example `tintri/csi`.\n## Usage\n\n### Dynamically provisioned volumes\n\nFor
          dynamic volume provisioning, the administrator needs to set up a _StorageClass_
          in the PV yaml (examples/nginx-dynamic-volume.yaml for this example) pointing
          to the driver.\nFor dynamically provisioned volumes, Kubernetes generates
          volume name automatically (for example `pvc-vmstore-cfc67950-fe3c-11e8-a3ca-005056b857f8`).\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\n```\n\n#### Example\n\nRun
          Nginx pod with dynamically provisioned volume:\n\n```bash\noc apply -f examples/nginx-dynamic-volume.yaml\n\n#
          to delete this pod:\noc delete -f examples/nginx-dynamic-volume.yaml\n```\n\n###
          Static (pre-provisioned) volumes\n\nThe driver can use already existing
          Vmstore filesystem,\nin this case, _StorageClass_, _PersistentVolume_ and
          _PersistentVolumeClaim_ should be configured.\n\n#### _StorageClass_ configuration\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc-nginx-persistent\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\nallowedTopologies:\n-
          matchLabelExpressions:\n  - key: topology.vmstore.csi.tintri.com/zone\n
          \   values:\n    - zone-1\nmountOptions:                         # list
          of options for `mount -o ...` command\n#  - noatime                          #\n```\n\n####
          _PersistentVolume_ configuration\n\n```yaml\napiVersion: v1\nkind: PersistentVolume\nmetadata:\n
          \ name: vmstore-csi-file-driver-pv-nginx-persistent\n  labels:\n    name:
          vmstore-csi-file-driver-pv-nginx-persistent\nspec:\n  storageClassName:
          vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n    - ReadWriteOnce\n
          \ capacity:\n    storage: 1Gi\n  csi:\n    driver: vmstore.csi.tintri.com\n
          \   volumeHandle: vmstore1/172.30.230.128:/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img\n```\n\n####
          _PersistentVolumeClaim_ (pointing to created _PersistentVolume_)\n\n```yaml\napiVersion:
          v1\nkind: PersistentVolumeClaim\nmetadata:\n  name: vmstore-csi-file-driver-pvc-nginx-persistent\nspec:\n
          \ storageClassName: vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n
          \   - ReadWriteOnce\n  resources:\n    requests:\n      storage: 1Gi\n  selector:\n
          \   matchLabels:\n      # to create 1-1 relationship for pod - persistent
          volume use unique labels\n      name: vmstore-csi-file-driver-pv-nginx-persistent\n```\n\n###
          Topology configuration\nIn order to configure CSI driver with kubernetes
          topology, use the `zone` parameter in driver config or storageClass parameters.
          Example config file with zones:\n  ```bash\n  vmstore_map:\n    vmstore1:\n
          \     mountPoint: /tintri-csi                            # mountpoint on
          the host where the tintri fs will be mounted\n      dataIp: 172.30.200.37\n
          \     managementAddress: tafzq23.tintri.com\n      vmstoreFS: /tintri/fs01\n
          \     zone: zone-1\n      username: admin\n      password: password\n\n
          \   vmstore2:\n      mountPoint: /tintri-csi                            #
          mountpoint on the host where the tintri fs will be mounted\n      dataIp:
          10.136.40.71\n      managementAddress: tafbm05.tintri.com\n      vmstoreFS:
          /tintri/csi\n      zone: zone-2\n      username: admin\n      password:
          password\n  ```\n\nThis will assign volumes to be created on Vmstore that
          correspond with the zones requested by allowedTopologies values.\n\n\n###
          Defaults/Configuration/Parameter options\n\n| Default  | Config               |
          \ Parameter    |  Desc                             |\n|----------|----------------------|---------------|-----------------------------------|\n|
          \  -      | dataIp               |   dataIp      | Data address used to
          mount Vmstore filesystem |\n| -    | mountPoint       |    -      | Mount
          point inside of the driver container where Vmstore filesystem will be mounted|\n|
          \  -      | username             |      -        | Vmstore API username
          |\n|   -      | password             |      -        | Vmstore API password
          |\n|   -      | zone                 |      -        | Zone to match topology.vmstore.csi.tintri.com/zone
          \           |\n|   -      | vmstoreFS            |   vmstoreFS   | Path
          to filesystem on Vmstore |\n|   -      | managementAddress    |      -        |
          FQDN for Vmstore management operations |\n|   -      | defaultMountOptions
          \ |   mountOptions| Mount options |\n|   0777   | mountPointPermissions|
          \  mountPointPermissions | File permissions to be set on mount point |\n\n####
          Example\n\nRun nginx server using PersistentVolume.\n\n**Note:** Pre-configured
          volume file should exist on the Vmstore:\n`/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img`.\n\n```bash\noc
          apply -f nginx-persistent-volume.yaml\n\n# to delete this pod and volume:\noc
          delete -f nginx-persistent-volume.yaml\n```\n\n## Snapshots\nTo use CSI
          snapshots, the snapshot CRDs along with the csi-snapshotter must be installed.\n```bash\noc
          apply -f deploy/kubernetes/snapshots/\n```\n\nAfter that the snapshot class
          for VMStore CSI must be created\n\n```bash\noc apply -f examples/snapshot-class.yaml\n```\n\nNow
          a snapshot can be created, for example\n```bash\noc apply -f examples/snapshot-from-dynamic.yaml\n```\n\nWe
          can create a volume using the snapshot\n```bash\noc apply -f examples/nginx-from-snapshot.yaml\n```\n\n##
          Troubleshooting\n### Driver logs\nTo collect all driver related logs, you
          can use the `oc logs` command.\nAll in on command:\n```bash\nmkdir vmstore-csi-logs\nfor
          name in $(oc get pod -owide | grep vmstore | awk '{print $1}'); do oc logs
          $name --all-containers > vmstore-csi-logs/$name; done\n```\n\nTo get logs
          from all containers of a single pod \n```bash\noc logs <pod_name> --all-containers\n```\n\nLogs
          from a single container of a pod\n```bash\noc logs <pod_name> -c driver\n```\n\n####
          Driver secret\n```bash\noc get secret vmstore-csi-file-driver-config -o
          json | jq '.data | map_values(@base64d)' > vmstore-csi-logs/vmstore-csi-file-driver-config\n```\n\n####
          PV/PVC/Pod data\n```bash\noc get pvc > vmstore-csi-logs/pvcs\noc get pv
          > vmstore-csi-logs/pvs\noc get pod > vmstore-csi-logs/pods\n```\n\n####
          Extended info about a PV/PVC/Pod:\n```bash\noc describe pvc <pvc_name> >
          vmstore-csi-logs/<pvc_name>\noc describe pv <pv_name> > vmstore-csi-logs/<pv_name>\noc
          describe pod <pod_name> > vmstore-csi-logs/<pod_name>\n```"
        displayName: Vmstore CSI file driver
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - CSI
        - Vmstore
        - storage
        - Tintri
        - DDN
        links:
        - name: DDN vmstore Filesystem
          url: https://www.ddn.com/products/lustre-file-system-vmstore/
        maintainers:
        - email: akhodos@tintri.com
          name: Alexey Khodos
        - email: rrethinam@tintri.com
          name: Ranjith Rethinam
        maturity: alpha
        provider:
          name: DataDirect Networks
          url: https://www.ddn.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/ddn/vmstore-csi-file-driver-operator@sha256:64df1c504a9d70bd2b8b34520c8d431a92bb54702d601a57a2a989c6dd982462
        - tintri/vmstore-csi-driver-operator-controller@sha256:749ceff2bd46c71d81338510c1cc5866d39c66345f971d3093f986ea0417311d
        version: 1.1.0
      entries:
      - name: vmstore-csi-driver-operator.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vmstore-csi-driver-operator
    provider:
      name: DataDirect Networks
      url: https://www.ddn.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle-rhmp@sha256:1f3d005224eada1b71d04f21d07a7c4fde159725c7a44d967d996152c802bede
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle-marketplace@sha256:7864f9f1cf3ad1d551b698b2e803302d207704b448b15f6c0013cbd9e854593c
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator-bundle@sha256:e284d0839795f763ca53402c2b050fa9ffedebccd0f3832fe7259368c3ad32f2
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-s3-controller@sha256:058a37a9a2fe6a42bbdddabb3fc7495450f1ebbd7f46194866c1491c40ce1a9f
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ClusterPulse
      provider-url: ""
    name: clusterpulse
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterpulse.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.clusterpulse.io/v1alpha1",
                "kind": "ClusterPulse",
                "metadata": {
                  "name": "clusterpulse-prod"
                },
                "spec": {
                  "api": {
                    "enabled": true,
                    "imagePullPolicy": "Always",
                    "logging": {
                      "json": true,
                      "level": "INFO"
                    },
                    "oauth": {
                      "enabled": true,
                      "headerEmail": "X-Forwarded-Email",
                      "headerUser": "X-Forwarded-User"
                    },
                    "rbac": {
                      "auditEnabled": true,
                      "cacheTTL": 0,
                      "defaultDeny": true,
                      "enabled": true,
                      "groupSyncInterval": 300
                    },
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "security": {
                      "allowedOrigins": [
                        "*"
                      ],
                      "sessionExpireMinutes": 1440,
                      "trustedHosts": [
                        "*"
                      ]
                    },
                    "server": {
                      "host": "0.0.0.0",
                      "port": 8080,
                      "reload": false,
                      "workers": 1
                    }
                  },
                  "clusterEngine": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "timing": {
                      "cacheTTL": "600",
                      "fullScanInterval": "300",
                      "reconciliationInterval": "30"
                    }
                  },
                  "frontend": {
                    "config": {
                      "features": {
                        "devTools": false,
                        "mockData": false,
                        "nodeDetails": true,
                        "operatorDetails": true,
                        "registryStatus": true
                      },
                      "refreshInterval": 30000,
                      "theme": {
                        "defaultTheme": "auto"
                      }
                    },
                    "enabled": true,
                    "oauth": {
                      "enabled": true,
                      "image": "quay.io/openshift/origin-oauth-proxy:4.17"
                    },
                    "replicas": 3,
                    "route": {
                      "enabled": true,
                      "tls": {
                        "termination": "reencrypt"
                      }
                    }
                  },
                  "policyEngine": {
                    "batch": {
                      "cacheClearBatchSize": 500,
                      "redisPipelineBatchSize": 1000,
                      "redisScanBatchSize": 100
                    },
                    "cache": {
                      "groupCacheTTL": 300,
                      "policyCacheTTL": 300
                    },
                    "enabled": true,
                    "limits": {
                      "maxPoliciesPerUser": 100,
                      "policyValidationInterval": 300
                    },
                    "logging": {
                      "level": "INFO"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "enabled": true,
                    "volumePermissions": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "ClusterConnection",
                "metadata": {
                  "name": "example-cluster"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "cluster-credentials"
                  },
                  "displayName": "Example Production Cluster",
                  "endpoint": "https://api.example.openshift.com:6443",
                  "monitoring": {
                    "interval": 30,
                    "timeout": 10
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "MonitorAccessPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "access": {
                    "effect": "Allow",
                    "enabled": true
                  },
                  "identity": {
                    "priority": 100,
                    "subjects": {
                      "groups": [
                        "default-group"
                      ],
                      "users": [
                        "default-user"
                      ]
                    }
                  },
                  "operations": {
                    "audit": {
                      "logAccess": false,
                      "requireReason": false
                    }
                  },
                  "scope": {
                    "clusters": {
                      "default": "none",
                      "rules": [
                        {
                          "permissions": {
                            "view": true,
                            "viewAuditInfo": true,
                            "viewCosts": true,
                            "viewMetadata": true,
                            "viewMetrics": true,
                            "viewSecrets": true,
                            "viewSensitive": true
                          },
                          "resources": {
                            "namespaces": {
                              "visibility": "all"
                            },
                            "nodes": {
                              "filters": {
                                "hideMasters": false
                              },
                              "visibility": "all"
                            },
                            "operators": {
                              "visibility": "none"
                            },
                            "pods": {
                              "visibility": "all"
                            }
                          },
                          "selector": {
                            "matchPattern": ".*"
                          }
                        }
                      ]
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "RegistryConnection",
                "metadata": {
                  "name": "dockerhub",
                  "namespace": "clusterpulse"
                },
                "spec": {
                  "displayName": "Docker Hub",
                  "endpoint": "https://registry-1.docker.io",
                  "healthCheckPaths": [
                    "/v2/"
                  ],
                  "insecure": false,
                  "labels": {
                    "environment": "production",
                    "provider": "dockerhub"
                  },
                  "monitoring": {
                    "checkCatalog": false,
                    "interval": 60,
                    "maxCatalogEntries": 100,
                    "timeout": 10
                  },
                  "skipTLSVerify": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: docker.io/clusterpulse/clusterpulse-operator:0.1.0
          createdAt: "2025-09-26T21:09:27Z"
          description: Provides OpenShift cluster health monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/ClusterPulse/operator
          support: None
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a connection to a monitored Kubernetes/OpenShift
              cluster
            displayName: Cluster Connection
            kind: ClusterConnection
            name: clusterconnections.clusterpulse.io
            version: v1alpha1
          - description: Manages the deployment and configuration of ClusterPulse
              monitoring platform
            displayName: ClusterPulse Instance
            kind: ClusterPulse
            name: clusterpulses.charts.clusterpulse.io
            version: v1alpha1
          - description: Defines fine-grained access control policies for cluster
              monitoring
            displayName: Monitor Access Policy
            kind: MonitorAccessPolicy
            name: monitoraccesspolicies.clusterpulse.io
            version: v1alpha1
          - description: Defines a connection to a container registry for monitoring
            displayName: Registry Connection
            kind: RegistryConnection
            name: registryconnections.clusterpulse.io
            version: v1alpha1
        description: |
          ## About ClusterPulse
          ClusterPulse is a comprehensive multi-cluster monitoring and management platform designed for OpenShift environments. It provides centralized visibility, policy-based access control, and real-time monitoring across multiple clusters through a unified dashboard. For more information, visit [https://github.com/ClusterPulse/operator](https://github.com/ClusterPulse/operator).

          ## Features and Capabilities
          - **Multi-Cluster Monitoring**: Connect and monitor multiple OpenShift/Kubernetes clusters from a single interface
          - **Policy-Based Access Control**: Fine-grained access policies using MonitorAccessPolicy CRDs
          - **Real-Time Metrics**: Live cluster health, node status, and resource utilization tracking
          - **Registry Monitoring**: Monitor connected container registries
          - **OAuth Integration**: Seamless OpenShift OAuth integration for authentication
          - **Redis-Backed Storage**: Scalable data storage using Redis for metrics and state
        displayName: ClusterPulse
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - clusterpulse
        links:
        - name: ClusterPulse
          url: https://github.com/ClusterPulse/operator
        maintainers:
        - email: asmoore@redhat.com
          name: Asa Moore
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: ClusterPulse
        relatedImages:
        - docker.io/clusterpulse/clusterpulse-operator:0.1.0
        - quay.io/community-operator-pipeline-prod/clusterpulse:0.1.0
        version: 0.1.0
      entries:
      - name: clusterpulse.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: clusterpulse
    provider:
      name: ClusterPulse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:61685ecb519d17160149a5dc16ae28d6fd96264652a6f19fef861addfa4ec6f4
        - quay.io/rasen708/ncn-operator:v0.2.2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:a99a5ca46c5bd3065905588f4d0e3cdc6409d435abdc02793be88e77c1a8f0f5
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:7d2ad8e800339c15efa5d5954106e01151d34135b25e9a23e30f0027ee23f193
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:eaa77ae0f1a1f3d32e269e73eb29a73d11012d5e920af1011d1842bd904d7293
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:4d9d003842e3ffc386f4e07ce7a81cfa85e676031fb6fdc96aefa9654244974a
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:a7641fc5f4c7aed4b73d7d98ccb58ef64cf14ab16d72378f56dba3ad4b8714c6
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:7dffd170b9b7d021b1615e6d746cbdfcbd819a1306ecdbaf042bcaf2202f7daa
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:111d1f4ab8f006386102d1bec4cd6a41b446d9ae401a43c821b2f318c8ba74c8
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:d7a4f9b8e5eb9bf6d11e00898eb168e86636f0633959eeb2b36b2d6e5e505f71
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        - registry.connect.redhat.com/loadcore/keysight-wap-diagnostics-bundle@sha256:4e8400a38a04c61dc20ece1ee7dac1ad8a56ac309621f3225e7389e84e20c870
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator-bundle@sha256:85e75ec797f70c5807ec1ea25208cf0bb1990daa05eca9fbc82bc38aa4a840ab
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
          createdAt: "2025-07-16T09:03:58Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <2.8.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk@sha256:99aec4a74ea4c9910b50cc035b4c0c85ed4711a4c9b456028690d0f0c7af2657
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/splunk/splunk-operator-bundle@sha256:433bec6334c4771e140a24ea8753e7bcbf358c4504c942658fd3364915ac86ac
        - splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
        version: 2.8.1
      entries:
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
          createdAt: "2025-09-19T19:24:03Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-mq-controller@sha256:702ffd931cb4baafb1eac49f9fda5588d8b02f3b8c696e477dd626b0efad9f54
        version: 1.1.1
      entries:
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ssm-controller@sha256:922fd6daa660d2250bad8a3c54733cceae845624a14651cb20e049e29bf90e1c
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        - quay.io/openshift-community-operators/community-trivy-operator@sha256:e3652e68e1e569d7f40432fd8cbb7bfdc6738a116ecc1011c71a36f7ca5663c6
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:3b538de0bbaa43d2368402ea1c0eaf767c8b9e5a28e112a309148a630dd3c052
        - registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:5b61484c6a2bbc377435093bd9abb31df980ece470325cae92f3106963b6482c
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:af8ccc7a075f11bf6f31310830704e3a27b23587094f56e518b1f264c92b0fe4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: kubernetes-nmstate-operator.4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: kubernetes-nmstate-operator.4.18.0-202508271223
        version: 4.18.0-202508271223
      - name: kubernetes-nmstate-operator.4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: kubernetes-nmstate-operator.4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: kubernetes-nmstate-operator.4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: kubernetes-nmstate-operator.4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: kubernetes-nmstate-operator.4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: kubernetes-nmstate-operator.4.18.0-202505280403
        version: 4.18.0-202505280403
      - name: kubernetes-nmstate-operator.4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: kubernetes-nmstate-operator.4.18.0-202505061007
        version: 4.18.0-202505061007
      - name: kubernetes-nmstate-operator.4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: kubernetes-nmstate-operator.4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: kubernetes-nmstate-operator.4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: kubernetes-nmstate-operator.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: kubernetes-nmstate-operator.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: kubernetes-nmstate-operator.4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: kubernetes-nmstate-operator.4.18.0-202503190933
        version: 4.18.0-202503190933
      - name: kubernetes-nmstate-operator.4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: kubernetes-nmstate-operator.4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: kubernetes-nmstate-operator.4.18.0-202502250802
        version: 4.18.0-202502250802
      - name: kubernetes-nmstate-operator.4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/cluster-observability-operator-bundle@sha256:8c27ac0dfae9e507601dc0a33ea19c8f757e744350a41f41b39c1cb8d60867b2
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:f5f025c7f56643fab9148bb5de8ee34a84d02b78a6abb4b5592306cf54296654
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        - quay.io/community-operator-pipeline-prod/percona-postgresql-operator@sha256:44e687b38a48f68bbbb919e6df0d3b87b75492aca7bd41e508a2aa0e80d024a9
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        - quay.io/community-operator-pipeline-prod/ack-elasticache-controller@sha256:a489c815e9d3777d775ccd65bbd0ecd5fe7c577de36c504b63cf1620531d4ccb
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:301c3b793d8068f9e50c855196da91b6c07111ef500b02a91fbb354c9fc1316b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:8edd2f89d7adda6c1a5aae7414f50e0d786ec8443c1db13834ea4924009b6419
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:227159c5a93966a43fa3a84cb69536ff0051cc261da068561241261d3837998a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:60635156d6b4e54529195af3bdd07329dcbe6239757db86e536dbe561aa77247
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:e4d49cf6a2fb8b3585ecfe3b3da3e9de090fb99de3dee2da0601b1f6b071173c
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:095e93f8dd31a0985d6fef6a6168da456630febe89ab9e0cb9480170f1036541
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:524d315efa226f97ff26b8668f9297e755c6f0c815b8f3385b317667553f0e62
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        - registry.connect.redhat.com/loadcore/keysight-rest-api-browser-v1-bundle@sha256:89ef7cf0dc8b27e58a63a82a701894a7cc4d0643bf0c8c8e20c9bb59d1a1b1bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-bundle@sha256:80ca5ad4634305808ac5a3e8ff656ce6ff69b3c16dccc5e03f42763fa89bc762
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:857c1bdff46d1a1d5abb51f75a80b7944bb461687b179b6f1622103947b8c3b1
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:d101e2b9ab4d8ca7c87c86ea4e3630dcc82a9c5a78885179c8b16773c8aac587
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:ecdd5c5fc609128e0af751f87b797b71e4ab0200a9abaaca835ff9a3f293fb10
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:be099e23851cb4b6d15e9c44c4da8fb6cba74aab005597d305757d1dae8b0122
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:96119f61731cec7eb037c3a4a7c2c8bfdf9ef11fd5b08abc25e3d24ddf342b06
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:2ea267aac67763b4f160c4e6876c0226d2fa135ee00614814a59b4de1d91f4c4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:218905f7902e6ac80366856d6c5ae37cd6d167c81e9521df7c3c3d6619f6c425
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:09fe022370abfcefcf3abfe8b6debc2f201a0ac8b8dacaf1278b8a6bee2b5273
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:5b865231214649649a2a9fb06934f931503f68ed740e150be1c7c8fcbb6f1317
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:16245de1e88228c231ab2edb4e9380753eff982bb72febadd00cdaad286a8eb3
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:491e7aea221eb0e1d2992310e53d50db287cbf32f62d090bd842236b603d38bb
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:7de09e5394985cdec87f5afab65b3343b8a37c0eb7f778774b21933ae69d691c
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle-marketplace@sha256:28f388286bc4b4981182ff1ab9e927f59744dcb755392907fe65cd3b96a44519
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:591fc4b5867dff26df1952653423b33889c0f56ba3f756057694d4c0e396eeb0
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:3225c49cf30a146784533496ce478466034d4549ff9721752958bac342fdec77
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:b0fa6b0356d83626509289b33dfa669bd7d32542aba986a18d296b85ce63389a
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:fe759127885502dc66423530ff03e37973355c5566a1c27e988b59d71528dd2b
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:c0336e5079f255ab6cd38e4e2e61d0cb4c6a35a976528fb44409bfde57e70dde
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/openshift-qiskit-operator@sha256:633ce4190b02034d70b52d7e624fc9d417ef997607f9710a36e73bbc0963fe4d
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/community-operator-pipeline-prod/openshift-node-upgrade-mutex-operator@sha256:5ac79c9193934440464a595434fe4241e038505d32650843afcad5f5bc2ce8a7
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        - quay.io/openshift-community-operators/mercury-operator@sha256:0c253fd002cb389db40b2438d4d1c9e953a84125dbf0f4b6f5e3a6efd87cf3a6
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:9db455e082454734d2a17811dcfec53d3d60c28df6e7fe8bc979a45b28a35061
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:8650430afd813fd749fcd80ad5f6fd0aeaab3df048e8fe34e6aa72a07ab801fc
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:ed063f507fa8e2f378d0adbde823c110484d0212d124cc72d67efa987e86444e
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.8.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:68b310ed3cfd65db893ba015ef1d5442365201c0ced006c1915e90edb99933ea
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:bcc5ed9560667c7347f3044c7fa6148e3867c17e93e83595a62794339a5a6b0d
        version: 0.8.1
      entries:
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/3scale-apicast-operator-bundle@sha256:29fe669a56a69adb537798c91b9359fa3b051da5ffa9f27789a2067a3a72330a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.7.9-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:ea2229948fea841343fcaa337bb7e2854acf5f563d3ffb0af8bda1c21ec2a91f
        - registry.redhat.io/3scale-mas/apicast-operator-bundle@sha256:d3a608b37aa089ea46e803696ba4957585b875f8a810f517e891f385ad105937
        - registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
        version: 0.7.9-mas
      entries:
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        - registry.connect.redhat.com/loadcore/keysight-wap-storage-bundle@sha256:0955dbda03bd7307364a1a458ab8d6ac5a8c9feb6150706dcede9676a6513b0d
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        - quay.io/openshift-community-operators/hedvig-operator@sha256:81d956c12057f18bf920044a4e39c27c99b2c23afdccc0ff9468f0484e1961d7
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a77d39ded03271f2deb8b438abd4328a1846ee0d4886cce14d869ba61f84d600
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a6f3a3a8aa52de4441ef3249a76dcff1b6b6defba417a9fdbba37eb458adee04
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:6208f64c7fe75f39ab7511e2b480c493d2ecc8d1e76a1ac498a93a6a24e8beb5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:11d64a921a98b7ec1a2126a03e446a055a5ae1705dc2005bb8b73cbbcf7ecf99
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatchlogs-controller@sha256:94145cae2b8cef1410acbb4f1c1eb571e71408341716af1742f27ddf4dc350b6
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/coherence-operator:3.5.4
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-operator-bundle@sha256:f40e7955ac649c383134ab6e77bdd521b75b21ecd9b476436405d0674c7fc855
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-operator-bundle@sha256:d04033c2cd2d236f6319a3e35880b89c8f0c60d53cd32f6f097d07d6d27e5f2c
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-ati-updates-bundle@sha256:3d06a58b2a277a67339c22f3fe10c1daac1c8eabaac8540b01177f6f5688784c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/volume-expander-operator@sha256:b775df99f891e2cee9c2169e9c948a8d61d65b4cc90f36752b46a56e2d4be602
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/apimatic-kubernetes-operator@sha256:a5f68205291caa9bb7903180afc974fe86252ba87e671978c8b49e0206976d06
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
        - quay.io/community-operator-pipeline-prod/ndb-operator@sha256:5f33c2a8181ea8e91844afd1f4042affbd9eb878266b6b2bcc6906f0829ed00b
        version: 1.6.0
      entries:
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - quay.io/openshift-community-operators/keda@sha256:50816f675c76d36dbc62846579150477018ef87df325861751f3fc02a70c8781
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/community-operator-pipeline-prod/alloydb-omni-operator@sha256:08fafb811f709780c0ccb3c22e3c6417cb14a70a9305e46a7b50457647a82102
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:ed530db9b4a29fb591ea208cd0c6b3b83b49537ff5a0a4f501b54ce667f9b303
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:31bdff225bd825d02ce0f468b1edc77541ce85b7b9111f14a583c418c984232f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:80198f65f7e3f9022a34e0342e9423807b855dcb7bda810172eebc9a27afb5c2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:f04d565eb5162fd5a9c78cd8443af6a3b184a660277537f1d6d575caa2e85441
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:773592dfb9964970f444871a8812e8b74db3f205faaa0bf148defa1a2b87b649
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:0b1771f697a700263c5b827dbde56a5c9bab610e9d7df55ef93885a4e54cc0e9
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8441322a0fce02df407573675b5b5d92d56de97c8aec72541b33d2dadc4050d5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:be93be0341649215fb001725aac0c3c5925343adbf1ad00c979b8c3d489512d6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b22190de3162d3e03481cae8904ce563912756eea7ac7e7feec0acbbdae5e6f6
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:c2f39be75173b4cc57498c5845b14c464bf9af92b49716df9661c8135ac2cf7a
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b2d58f4dc20eb7e17e09af0e8c5de55fa04f9f4f697e10c6be81a7a6a08381b2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:27efa0ab1d9c69047a924bc1dcd2b56108f5f390a6da56afb6ab7a37cb4578e2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:ed8230e90f0ab0586a743d5e760a7469345b1136971aad17ef675c1beccf1982
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:08995782514e496c24d888352e9d1e6b788dd87adb761e0c2e745be2975ffcaf
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:7b324e897fbcc881ececa529822d7b738f82cd4c3adf4cc54b66080b888f5d9c
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:e2a329fc74ed5f17e395367126c7e96197178e8f886d475b657a7910a4a14152
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:0218d68f97e89b48bbdd9c7cd750d4a666cf33af1cb8a76507c067a332a7b6bd
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:2bc56e149c14920662b549a5367f98daf9afad0df29b4746b195f6afc0c9c597
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
          createdAt: "2025-09-19T22:01:50Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
        - quay.io/community-operator-pipeline-prod/ack-eks-controller@sha256:120566a02e487aaa8823323423fccfaa377135037e2e1cb9f3ff14313cf3bfeb
        version: 1.9.2
      entries:
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: sonar-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        - quay.io/openshift-community-operators/sonar-operator@sha256:5e7a87f52955162f090c1d7fc69f6d0917444b553355a11c061ab60f289a17d7
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    defaultChannel: alpha
    packageName: sonar-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-documentdb-controller@sha256:8a991c76f281ec30258b6317bd623894b78a4f39d9790a1cbb32e6f1b7ff63f5
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        - quay.io/community-operator-pipeline-prod/dynatrace-operator@sha256:c475f9cc0f6caa2bb316ff42a35a40b1d4bf8ef7c8a3384d9c78eb205bda7076
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        - quay.io/community-operator-pipeline-prod/kaoto-operator@sha256:4b6cc43eba8efbe17ffec746431d6cf6b045886ff865673a08a0820abecebdcd
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:5ea2dd102074798d1ee005dc03fbbd24f4324df831f279355dc2dfc262c9d880
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:e4dec1f39f8e24043cf16703efcac2cc3b953d3f4c37b3956fce0c65ef829764
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/operator-bundle@sha256:bc8a47d5f8ffe3a2fe48819f08c82525677131cfb4f3c0facd82470c0f10acee
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        - registry.connect.redhat.com/loadcore/keysight-nimbusmosaic-dusim-bundle@sha256:97d5f906ef83272a8d7a2a249f38f8b9b186f8e9b64729a049a7eac0dbda8b1a
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        - quay.io/community-operator-pipeline-prod/pelorus-operator@sha256:f54c42a30675897a971cac278d5d8ea24edf3121570264c07cde04f11cc4e7a8
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:57af9e32c31dad769bda449ebe35e09ea135e8a7d06cddb42be4d1a09005f518
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ab3cb65b88c34fd58524a2a51cd510c611924e5860834965bad0b4418a8b6dd3
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
        - registry.redhat.io/openshift4/ose-local-storage-operator-bundle@sha256:d303fc63f5b849f073e34a7791ea92cdc3cff269cd1b593e27480ae96cac1a15
        version: 4.18.0-202509100344
      entries:
      - name: local-storage-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: local-storage-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: local-storage-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: local-storage-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: local-storage-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: local-storage-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: local-storage-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: local-storage-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: local-storage-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: local-storage-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: local-storage-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: local-storage-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: local-storage-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: local-storage-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: local-storage-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: local-storage-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
          createdAt: "2024-09-04T17:53:42Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:07cafb83ba90378c1fa60338cf951a9d7145ec245710637e90fdb508875dcc45
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:1d86075469312e84f3d16379efcd6130ec9e8b23a1343e3945f071d82beff6e3
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:dc10544e1bb24883f38373a0445944745282e5a69ed1744f428f98871a7a39c9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:31b3658dd79c18111c21e99aea52cc198f8e95b458a1a4f8d2bfc4d21e9bdacb
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:1b8008a16a0f53c09ecc5a219964194689a1aeb01acee8313aa49a9c26259429
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:a539064234088a4ee13b8a9666ca9507208fe38189ee9b8b673a38a58aa75599
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:b137e1345038c9e566767f24cce60c03ec278f5a2efc28fa08e0bd8bfa016486
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:c001dc96a1478a58f30461ab018e40bede13a78f3cdce869a70dcb8b9bba22b1
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:92a3162c705537147005a065cdddbb410e1863d80c08d8b0f27b7ff5588500cd
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:8fbfffb0bc3adfb4d72961dad381e1daba414bf4a7084f03b5208f6be4191ceb
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:035855c3d4b32a29232846d73dc398e81f941fd90a03e33241ca06f6e80b2204
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:f86e96edb0418811df328debdee93b4a51cf7f651134ee35e9c50b9dbfc29fd8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:084f1e54ef0d7f457d5a7cfc57b567ad8d85237c12f707c82f7a30a54e3652ba
        version: 1.7.17
      entries:
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:9f5479ac795eb50fa5ea597828b2d50ddaa4878faa1a07b3de0e611cb4af5fc6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
          createdAt: "2024-11-21T23:48:37Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVSNetwork is the Schema for the ovsnetworks API
            displayName: OVSNetwork
            kind: OVSNetwork
            name: ovsnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-operator-bundle@sha256:4c8271d0bf6d94061fe8df8110b041cb0e1a840fb6263315f552166fc5053e73
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:2e71d4125a35f6e7bd2a04cdd8cb6aae058d745eda372ae3d3fd37783778f486
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:57a57f4f3a0adb8b7234683f2968352988ec9a414824f4f444cca096825a5df5
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:45835aba95af72aa8e73b630d1d2b4c1079217531f025f4715a2eb8e8856205f
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:8b412541a797aaff1b49aa229e310957a8d7da03a3d38e6bbfd65f2ca36c2bdd
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:6eb244a301ae5395b69b1bfcd0baac37c094f4c38a1a330c0a70fb81e71f6157
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:3b87d49b200ac4fdeac97f0e97a46001e91c8f4680e2975e90bd94ec37864403
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:0f18db6a5bbc952bee8f3e4ac9e93e807d1e85cd132affe94b1f4005c1e128ce
        - registry.redhat.io/openshift4/ose-sriov-rdma-cni-rhel9@sha256:65864bc235689ff38b63d2a7b2c72874ee155241a493fa9429fbe87d5100cfcf
        version: 4.18.0-202509100344
      entries:
      - name: sriov-network-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: sriov-network-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: sriov-network-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: sriov-network-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: sriov-network-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: sriov-network-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: sriov-network-operator.v4.18.0-202505260733
        version: 4.18.0-202505260733
      - name: sriov-network-operator.v4.18.0-202505121333
        version: 4.18.0-202505121333
      - name: sriov-network-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: sriov-network-operator.v4.18.0-202504221235
        version: 4.18.0-202504221235
      - name: sriov-network-operator.v4.18.0-202504160002
        version: 4.18.0-202504160002
      - name: sriov-network-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: sriov-network-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: sriov-network-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: sriov-network-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: sriov-network-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: sriov-network-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle-marketplace@sha256:4145f7a04d7dbdda226b75af4e4cabca63a803f290b1581ea4ae0dc223eca7df
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        - registry.connect.redhat.com/loadcore/keysight-wap-appsec-bundle@sha256:e5049593a52fb26c124de7005b45919f26fa0430293f18283a20c8cde981b025
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cli-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager/cli-manager-rhel9-operator@sha256:be0fc230fef111cff36a8b846bbac9aaa98aae94ad9200416e841aa487121de4
        - registry.redhat.io/cli-manager/cli-manager-rhel9@sha256:a6b7857c1acd5173c770e6ba1d03947eb13804d2512f7f7cd6bb96e5e5cdaa17
        - registry.redhat.io/cli-manager/cli-manager-operator-bundle@sha256:74c88f1cfef4cc31b794fc5ed8dcc10d3b0e6869c228d808f1d98819e69554b6
        version: 0.1.1
      entries:
      - name: cli-manager-operator.v0.1.1
        version: 0.1.1
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:a0e4c432590e65818e7ebaab127d483d68fdc5ccae9a13820665a8aef122fde7
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:242d754d5fe1f2a07091ae93da284b8912e0f4b4c973200cf6d1241e6abb88a9
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:437a786a301ea720cdb5651dff878a9f51470bab46ba28daae4928c7074c9157
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/lws-operator
    name: leader-worker-set
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: leader-worker-set.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "LeaderWorkerSetOperator",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-04-14T08:12:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lws-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/lws-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LeaderWorkerSetOperator
            name: leaderworkersetoperators.operator.openshift.io
            version: v1
        description: "Leader Worker Set Operator is based on the [LWS](https://lws.sigs.k8s.io/docs/)
          open source project.\nLeader Worker Set: An API for deploying a group of
          pods as a unit of replication. \nIt aims to address common deployment patterns
          of AI/ML inference workloads, \nespecially multi-host inference workloads
          where the LLM will be sharded and run across multiple devices on multiple
          nodes.\n"
        displayName: Leader Worker Set Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - leader-worker-set
        links:
        - name: Leader Worker Set
          url: https://lws.sigs.k8s.io/docs/overview/
        - name: Operand Source Code
          url: https://github.com/openshift/kubernetes-sigs-lws
        - name: Operator Source Code
          url: https://github.com/openshift/lws-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/lws-operator
        relatedImages:
        - registry.redhat.io/leader-worker-set/lws-rhel9@sha256:affb303b1173c273231bb50fef07310b0e220d2f08bfc0aa5912d0825e3e0d4f
        - registry.redhat.io/leader-worker-set/lws-operator-bundle@sha256:f641cf3b2d6659e26eb0ae49c67a3416152a2fbc2f26f490889a2bd169b9ea06
        - registry.redhat.io/leader-worker-set/lws-rhel9-operator@sha256:c202bfa15626262ff22682b64ac57539d28dd35f5960c490f5afea75cef34309
        version: 1.0.0
      entries:
      - name: leader-worker-set.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: leader-worker-set
    provider:
      name: Red Hat
      url: https://github.com/openshift/lws-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:72a54c9457e311001b25c1b414b7001c6d7f9a34b89efb1b4198084586386f31
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        - registry.connect.redhat.com/questlabs/alloy-operator@sha256:fe4e83fc4a50e1aa04e6a9ba2a57acc737077b86681c1d7c597646fad337f221
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - registry.connect.redhat.com/kove/kove-operator@sha256:04bf970760c4dfe910465b0ef6a18e9c17fa70f6388131f12ae0aa0cd646f663
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vertical-pod-autoscaler.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "autoscaling.k8s.io/v1",
                "kind": "VerticalPodAutoscaler",
                "metadata": {
                  "name": "myapp-vpa"
                },
                "spec": {
                  "resourcePolicy": {
                    "containerPolicies": [
                      {
                        "containerName": "*",
                        "controlledResources": [
                          "cpu",
                          "memory"
                        ],
                        "maxAllowed": {
                          "cpu": 1,
                          "memory": "500Mi"
                        },
                        "minAllowed": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    ]
                  },
                  "targetRef": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "myapp-deployment"
                  }
                }
              },
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "minReplicas": 2,
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "safetyMarginFraction": 0.15
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
          createdAt: "2024-12-05T23:41:17Z"
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1beta2
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1beta2
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: Vertical Pod Autoscaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - verticalpodautoscaler
        - vpa
        - autoscaling
        - scaling
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-operator-bundle@sha256:f5df5d878ef0a0b494c9bafe99c35f3152d30f6c426be789fcd636dd13eff6e6
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:5da084c4dc81251a707db1bd9347c13b39a1e01660120aa2e6a47f9d1d981fe9
        version: 4.18.0-202509100344
      entries:
      - name: vertical-pod-autoscaler.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: vertical-pod-autoscaler.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: vertical-pod-autoscaler.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: vertical-pod-autoscaler.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: vertical-pod-autoscaler.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: vertical-pod-autoscaler.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: vertical-pod-autoscaler.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: vertical-pod-autoscaler.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: vertical-pod-autoscaler.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: vertical-pod-autoscaler.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: vertical-pod-autoscaler.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: vertical-pod-autoscaler.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: vertical-pod-autoscaler.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: vertical-pod-autoscaler.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: vertical-pod-autoscaler.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator-metadata@sha256:d6d0e71f5451e8255edfff87ef28ba7743fe7057efd53b25bbbd01bf9e1d0307
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:c6d6c3e0408b851cdc9589cc0ee642d77397cdfaa40f50cf42557c5aaf9c0690
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:ff314ab983699455c32ab0ef27bb552d00825974a2d7088df2000f09afd8a2e2
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:5f10cfc1bc8571f44a99d88755adade0743106219673bbdd2e4d3a04ee714778
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:9f90cd7ab8c0ab5da6eb12f7f0dbd177df62f07c38fb759a97989899cfcbe75d
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:71ff93b137b6569f8fdefbd446cf12cd95477e4294bdea98fb5318e5d7d1cdd6
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:7a04dfd687dfe1646bd59a39c9e88781ffb815085bbd15f6305f6eac5f59bc49
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:d1a9246def81537e3417a6d48d1e818459583db658b54a87ca5af6911c8a3f53
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:e88676c0346c245b315f1a3bf7f85b10515bae3620c786fedda0533b39d2a400
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:66bd3eba0b68c317fe128c6713488cb6374f9647095768d3f7eed8120bf02e77
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        - quay.io/openshift-community-operators/authorino-operator@sha256:8fe34f94f87c567401cc2bbbbcc3a3e719364b80a2b17851496f86c85af5c09a
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino:v0.20.0
        - quay.io/community-operator-pipeline-prod/authorino-operator@sha256:92ad93f27c34f1406fc05a248d04709b0109adf37f2f9424f48fc6a73d9d359e
        - quay.io/kuadrant/authorino-operator:v0.16.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/logging-operator@sha256:d3ce3322f7f25c7b8bfabdd6e9fd32b833239fc39592240e0ccff3a13757fb17
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:f252d6f89a7cdcc520c3af461ca57d0dfabf736b9c38277d77477c4c880e182c
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:97d03647567ddb6f04edea39244bc0814e8f7e557251060d9b241a5fabe04dfc
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:f3497acc3216d820014cce7332e9d0683f4ca6438782ace02210192f787ca097
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:23f2f66e9f6d79d05e2be5a10c394834f093f4677ab06d8935ed33980408e8d3
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kms-controller@sha256:a2c60484a6556949318aa1d730c601b5bb05b1795b945a607209ef3434cce72f
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - k8s.gcr.io/redis:v1
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - quay.io/openshift-community-operators/ripsaw@sha256:ff8d796d5bd6c8cb0e155b3ced631122397758eb472470223e30352eeddba480
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle@sha256:2a28b64f7321df23928c8226b3eb3161cf520d38cc57d875b3cbfc828c164edc
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        - quay.io/community-operator-pipeline-prod/hazelcast-platform-operator@sha256:1209876fc9103eb2d69c34e97e4e2e0b384a9a778ae7e6a53222b8552fbf2f26
        version: 5.13.0
      entries:
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        - registry.connect.redhat.com/paloalto/pcc-operator@sha256:c7200a8c652ee544b0e0bb4d9427eab3a3b2dee6ca61ee88abb8052316382673
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:68cdc37990bc7da8a6bbe3a4908378c8026e5ff980bf8c44c0277f689006fc51
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:97762fcd8bfeec64fe4d3d43a6e60bc705ff2c8773354357e545adf31c8f85c5
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:4fa7a30eb3f679a495aeb96794056ffd003237686b2b60e693c8cabf6bc15bae
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6dce519730a6eb73bc18bef8372cdde65c29af3cb52eb78afc808d8582526fd3
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:9f2c01e88b29154ba7a0f70a1c69a49c86a567808cf057cb5d115159278cf2cf
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:edec37e045212f74d160724ffc7b2c8ccf4e60eb95629aae80360494d78d5433
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:465f0bef1f167b1d16ba9ae5f2d9a1133d11786b6809698a84c5c962387dbf6f
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8be098242f971aefa61a4a24d2c6630ecaa7856ece842166c38887a66bce1bda
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6afbc7ffc98918e3a557b0b3d9b2c3aa80caa47b9a73205fac9e6572c66f4df7
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:eefdd05485167d490f0011d9e5d2bb6112db1daea25940c79e1f55b3af254548
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:05e66583757da895fd23c3af9830ae1ac6a84d107107e5b6bbaece57e4d79c9a
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:02f53d3395b2b98f92a96b466c83ca15c79a75d82451e2ed8ba921375e727c7f
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8abe0994052e2a328e54c9d30c920dc331e9b816a4606234a3023b9ef65ec4ed
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:ef3f85e1227e886a131d3ddec3fb431aef6d3daed46115b288246b9ace0e69e2
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8b9b06ead5e014b920c82034ded615413dca53cf36f804005408726c9bbda66f
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:51db956f1472ae0498c57b9488448c34cf7e24afab6b9997eb3689b001d73277
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:392f9299c75a3ca4f593eaebdfe515145696f12db9f990bb004cfa58dd6bf0ef
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:df4e4dce45f87e684e7e1044a7edb2456160141afb6ba57b9af28088ffbac407
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:fecad1a330421341e6946252f95735ed177a690eab6250be097413ea29ce9473
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:9dcf24fecf3ee0d999601c276ac15e8634c275590f2c42b8a6bcb83984bf1445
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a9950264bfefc536b4e0a0adc2348f92aa41da55dd50a2324ddad0253291b67c
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:4d59e60cc75de82bd3f3a063237cef8a5598d315f6d53e803d37589a94929b0e
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:40d17d4d0632c272779da152f710fb7c2de50acc3791eb0b9fe445a13c0cb102
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:8e321dc1bf74ba331b62d21bff4d8993956ab935cf5be32f01e7f7aa6897bf22
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:16fbdaaf303ab40202d49438737f3307926594a5259081eeeb39595b7ea41f8e
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-zero-trust-workload-identity-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zero-trust-workload-identity-manager.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "ZeroTrustWorkloadIdentityManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/part-of": "zero-trust-workload-identity-manager"
                  },
                  "name": "zerotrustworkloadidentitymanager-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
          createdAt: 2025-06-16T11:15:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: zero-trust-workload-identity-manager
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/zero-trust-workload-identity-manager
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterFederatedTrustDomain
            name: clusterfederatedtrustdomains.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterSPIFFEID
            name: clusterspiffeids.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterStaticEntry
            name: clusterstaticentries.spire.spiffe.io
            version: v1alpha1
          - kind: SpiffeCSIDriver
            name: spiffecsidrivers.operator.openshift.io
            version: v1alpha1
          - kind: SpireAgent
            name: spireagents.operator.openshift.io
            version: v1alpha1
          - kind: SpireOIDCDiscoveryProvider
            name: spireoidcdiscoveryproviders.operator.openshift.io
            version: v1alpha1
          - kind: SpireServer
            name: spireservers.operator.openshift.io
            version: v1alpha1
          - kind: ZeroTrustWorkloadIdentityManager
            name: zerotrustworkloadidentitymanagers.operator.openshift.io
            version: v1alpha1
        description: The Zero Trust Workload Identity Manager is an Openshift Day-2
          Operator designed to automate the setup and management of SPIFFE/SPIRE components
          (like spire-server, spire-agent, spiffe-csi-driver, and oidc-discovery-provider)
          within OpenShift clusters. It enables zero-trust security by dynamically
          issuing and rotating workload identities, enforcing strict identity-based
          authentication across workloads. This manager abstracts complex SPIRE configurations,
          streamlining workload onboarding with secure identities and improving the
          cluster's overall security posture.
        displayName: Zero Trust Workload Identity Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zero-trust-workload-identity-manager
        - ztwim
        - spire
        - spiffe-spire
        - security
        links:
        - name: Documentation
          url: https://github.com/openshift/zero-trust-workload-identity-manager/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-agent-rhel9@sha256:0281d6ab233a685ba0f739629584aacd9254e0b13a697fe8708efa41289e746e
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-operator-bundle@sha256:1974903a3e2b692eedab2602395fdc4a6e2ffbc79ed4e633705c05ce7771e1b7
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-helper-rhel9@sha256:a39c80c6e1495ea166a858c8ed03304c6455420c077d937c99bf8732a6829790
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-controller-manager-rhel9@sha256:67f0e6670195c84515e7d78139a2951e463db11445740fe1d9a449a7f338556c
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-oidc-discovery-provider-rhel9@sha256:5ba7e94b7ea06852d32002b055df556ff2a2fdc9a2e8f793b837c4856d773782
        - registry.access.redhat.com/ubi9@sha256:861e833044a903f689ecfa404424494a7e387ab39cf7949c54843285d13a9774
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:7d170ba9b3de2e88dc0d0a6b0c13630ca3b23eed96ea5e3c725516046d7c7fc7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-csi-driver-rhel9@sha256:a16ac0a4363f7cffada39a2f17aa42f5de3faf78b3688ae5ebf2ecf3a4759cb7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-server-rhel9@sha256:70f97c613feeaadfed99d37aabd62e1cb2247825292ad9be806bc2fd065c9166
        version: 0.1.0
      entries:
      - name: zero-trust-workload-identity-manager.v0.1.0
        version: 0.1.0
      name: tech-preview-v0.1
    defaultChannel: tech-preview-v0.1
    packageName: openshift-zero-trust-workload-identity-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-operator-bundle-rhel8@sha256:0fbbb0b32adadc5746251103d25bba6bd365b2ff60e289e4694061fc92c3ce6e
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        - quay.io/community-operator-pipeline-prod/ack-rds-controller@sha256:80b0d6e33fc94471d05e05f6d370be3a84fcb5c2044486e007d57360eb69d9d7
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-messaging-topology-operator@sha256:4c62fb6d686a20a28437905d4e1a9b93ac05853b4dbf29fcce34421a43268724
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/sap-hana-express-operator@sha256:0fa4f29bd7669690be87a08405bed95d3ead2293a125b28149de2b10b02d2cd6
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/community-operator-pipeline-prod/falcon-operator@sha256:3916adf5083c83838564b1ed039decbadb2efa4ec288699ef0137a66983a3cc4
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.connect.redhat.com/hitachi/hnas-csi-operator-bundle@sha256:79e6ba3db78aa55a9f3839a172843d284d91e9cdaae9cb2c8e7e705f3d03115d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:60b293c23587072b3376f83afee95b0207228c5c7b80483b025adda572c3551d
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:b10c121395f5bc46d61fdeddd35a0c8d9ae4e67a9a77d8ada90e0ea53666323e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle-marketplace@sha256:9d84d46dd1eb74afa18466c73766402b46c2885ef483600aff83065f8065d8d1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/jhipster-online-operator@sha256:5f2436297c026832e5c15241d60090da581daa5502e352ea33d77f22ef3e7a52
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        - quay.io/community-operator-pipeline-prod/argocd-operator@sha256:483854621b95dcefa03d18a970600b271726422ae3362479f73dcfca9085c92d
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        - registry.redhat.io/jboss-eap/eap-operator-bundle@sha256:0db70367917668964e2f4f5efdcef864ccc3f065848023e52caf4abdec728d9d
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:096946dd896b6d9d937eb6529c3a23a84ccb129649c5a0389d0ab83989c28b53
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-load-core-bundle@sha256:f62a7f8eb74d0d5309fede6c9c0201d1f0c3736f93f1aeec7446089d64fce28f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/openshift-community-operators/sigstore-helm-operator@sha256:ff4e548a8a003bd23e5076362bafb2e22f77b8445c5e3585d99cd5ce26fcaf56
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/mtr/mtr-operator-bundle@sha256:190a6fa2a6a9b6b182cdf9cab18814d05c203b54ca18dc533e553fa5dcca779e
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:2808a0397495982b4ea0001ede078803a043d5c9b0285662b08044fe4c11f243
        version: 1.0.2
      deprecation:
        message: |
          The preview-1.0 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:a91cee14f47824ce49759628d06bf4e48276e67dae00b50123d3233d78531720
        version: 1.1.0
      deprecation:
        message: |
          The preview-1.1 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:4477810692429ac498b6ec1b295e6ead0f8fea25e91f6ddf17255d7e349b82e1
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:ff2bb666c2696fed365df55de78141a02e372044647b8031e6d06e7583478af4
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        version: 1.2.2
      deprecation:
        message: |
          The stable-1.2 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:8da62ba1c19c905bc1b87a6233ead475b047a766dc2acb7569149ac5cfe7f0f1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        version: 1.3.3
      deprecation:
        message: |
          The stable-1.3 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:2cf058c819b40512094f84f4d374457455370ec4994d34cee6e39b9d87841f05
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        - quay.io/openshift-community-operators/camel-karavan-operator@sha256:6dc78ecdf6e1936872626c02d41986d6553c1500002d85692c97e86fa77c129d
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:5a5e2f775760aa68c85d0f69b700d23ef2a6be7bb415bd96401d03eb6077f90c
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:79a9191f0cd5d11af004ead87ba9eb5359c67fd658c346db4c3cea3b16b23987
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:54ea11add0a1d988e0c965968f9a65e011978190f76fa2f826f1378474bac1d4
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
          createdAt: "2024-10-11T14:05:47Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8337924489a1747d4bbca6dd6ac42210731cf721b549b5830846b271621e2ac1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:6d0aeffbccaf9df83b42c530bf82707aa7c1cc16f72a5ef30b66fbd562c01ab3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8c05a303be5c536f4051ef291475b578d4d38ca0946fb4b4dd8bca376ef432aa
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:e91c2e5473f3e24c185dd25e9843f333e6d7419d9c5f4d4396669717b5969683
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:c543dbbb66fcac02345e1d3c12333780721bf049dca0f0a27d4c55b15a5d3e14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:a65780c5680a6fbbd0e3d0eed423b14a9c9b3d7d17f801c1eb6eb6b777384ca5
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fd6fb83bbeb9010f7e91f0afc91beac0bf38cb79c2de8f533a2bd07b1b238262
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:936fd68d7659683de7789fcf34d2e5873bd4bd1534c01c2d80501647f7c3c442
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1dd8f0f08ff1323d910fea8adfa13cef64c2fd2f01968e4d9d55f71cf42835d3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:365115407964d182de7e7564438c6fd089e2c0bcc0f94ec48621ca0cc2bf0fe4
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4bea07e101740afc36a1617363a79b1c474efcacd584f6e475a53b3f2395a079
        version: 2.7.2
      entries:
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:3d0409c1a0b17c51d3dbd5aa0371edd597e3c5ba194aad987fe867dd804d1923
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    - currentCSV: mtv-operator.v2.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
          createdAt: "2025-09-09T18:06:51Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:41cc528bf93000095b4eb56aaff9a6dfb124d1711a2f31f0745163ff934ac96e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:0c8ebfdd52053dbcc221a3df7d603163a821dd9e0ac6f54977e081adc6cb4504
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a299acb18821691cddba52904140cef8bf12632fbf2e752b33118a574a5ff394
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:ad3b548405e8c4c075ef7d4c0eb0d0430684066ed7eab4563a1f8378eaf3d939
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4b37b1aecbbf087257be5ba2bb559f2de005aacb4cd371192cdde1a9891739cf
        - registry.redhat.io/migration-toolkit-virtualization/mtv-vsphere-xcopy-volume-populator-rhel9@sha256:c0a4a232ff8adfb7d57e0761251ad0d79a0f02d27c4bd813f62f6b943deb6585
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f90f4eaaaa02b289704b19dd171282f3d11f370ed6cea992e55964729ad5116e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:258efbf073f78f8b5b48fdda22ef055e2315046cfecc9644da51959289c836d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d57b39674260f3bb1ae4c164022c2c7be3cc9645f1c46a83e43f40965633451a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:aeaa9ffcb0ffbb0f69b2712a3031f7e2dfc0824273b5c9c0f6349bc701a0e89e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:e438c5fccc957ad26ecfc3633561342d3b6f2a6d5a4c7839109564008de62222
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:83a523392a9c715972bba7044c7c3d3944594cc89778449e1f59fbd6a2659baf
        version: 2.9.4
      entries:
      - name: mtv-operator.v2.9.4
        version: 2.9.4
      - name: mtv-operator.v2.9.3
        version: 2.9.3
      - name: mtv-operator.v2.9.2
        version: 2.9.2
      - name: mtv-operator.v2.9.1
        version: 2.9.1
      - name: mtv-operator.v2.9.0
        version: 2.9.0
      name: release-v2.9
    defaultChannel: release-v2.9
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:0ee0bd2449f183277af845e312257b4a1ced4de865831b15e1758a397d8d29eb
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-bundle@sha256:47a7bafff2715fbef6130ca0b84560b6c2358275821b72ae54ad46020e02bee5
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:f182078b0077b6df6d47ddcab0204e963c1475486bd61b8eccb73b46eb8b2995
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c8b3505ed491c1bf6a913f161715c8ac483d8d9adf1fd29916ea4edad987cd67
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:af5a13351f960c0938c68604b772f24743c0e997701838a7784c53406471e72a
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:0d9f5293d69eebd91fffd83ae371141e640cab8f9e2d559a83714ce0f94fee08
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:d6675617c19b4ac7c1913078eaf9ffbe3d03761b570dffa002d9de11f337f19f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:2a35a0ef852cb4d76e963df50449712a06ff13391c30675aaf931e21cb561b8a
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:00f7ce0148988e93679d4b53b45ae8d2ad54f308b908940e7b28504318f718c6
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:c352f5633beef2eaaa324dda66b9b484e5ea79efbbc3b781079f2e9d10c880a9
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        - quay.io/community-operator-pipeline-prod/klusterlet@sha256:9efef9a034c9f12505a0d81bc9697c4a3ab7e3d05d1e7aaa4180490d70ddf6a2
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:30264b9f49912cc0e6bed59001869e3168caace20b292c9dad4cc2d4947d289d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:d7cc6d2575eb252fb3e2468106e801f192da7844b0b153dc3c6a12cf7cc0a8fe
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:60e99bb91309e448c652898dac070a6d920fda164f27879d94632c88a9e61ffa
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hive-operator:1.2.4953-c248a51
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:be43412a2e4bad2cd9d15e4fbabe0292585c5bbbd65da12d519f30f17d2fe594
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:822fd8e65c241cebedd835bfebf67645199561cfdd6148d5afffb8b35e499ae3
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        - quay.io/openshift-community-operators/hive-operator@sha256:9a3cb682afc66ad662045b3d8ff214203cc00be60f66f74564820f6dcf99cd69
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        - registry.connect.redhat.com/fortanix/ccm-node-agent-dcap-operator-bundle@sha256:545724c273dfd6a1fa923d468b2d054f39ec3186d01586a8b3fe494c0f22f8dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/ndmspc-operator@sha256:b1f47ce0cd6930d902dc3538a1a87140dbe9257dd50773a518fc2350185f90cc
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patch-operator@sha256:437f5b7c661d3e9bc2c8ac13b0c8f3565f685b9e5ef52791312bc64d55e2e476
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:21f4648e0e71b33ef24b4921ab34b9e2b5e90eac25ade735d51331ac4f7c4cf3
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:2ed6bfdf37e36c63cde3633443553dee410201200ad351e6d91a5651f8e20ecc
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:11edce56874ff1ee666752ec5fa521f3f6ae78c52cf1f289ceb97a86e6c083aa
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:5e4c319ce871a0dcbab7fff23325e6233c9dbcb2592fdeb1a9ba7c57ac2a63b2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:d86bb21fdeec35b1fc9789023a558750da76e036e0790dcfce97dcd137dcb6fb
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:7d94521a5cbdd34f8307653a3e1b596ec86415b2165fabdbabc8ed8bed35473d
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:3d5ad015a953708cc53ac2ea92b1b1d99c870a801b26e790d41458adc0a958e5
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:fea5b977249013181616f60b95c5e6d224cd0f248717171d946ff0e4d6eed6c0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-memorydb-controller@sha256:7f2e65d817479fcc5658cac745cef325b2bfd6256eeed8b819f1713afd59fe68
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.connect.redhat.com/cpopen/ibm-licensing-operator@sha256:a9f68d204b6acb50ac5023e5f2314ad0d06e06483f87290381d1e92476a79579
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator-certified.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-operator-bundle@sha256:4512f786fed0d9b2285b5a2827ec3b597452f90c217e27398f2f0f552377d844
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - quay.io/helmoperators/cockroachdb:v5.0.4
        - quay.io/openshift-community-operators/cockroachdb@sha256:f42337e7b85a46d83c94694638e2312e10ca16a03542399a65ba783c94a32b63
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        - quay.io/openshift-community-operators/cockroachdb@sha256:d3016b1507515fc7712f9c47fd9082baf9ccb070aaab58ed0ef6e5abdedde8ba
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/netobserv-operator@sha256:73af4040707ce79baa3115deed1e9148436b248ef3570f1175a6d31fdf9596c4
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:394d122efcf0da16bb68bfef1292395cda1bdb21909ae55f8fcc54a16814f1cf
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:62f93c5989d1d8217a3bac076addac05c0906a1a1f1eeaacb87bcf256812fdf9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:ea8c060444aa8e89dfbd90d61d6ec3900ab4d6634f6aa8067face1e6974b9065
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        - registry.connect.redhat.com/loadcore/keysight-wap-notifications-bundle@sha256:9c2ebb98476bf3374b71a838c394017e7d31aaa045bd52b3a66924b6f6b7b067
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-wap-migration-bundle@sha256:3deff8b48ab1a6f0e5fd0f8315bc1e191da55525cb3cf0f315f2f7afdb47a6c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - registry.connect.redhat.com/noiro/aci-operator-bundle@sha256:80e25268ab5aa6ea418deae5ccc197fbffe2b9cdf4f27894f5beeb9315f6b123
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:48b37460dd7245cf661009267c3353962f0f195efd6a3007c9cc70b51b91a9e3
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:9949a581df3960a770117b6cf6b9fee1c63ea12ca09ac2d677f1cbe0a83fbd93
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - quay.io/akaris/sosreport-operator:0.0.2
        - quay.io/openshift-community-operators/sosreport-operator@sha256:f0d28a3e2ac2e1cbd0c307797b728ad89ac33375135d72054a7b765e274e4bea
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/node-maintenance-operator@sha256:80dfaab6f7a5c34cebb45c874ce134bf60e3697e14a95b69f1972dd98ecc172a
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=4.11.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:75eaf7b8b71e1a15e002f6cc5430e5e50039699d7b52c3150f5f71b1595f00c5
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-operator-bundle@sha256:41ece1ea773fd610cf8a57483a4d2846e0bb3ce81db701021a3e0dd91e13b7c8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:f7159e880ec55da49ac774db0c454800055e7cd3f37ba94fbcb629d08aac1146
        version: 4.18.0-202509100344
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-operator-bundle@sha256:4f8908ba97517826fda14cab3624ed42a6348fb96f0be4abf9cd2a7b68f7d3e9
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:20a4f2c56cb186fff4e8d31021d84b690512cc65c3c9629ffe623a0213ca9b0a
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        - quay.io/community-operator-pipeline-prod/mariadb-operator@sha256:3cbbff7ae75fb4f4629c7b8b58ca85ce67b23d5da558241301266d010e8b1188
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:3b683338f5849b887e1bc81ea078a9b799d272f824abd4247ff87e70ca7acbf2
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle@sha256:333a6f476d03adcbfaa4ab38802859719a7dd8857783f8487c99b008d06815c5
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acm-controller@sha256:4d63542b39f41c458dd1eec28e44edce3a919f4dd9137ff64a9452cac1a1f954
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - docker.io/absaoss/k8gb:v0.11.4
        - quay.io/openshift-community-operators/k8gb@sha256:9743f957b00e5ad344fed0c062d616d8222990139ac4afbf14871fd9813f0ef0
        - absaoss/k8s_crd:v0.0.11
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/ucs-ci-solutions-operator@sha256:e284253263183eb2cd313f6fb009bb830b9a79a4d4e2481468821c30a9cdce03
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-operator-bundle@sha256:67723df19ab7d7871a6b639da3975c230d95dd250fa2995a9f44bdbb4b141454
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:439aeb806c3264948a15e325f7b1e71037099a085c8d971646188ace77fcf5d3
        version: 4.18.0-202509100344
      entries:
      - name: secrets-store-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: secrets-store-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: secrets-store-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: secrets-store-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: secrets-store-csi-driver-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: secrets-store-csi-driver-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: secrets-store-csi-driver-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: secrets-store-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: secrets-store-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: secrets-store-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: secrets-store-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: secrets-store-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: secrets-store-csi-driver-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: secrets-store-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: secrets-store-csi-driver-operator.v4.18.0-202502280802
        version: 4.18.0-202502280802
      - name: secrets-store-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: secrets-store-csi-driver-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-09T14:17:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:c5bd4f10120088f285a3b65e85bba825f9f55964758901f3b5827ac31516675b
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:26935de5f21a7865aba484599136df6a251c009e8fa9781e1040da3b19e7dd33
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:e05074d46e062268461d2689f7048c1a5dc4ed89a039d93e382f5ebc0726a269
        version: 4.17.13-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-multicluster-orchestrator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-multicluster-orchestrator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-multicluster-orchestrator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-multicluster-orchestrator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-multicluster-orchestrator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-multicluster-orchestrator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-multicluster-orchestrator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-multicluster-orchestrator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-multicluster-orchestrator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-multicluster-orchestrator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-multicluster-orchestrator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-multicluster-orchestrator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-multicluster-orchestrator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-multicluster-orchestrator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-11T15:26:27Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:4b78c8df4f4aa4be805a442b059b5e2822464bb029826f2f0090cf8b6b747768
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:47c754d0e2782dd8d885758e68803173c32f1fe6319ac09c786dd24e5124df25
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:957805157e7690deedaae63856c9ce71e5c299610bca401524397a40e2820537
        version: 4.18.11-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-multicluster-orchestrator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-multicluster-orchestrator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-multicluster-orchestrator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-multicluster-orchestrator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-multicluster-orchestrator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-multicluster-orchestrator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-multicluster-orchestrator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-multicluster-orchestrator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-multicluster-orchestrator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-multicluster-orchestrator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-multicluster-orchestrator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1a156d67fd5beb9ffddecab724490acb7143f8141ff3b392c6f519e76c3ad374
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:645fcb18684884aef8ca4d5299f0d053d496c3a5d3f37cadd66c51d38514a680
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:5a73ce3a8cc8422aeda37fd9ef97f5aecdbf8c69fb6ff488df26342e989b5c57
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1dfe9180a9fca4f773c1282dfd0006daa0bb48dfbac87e352661e051107f9033
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-websocket-service-bundle@sha256:da100dd9c5b542ba46de7166e783e28a0f7e675f9ac896f211405e5a9e2082ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2cc940711539edd93a23355dcb06365821d140346e1df248386a77667c0f47ea
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2418015d00846dd0d7a8aca11927f1e89b4d8d525e6ae936360e3e3b3bd9e22f
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - registry.connect.redhat.com/a5f1f0cc7ac3db90370970f9b/tawon@sha256:3691c0948c8eb266ada348de504622227af6516f3b3d331eb09fd258dd0609a5
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    defaultChannel: alpha
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
          createdAt: "2025-09-19T13:48:41Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dapr-kubernetes-operator@sha256:8f2087df1987437f1c5e62aa9453fea220a41012077fec67b6ba8368ec267421
        - docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
        version: 0.0.11
      entries:
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-apigatewayv2-controller@sha256:1cf12140fcd67b0f3e8c5791a91c5b705bb3af8b2e03ea4ba2d4ce6f9f7a4809
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - registry.connect.redhat.com/mongodb/enterprise-operator-bundle@sha256:23c3c8387eb9377b52327d7193f59eeea5831f4050248b3cbb54cbcd9437fc71
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "rawDeploymentServiceConfig": "Headed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.34.0
          createdAt: "2025-09-03T09:59:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Argo
          Workflows\n* Model Mesh - ModelMesh Serving is the Controller for managing
          ModelMesh, a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks\n* Feast - Feature
          Store is the fastest path to manage existing infrastructure to productionize
          analytic data for model training and online inference.\n* Llama Stack –
          unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opendatahub-operator:2.34.0
        - quay.io/opendatahub/opendatahub-operator:v2.34.0
        version: 2.34.0
      entries:
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        - quay.io/community-operator-pipeline-prod/opendatahub-operator@sha256:d3e3d5ae02b834522481fa9046bf2ad3ea06d706c5cb06a32d32e61e229e4a6c
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:2992841660e8b1664fa2ce06da894aa828d4ea1be3eb5b27476efdd73731486a
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:bf85a9da8a100a90deb3745497bc1505852910ef23dd12c624fa17de51bd8377
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-licensing-bundle@sha256:2d3abd0263183bc92718316581675b92073f37c3786784fae24b3587a9441def
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:329d6eaeff787bcde5db0ba4f9403eee533a601b45dc2d5d349b318705452591
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:2cac5fdb55664c08928a7d924eacb18cc4c03c609261e3b6d812590e4d166375
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:b35da8aa4e6746952e5cfdbbac72950a48f7ad32a3d762a470ce62118c78643a
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:e5a0cb8af7879bf13d940df278ccdb8a5c400e5d7108f23b4abef94c964e40d6
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:91aae0500cd3cfc9a4a536241feb4fe50288e2a68f69ab5c602c02169e5f5b2e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:40c7d896355dfe67ff3616f1e1df86548bfcd0524f7e56d4d72ab4f5f1739f6a
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:3f69a124f9c31ebc0c78f249872a69b1289622bf4f5aad4c739f6e8caf9cb85b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:33a2e21a069ff24af4bf0ffc51e83ddf398df5d9735d0123120768e9dc2c27f9
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:826354c9c06a06f998e522a88d6a81239e003e1d6a2633e9b7b27b8a61d5fe38
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-operator-bundle@sha256:a59435a22c2f470b149ff721240f14144d2e56200b141e59a4f1011267a6c925
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:1346dcee82f2120bd2af22be9f041e057bdc32f9ad1c3ef717d599ba6519faeb
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:44a5c4cc3086e4eac42dc78a72b7492d13df0098017288278d9006e767a480a1
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:b9dbbce235b7a4f45f42184cf0b908e91e17094e971bc29a8710f35127a6b06f
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:026418bae8b2d425d45b2b3481f65c241d89e25f455e3fbfcf2f79536d978bca
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        - registry.connect.redhat.com/nec/webotx-operator-bundle@sha256:7e655933a7fab3ac3aa88eb5e4ca9ac3a71b98875ef1ef5bc49e87284f160ff8
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/federatorai-operator-bundle@sha256:d0ac978d749654ddf4cc20823805ba7d33ee62e9a7ec98fdd53d9926b697c347
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/community-operator-pipeline-prod/cert-utils-operator@sha256:92a0c0b893b56330593e0a255b8df61031e92a61d172aecc58be5805d847798a
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/community-operator-pipeline-prod/gitwebhook-operator@sha256:2d33e4f4f4b607ee2006481efc7ad9341049085ef988777d23e586f7406a5109
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - quay.io/bpfman/bpfman-operator:v0.5.6
        - quay.io/community-operator-pipeline-prod/bpfman-operator@sha256:6b3603c7c1917c67b5aa57202d1d2fdc0d9ac6b3a9373afa7172f322b3b3fa9b
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:74775c7851ac98ba2903435c71b9d02cc504ca8ee98b65ad7282d671392756a6
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.connect.redhat.com/cpopen/ibm-elasticsearch-operator@sha256:0bedcf95b31e28d41b124780f3b97c68a87db5626ecf1d5e81bb06b8b3fcd91a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle-marketplace@sha256:19c9d44cb9a0c7f52501fcb589750f3e85e270d066423bcadac593dd79e48ff4
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        - quay.io/openshift-community-operators/github-arc-operator@sha256:d8d97d19cea466c4c240097b79bfb7de90312d937ad4db63f9e90901686e54da
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cognitoidentityprovider-controller@sha256:093670fcaee6ef20e471c0df45c9ea7f40e5f0c43208c9447623324fdb39dbe0
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/self-node-remediation@sha256:549585ca18d331cfbd5911fa3e79912428a0be3308588be2438ece3247746bd8
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |2

          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler,
          **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: kaliu@redhat.com
          name: KaiYiLiu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-operator-bundle@sha256:5ebbc06088e99b937327cd457324461cafdda9649b2dd2db1bcb892be8556eed
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:141b069ff9925390607315b0cf675003abbdac9555b29700e3b999fadd6c89f2
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:971174692a1aae68ac07ffa132254eb3c018a52507eb32621b6dc286fd25ca0d
        version: 0.5.0
      entries:
      - name: power-monitoring-operator.v0.5.0
        version: 0.5.0
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.2-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
          createdAt: "2025-09-11T18:04:40Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.2'
          olm.substitutesFor: cryostat-operator.v4.0.2
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:b6a6e401042daf1780ae6be19aedc551ed176d1a13bf1f0f48359552c638a156
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:8fd30d86f5468227899560071a28e92197c5790ea820e918c47559f0695cfb8a
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:d2788c8732fec62c7a27e7bd1497fea1ab3ff605e7e913bbe17d8ef97567329b
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:550de94fb20462c5cc5c52166820319544e774b8d45fc7beed7a095f5c1300aa
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:bbee66a44c2e3a5525c2fa81db72e8528bccf40583d0bd665841644473308310
        - registry.redhat.io/ubi9/nginx-124@sha256:126903359331eb9f2c87950f952442db456750190956b9301d00acfcadbc19ed
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:67e313e50c6ef100d6938dc1a734728661f328ab1f9757667c7a06de11689967
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:d9720fefa3e34d05a5536fe1a66e23b30fe0845ae2b96b9c9be4ee705be9633b
        - registry.redhat.io/cryostat/cryostat-operator-bundle@sha256:455ecde240a152987fcfe0013abee992014510de9e89c014559a5d0665a56512
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:9ba156d802c9c3eafc948afef88e0eaca6b5e73bf32ba80ea7b0651ba7ce923d
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:2a5e1d2d39a7d9d2bc74316110e1a65e9e18f3a9fb7f0a498c7457dd9e6684ee
        version: 4.0.2+4
      entries:
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/openshift-community-operators/pulp-operator@sha256:1926e96b185443ff0c7a2861cefc2e3aded0c0493e0d7f7d78e45447e06750d4
        - quay.io/pulp/pulp-minimal:stable
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:1e75649538f13cf3bb2e6131160f038a70689d6f92a110bdcec348e865040c4a
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    - currentCSV: pulp-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 1
                  },
                  "custom_pulp_settings": "settings",
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "ingress_type": "nodeport",
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "example-pulp",
                  "postgres_configuration_secret": "example-pulp-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "example-pulp"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.1.0
          createdAt: "2025-09-18T18:00:01Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          metadata.annotations.support: Community
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://pulpproject.org/](https://pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://pulpproject.org/pulp-oci-images/docs/admin/reference/available-images/single-process-images/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://pulpproject.org/pulp_operator/](https://pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:92a3e95cb3b1e50100933ea08cb44074fc505f6c0e7fa24d028ec46fa9e95590
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.1.0
        - quay.io/pulp/pulp-web:stable
        version: 1.1.0
      entries:
      - name: pulp-operator.v1.1.0
        version: 1.1.0
      - name: pulp-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
          createdAt: "2024-11-27T00:00:00Z"
          description: The deployment validation operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        - registry.redhat.io/dvo/deployment-validation-operator-bundle@sha256:c04bae05ee0c24fe5bf94257b0dd2cb020f5c9b5f173ceffffdd041b6dff4ee9
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.11
        version: 0.7.11
      - name: deployment-validation-operator.0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.0.7.5
        version: 0.7.5
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CephCSI Community
      provider-url: ""
    name: cephcsi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cephcsi-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:12:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:b023d95637934b16ec8342994edd4e3198579517b68817ce6b4cc14958746b9b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:10cfe59ee2d625ed67f3804ba7e471f10fa1e12bc4f96de5e2f453e5cdefcf89
        version: 4.17.13-rhodf
      entries:
      - name: cephcsi-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: cephcsi-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: cephcsi-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: cephcsi-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: cephcsi-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      name: stable-4.17
    - currentCSV: cephcsi-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:20:52Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:903e47fa88119b5507a0d773d58318236bbce53201b5e6a18a16311f704b1120
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:562703a8b197f03b29d0564df329ecfda192a94b7818c6fd5f919329a5cc9bd7
        version: 4.18.11-rhodf
      entries:
      - name: cephcsi-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: cephcsi-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: cephcsi-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: cephcsi-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: cephcsi-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: cephcsi-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: cephcsi-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: cephcsi-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: cephcsi-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: cephcsi-operator
    provider:
      name: CephCSI Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        - registry.connect.redhat.com/hazelcast/hazelcast-platform-operator-bundle@sha256:0aabd3e7f5f993c1288450678d55e8f7b36c8b30894e5d844521a474aa6b2714
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudtrail-controller@sha256:20fe3c3ba7688906de40d7fd36c00552be280595708935ff800ad5228829f781
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:fbf847c41e1736fb6efab7a6acdc03acc562fa7e667306773880b277883bc4e5
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-operator-bundle@sha256:7c808f6e4ea469d88703300bdcc6bdce160634833cbd10bd110dc69d05be588c
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
        version: 4.18.0-202509100344
      entries:
      - name: clusterresourceoverride-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: clusterresourceoverride-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: clusterresourceoverride-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: clusterresourceoverride-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: clusterresourceoverride-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: clusterresourceoverride-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: clusterresourceoverride-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: clusterresourceoverride-operator.v4.18.0-202505070803
        version: 4.18.0-202505070803
      - name: clusterresourceoverride-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: clusterresourceoverride-operator.v4.18.0-202504251302
        version: 4.18.0-202504251302
      - name: clusterresourceoverride-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: clusterresourceoverride-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: clusterresourceoverride-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: clusterresourceoverride-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: clusterresourceoverride-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: clusterresourceoverride-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: clusterresourceoverride-operator.v4.18.0-202501231202
        version: 4.18.0-202501231202
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator-metadata@sha256:a2f0513cd8b96d71855464384cad7f1cd8c39f8d445be1ff1c1b91f22008d16d
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-operator-bundle@sha256:c2229c2cfad9bbbca9179cb6617bce0033f5be40731641a69b41f9a90e034c8b
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/keycloak-operator@sha256:a4da3bcbb4e99d60b8217d9e62742eeb83eef163dfefb2f7efda131f312e46e3
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        - quay.io/openshift-community-operators/keycloak-operator@sha256:65fe00f310289b70538a6dde897ef663731d0da44a7bb1d0f55a3001a50e53fa
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/keycloak-operator@sha256:70bbd44f6ecf86aa8e9710175a96dd3c1d813aec40667a09800241ce30afd93f
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/smart-gateway-operator-bundle@sha256:f12b2f1dd34c73b9c8a5ad461e771c0ce56c9ea4eb88f03225be1462a42c5a78
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:0cda4be8f74da7524a72c961c8762927c66e145d82e59af92073a54179bc9936
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:4bab095c5dd6f8c95a17a8ff2bf70ff3f53faea48b56eea8baa9cf9a42a5227a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator-bundle@sha256:8d130a5f2407931dbc73e45a53610b95ceb04cb507b8f6519f630cada1c0da45
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:9009198860c7a1e0de4f429261d442b75b5b2893a943bb77914b14d643db8829
        - registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
          createdAt: "2025-06-19T11:07:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-zookeeper@sha256:c8487e18318a92a64e84a5ab29f844a0973f3b8b2aa291723014731471459176
        - docker.io/confluentinc/cp-server@sha256:072719eef75ee61294027ba77f7b9cd821437158761495a1644b5687af9305a8
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:7a271565db9c500f8c5755de397e8093251673d2fc6b8a10974f927f382b4704
        - confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2520a514b848b53338c411f0e0dfa68491596761f8bbbb090d652f71eda010af
        - docker.io/confluentinc/cp-kafka-rest@sha256:9c19e7900669e555810d26676393737a7d573b39c31e284c0ad78807bfe05ce3
        - docker.io/confluentinc/cp-server-connect@sha256:05e77fe8e19747521e1f1d5630bbf384b0fe0ec9efe8dc798de6e4a3eb0bfa2d
        - docker.io/confluentinc/cp-schema-registry@sha256:5b6bf5b971dd8fe2c821e227391858f58f6ae2128f1ae1faa988c3ea59b86801
        - docker.io/confluentinc/confluent-init-container@sha256:550a6b92fc5aca81ad8e0d3d47c0063af9da877e45c1ecda1a9cc44a70e5d87a
        - docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df1f6a99ecfe8d8acc0af222848aad13349f06071fc2bf1360d6d20f99769394
        version: 2.10.2
      entries:
      - name: confluent-for-kubernetes.v2.10.2
        version: 2.10.2
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.2",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
          createdAt: "2025-06-19T11:14:20Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:b7cde5736b5275abafafbb81f4c917890abfe8f400f436c3a1f712cf200832f8
        - docker.io/confluentinc/cp-schema-registry@sha256:4f6cabeea11ae532cb47905eadaa3192d7113ea6d0a0da310cdfd5eb7142fc45
        - docker.io/confluentinc/confluent-init-container@sha256:37161a90b89dfad625bf61ba7d299707d7b2a96271e9667044c3eaba3ffc5f66
        - docker.io/confluentinc/cp-server@sha256:dfed43a1618de1fea6d202921ca4cc2168f7455c3d37fe046f3e64febf042a87
        - docker.io/confluentinc/cp-zookeeper@sha256:f1e6d3d698a6e2757e3c225b5c6f8f007251838226c46f652a4cbb4314b50277
        - confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:2baa0cb5df8017d9d14c2f0c7e02ce371ba92462d8d1e93a3ca1bc548c20a74e
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:507e1538a64c4232ffda86504594def45cb20bbcb6b13fc6196c1f41e71f93bf
        - docker.io/confluentinc/cp-ksqldb-server@sha256:744f5d844ac223c857116f3994c5be2105898ea3896d53026eb62e76c7240c4f
        - docker.io/confluentinc/cp-server-connect@sha256:73822ffff10d386b98c9ca2bb295b4d079deb210116c8e4283cdc26ce304526f
        version: 2.11.2
      entries:
      - name: confluent-for-kubernetes.v2.11.2
        version: 2.11.2
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.0",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
          createdAt: "2025-06-13T13:02:11Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:ba49d46654770994c302d586a609afdfc58a7cc05f204aa090b643f0680a5014
        - docker.io/confluentinc/cp-server-connect@sha256:e4cbc1836402497ea42b2c961c5d6b4420c25b3c6a5f6f71be3e99fcc0f09702
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:a283f2a72b6346febaf711e71dee1e0e3666d7265320f3f69307fc2c0b95d5ef
        - docker.io/confluentinc/cp-schema-registry@sha256:3dcb9652a4c83d4566a7334041c1ac3ec359665a9450a70c5d7b41adaea55c90
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:35f2799f71c308d87e6a7dada00263abb9e08e895b10f10c815e7a9edf19dd49
        - docker.io/confluentinc/confluent-init-container@sha256:82277228368a783ba84367c920a9fa15fad6844723df1e16b07c9dcc829a54ad
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-server@sha256:ffd5961412da42c81c9b20f35e334aa8bdf351dcb452610e84156ed511116ef5
        version: 3.0.0
      entries:
      - name: confluent-for-kubernetes.v3.0.0
        version: 3.0.0
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator-bundle@sha256:21b69b5eb896a37691a8e086f3a51c0f68086a0ff9370fd66667acc6c935a850
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/mta/mta-operator-bundle@sha256:65c0932bfda32b1e62be9edd0f96bb4d6e41e302a2e8080fc6cc794ee5c52276
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-operator-bundle@sha256:f3d585e24d65905b2e9f594dcdb67721aa95057ab7587ae531c3487aa2f94e53
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/mta/mta-operator-bundle@sha256:a13643117c2867351718a872f7f1b2350c67855ca73a727a1dc140754ffe6589
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/mta/mta-operator-bundle@sha256:9eeb43af2bcab84f5261d1575f7c897903a7696dba011d256abffa1fe850eba2
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-operator-bundle@sha256:6aef520983fe01351886e108d32b998b20e74116fc0280c754a914bfad03d0bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ed10e35f45a140e21d826bee71b3ab96e67f1b4dc8c2f7f1676380fa51b67fe
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        - registry.redhat.io/mta/mta-operator-bundle@sha256:68953a2a721309fd103e09f3171d0c9d50020413631b575f1a44602a5d2ac855
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ec977b0c9c26551db110dcf2d5069647b1a5495f869ff48a0319cb496b84cdc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - registry.connect.redhat.com/eginnovations/universal-agent-operator@sha256:65f851074e0e1db57a62b559c0f10f9d30543011727aec3b832f9a35468139e9
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - registry.connect.redhat.com/harness/gitops-operator@sha256:f97cdc2245b7133a04235c55540d3a87165e080b1d5b9572a663731f6ccbe597
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        - quay.io/openshift-community-operators/tf-controller@sha256:cdaeb470d9b331a3fa08167e5c904befb00be983e1354651670e21e13560bf9b
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/openshift-community-operators/sosivio@sha256:110d7f4fb993d65ff497eca57c42206faab07dc5f5b1f43b1b682387e7cd3a10
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        - quay.io/openshift-community-operators/l5-operator@sha256:e4cb13801a80753c99a47f009ca0a29416b3e04afe5314d07d97c2464a6f0d75
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:712d31214423b326ab8f107435c85431ce8d99c2e52fd8d615005a4274e0ff98
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/operator@sha256:2cb5ec186af3264cba42333245616dd01f0f9868a849129b1a31afbc51ee1410
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-applicationautoscaling-controller@sha256:4f1b92eac30d72a66d8fe8c2c44339c7a5fd6af65d48c0d8697459c7af290741
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.18.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
          - description: L2ServiceStatus is the Schema for the l2servicestatus API
            kind: ServiceL2Status
            name: servicel2statuses.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-metallb-operator-bundle@sha256:c80c3467cf52fc4b2a55e444fba9fdc8cf3b1efedad5d1276a889d07aed44e68
        - registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:6f23dd225f2bccfe39d05428a51de990a4a9338c49da079384eb30fb470c04f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509100344
      entries:
      - name: metallb-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: metallb-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: metallb-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: metallb-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: metallb-operator.v4.18.0-202506251003
        version: 4.18.0-202506251003
      - name: metallb-operator.v4.18.0-202506070804
        version: 4.18.0-202506070804
      - name: metallb-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: metallb-operator.v4.18.0-202505081435
        version: 4.18.0-202505081435
      - name: metallb-operator.v4.18.0-202504291304
        version: 4.18.0-202504291304
      - name: metallb-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: metallb-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: metallb-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: metallb-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: metallb-operator.v4.18.0-202503181802
        version: 4.18.0-202503181802
      - name: metallb-operator.v4.18.0-202503110933
        version: 4.18.0-202503110933
      - name: metallb-operator.v4.18.0-202502261233
        version: 4.18.0-202502261233
      - name: metallb-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: metallb-operator.v4.18.0-202502041632
        version: 4.18.0-202502041632
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:latest"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:latest
          createdAt: "2025-01-03T19:50:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The eBPF Manager Operator is a Kubernetes Operator for deploying
          [bpfman](https://bpfman.netlify.app/), a system daemon\nfor managing eBPF
          programs. It deploys bpfman itself along with CRDs to make deploying\neBPF
          programs in Kubernetes much easier.\n\n## Quick Start\n\nTo get bpfman up
          and running quickly simply click 'install' to deploy the bpfman-operator
          in the bpfman namespace via operator-hub.\n## Configuration\n\nThe `bpfman-config`
          configmap is automatically created in the `bpfman` namespace and used to
          configure the bpfman deployment.\n\nTo edit the config simply run\n\n```bash\nkubectl
          edit cm bpfman-config\n```\n\nThe following fields are adjustable\n\n- `bpfman.agent.image`:
          The image used for the bpfman-agent`\n- `bpfman.image`: The image used for
          bpfman`\n- `bpfman.log.level`: the log level for bpfman, currently supports
          `debug`, `info`, `warn`, `error`, and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`:
          the log level for the bpfman-agent currently supports `info`, `debug`, and
          `trace` \n\nThe bpfman operator deploys eBPF programs via CRDs. The following
          CRDs are currently available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n
          - KprobeProgram\n- TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n-
          XdpProgram\n - BpfNsApplication\n- TcProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n\n ## More information\n\nPlease checkout the [bpfman community
          website](https://bpfman.io/) for more information."
        displayName: eBPF Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - registry.redhat.io/bpfman/bpfman-operator-bundle@sha256:b6d43f1fec6ba7336c23d1fbceab15ebb2ed9e9e2eebb08c78d7541ee47b4204
        - registry.redhat.io/bpfman/bpfman-rhel9-operator@sha256:c3d878d77091bed03d99f72943d9b78e210e2f6194ea4b5cd2aa23f8d48f3f61
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      name: latest
    defaultChannel: latest
    packageName: bpfman-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - quay.io/community-operator-pipeline-prod/gitlab-runner-operator@sha256:3e0314a2695c071fe0a34f3dc7684b43b5a31b6705f42982b4dbaacabe360d2b
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Agents",
                "metadata": {
                  "name": "loadcore-agents-sample",
                  "namespace": "keysight-loadcore-agents"
                },
                "spec": {
                  "agents": [
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent1",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "replicas": 2,
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    },
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent2",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    }
                  ],
                  "image": {
                    "registry": "prod.harbor.keysight.digital",
                    "repository": "loadcore/loadcore-agent",
                    "tag": "UBI8-0.1-7d990d7254"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "regcred"
                    }
                  ],
                  "networks": [
                    {
                      "config": "{ \"cniVersion\": \"0.3.1\", \"type\": \"ipvlan\", \"master\": \"ens192\", \"mode\": \"l2\", \"ipam\": { \"type\": \"whereabouts\", \"range\": \"192.168.1.0/24\", \"routes\": [ { \"dst\": \"0.0.0.0/0\" } ] } }",
                      "name": "loadcore-agent-network"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:53:55Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:f1d1e9f2aca11505407785bc33905744b667bc8aea49df73fcc50db379c0b375
        - registry.connect.redhat.com/loadcore/keysight-loadcore-agents-bundle@sha256:f16fcea424cb7b1eb647f3e9d3c08bd93c8a1b8bb7eab71a4e756da72a7a56a0
        version: 0.1.21
      entries:
      - name: keysight-loadcore-agents.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:a58c9c8454ac337c5a359dfe5b7c6b06983b7050ad62a0d11054f06fc17e92c1
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.89.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.89.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:f6c37a0808c137f226b6afb6cac1e2ce05184d6a01b9234969ce63056571e991
        version: 42.89.0
      entries:
      - name: t8c-operator.v42.89.0
        version: 42.89.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - quay.io/openshift-community-operators/odoo-operator@sha256:616e1067ebf8deb66636d57cae2ac1f9981b83419d8e7f023bc7df1c1b745536
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/docling-operator@sha256:9a3717ee195572b29d1ee31b96bb95561f5521b032dbaa09a000413b227b5f12
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-operator-bundle@sha256:2bab9fa63b128a3ba0a6031f70f0342f95d17dd53a3ec2930b19940b5b2e6f47
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - registry.connect.redhat.com/eweb/cloudnfv@sha256:0bb427954fb8df5de197353908a5189922f514e121ae8cdb8242d125b1b3807d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        - quay.io/community-operator-pipeline-prod/lbconfig-operator@sha256:b0bb724a5d9905633962a3e8267e75d371f754bce866a29cba3ac0d0c671ebff
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        - quay.io/community-operator-pipeline-prod/apicurio-registry-3@sha256:e207a4f7f53486b78f95ad958f7e0053a8a8561c3dcb1463a7176f27eb9ea082
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
          createdAt: "2025-09-19T21:59:41Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatch-controller@sha256:254df5f0e6821a65277bc3d029e7af87b10800b192f9724997996224525a9943
        version: 1.2.1
      entries:
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        - registry.connect.redhat.com/scaleops-sh/scaleops-operator-bundle@sha256:efd483d9c826e25b0d1cd79deff4263772260c72da79bad149ef7985cc00a44b
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:54bb383ac6c5f57ae77dcd96cdc33cea5c4c1aff9f9d356c1721591aa820e82d
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:948ad5eb47b7809ba5dd78cb6bb30b1bc64163617f229dc33cec251c1d5bb20e
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:12d6ae3f139daf33dbf104499c195d9915470b957f401b5f622cfde92350a6c1
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:965ec7875eccc6a5fc56c3b8f0da668d826d34580117ba1438a607a9d0d13b73
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:f8a8cfcebba4b17fa6cfa99b3ef430af272c5237476dba4665803a90adfd5d2a
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:610a5e66cf1b6c7f5451c9a0f5e0a6cc9435f2548335fbd50f44b8656adadd99
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:bc0d8b62d5814407e08817ece44626400cc3d9d29b14aa35dfe711b907fddf21
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.connect.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:15e8cbce994b783fa6dac737ac28f1b713d1c75891dbce4987d2c09ff295376b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:d9930900b0d2f92ae813246fc322290234ad8bbd165b10e36671b53852228718
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:87448055a162e215a4d4982a8494fec6e7cc59fe6704494ddfdcfea076a2100c
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:a7c6f1c9260c79ebbdfb25ea996516ba1e2bcbedf25401e9e6848b955a2d3d78
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-operator-bundle@sha256:9ab3d5eb70c654270fef9e8e9d4e5b68e1ac3c86bd0e0dcd36cc6f8741be1603
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:36561e5888cd34f8fb83ceff2910141fceaa6ea290bd246125930fe183e2ebc3
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:9e8bcbff46b97f3bc1071146b8db233abd1630016e6fc0efb679df8243c6fd03
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
          createdAt: "2024-07-10 00:59:51"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:518691d157b3e3b5031a2b1e4cc3f415839536079bb8570ae8690ac92d302ba0
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:a64f67f29a4eb8ce426ed842f1a5a5fcd83fd9d91acebb064cf65c89c9d8bfb6
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        version: 3.74.9
      deprecation:
        message: |
          The `latest` channel is no longer supported.  Please switch to the `stable` channel.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:7fd7595e6a61352088f9a3a345be03a6c0b9caa0bbc5ddd8c61ba1d38b2c3b8e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        version: 3.62.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:5187b04f8010a274d6bc017fc82f1b00d886124e520c1b7de8b4da8cd2f37bc0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        version: 3.64.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:d3bf33899ca8b199ca8782978618ca343164e43fd0ae2ac6a0cfbf697efa6428
        version: 3.65.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6536ee795e8aaeab9321ee4820e55933195983a768ba24fab40508d03e50a6e4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        version: 3.66.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:ad88773ab141f45fe24c61e83ec736aab27f652fb82ca2a85fee6f885568c5b9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        version: 3.67.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6a10a4859527946f61f497ff84b80f973cb7be5982d428423836b5aca81a1f5d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:297e7f863106c6b41f882bbd5b691a12c797c24d832473a04e604d4639e4a68e
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        version: 3.69.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:9b35ae6df2c19935783735d213483f3fcecead2f49f873a4f7844ef9fb5c7f1a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        version: 3.70.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e001b9f844a6c298dd74f1ab683f3dbbf0ad66044cec419c373ef4f97d47be73
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        version: 3.71.3
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:878ed77d07d3f40f049a594af44376797305786a2a264dae3d0c50401643bfa5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        version: 3.72.4
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:459a8d91beff91638872d377a8b82c279ad2905ce3f4ad4bfc00f6822418ac57
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        version: 3.73.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        version: 3.74.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:a0d9cd31d953aaf7e84f5a39f405c5aa2ea62e68d6c728a936ca1ffc4dc1b93e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        version: 4.0.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:b26133843687cc50e2c52d6067d8b3369b5cb28dfbfd15dd4b6062096673195a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        version: 4.1.6
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:08aebe6e8275d3314b6756d522ec568ccdfb9c52df98c9822f187f6c5a3af19c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        version: 4.2.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:30f9384244819da013358325cec364bb210b3b33b4cf9680698531fe50562aac
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        version: 4.3.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:123c0108bd7a190dd2accfd21786600f787147f040e57fb5f840861e2caa441d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        version: 4.4.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:41351e88688d23a443ad4f141caf0433b2c2789e6a1a85b4f5b36abeb5656bb1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        version: 4.5.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6d6b2696a12d49cf3403e338a0e028187473b6f43ef2b0bd72c7701200ac8e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:c7f894ab773d1d389656d13ed88fbf575faddd95c5b8396c6593ab3f993f331e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-bundle-rhm@sha256:5d14e0c44ffb543b7cf59817fe587c8e819542d8edfa261a44e58aea0ecdb79b
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        - registry.connect.redhat.com/randoli/app-insights-operator@sha256:21ed3d9b8b8466665c1cb784d0bc244279d0730f6319625caf46af111c25d333
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
          createdAt: "2025-05-09T05:49:47Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-operator-bundle@sha256:f49b7a2c60e79edaba1a5faabc3983f87f9123b6d84121a438d179ccbb8102e3
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
        version: 1.1.1
      entries:
      - name: multiarch-tuning-operator.v1.1.1
        version: 1.1.1
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.9-r3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
          createdAt: "2025-08-05"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:327a9899adfe89af9aebdcfac651a9eddec9ecf8a345caee1cef1b4a1fb279d5
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:df0493f7edbdedf754c70e4efc256fa4d246a24d6c8882c562e56ef3844278c1
        - registry.redhat.io/apicurio/apicurio-registry-3-operator-bundle@sha256:728354aee3729a1fba233c2aa92cacb21d27efd4869706ddd214a2d3688e33b2
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
        version: 3.0.9-r3
      entries:
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        - quay.io/openshift-community-operators/aws-efs-operator@sha256:6372847df81faddc23c938928157c3ac280e537a4eec9f7edd599480260bc566
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/cluster-impairment-operator@sha256:6fd114eb8f525f670e505373c2999a4e66b53beda988e99b267204d2fe0e1b4f
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/keepalived-operator@sha256:5a46fcb72f02a1f0d9cbc72f44948ef4b09149a4b3fe2e45fd31df90d20b3c56
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-keyspaces-controller@sha256:6e56400162867a4f55a70c340aeeb074726fd3dca13b6da60c395d6f35a48b48
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/automotive-infra@sha256:6d4180ff6899d4de8ba2cd4a06ece234bbd8f0ec1bd0b4b0ae4b7a0c047121da
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:9d21b738b0d357ef064690eae66bd4b3810664bcde57237822f1e90e9fd6ad03
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - quay.io/openshift-community-operators/reportportal-operator@sha256:67a7ffaf6a303b92541170f266d7b2871f5c2fe73f6b6c9a2b7b33dbc4555fa5
        - quay.io/waynesun09/rp5-operator:v0.0.8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:8197e4a96ac8949d43cda62ac2257a9b1e71f0b54f42283ee4925f88c369f873
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:728074151b245f5e0d65bd8956149f05ba22a1b6438120ebca2d85b0d7d0c837
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:4109698147632b858fe2787964772a1339cdebee72b75b42d9b1c5dbbbecb45c
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:088d0eebf60c8f6961ca72c268f61d12648d575aa41552ae6398fefc6d9700cb
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d6d444e642bfd567eb40c7e084f5df743c23b47935e52e35ef541c67fcc00433
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:b3469cd953ef2015fad2f7f96d960cee5ecaa6c19df631a9c882154ab5b3de98
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:f49d3eddbfcec1018fd3716cd0ec2addef9961d2d308099636541868566979b0
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        - quay.io/openshift-community-operators/t8c@sha256:403b29a6262ba0e06de117f788aea7224268b798868c3e5b738b05e96df20456
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4bfaf97ab09a1aff1316536b99957b1f82763b009dde6d08fc9eea1fdd290fbf
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        - quay.io/openshift-community-operators/t8c@sha256:61f1ea9b3851d7272703f5bc21203ff03bb35fd52ed9780d0cf996379c78a00b
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:48a5a9fbfa92b4b87b3653613467a3fcceacadbbabd274e367b40f6871039cc8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d65834d8f9946c2764d45666c90c50f428b74afe7b066d666c18e9545994c9a8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4964196aba839fde569ce58ab7e4eef438c9638ae4057194f2d7a58eaa5a98d9
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        - quay.io/openshift-community-operators/t8c@sha256:46c815fa635d58820757f41b1d8f04333638c39471f4d6c901a6500379680b57
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d8a575a84dd86d241acaaec53af1e164bc1ff1b95b5eb710851ae698bbed6bf5
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:dfdc8fbdfc0ee76b2d5c62f6fd5deac2c50756661871d8b4357e25043e5ce66a
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:081d7f4f20332599fdbbd07f22cc18eca6c5625b619efeae6e861239cde4b12a
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:cf7ccc1a60d6787ebcd1d76415cca4cc55c2c3ee18d52c0511e66ca6bf9310f8
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: loki-operator.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
          createdAt: "2025-06-11T13:21:40Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b311677919183c6c91c6da40b0ddb2529f92ed52b2c46a07e399a4bad5dea280
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:fce287241be64b7b2c20a971c4f2d468c079ae81556aaab756572ebad27a17f5
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:ff98532bd8f21ecf375062d465f2aee4cdcd70b647d258c1e09f7e2286f998c9
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:ea2d995744cd4a4bd6e7467c2d6927bfc89a47d551ef0d6b048c0c57652f81be
        version: 6.3.1
      entries:
      - name: loki-operator.v6.3.1
        version: 6.3.1
      - name: loki-operator.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b793aef88881aa75d47a22377d8932efc0b497bb01b2f3f376d64891d1dace54
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-operator-bundle@sha256:722cf3c05eb5663a737cb1385b971b6d1f0c8b2595d5ac425f7b100c25e2fd26
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-operator-bundle@sha256:ed97b11fa0f0f7aadf193810a5a599bf26015a0af8ee6fd90061110dcb0d50f4
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f19ae051041c46fe7587f903a9c6221bcbe8be2c4cc0235739f7592eb93ae164
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:28e37ec35de1307405264cd4f79aaa6ce4d4ecbd6cbc27f4fef5db415c678d64
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        - registry.redhat.io/quay/quay-operator-bundle@sha256:4a004584fa6c7187481e8c8480f17e0e4edd18c32f52a2e271c68d02f1a9655f
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f16c0a135dd7380e2a489f963b38bd76a63a973f482d06ece8d3656aff223703
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b499eebed32c0261db414e307f96f8aef9e743035544c80c048526e7ba4fbdc2
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-operator-bundle@sha256:55a5e6297109aea029bd40fac09a9fbc10c28a8c5558f9afc9b9b86c5dc7a4d4
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-operator-bundle@sha256:e3f63ea56de7ca5388a20c17ac63e16f07935ab09bcdcd38a86639c9d59c8a7d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-operator-bundle@sha256:53fcddadaa31a5f7d74c38f0bbd0d80b4c09796a580ef8494f60ff82fb6fe6c1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-bundle@sha256:9d0ed6d279800a189c1466e0b32f80af2f8bad96a65b7dc0322c64337dddf39c
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:bbfddd7cb4a6cb7a131a0cef174054890fd39956d63e737c92fe8233ad8f98e4
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5a35593357fbf280561a1d4a64ecc0d5c2044409acff3ce6cb490a4a61462a4e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:791ac512370af5cde1e089de3def760e7dbce69badc03137051520ce316d63e2
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5dc66b7c486d1604c4bbb04c85544147c43264840e597958e6c5ca83e7e34c0e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:ddfdf850c9238a6988889e1840efe3175f03dad2489a5651e21cac62ac44e23e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/application-services-metering-operator@sha256:c3600cc8cd789df1f3c8f48bb3b1bc4c9cbf0285b6f6ce06f1d563c52f50915e
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/openshift-community-operators/etcd@sha256:e42581eacbc8349f40c6a4ed6d7d82d6168518e15b1a24cba36817205bd8688f
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:1d308347e5e290ac4bb7dd7aab0a1c0659d661153541d81dd30733b52d1fbf83
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:f1757deff8c4048ae42cb033f3ddd7686dec2968e5a4cd61ff8f74fda69a8a09
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - docker.io/grafana/tempo-query:2.8.2
        - docker.io/grafana/tempo:2.8.2
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - quay.io/community-operator-pipeline-prod/tempo-operator@sha256:14282a0ca95401ab3e4c11dc2b611a6f57c00c461254f52ddf03ef74c725767b
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/jboss-webserver-5/jws5-operator-bundle@sha256:ce4060138117dbc8fc21abb60f0b8ef1bfa5ebc0dd8eb803c0e79d7447341964
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9y
            dCgzMzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQ
            LmFkZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAg
            ICAgICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24i
            LAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJh
            c2gpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3Zp
            ZGVkIGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAog
            ICAgICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmls
            ZXN5c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IG1ldGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBk
            ZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2Vw
            aGZzIGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1l
            bnQoCiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92
            aWRlcyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9v
            bC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRh
            IHBvb2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBp
            biB0aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAg
            KQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1l
            bmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZh
            bHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+
            OjxQT1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCki
            LAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAg
            ICAiLS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAg
            ICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBv
            aW50IFRMUyBjZXJ0aWZpY2F0ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRf
            YXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1za2lwLXRscyIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJJ
            Z25vcmUgVExTIGNlcnRpZmljYXRpb24gdmFsaWRhdGlvbiB3aGVuIGEgc2VsZi1zaWduZWQgY2Vy
            dGlmaWNhdGUgaXMgcHJvdmlkZWQgKE5PVCBSRUNPTU1FTkRFRCIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW1vbml0b3JpbmctZW5k
            cG9pbnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgZW5k
            cG9pbnRzIChjb21tYSBzZXBhcmF0ZWQgbGlzdCBvZiAoZm9ybWF0IGA8SVB2ND5gIG9yIGA8W0lQ
            djZdPmAgb3IgYDxGUUROPmApIGVudHJpZXMgb2YgYWN0aXZlIGFuZCBzdGFuZGJ5IG1ncnMpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tbW9uaXRvcmluZy1lbmRwb2ludC1wb3J0IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAg
            ICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkNlcGggTWFuYWdlciBw
            cm9tZXRoZXVzIGV4cG9ydGVyIHBvcnQiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1za2lwLW1vbml0b3JpbmctZW5kcG9pbnQiLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBhY3Rpb249InN0b3JlX3RydWUi
            LAogICAgICAgICAgICBoZWxwPSJEbyBub3QgY2hlY2sgZm9yIGEgbW9uaXRvcmluZyBlbmRwb2lu
            dCBmb3IgdGhlIENlcGggY2x1c3RlciIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgZXJhc3VyZSBjb2RlZCBSQkQgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWRyeS1ydW4iLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAg
            ICAgICBhY3Rpb249InN0b3JlX3RydWUiLAogICAgICAgICAgICBoZWxwPSJEcnkgcnVuIHByaW50
            cyB0aGUgZXhlY3V0ZWQgY29tbWFuZHMgd2l0aG91dCBydW5uaW5nIHRoZW0iLAogICAgICAgICkK
            ICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yYWRvcy1u
            YW1lc3BhY2UiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1G
            YWxzZSwKICAgICAgICAgICAgaGVscD0iRGl2aWRlcyBhIHBvb2wgaW50byBzZXBhcmF0ZSBsb2dp
            Y2FsIG5hbWVzcGFjZXMsIHVzZWQgZm9yIGNyZWF0aW5nIFJCRCBQVkMgaW4gYSBDZXBoQmxvY2tQ
            b29sUmFkb3NOYW1lc3BhY2UgKHNob3VsZCBiZSBsb3dlciBjYXNlKSIsCiAgICAgICAgKQogICAg
            ICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXN1YnZvbHVtZS1n
            cm91cCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgc3Vidm9sdW1lIGdy
            b3VwIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tcmd3LXJlYWxtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAg
            ICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhlIG5hbWUg
            b2YgdGhlIHJndy1yZWFsbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXJndy16b25lLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlk
            ZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cC1uYW1lIiwKICAg
            ICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAg
            ICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSByZ3ctem9uZWdyb3VwIiwKICAgICAg
            ICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9w
            b2xvZ3ktcG9vbHMiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJl
            ZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdG9wb2xv
            Z3ktY29uc3RyYWluZWQgcmJkIHBvb2xzIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3Vw
            LmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9wb2xvZ3ktZmFpbHVyZS1kb21haW4tbGFi
            ZWwiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwK
            ICAgICAgICAgICAgaGVscD0iazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gbGFiZWwgKGV4YW1w
            bGU6IHpvbmUsIHJhY2ssIG9yIGhvc3QpIGZvciB0aGUgdG9wb2xvZ3ktcG9vbHMgdGhhdCBtYXRj
            aCB0aGUgY2VwaCBkb21haW4iLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdGhlIGs4cyBjbHVzdGVyIGZhaWx1
            cmUgZG9tYWluIHZhbHVlcyBjb3JyZXNwb25kaW5nIHRvIGVhY2ggb2YgdGhlIHBvb2xzIGluIHRo
            ZSBgdG9wb2xvZ3ktcG9vbHNgIGxpc3QiLAogICAgICAgICkKCiAgICAgICAgdXBncmFkZV9ncm91
            cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikKICAgICAgICB1cGdyYWRlX2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFkZSIsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAg
            IGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9yIGV4YW1wbGU6IGNsaWVudC5j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0aGNoZWNrZXIgY2VwaCB1c2Vy
            cyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBuZXcgY2x1c3RlciB2ZXJzaW9u
            IGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBwbGllZC4iCiAgICAgICAgICAg
            ICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rl
            ci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVwZ3JhZGUgYWxsIHRoZSBkZWZh
            dWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAgICAgICsgIkZvciByZXN0cmlj
            dGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVu
            c2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2luZyAtLXJlc3RyaWN0ZWQtYXV0
            aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9yeSBmbGFncyIKICAgICAgICAg
            ICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29sLW5hbWUsIC0tazhzLWNsdXN0
            ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUgdXBncmFkaW5nIgogICAgICAg
            ICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3UgaGF2ZSBwYXNzZWQgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcgdXNlciB0aGVuIHdoaWxlIHVw
            Z3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAgICAgICAgICsgIlNhbXBsZSBy
            dW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMu
            cHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxpY2Fwb29sIC0tazhzLWNsdXN0
            ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGllbnQuY3NpLXJiZC1ub2RlLXJv
            b2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAiUFM6IEFuIGV4aXN0aW5nIG5v
            bi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0byBhIHJlc3RyaWN0ZWQgdXNl
            ciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRlIGZsYWcgc2hvdWxkIG9ubHkg
            YmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVzZXJzLCBpdCBzaG91bGRuJ3Qg
            YmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxpZWQgcGVybWlzc2lvbiwgZm9y
            IGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2ggcG9vbCB1c2VyIGhhcyBhY2Nl
            c3MiLAogICAgICAgICkKCiAgICAgICAgIyBBZGQgY29tbWFuZC1saW5lIGFyZ3VtZW50cwogICAg
            ICAgIGNvbmZpZ19ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb25maWciKQogICAg
            ICAgIGNvbmZpZ19ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWNvbmZpZy1maWxl
            IiwKICAgICAgICAgICAgdHlwZT1zdHIsCiAgICAgICAgICAgIGhlbHA9IlBhdGggdG8gdGhlIGNv
            bmZpZ3VyYXRpb24gZmlsZSwgUHJpb3JpdHk6IGNvbW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmlu
            aSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgKQoKICAgICAgICBpZiBhcmdzX3Rv
            X3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAogICAgICAgICAgICAgICAgdHlwZShhcmdzX3Rv
            X3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICksICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJz
            ZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgYXJnc190b19w
            YXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJldHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190
            b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6
            CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCgogICAgICAgICAgICBpZiByYmRfcG9v
            bF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBpZiByYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0
            YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAg
            KQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQi
            OiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            IG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCdd
            fSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21z
            ZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZh
            bHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtleSBpbiBqc29uX291dFsicG9vbHMiXToKICAg
            ICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wg
            bmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBvb2wKICAgICAgICAgICAgICAgIGlmICgKICAg
            ICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAg
            ICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bv
            b2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxl
            IGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAg
            ICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29s
            X25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgICAgIHBvb2xfZXhpc3Qg
            PSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAg
            ICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9wb29sIG5hbWUsIgogICAgICAgICAgICAgICAg
            ICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBwb29sX2V4aXN0OgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            IlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwge3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBl
            eGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihzZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIg
            KyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoOgogICAgICAg
            ICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rp
            bmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAgICAgY29udGVudHMgPSBmLnJlYWQoKQogICAg
            ICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJzdHJpcCgpCgogICAgZGVmIF9jaGVja19jb25m
            bGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJF
            aXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRhdGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBz
            aG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAgICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBv
            aW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAgICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdl
            dHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRlbGltaXRlcgogICAgICAgIHRyeToKICAgICAg
            ICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEp
            CiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikK
            CiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgog
            ICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4o
            ZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAgICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAg
            ICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNzKGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRB
            ZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAgICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJv
            cjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUUROIgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGln
            aXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQg
            bm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGludFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBp
            ZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioqMTYgLSAxOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9y
            dH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoKICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYs
            IGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMg
            aWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0
            IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBvaW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShz
            ZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gIiIsICIiLCAiIgog
            ICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJz
            cGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            Ik5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSAhPSAiWyI6CiAgICAgICAgICAgICAg
            ICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRwb2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAg
            ICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2VuZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3Ry
            X3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBbImh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJl
            c3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZvciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlcCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRf
            c3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9IE5vbmUKICAgICAgICAgICAgICAgICMgSWYg
            dmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBkaXJlY3RvcnksCiAgICAgICAgICAgICAgICAj
            IHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4gcHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFz
            aCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNTTC4KICAgICAgICAgICAgICAgIGlmIHByZWZp
            eCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RsczoKICAgICAgICAg
            ICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1GYWxzZSkKICAgICAgICAgICAgICAg
            IGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNlcnQ6CiAgICAgICAgICAgICAgICAgICAgdmVy
            aWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1l
            b3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAgICAgICAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10aW1lb3V0KQogICAgICAg
            ICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAyMDA6CiAgICAgICAgICAgICAgICAgICAgcmV0
            dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAgICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVy
            cjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vycm9yID0gZXJyCiAgICAgICAgICAgICAgICBj
            b250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRv
            IGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNw
            b25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICIiLAog
            ICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCItMSIpLAogICAgICAgICkKCiAgICBkZWYgcGFy
            c2VfY29uZmlnX2ZpbGUoc2VsZiwgY29uZmlnX2ZpbGUpOgogICAgICAgIGNvbmZpZyA9IGNvbmZp
            Z3BhcnNlci5Db25maWdQYXJzZXIoKQogICAgICAgIGNvbmZpZy5yZWFkKGNvbmZpZ19maWxlKQog
            ICAgICAgIGZvciBhcmcgaW4gbGlzdCh2YXJzKHNlbGYuX2FyZ19wYXJzZXIpKToKICAgICAgICAg
            ICAgIyBweXRob24gdHJlYXRzIGZsYWctbmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28g
            Y29udmVydGluZyBiYWNrIHRvIGZsYWctbmFtZSwKICAgICAgICAgICAgIyBzbyB3ZSBjYW4gZ2V0
            IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAgIGFyZ3VtZW50ID0gYXJn
            LnJlcGxhY2UoIl8iLCAiLSIpCiAgICAgICAgICAgIGFyZ3VtZW50VmFsdWUgPSBzdHIoZ2V0YXR0
            cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcpKQogICAgICAgICAgICBjb25maWdfdmFsdWUgPSBjb25m
            aWcuZ2V0KCJDb25maWd1cmF0aW9ucyIsIGFyZ3VtZW50LCBmYWxsYmFjaz1Ob25lKQogICAgICAg
            ICAgICAjIGdpdmUgcHJpb3JpdHkgdG8gY29tbWFuZCBsaW5lIGFyZ3VtZW50LCBpZiBjb21tYW5k
            IGxpbmUgYXJndW1lbnQgaXMgbm90IHByZXNlbnQgdXNlIGNvbmZpZy5pbmkgdmFsdWUsCiAgICAg
            ICAgICAgIGlmIChzdHIoc3lzLmFyZ3YpLmZpbmQoYXJndW1lbnQpID09IC0xKSBhbmQgKAogICAg
            ICAgICAgICAgICAgKGNvbmZpZ192YWx1ZSAhPSBOb25lKSBhbmQgKGNvbmZpZ192YWx1ZSAhPSAi
            IikKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuX19zZXRh
            dHRyX18oYXJnLCBjb25maWdfdmFsdWUpCgogICAgICAgIHJldHVybiBjb25maWcKCiAgICBkZWYg
            X19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAgICAgICAgc2VsZi5vdXRfbWFwID0ge30K
            ICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0KCkKICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3RvX3BhcnNlPWFyZ19saXN0KQogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGU6CiAgICAgICAgICAgIHNlbGYuY29uZmln
            ID0gc2VsZi5wYXJzZV9jb25maWdfZmlsZShzZWxmLl9hcmdfcGFyc2VyLmNvbmZpZ19maWxlKQog
            ICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoKQogICAgICAgIHNlbGYucnVu
            X2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRw
            dXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAgICAgc2VsZi5jZXBoX2NvbmYg
            PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNlbGYuY2VwaF9rZXlyaW5nID0g
            c2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1c2VyIG5vdCBwcm92aWRlZCwg
            Z2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxmLnJ1bl9hc191c2VyIGFuZCBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLnJ1bl9hc191c2Vy
            ID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGU6CiAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gc2VsZi5ERUZBVUxUX1JH
            V19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25mOgogICAgICAgICAgICBrd2Fy
            Z3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmluZzoKICAgICAgICAgICAgICAg
            IGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBoX2tleXJpbmd9CiAgICAgICAg
            ICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxlPXNlbGYuY2VwaF9jb25mLCAq
            Kmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIgPSByYWRvcy5S
            YWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3JlYWRfZmlsZSgpCiAgICAgICAg
            c2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToKICAgICAgICBp
            ZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAgICAgICAgICAgIHNlbGYuY2x1
            c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2Vw
            aCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rlci5nZXRfZnNpZCgpKQoKICAg
            IGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNvbik6CiAgICAgICAgY21kID0g
            anNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAgICAgICAgcmV0X3ZhbCwgY21k
            X291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFuZChjbWQsIGIiIikKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAgICAgIHByaW50KGYiQ29tbWFu
            ZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgIGYiUmV0
            dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2NtZF9vdXR9XG4iCiAgICAgICAg
            ICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0tLS0tLS0tLVxuIgogICAgICAg
            ICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJy
            b3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQnIGlzIG5vdCBlbXB0eQogICAg
            ICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24gb3V0cHV0CiAgICAgICAgaWYg
            cmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBqc29uX291dCA9IGpzb24ubG9h
            ZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cKCiAg
            ICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6CiAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9y
            dW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            IyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9ydW1fc3RhdHVzJyBjb21tYW5k
            IGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwg
            IT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICApCiAgICAgICAg
            cV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVyX25hbWUiXQogICAgICAgIHFf
            bGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21hdGNoaW5nX2xpc3QgPSBbCiAg
            ICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJdWyJtb25zIl0gaWYgbFsibmFt
            ZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQiKQogICAgICAgIHFfbGVhZGVy
            X2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAgICAgICAgIyBnZXQgdGhlIGFk
            ZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAgICAgcV9sZWFkZXJfYWRkcnZl
            YyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMiLCB7fSkuZ2V0KCJhZGRydmVj
            IiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9kZXRhaWxzWyJwdWJsaWNfYWRk
            ciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudjJfcG9ydF9l
            bmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIi
            OgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bImFkZHIiXQog
            ICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+IDEgYW5kIHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVh
            ZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAg
            c3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUg
            bm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAg
            ICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9h
            ZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBw
            b3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15
            IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRo
            b2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNp
            bnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYg
            aXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91
            bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUg
            aG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlw
            CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1h
            dCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlz
            IGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBm
            YWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3Zh
            bCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBh
            ZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28g
            dmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRw
            b2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAg
            ICAgICAgaXAgPSBbXQogICAgICAgICAgICAjIGV4YW1wbGUgb3V0cHV0IFsoPEFkZHJlc3NGYW1p
            bHkuQUZfSU5FVDogMj4sIDxTb2NrZXRLaW5kLlNPQ0tfU1RSRUFNOiAxPiwgNiwgJycsICgnOTMu
            MTg0LjIxNi4zNCcsIDgwKSksIC4uLl0KICAgICAgICAgICAgIyB3ZSBuZWVkIHRvIGdldCA5My4x
            ODQuMjE2LjM0IHNvIGl0IHdvdWxkIGJlIGlwWzBdWzRdWzBdCiAgICAgICAgICAgIGlmIGhvc3Rf
            aXBfdHlwZSA9PSAiSVB2NiI6CiAgICAgICAgICAgICAgICBpcCA9IHNvY2tldC5nZXRhZGRyaW5m
            bygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZhbWlseT1zb2NrZXQuQUZf
            SU5FVDYsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbGlmIGhvc3RfaXBfdHlwZSA9PSAiSVB2NCI6CiAgICAgICAgICAgICAgICBpcCA9IHNv
            Y2tldC5nZXRhZGRyaW5mbygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZh
            bWlseT1zb2NrZXQuQUZfSU5FVCwgcHJvdG89c29ja2V0LklQUFJPVE9fVENQCiAgICAgICAgICAg
            ICAgICApCiAgICAgICAgICAgIGRlbCBzb2NrZXQKICAgICAgICAgICAgcmV0dXJuIGlwWzBdWzRd
            WzBdCiAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQoKICAgIGRlZiBnZXRfYWN0aXZlX2FuZF9zdGFu
            ZGJ5X21ncnMoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAg
            ICAgICAgICByZXR1cm4gIiIsIHNlbGYuZHJ5X3J1bigiY2VwaCBzdGF0dXMiKQogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gc2VsZi5fYXJn
            X3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50CiAgICAgICAgc3RhbmRieV9tZ3JzID0gW10KICAg
            ICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0OgogICAgICAgICAgICBjbWRf
            anNvbiA9IHsicHJlZml4IjogInN0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1w
            dCwKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAg
            ICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICAgICAgIidtZ3Igc2VydmljZXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAg
            ICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlf
            T1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtb25pdG9yaW5nX2Vu
            ZHBvaW50ID0gKAogICAgICAgICAgICAgICAganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0
            KCJzZXJ2aWNlcyIsIHt9KS5nZXQoInByb21ldGhldXMiLCAiIikKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgImNhbid0IGZp
            bmQgbW9uaXRvcmluZ19lbmRwb2ludCwgcHJvbWV0aGV1cyBtb2R1bGUgbWlnaHQgbm90IGJlIGVu
            YWJsZWQsICIKICAgICAgICAgICAgICAgICAgICAiZW5hYmxlIHRoZSBtb2R1bGUgYnkgcnVubmlu
            ZyAnY2VwaCBtZ3IgbW9kdWxlIGVuYWJsZSBwcm9tZXRoZXVzJyIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgIyBub3cgY2hlY2sgdGhlIHN0YW5kLWJ5IG1nci1zCiAgICAgICAgICAgIHN0
            YW5kYnlfYXJyID0ganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0KCJzdGFuZGJ5cyIsIFtd
            KQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5IGluIHN0YW5kYnlfYXJyOgogICAgICAgICAg
            ICAgICAgaWYgIm5hbWUiIGluIGVhY2hfc3RhbmRieS5rZXlzKCk6CiAgICAgICAgICAgICAgICAg
            ICAgc3RhbmRieV9tZ3JzLmFwcGVuZChlYWNoX3N0YW5kYnlbIm5hbWUiXSkKICAgICAgICAgICAg
            dHJ5OgogICAgICAgICAgICAgICAgcGFyc2VkX2VuZHBvaW50ID0gdXJscGFyc2UobW9uaXRvcmlu
            Z19lbmRwb2ludCkKICAgICAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYi
            aW52YWxpZCBlbmRwb2ludDoge21vbml0b3JpbmdfZW5kcG9pbnR9IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBwYXJzZWRfZW5kcG9p
            bnQuaG9zdG5hbWUKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoK
            ICAgICAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHN0cihwYXJzZWRfZW5k
            cG9pbnQucG9ydCkKCiAgICAgICAgIyBpZiBtb25pdG9yaW5nIGVuZHBvaW50IHBvcnQgaXMgbm90
            IHNldCwgcHV0IGEgZGVmYXVsdCBtb24gcG9ydAogICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQ6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYu
            REVGQVVMVF9NT05JVE9SSU5HX0VORFBPSU5UX1BPUlQKCiAgICAgICAgIyB1c2VyIGNvdWxkIGdp
            dmUgY29tbWEgYW5kIHNwYWNlIHNlcGFyYXRlZCBpbnB1dHMgKGxpa2UgLS1tb25pdG9yaW5nLWVu
            ZHBvaW50PSI8aXAxPiwgPGlwMj4iKQogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlz
            dCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdC5yZXBsYWNlKCIsIiwgIiAiKQogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCA9IG1vbml0b3JpbmdfZW5kcG9pbnRf
            aXBfbGlzdC5zcGxpdCgpCiAgICAgICAgIyBpZiBtb25pdG9yaW5nLWVuZHBvaW50IGNvdWxkIG5v
            dCBiZSBmb3VuZCwgcmFpc2UgYW4gZXJyb3IKICAgICAgICBpZiBsZW4obW9uaXRvcmluZ19lbmRw
            b2ludF9pcF9saXN0X3NwbGl0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKCJObyAnbW9uaXRvcmluZy1lbmRwb2ludCcgZm91bmQiKQogICAgICAgICMg
            Zmlyc3QgaXAgaXMgdHJlYXRlZCBhcyB0aGUgbWFpbiBtb25pdG9yaW5nLWVuZHBvaW50CiAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9z
            cGxpdFswXQogICAgICAgICMgcmVzdCBvZiB0aGUgaXAtcyBhcmUgYWRkZWQgdG8gdGhlICdzdGFu
            ZGJ5X21ncnMnIGxpc3QKICAgICAgICBzdGFuZGJ5X21ncnMuZXh0ZW5kKG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXBfbGlzdF9zcGxpdFsxOl0pCiAgICAgICAgZmFpbGVkX2lwID0gbW9uaXRvcmluZ19l
            bmRwb2ludF9pcAoKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50ID0gIjoiLmpvaW4oCiAgICAg
            ICAgICAgIFttb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnRd
            CiAgICAgICAgKQogICAgICAgIGlwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KG1vbml0
            b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBv
            aW50X2lwID0gc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQog
            ICAgICAgICAgICApCiAgICAgICAgICAgICMgY29sbGVjdCBhbGwgdGhlICdzdGFuZC1ieScgbWdy
            IGlwcwogICAgICAgICAgICBtZ3JfaXBzID0gW10KICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRi
            eV9tZ3IgaW4gc3RhbmRieV9tZ3JzOgogICAgICAgICAgICAgICAgZmFpbGVkX2lwID0gZWFjaF9z
            dGFuZGJ5X21ncgogICAgICAgICAgICAgICAgbWdyX2lwcy5hcHBlbmQoCiAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgICAgICAg
            ICAgZWFjaF9zdGFuZGJ5X21nciwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAg
            ICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiQ29udmVyc2lvbiBvZiBob3N0OiB7ZmFpbGVkX2lwfSB0byBJUCBmYWlsZWQuICIKICAgICAg
            ICAgICAgICAgICJQbGVhc2UgZW50ZXIgdGhlIElQIGFkZHJlc3NlcyBvZiBhbGwgdGhlIGNlcGgt
            bWdycyB3aXRoIHRoZSAnLS1tb25pdG9yaW5nLWVuZHBvaW50JyBmbGFnIgogICAgICAgICAgICAp
            CgogICAgICAgIF8sIF8sIGVyciA9IHNlbGYuZW5kcG9pbnRfZGlhbChtb25pdG9yaW5nX2VuZHBv
            aW50LCBpcF90eXBlKQogICAgICAgIGlmIGVyciA9PSAiLTEiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGVycikKICAgICAgICAjIGFkZCB0aGUgdmFsaWRhdGVk
            IGFjdGl2ZSBtZ3IgSVAgaW50byB0aGUgZmlyc3QgaW5kZXgKICAgICAgICBtZ3JfaXBzLmluc2Vy
            dCgwLCBtb25pdG9yaW5nX2VuZHBvaW50X2lwKQogICAgICAgIGFsbF9tZ3JfaXBzX3N0ciA9ICIs
            Ii5qb2luKG1ncl9pcHMpCiAgICAgICAgcmV0dXJuIGFsbF9tZ3JfaXBzX3N0ciwgbW9uaXRvcmlu
            Z19lbmRwb2ludF9wb3J0CgogICAgZGVmIGNoZWNrX3VzZXJfZXhpc3Qoc2VsZiwgdXNlcik6CiAg
            ICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJhdXRoIGdldCIsICJlbnRpdHkiOiBmInt1c2Vy
            fSIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgIGlmIHJldF92YWwgIT0gMCBv
            ciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIHJldHVy
            biBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaGZzX3Byb3Zpc2lvbmVy
            X2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAi
            YWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRh
            dGE9KiIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFk
            YXRhPXtjZXBoZnNfZmlsZXN5c3RlbX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAg
            ICBkZWYgZ2V0X2NlcGhmc19ub2RlX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRp
            dHkgPSAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIKICAgICAgICBjYXBzID0gewogICAgICAgICAg
            ICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAg
            ICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAiYWxsb3cgcncgdGFn
            IGNlcGhmcyAqPSoiLAogICAgICAgICAgICAibWRzIjogImFsbG93IHJ3IiwKICAgICAgICB9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAi
            azhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVy
            LW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lz
            dGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAg
            IGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7
            ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVz
            eXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgKj17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5Cgog
            ICAgZGVmIGdldF9lbnRpdHkoCiAgICAgICAgc2VsZiwKICAgICAgICBlbnRpdHksCiAgICAgICAg
            cmJkX3Bvb2xfbmFtZSwKICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgIGs4c19j
            bHVzdGVyX25hbWUsCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgKToKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgIG9y
            IHJiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgIG9yIGFsaWFzX3JiZF9w
            b29sX25hbWUgIT0gIiIKICAgICAgICAgICAgIyBjaGVja2luZyBhbGlhc19yYmRfcG9vbF9uYW1l
            IGlzIG5vdCBlbXB0eSBhcyB0aGVyZSBtYXliZSBhIHNwZWNpYWwgY2hhcmFjdGVyIHVzZWQgb3Ro
            ZXIgdGhhbiAuIG9yIF8KICAgICAgICApOgogICAgICAgICAgICBpZiBhbGlhc19yYmRfcG9vbF9u
            YW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICAgICAicGxlYXNlIHNldCB0aGUgJy0tYWxpYXMtcmJkLWRhdGEt
            cG9vbC1uYW1lJyBmbGFnIGFzIHRoZSByYmQgZGF0YSBwb29sIG5hbWUgY29udGFpbnMgJy4nIG9y
            ICdfJyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAg
            YWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgICAgIG9yIGFs
            aWFzX3JiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgICk6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICInLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgdmFsdWUgc2hvdWxkIG5vdCBj
            b250YWluICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9
            IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXthbGlhc19yYmRfcG9vbF9uYW1lfSIKICAg
            ICAgICBlbHNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfS17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZToKICAgICAg
            ICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICByZXR1
            cm4gZW50aXR5CgogICAgZGVmIGdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNf
            Y2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJz
            ZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJi
            ZC1kYXRhLXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYg
            azhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAg
            ICAgICAgZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAg
            ICAgIGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1l
            LAogICAgICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmIHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFt
            ZX0gbmFtZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBv
            b2w9e3JiZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVm
            IGdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktcmJkLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            b3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9
            IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xfbmFtZQog
            ICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rl
            cl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFk
            b3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgazhzX2Ns
            dXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBmb3VuZCwg
            cGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAgICAgICAg
            ZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAg
            ICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAgICAgIGlm
            IHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gKAog
            ICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0gbmFt
            ZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBvb2w9e3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9k
            ZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gc2VsZi5y
            dW5fYXNfdXNlcgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLAogICAg
            ICAgICAgICAibWdyIjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwKICAgICAgICAgICAgIm9zZCI6
            IGYicHJvZmlsZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD17c2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5yb290LCBhbGxv
            dyBydyBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2ws
            IGFsbG93IHJ4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubG9n
            LCBhbGxvdyB4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuYnVj
            a2V0cy5pbmRleCIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQoKICAg
            ICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jYXBzX2FuZF9lbnRpdHkoc2Vs
            ZiwgdXNlcl9uYW1lKToKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIi
            IGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lv
            bmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3Ry
            aWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRf
            Y2VwaGZzX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBz
            ZWxmLmdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            Y3NpLXJiZC1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50
            LmNzaS1yYmQtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KCkKICAgICAg
            ICBpZiAiY2xpZW50LmNzaS1yYmQtbm9kZSIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAi
            Y2xpZW50LmNzaS1yYmQtbm9kZSIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAg
            cmV0dXJuIHNlbGYuZ2V0X3JiZF9ub2RlX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2VudGl0eSgpCgogICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgIGYibm8gdXNlciBmb3VuZCB3aXRo
            IHVzZXJfbmFtZToge3VzZXJfbmFtZX0sICIKICAgICAgICAgICAgImdldF9jYXBzX2FuZF9lbnRp
            dHkgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcihzZWxmLCB1c2VyKToKICAgICAgICAiIiIKICAgICAgICBjb21tYW5kOiBjZXBo
            IGF1dGggZ2V0LW9yLWNyZWF0ZSBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciBtb24gJ2Fs
            bG93IHInIG1nciAnYWxsb3cgcncnIG9zZCAnYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            JwogICAgICAgICIiIgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2Vu
            dGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1
            dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAg
            ICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBj
            YXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0K
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4g
            KAogICAgICAgICAgICAgICAgc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgICAgICJjZXBo
            ICIKICAgICAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAg
            ICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50aXR5Il0KICAg
            ICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21k
            X2pzb25bImNhcHMiXSkKICAgICAgICAgICAgICAgICksCiAgICAgICAgICAgICAgICAiIiwKICAg
            ICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAg
            dXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJf
            a2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXksIGYie2VudGl0eS5zcGxpdCgn
            LicsIDEpWzFdfSIKICAgICAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4nLDEpWzFdIHRvIHJlbmFt
            ZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1ub2RlCgogICAgICAgIHJl
            dF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRf
            anNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAg
            ICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGdl
            dC1vci1jcmVhdGUge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYi
            RXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJ
            U1R9IgogICAgICAgICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0p
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4n
            LDEpWzFdIHRvIHJlbmFtZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1u
            b2RlCgogICAgZGVmIGdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoc2VsZik6CiAgICAgICAg
            Y21kX2pzb24gPSB7InByZWZpeCI6ICJmcyBscyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            aWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlf
            cnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsIHJlcG9ydCBhbiBlcnJvcgog
            ICAgICAgIGlmIHJldF92YWwgIT0gMDoKICAgICAgICAgICAgIyBpZiBmcyBhbmQgZGF0YV9wb29s
            IGFyZ3VtZW50cyBhcmUgbm90IHNldCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSA9
            PSAiIgogICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9v
            bF9uYW1lID09ICIiCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICByZXR1cm4KICAgICAg
            ICAgICAgIyBpZiB1c2VyIGhhcyBwcm92aWRlZCBhbnkgb2YgdGhlCiAgICAgICAgICAgICMgJy0t
            Y2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgb3IgJy0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lJyBhcmd1
            bWVudHMsCiAgICAgICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIGFzIHdlIGFyZSB1bmFibGUg
            dG8gdmVyaWZ5IHRoZSBhcmdzCiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIidmcyBscycgY2VwaCBjYWxsIGZhaWxlZCB3aXRoIGVy
            cm9yOiB7ZXJyX21zZ30iCiAgICAgICAgICAgICkKCiAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQg
            PSB7fQogICAgICAgICMgaWYgJy0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgYXJndW1lbnQgaXMg
            cHJvdmlkZWQsCiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBwcm92aWRlZCBmaWxlc3lzdGVt
            LW5hbWUgZXhpc3RzIG9yIG5vdAogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZToKICAgICAgICAgICAgIyBnZXQgdGhlIG1hdGNoaW5nIGxpc3QKICAgICAg
            ICAgICAgbWF0Y2hpbmdfanNvbl9vdXRfbGlzdCA9IFsKICAgICAgICAgICAgICAgIG1hdGNoZWQK
            ICAgICAgICAgICAgICAgIGZvciBtYXRjaGVkIGluIGpzb25fb3V0CiAgICAgICAgICAgICAgICBp
            ZiBzdHIobWF0Y2hlZFsibmFtZSJdKSA9PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWUKICAgICAgICAgICAgXQogICAgICAgICAgICAjIHVuYWJsZSB0byBmaW5kIGEgbWF0
            Y2hpbmcgZnMtbmFtZSwgcmFpc2UgYW4gZXJyb3IKICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5n
            X2pzb25fb3V0X2xpc3QpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiRmlsZXN5c3RlbSBwcm92aWRlZCwg
            J3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWV9JywgIgogICAgICAgICAg
            ICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBmcy1saXN0OiB7W3N0cih4WyduYW1lJ10p
            IGZvciB4IGluIGpzb25fb3V0XX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1hdGNo
            aW5nX2pzb25fb3V0ID0gbWF0Y2hpbmdfanNvbl9vdXRfbGlzdFswXQogICAgICAgICMgaWYgY2Vw
            aGZzIGZpbGVzeXN0ZW0gbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgIyB0cnkgdG8gZ2V0
            IGEgZGVmYXVsdCBmcyBuYW1lIGJ5IGRvaW5nIHRoZSBmb2xsb3dpbmcKICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAjIGEuIGNoZWNrIGlmIHRoZXJlIGlzIG9ubHkgb25lIGZpbGVzeXN0ZW0gaXMg
            cHJlc2VudAogICAgICAgICAgICBpZiBsZW4oanNvbl9vdXQpID09IDE6CiAgICAgICAgICAgICAg
            ICBtYXRjaGluZ19qc29uX291dCA9IGpzb25fb3V0WzBdCiAgICAgICAgICAgICMgYi4gb3IgZWxz
            ZSwgY2hlY2sgaWYgZGF0YV9wb29sIG5hbWUgaXMgcHJvdmlkZWQKICAgICAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICMg
            YW5kIGlmIHByZXNlbnQsIGNoZWNrIHdoZXRoZXIgdGhlcmUgZXhpc3RzIGEgZnMgd2hpY2ggaGFz
            IHRoZSBkYXRhX3Bvb2wKICAgICAgICAgICAgICAgIGZvciBlYWNoSiBpbiBqc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSBpbiBlYWNoSlsiZGF0YV9wb29scyJdOgogICAgICAgICAgICAgICAgICAgICAgICBtYXRjaGlu
            Z19qc29uX291dCA9IGVhY2hKCiAgICAgICAgICAgICAgICAgICAgICAgIGJyZWFrCiAgICAgICAg
            ICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIG1hdGNoaW5nIGZzIGV4aXN0cywgdGhhdCBtZWFucyBw
            cm92aWRlZCBkYXRhX3Bvb2wgbmFtZSBpcyBpbnZhbGlkCiAgICAgICAgICAgICAgICBpZiBub3Qg
            bWF0Y2hpbmdfanNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhX3Bv
            b2wgbmFtZSwge3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfSwiCiAgICAg
            ICAgICAgICAgICAgICAgICAgICIgZG9lcyBub3QgZXhpc3RzIgogICAgICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgIyBjLiBpZiBub3RoaW5nIGlzIHNldCBhbmQgY291bGRuJ3QgZmluZCBh
            IGRlZmF1bHQsCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGp1c3QgcmV0dXJu
            IHNpbGVudGx5CiAgICAgICAgICAgICAgICByZXR1cm4KCiAgICAgICAgaWYgbWF0Y2hpbmdfanNv
            bl9vdXQ6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSA9IHN0cihtYXRjaGluZ19qc29uX291dFsibmFtZSJdKQogICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19tZXRhZGF0YV9wb29sX25hbWUgPSBzdHIoCiAgICAgICAgICAgICAgICBt
            YXRjaGluZ19qc29uX291dFsibWV0YWRhdGFfcG9vbCJdCiAgICAgICAgICAgICkKCiAgICAgICAg
            aWYgaXNpbnN0YW5jZShtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdLCBsaXN0KToKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBoYXMgYWxyZWFkeSBwcm92aWRlZCBkYXRhLXBvb2wtbmFt
            ZSwKICAgICAgICAgICAgIyB0aHJvdWdoIC0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lCiAgICAgICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAg
            ICAgICAgIyBpZiB0aGUgcHJvdmlkZWQgbmFtZSBpcyBub3QgbWF0Y2hpbmcgd2l0aCB0aGUgb25l
            IGluIHRoZSBsaXN0CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgICAgICAgICBu
            b3QgaW4gbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXQogICAgICAgICAgICAgICAgKToK
            ICAgICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lOiAne3NlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfScsICIKICAgICAgICAgICAgICAgICAgICAg
            ICAgImRvZXNuJ3QgbWF0Y2ggZnJvbSB0aGUgZGF0YS1wb29scyBsaXN0OiAiCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGYie1tzdHIoeCkgZm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFf
            cG9vbHMnXV19IgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiBkYXRhX3Bv
            b2wgbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgICAgICMgdGhlbiB0cnkgdG8gZmluZCBh
            IGRlZmF1bHQgZGF0YSBwb29sIG5hbWUKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAg
            ICMgaWYgbm8gZGF0YV9wb29scyBleGlzdCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgICAg
            ICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPT0gMDoKICAgICAgICAg
            ICAgICAgICAgICByZXR1cm4KICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25f
            b3V0WyJkYXRhX3Bvb2xzIl1bMF0KICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiB0
            aGVyZSBhcmUgbW9yZSB0aGFuIG9uZSAnZGF0YV9wb29scycgZXhpc3QsCiAgICAgICAgICAgICMg
            dGhlbiB3YXJuIHRoZSB1c2VyIHRoYXQgd2UgYXJlIHVzaW5nIHRoZSBzZWxlY3RlZCBuYW1lCiAg
            ICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdKSA+IDE6CiAg
            ICAgICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgICAgICAiV0FSTklORzogTXVsdGlw
            bGUgZGF0YSBwb29scyBkZXRlY3RlZDogIgogICAgICAgICAgICAgICAgICAgIGYie1tzdHIoeCkg
            Zm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFfcG9vbHMnXV19XG4iCiAgICAgICAgICAg
            ICAgICAgICAgZiJVc2luZyB0aGUgZGF0YS1wb29sOiAne3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lfSdcbiIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2No
            ZWNrZXJLZXkoc2VsZiwgdXNlcik6CiAgICAgICAgY2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fw
            c19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21kX2pzb24gPSB7CiAgICAgICAgICAgICJwcmVm
            aXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAgICAgICAgICAgImVudGl0eSI6IGVudGl0eSwK
            ICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBjYXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMo
            KSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAg
            ICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oCiAgICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAg
            ICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAg
            ICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAgICAgICApCiAgICAg
            ICAgIyBjaGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYu
            Y2hlY2tfdXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAg
            ICAgICAgIHJldHVybiB1c2VyX2tleQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21z
            ZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVy
            ZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHtzZWxmLnJ1bl9h
            c191c2VyfScgY29tbWFuZCBmYWlsZWRcbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJf
            bXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAg
            ICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0pCgogICAgZGVmIGdl
            dF9jZXBoX2Rhc2hib2FyZF9saW5rKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgi
            OiAibWdyIHNlcnZpY2VzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiAr
            IGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1
            bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25f
            b3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIGlmICJkYXNoYm9hcmQi
            IG5vdCBpbiBqc29uX291dDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICByZXR1cm4g
            anNvbl9vdXRbImRhc2hib2FyZCJdCgogICAgZGVmIGNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIo
            c2VsZik6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAg
            ICAgICAgICJ1c2VyIiwKICAgICAgICAgICAgImNyZWF0ZSIsCiAgICAgICAgICAgICItLXVpZCIs
            CiAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAg
            ICAgICAgICItLWRpc3BsYXktbmFtZSIsCiAgICAgICAgICAgICJSb29rIFJHVyBBZG1pbiBPcHMg
            dXNlciIsCiAgICAgICAgICAgICItLWNhcHMiLAogICAgICAgICAgICAiYnVja2V0cz0qO3VzZXJz
            PSo7dXNhZ2U9cmVhZDttZXRhZGF0YT1yZWFkO3pvbmU9cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgIiAiLmpvaW4oY21kKSkKICAgICAgICB0cnk6CiAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBhbHJlYWR5IGV4aXN0cywgd2UganVzdCBxdWVyeSBpdAog
            ICAgICAgICAgICBpZiBleGVjRXJyLnJldHVybmNvZGUgPT0gZXJybm8uRUVYSVNUOgogICAgICAg
            ICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAg
            ICAgICAgICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICAgICAgICAgImluZm8iLAogICAg
            ICAgICAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5FWFRFUk5B
            TF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUs
            CiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICAgICAg
            ICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICAgICAgICAgIF0KICAgICAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVy
            cj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxs
            ZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgICAgICBlcnJfbXNnID0g
            KAogICAgICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2Nt
            ZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShl
            cnJfbXNnKQogICAgICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0x
            IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4
            ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0
            dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0
            dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5m
            bz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19j
            YXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAg
            ICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAg
            ICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Jl
            YWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUi
            LAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQog
            ICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQo
            Y21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNh
            bGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZl
            cnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRh
            dGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2Fw
            czogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4
            ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2Fw
            X3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJf
            bXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7
            Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJD
            b2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAg
            ICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNv
            bm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAg
            IGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0
            cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9y
            dGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0
            cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3twb29sX25hbWV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYgaW5pdF9yYmRfcG9vbChzZWxmLCBy
            YmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlS
            YWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9w
            ZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAg
            ICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkKCiAgICBkZWYgdmFsaWRhdGVfcmFk
            b3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmJkX21l
            dGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQogICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYgcmFkb3NfbmFt
            ZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRvc19uYW1lc3Bh
            Y2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNl
            LCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRhaW5zIHVwcGVy
            IGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAgICAg
            IGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICBp
            ZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3BhY2UpIGlzIEZh
            bHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFj
            ZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9vbCAne3JiZF9w
            b29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRlX3N1YnZvbHVt
            ZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUpOgog
            ICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAg
            ICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIsCiAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19v
            dXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9j
            ZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAg
            ICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAgICJzdWJ2b2x1
            bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgICAgIGNlcGhm
            c19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3Mu
            Y2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmInN1YnZvbHVt
            ZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3JlYXRlZCIKICAg
            ICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBzZWxmLCBzdWJ2
            b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBwaW5fc2V0dGlu
            ZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAgICAgICAgICAg
            ImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgInBpbiIsCiAg
            ICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9n
            cm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0aW5nLAogICAg
            ICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1
            dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3Mu
            Q2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZvbHVtZV9ncm91
            cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRf
            cmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNzX2tleSA9IHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAgICAgIHNlY3Jl
            dF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0KICAg
            ICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAg
            IGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIvYWRtaW4vaW5m
            bz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmVxdWVzdF91
            cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToKICAgICAgICAg
            ICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJsLAogICAgICAg
            ICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3X2VuZHBvaW50
            KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkKICAgICAgICBl
            eGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90IGFibGUgdG8g
            Y2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAgIGlmIHIxIGlz
            IE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAgICAgICAgKQog
            ICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICAgICAi
            LTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJzdG9yYWdlX2Jh
            Y2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X2VuZHBv
            aW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXIn
            IGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91
            dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVt
            bXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJndyBlbmRwb2lu
            dCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9IHNlbGYuX2lu
            dmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cg
            ZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBv
            aW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd190
            bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2VsZi5lbmRwb2lu
            dF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAgIGlmIGVyciAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBpZiB0aGUgcmd3
            IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMgb25seSBjaGVj
            ayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2FwX3N1cHBvcnRl
            ZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAgICByZ3dfZnNp
            ZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAgICAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAgICAgICAgICBp
            ZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAg
            ICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7c2VsZi5fYXJn
            X3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlkYXRpbmcgYnkg
            Y2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFuZCB2YWxpZGF0
            aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUgY2VwaCBjbHVz
            dGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9vbCBleGlzdAog
            ICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFzc2VkIGVsc2Ug
            aXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0aGVzZSBkZWZh
            dWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggIT0g
            ImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBbCiAgICAgICAg
            ICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3Lm1ldGEiLAog
            ICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAgICAgICAgZiJ7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAgICAgICAgICBd
            CiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bvb2xzX3RvX3Zh
            bGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0cyhf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2wsICd7X3Jnd19w
            b29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4gIiIKCiAgICBk
            ZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1l
            LCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0ZV9jb25maWcg
            IT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWlu
            IiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAgICAgICAgICAg
            ImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVfY29uZmlnLAog
            ICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAgICAgICAgXQog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgIGVycl9t
            c2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtj
            bWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNv
            ZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAg
            ICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAgZGVmIGNvbnZl
            cnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAgICByZXR1cm4g
            dmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5
            X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAgICgKICAgICAg
            ICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIKICAgICAgICAg
            ICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAgIG9yIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIiCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICAgICAg
            ICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxhZ3MgLS10b3Bv
            bG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10b3BvbG9neS1m
            YWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfdG9w
            b2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6CiAgICAgICAg
            aWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJv
            dmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIKICAgICAgICAg
            ICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9JywiCiAgICAg
            ICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3BvbG9neV9wb29s
            cyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAgICAgICAgICkK
            ICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYs
            IHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bv
            b2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgogICAgZGVmIGlu
            aXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAg
            Zm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLmluaXRfcmJk
            X3Bvb2wocG9vbCkKCiAgICAjIHRoaXMgd2lsbCByZXR1cm4gdGhlIGZpbmFsIGFyZ3MgdGhhdCBz
            Y3JpcHQgdXNlcyB0byBwcm9jZXNzCiAgICAjIHRoZSBwcmlvcml0eSB0byBzZXQgYSBwYXJ0aWN1
            bGFyIHZhbHVlIGlzLAogICAgIyBjb21tYW5kLWxpbmUtYXJncyA+IGNvbmZpZy5pbmkgZmlsZSB2
            YWx1ZXMgPiBkZWZhdWx0IHZhbHVlcwogICAgZGVmIGdldEZpbmFsVXNlZEFyZ3Moc2VsZik6CiAg
            ICAgICAgYXJndW1lbnQgPSBmIltDb25maWd1cmF0aW9uc11cbiIKICAgICAgICBmb3IgYXJnIGlu
            IHZhcnMoc2VsZi5fYXJnX3BhcnNlcik6CiAgICAgICAgICAgIGlmIHN0cihnZXRhdHRyKHNlbGYu
            X2FyZ19wYXJzZXIsIGFyZykpOgogICAgICAgICAgICAgICAgIyBweXRob24gdHJlYXRzIGZsYWct
            bmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28gY29udmVydGluZyBiYWNrIHRvIGZsYWct
            bmFtZSwKICAgICAgICAgICAgICAgICMgc28gd2UgY2FuIGdldCB0aG9zZSB2YWx1ZXMgZnJvbSBj
            b25maWcgZmlsZQogICAgICAgICAgICAgICAgYXJnVmFsdWUgPSBhcmcucmVwbGFjZSgiXyIsICIt
            IikKICAgICAgICAgICAgICAgICMgZG8gbm90IGFkZCB0aGUgY29uZmlnLWZpbGUgZmxhZyBhbmQg
            YWxzbyBub3QgYWRkIHRoZSBib29sZWFuIGZsYWdzIHdoaWNoIGFyZSBzZXQgdG8gRmFsc2UKICAg
            ICAgICAgICAgICAgICMgYmVjYXVzZSBjb25maWcuaW5pIGZpbGUgdHJlYXRzIGJvb2xlYW4gZmxh
            Z3MgYXMgVHJ1ZSBhbHdheXMKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBhcmdWYWx1ZSAhPSAiY29uZmlnLWZpbGUiCiAgICAgICAgICAgICAgICAgICAgYW5kIGdldGF0
            dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSAhPSBGYWxzZQogICAgICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgICAgICBhcmd1bWVudCArPSBmInthcmdWYWx1ZX0gPSB7c3RyKGdldGF0dHIo
            c2VsZi5fYXJnX3BhcnNlciwgYXJnKSl9XG4iCiAgICAgICAgcmV0dXJuIGFyZ3VtZW50CgogICAg
            ZGVmIF9nZW5fb3V0cHV0X21hcChzZWxmKToKICAgICAgICBpZiBzZWxmLm91dF9tYXA6CiAgICAg
            ICAgICAgIHJldHVybgogICAgICAgICMgc3VwcG9ydCBsZWdhY3kgZmxhZyB3aXRoIHVwZ3JhZGVz
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWU6CiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuY2x1c3Rl
            cl9uYW1lCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gKAogICAg
            ICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUubG93ZXIoKQogICAgICAg
            ICkgICMgYWx3YXlzIGNvbnZlcnQgY2x1c3RlciBuYW1lIHRvIGxvd2VyY2FzZSBjaGFyYWN0ZXJz
            CiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZSkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2woc2VsZi5fYXJnX3BhcnNlci5y
            YmRfZGF0YV9wb29sX25hbWUpCiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uo
            KQogICAgICAgIHNlbGYuX2V4Y2x1ZGVkX2tleXMuYWRkKCJLOFNfQ0xVU1RFUl9OQU1FIikKICAg
            ICAgICBzZWxmLmdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoKQogICAgICAgICMgZG91Ymxl
            IHN0cmluZyBuZWVkZWQgZm9yIHVwc3RyZWFtIGV4cG9ydHMgb2YgZmxhZ3MKICAgICAgICBzZWxm
            Lm91dF9tYXBbIkFSR1MiXSA9IGYnIntzZWxmLmdldEZpbmFsVXNlZEFyZ3MoKX0iJwogICAgICAg
            IHNlbGYub3V0X21hcFsiTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFyc2VyLm5hbWVzcGFjZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiSzhTX0NMVVNURVJfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0ZT
            SUQiXSA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSTkFNRSJdID0gc2VsZi5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0gPSBzZWxmLmdldF9jZXBoX2V4dGVybmFsX21vbl9k
            YXRhKCkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSA9
            IHNlbGYuY3JlYXRlX2NoZWNrZXJLZXkoCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FS
            RF9MSU5LIl0gPSBzZWxmLmdldF9jZXBoX2Rhc2hib2FyZF9saW5rKCkKICAgICAgICAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICBz
            ZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxm
            LmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1ub2RlIikKICAgICAg
            ICAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVu
            dC5jc2ktcmJkLXByb3Zpc2lvbmVyIikKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09M
            X05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDRVBIRlNfTUVUQURBVEFfUE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAg
            IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRhX3Bvb2xfbmFtZQogICAgICAgICkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVU
            SF9QRVJNSVNTSU9OIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3Rl
            ZF9hdXRoX3Blcm1pc3Npb24KICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQURPU19O
            QU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXSA9IHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1l
            X2dyb3VwCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0gPSAi
            IgogICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXSA9
            ICIiCiAgICAgICAgIyBjcmVhdGUgQ2VwaEZTIG5vZGUgYW5kIHByb3Zpc2lvbmVyIGtleXJpbmcg
            b25seSB3aGVuIE1EUyBleGlzdHMKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19O
            QU1FIl0gYW5kIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAo
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwK
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1F
            Il0sCiAgICAgICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGll
            bnQuY3NpLWNlcGhmcy1ub2RlIikKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLWNl
            cGhmcy1wcm92aXNpb25lciIpCiAgICAgICAgICAgICMgY3JlYXRlIHRoZSBkZWZhdWx0ICJjc2ki
            IHN1YnZvbHVtZWdyb3VwCiAgICAgICAgICAgIHNlbGYuZ2V0X29yX2NyZWF0ZV9zdWJ2b2x1bWVf
            Z3JvdXAoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmls
            ZXN5c3RlbV9uYW1lCiAgICAgICAgICAgICkKICAgICAgICAgICAgIyBwaW4gdGhlIGRlZmF1bHQg
            ImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5waW5fc3Vidm9sdW1lKAogICAg
            ICAgICAgICAgICAgImNzaSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSwgImRpc3RyaWJ1dGVkIiwgIjEiCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXToKICAgICAgICAgICAgICAgIHNlbGYuZ2V0X29yX2Ny
            ZWF0ZV9zdWJ2b2x1bWVfZ3JvdXAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAg
            c2VsZi5waW5fc3Vidm9sdW1lKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            c3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAiZGlzdHJpYnV0ZWQiLAog
            ICAgICAgICAgICAgICAgICAgICIxIiwKICAgICAgICAgICAgICAgICkKICAgICAgICBzZWxmLm91
            dF9tYXBbIlJHV19UTFNfQ0VSVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJ
            TkdfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBP
            SU5UX1BPUlQiXSA9ICIiCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIuc2tpcF9tb25p
            dG9yaW5nX2VuZHBvaW50OgogICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIk1PTklUT1JJTkdfRU5EUE9JTlQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsi
            TU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICkgPSBzZWxmLmdldF9hY3Rp
            dmVfYW5kX3N0YW5kYnlfbWdycygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1F
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoKQogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzCiAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJd
            ID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWlu
            X2xhYmVsCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lf
            ZmFpbHVyZV9kb21haW5fdmFsdWVzCiAgICAgICAgKQogICAgICAgIGlmICgKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICBhbmQgc2Vs
            Zi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAgICAg
            ICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMg
            IT0gIiIKICAgICAgICApOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3ZhbHVl
            cygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXko
            c2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKSwKICAgICAgICAgICAgICAgIHNlbGYuY29u
            dmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXQogICAgICAgICAgICAgICAg
            KSwKICAgICAgICAgICAgKQogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3JiZF9w
            b29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJy
            YXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIHNlbGYuaW5pdF90b3BvbG9neV9yYmRfcG9vbHMoCiAgICAgICAgICAgICAgICBzZWxmLmNv
            bnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSkKICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHNlbGYucmFpc2Vf
            ZXhjZXB0aW9uX2lmX2FueV90b3BvbG9neV9mbGFnX2lzX21pc3NpbmcoKQoKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJHV19QT09MX1BSRUZJWCJdID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXgKICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gIiIKICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludDoKICAgICAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICAgICAgc2VsZi5jcmVhdGVfcmd3X2FkbWluX29w
            c191c2VyKCkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAh
            PSAiIgogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZh
            bGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUsICJyZWFsbSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwgInpv
            bmVncm91cCIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgZXJyID0g
            c2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsICJ6b25lIgogICAgICAgICAgICAgICAgICAgICkK
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgICkgb3Ig
            KAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgIT0g
            IiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3Vw
            X25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgICgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19B
            RE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIGluZm9f
            Y2FwX3N1cHBvcnRlZCwKICAgICAgICAgICAgICAgICAgICAgICAgZXJyLAogICAgICAgICAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICAg
            ICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50KGluZm9fY2FwX3N1cHBvcnRl
            ZCkKICAgICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9w
            YXRoOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9p
            bnRfdGxzX2NlcnQoKQogICAgICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciwgc2V0IHRoZSBSR1dfRU5EUE9JTlQKICAgICAgICAg
            ICAgICAgICAgICBpZiBlcnIgIT0gIi0xIjoKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJSR1dfRU5EUE9JTlQiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIGVyciA9ICJQbGVhc2UgcHJv
            dmlkZSBhbGwgdGhlIFJHVyBtdWx0aXNpdGUgcGFyYW1ldGVycyBvciBub25lIG9mIHRoZW0iCiAg
            ICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnIpCgogICAgZGVmIGdlbl9zaGVs
            bF9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAgICAgIHNoT3V0
            SU8gPSBTdHJpbmdJTygpCiAgICAgICAgZm9yIGssIHYgaW4gc2VsZi5vdXRfbWFwLml0ZW1zKCk6
            CiAgICAgICAgICAgIGlmIHYgYW5kIGsgbm90IGluIHNlbGYuX2V4Y2x1ZGVkX2tleXM6CiAgICAg
            ICAgICAgICAgICBzaE91dElPLndyaXRlKGYiZXhwb3J0IHtrfT17dn17TElORVNFUH0iKQogICAg
            ICAgIHNoT3V0ID0gc2hPdXRJTy5nZXR2YWx1ZSgpCiAgICAgICAgc2hPdXRJTy5jbG9zZSgpCiAg
            ICAgICAgcmV0dXJuIHNoT3V0CgogICAgZGVmIGdlbl9qc29uX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICBqc29uX291dCA9IFsKICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgIm5hbWUiOiAiZXh0ZXJuYWwtY2x1c3Rlci11c2VyLWNvbW1hbmQi
            LAogICAgICAgICAgICAgICAgImtpbmQiOiAiQ29uZmlnTWFwIiwKICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICJhcmdzIjogc2VsZi5vdXRfbWFwWyJBUkdTIl0s
            CiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAg
            ICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uLWVuZHBvaW50cyIsCiAgICAgICAgICAgICAg
            ICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAg
            ICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05f
            REFUQSJdLAogICAgICAgICAgICAgICAgICAgICJtYXhNb25JZCI6ICIwIiwKICAgICAgICAgICAg
            ICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9
            LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwK
            ICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4tc2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAg
            ICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJ
            RCJdLAogICAgICAgICAgICAgICAgICAgICJtb24tc2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9yLWNyZWRzIiwKICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxf
            VVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAg
            XQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBh
            ZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENsdXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAg
            ICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBz
            ZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9pbnQiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdFbmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAgICAgICAiTW9uaXRvcmluZ1Bv
            cnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgIyBp
            ZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1y
            YmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lf
            UkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJLZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAg
            ICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JC
            RF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJk
            LXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNS
            RVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJJRCI6
            IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQog
            ICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQn
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNy
            ZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSdd
            fSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAg
            ICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lP
            TkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAg
            ICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfQ0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMs
            IHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhmcy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAg
            ICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewog
            ICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBI
            RlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2Vj
            cmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAg
            ICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAg
            ICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRf
            TElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNl
            Y3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElO
            SyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxp
            bmsiLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0Un
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJBRE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAg
            ICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAg
            ICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0K
            ICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAg
            ICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJyYWRv
            c05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAg
            ICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09M
            X05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92
            aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9y
            YWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5v
            dXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            IGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xf
            TkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6
            IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXIt
            c2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9T
            RUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9t
            YXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICAgICAgfQogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAog
            ICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2Vw
            aC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1l
            IjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUn
            XX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgICAgICkKCiAgICAg
            ICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwn
            LCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhl
            biBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VDbGFzcwogICAgICAgIGlmICgKICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91
            dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTIl0KICAgICAgICApOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9wb2xvZ3kiLAogICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYu
            b3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RP
            TUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAgICAgICAgXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZhbHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAg
            ICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAg
            ICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhw
            YW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycgU3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2Vs
            Zi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwK
            ICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5v
            dXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6
            IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAg
            ICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtz
            ZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAg
            ICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5k
            LXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBz
            ZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ci
            LAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJlbmRwb2ludCI6IHNl
            bGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            UHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZ
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJzZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2VwaC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0
            LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJj
            ZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImNl
            cnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdLAogICAgICAgICAgICAgICAgICAgIH0s
            CiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVt
            cHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRlZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25z
            KHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMt
            bm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAg
            ICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwKICAgICAgICAgICAgImNsaWVudC5jc2ktcmJk
            LXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAg
            ICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNlciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNl
            ciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVzZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2Vy
            KQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgogICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNl
            cl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBnZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBn
            cmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAgICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRo
            Y2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBhZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBw
            b29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAgICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBu
            YW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIgdXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcg
            Y29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdldCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMg
            cmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAg
            ICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0oLio/KVwucmd3XC5tZXRhIgogICAgICAgICAg
            ICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVybiwgY2FwcykKICAgICAgICAgICAgICAgIGlm
            IG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wgbmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAg
            ICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2Vy
            KToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhlIGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAog
            ICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNl
            cn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNn
            ID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBwcmludChmInVzZXIge3Vz
            ZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIpCiAgICAgICAgICAgIHJldHVybgogICAgICAg
            IGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsiY2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3
            X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2VyLCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAg
            ICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNh
            cF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIsICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQog
            ICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlzOgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVl
            cyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBuZXdfY2FwX3Zh
            bHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBjdXJfY2FwX3Blcm1f
            bGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVl
            cy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBdCiAgICAgICAgICAgIG5ld19j
            YXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAgICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19j
            YXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJpcCgpCiAgICAgICAgICAgIF0KICAgICAgICAg
            ICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUg
            b3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2Nh
            cF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxpbWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1
            c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCByZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3Qg
            YW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIKICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0g
            W10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5hcHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9w
            ZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9saXN0XQogICAgICAgICAgICBleGlzdGluZ19j
            YXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19jYXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhp
            c3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAgICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXAp
            CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChleGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAg
            ICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1dGggY2FwcyIsCiAgICAg
            ICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAgICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAg
            ICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQogICAgICAgIHJldF92YWwsIGpzb25fb3V0
            LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBp
            ZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNhcHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5c
            biBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVk
            IHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoKICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAg
            IGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFk
            ZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBl
            bGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJh
            dGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19w
            YXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNl
            bGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0
            OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBzZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0
            ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0gb3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2Rl
            PSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAgICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRf
            b3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3NlKCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19tYWluX18iOgogICAgcmpPYmogPSBSYWRvc0pT
            T04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1haW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAgIHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDog
            e2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAgZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAg
            ICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9z
            RXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hpbGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0
            YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAgICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:35e88716ace3576683a9c4aa686df5f7a8b5bb6ee7a144fd2d87e3c9f5bf3961
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        version: 4.17.13-rhodf
      entries:
      - name: rook-ceph-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: rook-ceph-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: rook-ceph-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: rook-ceph-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: rook-ceph-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: rook-ceph-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: rook-ceph-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: rook-ceph-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: rook-ceph-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: rook-ceph-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: rook-ceph-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: rook-ceph-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: rook-ceph-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: rook-ceph-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: rook-ceph-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9VHJ1ZSwgICMgZGVmYXVsdCBpcyBUcnVlIHRvIGVuYWJsZSB2MiBtb24g
            cG9ydCBpbiBkb3duc3RyZWFtCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9ydCgz
            MzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQLmFk
            ZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAgICAg
            ICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24iLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJhc2gp
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3ZpZGVk
            IGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmlsZXN5
            c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAg
            ICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZzIG1l
            dGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50
            KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRl
            cyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9vbC1u
            YW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2Us
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRhIHBv
            b2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBpbiB0
            aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAgKQog
            ICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1lbmRw
            b2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+OjxQ
            T1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAg
            IHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50
            IFRMUyBjZXJ0aWZpY2F0ZSAob3IgaW50ZXJtZWRpYXRlIHNpZ25pbmcgY2VydGlmaWNhdGUpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAg
            IGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBUTFMgY2VydGlmaWNhdGlv
            biB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0ZSBpcyBwcm92aWRlZCAo
            Tk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRl
            ZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJD
            ZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMgKGNvbW1hIHNlcGFyYXRl
            ZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBvciBgPEZRRE4+YCkgZW50
            cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAgICkKICAgICAgICBvdXRw
            dXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25pdG9yaW5nLWVuZHBvaW50
            LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgcG9y
            dCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFs
            c2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRv
            IG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0aGUgQ2VwaCBjbHVzdGVy
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0
            aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29sIiwKICAgICAgICApCiAg
            ICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tZHJ5LXJ1biIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1
            ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBleGVjdXRlZCBjb21tYW5k
            cyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAg
            IGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxw
            PSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFtZXNwYWNlcywgdXNlZCBm
            b3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSAoc2hv
            dWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwKICAgICAgICAgICAgZGVm
            YXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9InBy
            b3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAogICAgICAgICkKICAgICAg
            ICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0tbmFt
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3LXJlYWxtIiwKICAg
            ICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0t
            cmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3
            LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhl
            IG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1wb29scyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25zdHJhaW5lZCByYmQgcG9v
            bHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAg
            ICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJrOHMg
            Y2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9uZSwgcmFjaywgb3IgaG9z
            dCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBjZXBoIGRvbWFpbiIsCiAg
            ICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICIt
            LXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBh
            cmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gdmFsdWVzIGNvcnJl
            c3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3BvbG9neS1wb29sc2AgbGlz
            dCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJnUC5hZGRfYXJndW1lbnRf
            Z3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAg
            ICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iVXBncmFkZXMgdGhlIGNl
            cGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIp
            IGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGggbmV3IHBlcm1pc3Npb25z
            IG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9sZGVyIHBlcm1pc3Npb24g
            d2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhv
            bjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3Jh
            ZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3NpIHVzZXJzKG5vbi1yZXN0
            cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNlcnMoRm9yIGV4YW1wbGU6
            IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMpLCB1
            c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24gZmxhZyBuZWVk
            IHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJtYW5kYXRvcnkgZmxhZ3M6
            ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1lIGFuZCAtLXJ1bi1hcy11
            c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsgImluIGNhc2Ugb2YgY2Vw
            aGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUgZmxh
            ZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5nIGl0IHdpbGwgYmUgbWFu
            ZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBo
            L2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3JhZGUgLS1yYmQtZGF0
            YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlICAt
            LXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3JhZ2UtcmVwbGljYXBvb2xg
            IgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3RyaWN0ZWQgdXNlciBjYW5u
            b3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVwZ3JhZGluZy4iCiAgICAg
            ICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2VkIHRvIGFwcGVuZCBuZXcg
            cGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2VkIGZvciBjaGFuZ2luZyB1
            c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBsZSB5b3Ugc2hvdWxkbid0
            IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAgICAgICAgKQoKICAgICAg
            ICAjIEFkZCBjb21tYW5kLWxpbmUgYXJndW1lbnRzCiAgICAgICAgY29uZmlnX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoImNvbmZpZyIpCiAgICAgICAgY29uZmlnX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY29uZmlnLWZpbGUiLAogICAgICAgICAgICB0eXBlPXN0
            ciwKICAgICAgICAgICAgaGVscD0iUGF0aCB0byB0aGUgY29uZmlndXJhdGlvbiBmaWxlLCBQcmlv
            cml0eTogY29tbWFuZC1saW5lLWFyZ3MgPiBjb25maWcuaW5pIGZpbGUgdmFsdWVzID4gZGVmYXVs
            dCB2YWx1ZXMiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICApCgogICAgICAgIGlmIGFyZ3NfdG9fcGFyc2U6CiAgICAgICAgICAgIGFz
            c2VydCAoCiAgICAgICAgICAgICAgICB0eXBlKGFyZ3NfdG9fcGFyc2UpID09IGxpc3QKICAgICAg
            ICAgICAgKSwgIkFyZ3VtZW50IHRvICdnZW5fYXJnX3BhcnNlcicgc2hvdWxkIGJlIGEgbGlzdCIK
            ICAgICAgICBlbHNlOgogICAgICAgICAgICBhcmdzX3RvX3BhcnNlID0gc3lzLmFyZ3ZbMTpdCiAg
            ICAgICAgcmV0dXJuIGFyZ1AucGFyc2VfYXJncyhhcmdzX3RvX3BhcnNlKQoKICAgIGRlZiB2YWxp
            ZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xf
            bmFtZQogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0
            YV9wb29sX25hbWUKCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJGbGFnICctLXJiZC1kYXRhLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAg
            ICAgICAgICAgICAgICkKCiAgICAgICAgICAgIGlmIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUg
            PT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAog
            ICAgICAgICAgICAgICAgICAgICJGbGFnICctLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBjbWRfanNv
            biA9IHsicHJlZml4IjogIm9zZCBkdW1wIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAg
            cmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNt
            ZF9qc29uKQogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAw
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICBmIntjbWRfanNvblsncHJlZml4J119IGNvbW1hbmQgZmFpbGVkLlxuIgog
            ICAgICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNl
            IHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBt
            ZXRhZGF0YV9wb29sX2V4aXN0LCBwb29sX2V4aXN0ID0gRmFsc2UsIEZhbHNlCgogICAgICAgICAg
            ICBmb3Iga2V5IGluIGpzb25fb3V0WyJwb29scyJdOgogICAgICAgICAgICAgICAgIyBpZiBlcmFz
            dXJlX2NvZGVfcHJvZmlsZSBpcyBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IHJl
            cGxpY2EgcG9vbAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGtleVsi
            ZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBrZXlb
            InBvb2xfbmFtZSJdID09IHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgICAg
            ICk6CiAgICAgICAgICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCA9IFRydWUKICAgICAg
            ICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgbm90IGVtcHR5IGFuZCBwb29s
            IG5hbWUgZXhpc3RzIHRoZW4gaXQgaXMgZWMgcG9vbAogICAgICAgICAgICAgICAgaWYga2V5WyJl
            cmFzdXJlX2NvZGVfcHJvZmlsZSJdIGFuZCBrZXlbInBvb2xfbmFtZSJdID09IHJiZF9wb29sX25h
            bWU6CiAgICAgICAgICAgICAgICAgICAgcG9vbF9leGlzdCA9IFRydWUKCiAgICAgICAgICAgIGlm
            IG5vdCBtZXRhZGF0YV9wb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9u
            RmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiUHJvdmlkZWQgcmJkX2VjX21l
            dGFkYXRhX3Bvb2wgbmFtZSwiCiAgICAgICAgICAgICAgICAgICAgZiIge3JiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgaWYgbm90IHBvb2xfZXhpc3Q6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgcmJkX2RhdGFfcG9v
            bCBuYW1lLCB7cmJkX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICByZXR1cm4gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQoKICAgIGRlZiBk
            cnlfcnVuKHNlbGYsIG1zZyk6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICBwcmludCgiRXhlY3V0ZTogIiArICInIiArIG1zZyArICInIikKCiAgICBkZWYg
            dmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgIHdpdGggb3BlbihzZWxmLl9h
            cmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoLCBlbmNvZGluZz0idXRmOCIpIGFzIGY6CiAgICAg
            ICAgICAgICAgICBjb250ZW50cyA9IGYucmVhZCgpCiAgICAgICAgICAgICAgICByZXR1cm4gY29u
            dGVudHMucnN0cmlwKCkKCiAgICBkZWYgX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoc2VsZik6
            CiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZSBhbmQgbm90IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIkVpdGhlciAnLS11cGdyYWRlJyBvciAn
            LS1yYmQtZGF0YS1wb29sLW5hbWUgPHBvb2xfbmFtZT4nIHNob3VsZCBiZSBzcGVjaWZpZWQiCiAg
            ICAgICAgICAgICkKCiAgICBkZWYgX2ludmFsaWRfZW5kcG9pbnQoc2VsZiwgZW5kcG9pbnRfc3Ry
            KToKICAgICAgICAjIGV4dHJhY3QgdGhlIHBvcnQgYnkgZ2V0dGluZyB0aGUgbGFzdCBzcGxpdCBv
            biBgOmAgZGVsaW1pdGVyCiAgICAgICAgdHJ5OgogICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAs
            IHBvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICBleGNlcHQgVmFsdWVF
            cnJvcjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk5vdCBh
            IHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iKQoKICAgICAgICB0cnk6CiAgICAgICAg
            ICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSA9PSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2lu
            dF9zdHJfaXAgPSBlbmRwb2ludF9zdHJfaXBbMSA6IGxlbihlbmRwb2ludF9zdHJfaXApIC0gMV0K
            ICAgICAgICAgICAgaXBfdHlwZSA9ICgKICAgICAgICAgICAgICAgICJJUHY0IiBpZiB0eXBlKGlw
            X2FkZHJlc3MoZW5kcG9pbnRfc3RyX2lwKSkgaXMgSVB2NEFkZHJlc3MgZWxzZSAiSVB2NiIKICAg
            ICAgICAgICAgKQogICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICBpcF90eXBl
            ID0gIkZRRE4iCiAgICAgICAgaWYgbm90IHBvcnQuaXNkaWdpdCgpOgogICAgICAgICAgICByYWlz
            ZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiUG9ydCBub3QgdmFsaWQ6IHtwb3J0fSIpCiAg
            ICAgICAgaW50UG9ydCA9IGludChwb3J0KQogICAgICAgIGlmIGludFBvcnQgPCAxIG9yIGludFBv
            cnQgPiAyKioxNiAtIDE6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZiJPdXQgb2YgcmFuZ2UgcG9ydCBudW1iZXI6IHtwb3J0fSIpCgogICAgICAgIHJldHVybiBp
            cF90eXBlCgogICAgZGVmIGVuZHBvaW50X2RpYWwoc2VsZiwgZW5kcG9pbnRfc3RyLCBpcF90eXBl
            LCB0aW1lb3V0PTMsIGNlcnQ9Tm9uZSk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXInIGluc3Rh
            bmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91dCB0byB0
            aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlSYWRv
            cyk6CiAgICAgICAgICAgIHJldHVybiAiIiwgIiIsICIiCiAgICAgICAgaWYgaXBfdHlwZSA9PSAi
            SVB2NiI6CiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCwg
            ZW5kcG9pbnRfc3RyX3BvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICAg
            ICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiTm90IGEgcHJvcGVyIGVuZHBvaW50
            OiB7ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgZW5kcG9p
            bnRfc3RyX2lwWzBdICE9ICJbIjoKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9ICJb
            IiArIGVuZHBvaW50X3N0cl9pcCArICJdIgogICAgICAgICAgICBlbmRwb2ludF9zdHIgPSAiOiIu
            am9pbihbZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydF0pCgogICAgICAgIHByb3Rv
            Y29scyA9IFsiaHR0cCIsICJodHRwcyJdCiAgICAgICAgcmVzcG9uc2VfZXJyb3IgPSBOb25lCiAg
            ICAgICAgZm9yIHByZWZpeCBpbiBwcm90b2NvbHM6CiAgICAgICAgICAgIHRyeToKICAgICAgICAg
            ICAgICAgIGVwID0gZiJ7cHJlZml4fTovL3tlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAg
            dmVyaWZ5ID0gTm9uZQogICAgICAgICAgICAgICAgIyBJZiB2ZXJpZnkgaXMgc2V0IHRvIGEgcGF0
            aCB0byBhIGRpcmVjdG9yeSwKICAgICAgICAgICAgICAgICMgdGhlIGRpcmVjdG9yeSBtdXN0IGhh
            dmUgYmVlbiBwcm9jZXNzZWQgdXNpbmcgdGhlIGNfcmVoYXNoIHV0aWxpdHkgc3VwcGxpZWQgd2l0
            aCBPcGVuU1NMLgogICAgICAgICAgICAgICAgaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3NraXBfdGxzOgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9IEZh
            bHNlCiAgICAgICAgICAgICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGlt
            ZW91dCwgdmVyaWZ5PUZhbHNlKQogICAgICAgICAgICAgICAgZWxpZiBwcmVmaXggPT0gImh0dHBz
            IiBhbmQgY2VydDoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBjZXJ0CiAgICAgICAgICAg
            ICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCwgdmVyaWZ5PWNl
            cnQpCiAgICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQpCiAgICAgICAgICAgICAgICBpZiByLnN0YXR1c19j
            b2RlID09IDIwMDoKICAgICAgICAgICAgICAgICAgICByZXR1cm4gcHJlZml4LCB2ZXJpZnksICIi
            CiAgICAgICAgICAgIGV4Y2VwdCBFeGNlcHRpb24gYXMgZXJyOgogICAgICAgICAgICAgICAgcmVz
            cG9uc2VfZXJyb3IgPSBlcnIKICAgICAgICAgICAgICAgIGNvbnRpbnVlCiAgICAgICAgc3lzLnN0
            ZGVyci53cml0ZSgKICAgICAgICAgICAgZiJ1bmFibGUgdG8gY29ubmVjdCB0byBlbmRwb2ludDog
            e2VuZHBvaW50X3N0cn0sIGZhaWxlZCBlcnJvcjoge3Jlc3BvbnNlX2Vycm9yfSIKICAgICAgICAp
            CiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICIiLAogICAgICAg
            ICAgICAoIi0xIiksCiAgICAgICAgKQoKICAgIGRlZiBwYXJzZV9jb25maWdfZmlsZShzZWxmLCBj
            b25maWdfZmlsZSk6CiAgICAgICAgY29uZmlnID0gY29uZmlncGFyc2VyLkNvbmZpZ1BhcnNlcigp
            CiAgICAgICAgY29uZmlnLnJlYWQoY29uZmlnX2ZpbGUpCiAgICAgICAgZm9yIGFyZyBpbiBsaXN0
            KHZhcnMoc2VsZi5fYXJnX3BhcnNlcikpOgogICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxh
            Zy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRlcm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxh
            Zy1uYW1lLAogICAgICAgICAgICAjIHNvIHdlIGNhbiBnZXQgdGhvc2UgdmFsdWVzIGZyb20gY29u
            ZmlnIGZpbGUKICAgICAgICAgICAgYXJndW1lbnQgPSBhcmcucmVwbGFjZSgiXyIsICItIikKICAg
            ICAgICAgICAgYXJndW1lbnRWYWx1ZSA9IHN0cihnZXRhdHRyKHNlbGYuX2FyZ19wYXJzZXIsIGFy
            ZykpCiAgICAgICAgICAgIGNvbmZpZ192YWx1ZSA9IGNvbmZpZy5nZXQoIkNvbmZpZ3VyYXRpb25z
            IiwgYXJndW1lbnQsIGZhbGxiYWNrPU5vbmUpCiAgICAgICAgICAgICMgZ2l2ZSBwcmlvcml0eSB0
            byBjb21tYW5kIGxpbmUgYXJndW1lbnQsIGlmIGNvbW1hbmQgbGluZSBhcmd1bWVudCBpcyBub3Qg
            cHJlc2VudCB1c2UgY29uZmlnLmluaSB2YWx1ZSwKICAgICAgICAgICAgaWYgKHN0cihzeXMuYXJn
            dikuZmluZChhcmd1bWVudCkgPT0gLTEpIGFuZCAoCiAgICAgICAgICAgICAgICAoY29uZmlnX3Zh
            bHVlICE9IE5vbmUpIGFuZCAoY29uZmlnX3ZhbHVlICE9ICIiKQogICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5fX3NldGF0dHJfXyhhcmcsIGNvbmZpZ192YWx1
            ZSkKCiAgICAgICAgcmV0dXJuIGNvbmZpZwoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhcmdfbGlz
            dD1Ob25lKToKICAgICAgICBzZWxmLm91dF9tYXAgPSB7fQogICAgICAgIHNlbGYuX2V4Y2x1ZGVk
            X2tleXMgPSBzZXQoKQogICAgICAgIHNlbGYuX2FyZ19wYXJzZXIgPSBzZWxmLmdlbl9hcmdfcGFy
            c2VyKGFyZ3NfdG9fcGFyc2U9YXJnX2xpc3QpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            b25maWdfZmlsZToKICAgICAgICAgICAgc2VsZi5jb25maWcgPSBzZWxmLnBhcnNlX2NvbmZpZ19m
            aWxlKHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGUpCiAgICAgICAgc2VsZi5fY2hlY2tfY29u
            ZmxpY3Rpbmdfb3B0aW9ucygpCiAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuX2FyZ19w
            YXJzZXIucnVuX2FzX3VzZXIKICAgICAgICBzZWxmLm91dHB1dF9maWxlID0gc2VsZi5fYXJnX3Bh
            cnNlci5vdXRwdXQKICAgICAgICBzZWxmLmNlcGhfY29uZiA9IHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aF9jb25mCiAgICAgICAgc2VsZi5jZXBoX2tleXJpbmcgPSBzZWxmLl9hcmdfcGFyc2VyLmtleXJp
            bmcKICAgICAgICAjIGlmIHVzZXIgbm90IHByb3ZpZGVkLCBnaXZlIGEgZGVmYXVsdCB1c2VyCiAg
            ICAgICAgaWYgbm90IHNlbGYucnVuX2FzX3VzZXIgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGU6CiAgICAgICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLkVYVEVSTkFMX1VTRVJf
            TkFNRQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCBhbmQg
            bm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggPSBzZWxmLkRFRkFVTFRfUkdXX1BPT0xfUFJFRklYCiAgICAgICAg
            aWYgc2VsZi5jZXBoX2NvbmY6CiAgICAgICAgICAgIGt3YXJncyA9IHt9CiAgICAgICAgICAgIGlm
            IHNlbGYuY2VwaF9rZXlyaW5nOgogICAgICAgICAgICAgICAga3dhcmdzWyJjb25mIl0gPSB7Imtl
            eXJpbmciOiBzZWxmLmNlcGhfa2V5cmluZ30KICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFk
            b3MuUmFkb3MoY29uZmZpbGU9c2VsZi5jZXBoX2NvbmYsICoqa3dhcmdzKQogICAgICAgIGVsc2U6
            CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKCkKICAgICAgICAgICAgc2Vs
            Zi5jbHVzdGVyLmNvbmZfcmVhZF9maWxlKCkKICAgICAgICBzZWxmLmNsdXN0ZXIuY29ubmVjdCgp
            CgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIGlmIHNlbGYuY2x1c3Rlci5zdGF0ZSA9
            PSAiY29ubmVjdGVkIjoKICAgICAgICAgICAgc2VsZi5jbHVzdGVyLnNodXRkb3duKCkKCiAgICBk
            ZWYgZ2V0X2ZzaWQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoIGZzaWQiKQogICAgICAgIHJldHVy
            biBzdHIoc2VsZi5jbHVzdGVyLmdldF9mc2lkKCkpCgogICAgZGVmIF9jb21tb25fY21kX2pzb25f
            Z2VuKHNlbGYsIGNtZF9qc29uKToKICAgICAgICBjbWQgPSBqc29uLmR1bXBzKGNtZF9qc29uLCBz
            b3J0X2tleXM9VHJ1ZSkKICAgICAgICByZXRfdmFsLCBjbWRfb3V0LCBlcnJfbXNnID0gc2VsZi5j
            bHVzdGVyLm1vbl9jb21tYW5kKGNtZCwgYiIiKQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            dmVyYm9zZToKICAgICAgICAgICAgcHJpbnQoZiJDb21tYW5kIElucHV0OiB7Y21kfSIpCiAgICAg
            ICAgICAgIHByaW50KAogICAgICAgICAgICAgICAgZiJSZXR1cm4gVmFsOiB7cmV0X3ZhbH1cbkNv
            bW1hbmQgT3V0cHV0OiB7Y21kX291dH1cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3IgTWVzc2Fn
            ZToge2Vycl9tc2d9XG4tLS0tLS0tLS0tXG4iCiAgICAgICAgICAgICkKICAgICAgICBqc29uX291
            dCA9IHt9CiAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciAoaS5lOyByZXRfdmFsIGlzIFpF
            Uk8pIGFuZCAnY21kX291dCcgaXMgbm90IGVtcHR5CiAgICAgICAgIyB0aGVuIGNvbnZlcnQgJ2Nt
            ZF9vdXQnIHRvIGEganNvbiBvdXRwdXQKICAgICAgICBpZiByZXRfdmFsID09IDAgYW5kIGNtZF9v
            dXQ6CiAgICAgICAgICAgIGpzb25fb3V0ID0ganNvbi5sb2FkcyhjbWRfb3V0KQogICAgICAgIHJl
            dHVybiByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZwoKICAgIGRlZiBnZXRfY2VwaF9leHRlcm5h
            bF9tb25fZGF0YShzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogInF1b3J1bV9z
            dGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5
            X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25b
            InByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29t
            bW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3Vj
            Y2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQp
            ID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAiJ3F1b3J1bV9zdGF0dXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAg
            ICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9P
            VVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICBxX2xlYWRlcl9uYW1lID0ganNvbl9v
            dXRbInF1b3J1bV9sZWFkZXJfbmFtZSJdCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHt9CiAg
            ICAgICAgcV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCA9IFsKICAgICAgICAgICAgbCBmb3IgbCBpbiBq
            c29uX291dFsibW9ubWFwIl1bIm1vbnMiXSBpZiBsWyJuYW1lIl0gPT0gcV9sZWFkZXJfbmFtZQog
            ICAgICAgIF0KICAgICAgICBpZiBsZW4ocV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCkgPT0gMDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gbWF0Y2hpbmcgJ21v
            bicgZGV0YWlscyBmb3VuZCIpCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3RbMF0KICAgICAgICAjIGdldCB0aGUgYWRkcmVzcyB2ZWN0b3Igb2YgdGhlIHF1
            b3J1bS1sZWFkZXIKICAgICAgICBxX2xlYWRlcl9hZGRydmVjID0gcV9sZWFkZXJfZGV0YWlscy5n
            ZXQoInB1YmxpY19hZGRycyIsIHt9KS5nZXQoImFkZHJ2ZWMiLCBbXSkKICAgICAgICBpcF9hZGRy
            ID0gc3RyKHFfbGVhZGVyX2RldGFpbHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYg
            cV9sZWFkZXJfYWRkcnZlY1swXVsidHlwZSJdID09ICJ2MiI6CiAgICAgICAgICAgICAgICBpcF9h
            ZGRyID0gcV9sZWFkZXJfYWRkcnZlY1swXVsiYWRkciJdCiAgICAgICAgICAgIGVsaWYgbGVuKHFf
            bGVhZGVyX2FkZHJ2ZWMpID4gMSBhbmQgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09ICJ2
            MiI6CiAgICAgICAgICAgICAgICBpcF9hZGRyID0gcV9sZWFkZXJfYWRkcnZlY1sxXVsiYWRkciJd
            CiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAg
            ICAgICAgICAgICAgICAgICIndjInIGFkZHJlc3MgdHlwZSBub3QgcHJlc2VudCwgYW5kICd2Mi1w
            b3J0LWVuYWJsZScgZmxhZyBpcyBwcm92aWRlZCIKICAgICAgICAgICAgICAgICkKCiAgICAgICAg
            cmV0dXJuIGYie3N0cihxX2xlYWRlcl9uYW1lKX09e2lwX2FkZHJ9IgoKICAgIGRlZiBfY29udmVy
            dF9ob3N0bmFtZV90b19pcChzZWxmLCBob3N0X25hbWUsIHBvcnQsIGlwX3R5cGUpOgogICAgICAg
            ICMgaWYgJ2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgdHlwZSwKICAgICAgICAjIGNhbGwg
            dGhlIGR1bW15IGluc3RhbmNlJ3MgImNvbnZlcnQiIG1ldGhvZAogICAgICAgIGlmIG5vdCBob3N0
            X25hbWU6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIkVtcHR5
            IGhvc3RuYW1lIHByb3ZpZGVkIikKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3Rlciwg
            RHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybiBzZWxmLmNsdXN0ZXIuX2NvbnZlcnRfaG9z
            dG5hbWVfdG9faXAoaG9zdF9uYW1lKQoKICAgICAgICBpZiBpcF90eXBlID09ICJGUUROIjoKICAg
            ICAgICAgICAgIyBjaGVjayB3aGljaCBpcCBGUUROIHNob3VsZCBiZSBjb252ZXJ0ZWQgdG8sIElQ
            djQgb3IgSVB2NgogICAgICAgICAgICAjIGNoZWNrIHRoZSBob3N0IGlwLCB0aGUgZW5kcG9pbnQg
            aXAgdHlwZSB3b3VsZCBiZSBzaW1pbGFyIHRvIGhvc3QgaXAKICAgICAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJvcmNoIGhvc3QgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVt
            cHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICInb3JjaCBob3N0IGxzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAg
            ICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZ
            X09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaG9zdF9hZGRyID0g
            anNvbl9vdXRbMF1bImFkZHIiXQogICAgICAgICAgICAjIGFkZCA6ODAgc2FtcGxlIHBvcnQgaW4g
            aXBfdHlwZSwgYXMgX2ludmFsaWRfZW5kcG9pbnQgYWxzbyB2ZXJpZnkgcG9ydAogICAgICAgICAg
            ICBob3N0X2lwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KGhvc3RfYWRkciArICI6ODAi
            KQogICAgICAgICAgICBpbXBvcnQgc29ja2V0CgogICAgICAgICAgICBpcCA9IFtdCiAgICAgICAg
            ICAgICMgZXhhbXBsZSBvdXRwdXQgWyg8QWRkcmVzc0ZhbWlseS5BRl9JTkVUOiAyPiwgPFNvY2tl
            dEtpbmQuU09DS19TVFJFQU06IDE+LCA2LCAnJywgKCc5My4xODQuMjE2LjM0JywgODApKSwgLi4u
            XQogICAgICAgICAgICAjIHdlIG5lZWQgdG8gZ2V0IDkzLjE4NC4yMTYuMzQgc28gaXQgd291bGQg
            YmUgaXBbMF1bNF1bMF0KICAgICAgICAgICAgaWYgaG9zdF9pcF90eXBlID09ICJJUHY2IjoKICAg
            ICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAgICAgICAgICAgICAgICAg
            IGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVUNiwgcHJvdG89c29ja2V0LklQ
            UFJPVE9fVENQCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsaWYgaG9zdF9pcF90eXBl
            ID09ICJJUHY0IjoKICAgICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAg
            ICAgICAgICAgICAgICAgIGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVULCBw
            cm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZGVs
            IHNvY2tldAogICAgICAgICAgICByZXR1cm4gaXBbMF1bNF1bMF0KICAgICAgICByZXR1cm4gaG9z
            dF9uYW1lCgogICAgZGVmIGdldF9hY3RpdmVfYW5kX3N0YW5kYnlfbWdycyhzZWxmKToKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIiwgc2Vs
            Zi5kcnlfcnVuKCJjZXBoIHN0YXR1cyIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0
            ID0gc2VsZi5fYXJnX3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5k
            cG9pbnQKICAgICAgICBzdGFuZGJ5X21ncnMgPSBbXQogICAgICAgIGlmIG5vdCBtb25pdG9yaW5n
            X2VuZHBvaW50X2lwX2xpc3Q6CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAic3Rh
            dHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVy
            cl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAj
            IGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRf
            dmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ21nciBzZXJ2aWNl
            cycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9t
            c2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnQgPSAoCiAgICAgICAgICAg
            ICAgICBqc29uX291dC5nZXQoIm1ncm1hcCIsIHt9KS5nZXQoInNlcnZpY2VzIiwge30pLmdldCgi
            cHJvbWV0aGV1cyIsICIiKQogICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9y
            aW5nX2VuZHBvaW50OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiY2FuJ3QgZmluZCBtb25pdG9yaW5nX2VuZHBvaW50
            LCBwcm9tZXRoZXVzIG1vZHVsZSBtaWdodCBub3QgYmUgZW5hYmxlZCwgIgogICAgICAgICAgICAg
            ICAgICAgICJlbmFibGUgdGhlIG1vZHVsZSBieSBydW5uaW5nICdjZXBoIG1nciBtb2R1bGUgZW5h
            YmxlIHByb21ldGhldXMnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIG5vdyBjaGVj
            ayB0aGUgc3RhbmQtYnkgbWdyLXMKICAgICAgICAgICAgc3RhbmRieV9hcnIgPSBqc29uX291dC5n
            ZXQoIm1ncm1hcCIsIHt9KS5nZXQoInN0YW5kYnlzIiwgW10pCiAgICAgICAgICAgIGZvciBlYWNo
            X3N0YW5kYnkgaW4gc3RhbmRieV9hcnI6CiAgICAgICAgICAgICAgICBpZiAibmFtZSIgaW4gZWFj
            aF9zdGFuZGJ5LmtleXMoKToKICAgICAgICAgICAgICAgICAgICBzdGFuZGJ5X21ncnMuYXBwZW5k
            KGVhY2hfc3RhbmRieVsibmFtZSJdKQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBw
            YXJzZWRfZW5kcG9pbnQgPSB1cmxwYXJzZShtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgICAg
            ICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50OiB7bW9u
            aXRvcmluZ19lbmRwb2ludH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3Jp
            bmdfZW5kcG9pbnRfaXBfbGlzdCA9IHBhcnNlZF9lbmRwb2ludC5ob3N0bmFtZQogICAgICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICAgICAgbW9uaXRv
            cmluZ19lbmRwb2ludF9wb3J0ID0gc3RyKHBhcnNlZF9lbmRwb2ludC5wb3J0KQoKICAgICAgICAj
            IGlmIG1vbml0b3JpbmcgZW5kcG9pbnQgcG9ydCBpcyBub3Qgc2V0LCBwdXQgYSBkZWZhdWx0IG1v
            biBwb3J0CiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoKICAgICAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0ID0gc2VsZi5ERUZBVUxUX01PTklUT1JJTkdfRU5E
            UE9JTlRfUE9SVAoKICAgICAgICAjIHVzZXIgY291bGQgZ2l2ZSBjb21tYSBhbmQgc3BhY2Ugc2Vw
            YXJhdGVkIGlucHV0cyAobGlrZSAtLW1vbml0b3JpbmctZW5kcG9pbnQ9IjxpcDE+LCA8aXAyPiIp
            CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcF9saXN0LnJlcGxhY2UoIiwiLCAiICIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0X3NwbGl0ID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0LnNwbGl0KCkKICAgICAg
            ICAjIGlmIG1vbml0b3JpbmctZW5kcG9pbnQgY291bGQgbm90IGJlIGZvdW5kLCByYWlzZSBhbiBl
            cnJvcgogICAgICAgIGlmIGxlbihtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQpID09
            IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vICdtb25p
            dG9yaW5nLWVuZHBvaW50JyBmb3VuZCIpCiAgICAgICAgIyBmaXJzdCBpcCBpcyB0cmVhdGVkIGFz
            IHRoZSBtYWluIG1vbml0b3JpbmctZW5kcG9pbnQKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X2lwID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzBdCiAgICAgICAgIyByZXN0
            IG9mIHRoZSBpcC1zIGFyZSBhZGRlZCB0byB0aGUgJ3N0YW5kYnlfbWdycycgbGlzdAogICAgICAg
            IHN0YW5kYnlfbWdycy5leHRlbmQobW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzE6
            XSkKICAgICAgICBmYWlsZWRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwCgogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnQgPSAiOiIuam9pbigKICAgICAgICAgICAgW21vbml0b3JpbmdfZW5k
            cG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydF0KICAgICAgICApCiAgICAgICAgaXBf
            dHlwZSA9IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQobW9uaXRvcmluZ19lbmRwb2ludCkKICAgICAg
            ICB0cnk6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBzZWxmLl9jb252ZXJ0
            X2hvc3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCwg
            bW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgIyBjb2xsZWN0IGFsbCB0aGUgJ3N0YW5kLWJ5JyBtZ3IgaXBzCiAgICAgICAgICAgIG1ncl9p
            cHMgPSBbXQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5X21nciBpbiBzdGFuZGJ5X21ncnM6
            CiAgICAgICAgICAgICAgICBmYWlsZWRfaXAgPSBlYWNoX3N0YW5kYnlfbWdyCiAgICAgICAgICAg
            ICAgICBtZ3JfaXBzLmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9jb252ZXJ0X2hv
            c3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgICAgICAgICBlYWNoX3N0YW5kYnlfbWdyLCBt
            b25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgZXhjZXB0OgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJDb252ZXJzaW9uIG9mIGhvc3Q6
            IHtmYWlsZWRfaXB9IHRvIElQIGZhaWxlZC4gIgogICAgICAgICAgICAgICAgIlBsZWFzZSBlbnRl
            ciB0aGUgSVAgYWRkcmVzc2VzIG9mIGFsbCB0aGUgY2VwaC1tZ3JzIHdpdGggdGhlICctLW1vbml0
            b3JpbmctZW5kcG9pbnQnIGZsYWciCiAgICAgICAgICAgICkKCiAgICAgICAgXywgXywgZXJyID0g
            c2VsZi5lbmRwb2ludF9kaWFsKG1vbml0b3JpbmdfZW5kcG9pbnQsIGlwX3R5cGUpCiAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZXJyKQogICAgICAgICMgYWRkIHRoZSB2YWxpZGF0ZWQgYWN0aXZlIG1nciBJUCBpbnRvIHRo
            ZSBmaXJzdCBpbmRleAogICAgICAgIG1ncl9pcHMuaW5zZXJ0KDAsIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXApCiAgICAgICAgYWxsX21ncl9pcHNfc3RyID0gIiwiLmpvaW4obWdyX2lwcykKICAgICAg
            ICByZXR1cm4gYWxsX21ncl9pcHNfc3RyLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKCiAgICBk
            ZWYgY2hlY2tfdXNlcl9leGlzdChzZWxmLCB1c2VyKToKICAgICAgICBjbWRfanNvbiA9IHsicHJl
            Zml4IjogImF1dGggZ2V0IiwgImVudGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0K
            ICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoK
            ICAgICAgICAgICAgcmV0dXJuICIiCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5
            Il0pCgogICAgZGVmIGdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAg
            ICJvc2QiOiAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwKICAgICAgICAgICAgIm1k
            cyI6ICJhbGxvdyAqIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0
            cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rl
            cl9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiazhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNf
            ZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgog
            ICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAg
            ICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNf
            Y2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVzeXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1si
            b3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9e2NlcGhmc19maWxlc3lzdGVt
            fSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2VwaGZzX25vZGVf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3
            IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsCiAgICAgICAg
            ICAgICJtZHMiOiAiYWxsb3cgcnciLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4
            c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0g
            PT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0
            eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAg
            ICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyAqPXtjZXBoZnNfZmlsZXN5c3Rl
            bX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2VudGl0eSgKICAg
            ICAgICBzZWxmLAogICAgICAgIGVudGl0eSwKICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAg
            IGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICBy
            YWRvc19uYW1lc3BhY2UsCiAgICApOgogICAgICAgIGlmICgKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgb3IgcmJkX3Bvb2xfbmFtZS5jb3VudCgi
            XyIpICE9IDAKICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZSAhPSAiIgogICAgICAg
            ICAgICAjIGNoZWNraW5nIGFsaWFzX3JiZF9wb29sX25hbWUgaXMgbm90IGVtcHR5IGFzIHRoZXJl
            IG1heWJlIGEgc3BlY2lhbCBjaGFyYWN0ZXIgdXNlZCBvdGhlciB0aGFuIC4gb3IgXwogICAgICAg
            ICk6CiAgICAgICAgICAgIGlmIGFsaWFzX3JiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAg
            ICJwbGVhc2Ugc2V0IHRoZSAnLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgYXMgdGhl
            IHJiZCBkYXRhIHBvb2wgbmFtZSBjb250YWlucyAnLicgb3IgJ18nIgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLmNv
            dW50KCIuIikgIT0gMAogICAgICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3Vu
            dCgiXyIpICE9IDAKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIictLWFsaWFzLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyB2YWx1ZSBzaG91bGQgbm90IGNvbnRhaW4gJy4nIG9yICdfJyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0
            ZXJfbmFtZX0te2FsaWFzX3JiZF9wb29sX25hbWV9IgogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtyYmRfcG9vbF9uYW1lfSIK
            CiAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgIHJldHVybiBlbnRpdHkKCiAgICBkZWYgZ2V0
            X3JiZF9wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0g
            ImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAg
            ICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAg
            ICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSBy
            YmQiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNl
            ci5yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19j
            bHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAg
            ICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAg
            ICAgICAgICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBm
            bGFnIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFn
            IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIi
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAn
            LS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bnRpdHkgPSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAg
            ICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFt
            ZSwKICAgICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRv
            c19uYW1lc3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNl
            ICE9ICIiOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25h
            bWVzcGFjZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBjYXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgog
            ICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X3JiZF9ub2RlX2NhcHNfYW5k
            X2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtbm9kZSIKICAg
            ICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSByYmQiLAog
            ICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJt
            aXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19jbHVzdGVy
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgcmFk
            b3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICAg
            ICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5v
            dCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIgogICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAg
            ICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAg
            ICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMt
            Y2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkg
            PSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAgICAgICAg
            ICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAg
            ICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRvc19uYW1l
            c3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlICE9ICIi
            OgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgZiJw
            cm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25hbWVzcGFj
            ZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBj
            YXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAg
            IHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2Vu
            dGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAgY2Fw
            cyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9z
            dGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLAogICAgICAgICAgICAib3NkIjogZiJwcm9maWxlIHJiZC1yZWFkLW9u
            bHksIGFsbG93IHJ3eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9e3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2csIGFsbG93IHggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5idWNrZXRzLmluZGV4IiwKICAgICAgICAg
            ICAgIm1kcyI6ICJhbGxvdyAqIiwKICAgICAgICB9CgogICAgICAgIHJldHVybiBjYXBzLCBlbnRp
            dHkKCiAgICBkZWYgZ2V0X2NhcHNfYW5kX2VudGl0eShzZWxmLCB1c2VyX25hbWUpOgogICAgICAg
            IGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAg
            ICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fw
            c19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgaW4gdXNl
            cl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgIT0gdXNlcl9u
            YW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVy
            bWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19ub2RlX2Nh
            cHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiBp
            biB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X3JiZF9w
            cm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1u
            b2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1ub2RlIiAh
            PSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRf
            YXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX25v
            ZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiIGlu
            IHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5oZWFsdGhjaGVja2VyIiAhPSB1c2Vy
            X25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9w
            ZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfZGVmYXVsdFVzZXJf
            Y2Fwc19hbmRfZW50aXR5KCkKCiAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgZiJubyB1c2VyIGZvdW5kIHdpdGggdXNlcl9uYW1lOiB7dXNlcl9uYW1l
            fSwgIgogICAgICAgICAgICAiZ2V0X2NhcHNfYW5kX2VudGl0eSBjb21tYW5kIGZhaWxlZC5cbiIK
            ICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKHNlbGYsIHVzZXIp
            OgogICAgICAgICIiIgogICAgICAgIGNvbW1hbmQ6IGNlcGggYXV0aCBnZXQtb3ItY3JlYXRlIGNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIG1vbiAnYWxsb3cgcicgbWdyICdhbGxvdyBydycg
            b3NkICdhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSonCiAgICAgICAgIiIiCiAgICAgICAg
            Y2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fwc19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21k
            X2pzb24gPSB7CiAgICAgICAgICAgICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAg
            ICAgICAgICAgImVudGl0eSI6IGVudGl0eSwKICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBj
            YXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMoKSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAg
            ICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICBzZWxm
            LmRyeV9ydW4oCiAgICAgICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAgICAg
            ICsgY21kX2pzb25bInByZWZpeCJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgICAgICsgIiAi
            CiAgICAgICAgICAgICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAg
            ICAgICAgICAgKSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCiAgICAgICAgIyBj
            aGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYuY2hlY2tf
            dXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAgICAgICAg
            IHJldHVybiB1c2VyX2tleSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICAg
            ICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0eShjbGllbnQuY3NpLXJi
            ZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9t
            c2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhl
            cmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9y
            IGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7dXNlcn0nIGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0
            X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAg
            ICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSksIGYie2VudGl0eS5zcGxpdCgnLics
            IDEpWzFdfSIKICAgICAgICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0
            eShjbGllbnQuY3NpLXJiZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICBkZWYgZ2V0X2NlcGhm
            c19kYXRhX3Bvb2xfZGV0YWlscyhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4Ijog
            ImZzIGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29u
            WyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2Nv
            bW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1
            Y2Nlc3NmdWwgYXR0ZW1wdCwgcmVwb3J0IGFuIGVycm9yCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            OgogICAgICAgICAgICAjIGlmIGZzIGFuZCBkYXRhX3Bvb2wgYXJndW1lbnRzIGFyZSBub3Qgc2V0
            LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPT0gIiIKICAgICAgICAg
            ICAgKToKICAgICAgICAgICAgICAgIHJldHVybgogICAgICAgICAgICAjIGlmIHVzZXIgaGFzIHBy
            b3ZpZGVkIGFueSBvZiB0aGUKICAgICAgICAgICAgIyAnLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1l
            JyBvciAnLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUnIGFyZ3VtZW50cywKICAgICAgICAgICAgIyBy
            YWlzZSBhbiBleGNlcHRpb24gYXMgd2UgYXJlIHVuYWJsZSB0byB2ZXJpZnkgdGhlIGFyZ3MKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiJ2ZzIGxzJyBjZXBoIGNhbGwgZmFpbGVkIHdpdGggZXJyb3I6IHtlcnJfbXNnfSIKICAgICAg
            ICAgICAgKQoKICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IHt9CiAgICAgICAgIyBpZiAnLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lJyBhcmd1bWVudCBpcyBwcm92aWRlZCwKICAgICAgICAjIGNo
            ZWNrIHdoZXRoZXIgdGhlIHByb3ZpZGVkIGZpbGVzeXN0ZW0tbmFtZSBleGlzdHMgb3Igbm90CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lOgogICAgICAg
            ICAgICAjIGdldCB0aGUgbWF0Y2hpbmcgbGlzdAogICAgICAgICAgICBtYXRjaGluZ19qc29uX291
            dF9saXN0ID0gWwogICAgICAgICAgICAgICAgbWF0Y2hlZAogICAgICAgICAgICAgICAgZm9yIG1h
            dGNoZWQgaW4ganNvbl9vdXQKICAgICAgICAgICAgICAgIGlmIHN0cihtYXRjaGVkWyJuYW1lIl0p
            ID09IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgdW5hYmxlIHRvIGZpbmQgYSBtYXRjaGluZyBmcy1uYW1lLCByYWlzZSBh
            biBlcnJvcgogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRfbGlzdCkgPT0gMDoK
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgZiJGaWxlc3lzdGVtIHByb3ZpZGVkLCAne3NlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgZiJpcyBub3QgZm91
            bmQgaW4gdGhlIGZzLWxpc3Q6IHtbc3RyKHhbJ25hbWUnXSkgZm9yIHggaW4ganNvbl9vdXRdfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBtYXRjaGlu
            Z19qc29uX291dF9saXN0WzBdCiAgICAgICAgIyBpZiBjZXBoZnMgZmlsZXN5c3RlbSBuYW1lIGlz
            IG5vdCBwcm92aWRlZCwKICAgICAgICAjIHRyeSB0byBnZXQgYSBkZWZhdWx0IGZzIG5hbWUgYnkg
            ZG9pbmcgdGhlIGZvbGxvd2luZwogICAgICAgIGVsc2U6CiAgICAgICAgICAgICMgYS4gY2hlY2sg
            aWYgdGhlcmUgaXMgb25seSBvbmUgZmlsZXN5c3RlbSBpcyBwcmVzZW50CiAgICAgICAgICAgIGlm
            IGxlbihqc29uX291dCkgPT0gMToKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0g
            anNvbl9vdXRbMF0KICAgICAgICAgICAgIyBiLiBvciBlbHNlLCBjaGVjayBpZiBkYXRhX3Bvb2wg
            bmFtZSBpcyBwcm92aWRlZAogICAgICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgIyBhbmQgaWYgcHJlc2VudCwgY2hlY2sg
            d2hldGhlciB0aGVyZSBleGlzdHMgYSBmcyB3aGljaCBoYXMgdGhlIGRhdGFfcG9vbAogICAgICAg
            ICAgICAgICAgZm9yIGVhY2hKIGluIGpzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lIGluIGVhY2hKWyJkYXRhX3Bvb2xz
            Il06CiAgICAgICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0gZWFjaEoKICAg
            ICAgICAgICAgICAgICAgICAgICAgYnJlYWsKICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            bm8gbWF0Y2hpbmcgZnMgZXhpc3RzLCB0aGF0IG1lYW5zIHByb3ZpZGVkIGRhdGFfcG9vbCBuYW1l
            IGlzIGludmFsaWQKICAgICAgICAgICAgICAgIGlmIG5vdCBtYXRjaGluZ19qc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGFfcG9vbCBuYW1lLCB7c2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9LCIKICAgICAgICAgICAgICAgICAgICAgICAgIiBk
            b2VzIG5vdCBleGlzdHMiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGMuIGlm
            IG5vdGhpbmcgaXMgc2V0IGFuZCBjb3VsZG4ndCBmaW5kIGEgZGVmYXVsdCwKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgICMganVzdCByZXR1cm4gc2lsZW50bHkKICAgICAgICAgICAg
            ICAgIHJldHVybgoKICAgICAgICBpZiBtYXRjaGluZ19qc29uX291dDoKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID0gc3RyKG1hdGNoaW5nX2pzb25f
            b3V0WyJuYW1lIl0pCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0WyJtZXRh
            ZGF0YV9wb29sIl0KICAgICAgICAgICAgKQoKICAgICAgICBpZiBpc2luc3RhbmNlKG1hdGNoaW5n
            X2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0sIGxpc3QpOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGhhcyBhbHJlYWR5IHByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lLAogICAgICAgICAgICAjIHRocm91
            Z2ggLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUKICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGlmIHRoZSBwcm92aWRl
            ZCBuYW1lIGlzIG5vdCBtYXRjaGluZyB3aXRoIHRoZSBvbmUgaW4gdGhlIGxpc3QKICAgICAgICAg
            ICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19k
            YXRhX3Bvb2xfbmFtZQogICAgICAgICAgICAgICAgICAgIG5vdCBpbiBtYXRjaGluZ19qc29uX291
            dFsiZGF0YV9wb29scyJdCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYi
            UHJvdmlkZWQgZGF0YS1wb29sLW5hbWU6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWV9JywgIgogICAgICAgICAgICAgICAgICAgICAgICAiZG9lc24ndCBtYXRjaCBmcm9t
            IHRoZSBkYXRhLXBvb2xzIGxpc3Q6ICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4
            KSBmb3IgeCBpbiBtYXRjaGluZ19qc29uX291dFsnZGF0YV9wb29scyddXX0iCiAgICAgICAgICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGlmIGRhdGFfcG9vbCBuYW1lIGlzIG5vdCBwcm92aWRl
            ZCwKICAgICAgICAgICAgIyB0aGVuIHRyeSB0byBmaW5kIGEgZGVmYXVsdCBkYXRhIHBvb2wgbmFt
            ZQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBpZiBubyBkYXRhX3Bvb2xzIGV4
            aXN0LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dFsiZGF0YV9wb29scyJdKSA9PSAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPSBzdHIo
            CiAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXVswXQog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGlmIHRoZXJlIGFyZSBtb3JlIHRoYW4gb25l
            ICdkYXRhX3Bvb2xzJyBleGlzdCwKICAgICAgICAgICAgIyB0aGVuIHdhcm4gdGhlIHVzZXIgdGhh
            dCB3ZSBhcmUgdXNpbmcgdGhlIHNlbGVjdGVkIG5hbWUKICAgICAgICAgICAgaWYgbGVuKG1hdGNo
            aW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID4gMToKICAgICAgICAgICAgICAgIHByaW50KAog
            ICAgICAgICAgICAgICAgICAgICJXQVJOSU5HOiBNdWx0aXBsZSBkYXRhIHBvb2xzIGRldGVjdGVk
            OiAiCiAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4KSBmb3IgeCBpbiBtYXRjaGluZ19qc29u
            X291dFsnZGF0YV9wb29scyddXX1cbiIKICAgICAgICAgICAgICAgICAgICBmIlVzaW5nIHRoZSBk
            YXRhLXBvb2w6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9J1xuIgog
            ICAgICAgICAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2hlY2tlcktleShzZWxmLCB1c2VyKToK
            ICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAg
            ICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVh
            dGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtj
            YXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0
            XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigK
            ICAgICAgICAgICAgICAgICJjZXBoICIKICAgICAgICAgICAgICAgICsgY21kX2pzb25bInByZWZp
            eCJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50
            aXR5Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICArICIgIi5qb2luKGNt
            ZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            CgogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBh
            dHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmIidhdXRoIGdldC1vci1jcmVhdGUge3NlbGYucnVuX2FzX3VzZXJ9JyBjb21tYW5kIGZhaWxl
            ZFxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVs
            c2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            c2VsZik6CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJtZ3Igc2VydmljZXMiLCAiZm9y
            bWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQog
            ICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAog
            ICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAg
            IHJldHVybiBOb25lCiAgICAgICAgaWYgImRhc2hib2FyZCIgbm90IGluIGpzb25fb3V0OgogICAg
            ICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIHJldHVybiBqc29uX291dFsiZGFzaGJvYXJkIl0K
            CiAgICBkZWYgY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcihzZWxmKToKICAgICAgICBjbWQgPSBb
            CiAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAgICAgICAgICAgInVzZXIiLAogICAgICAg
            ICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgIi0tdWlkIiwKICAgICAgICAgICAgc2VsZi5FWFRF
            Uk5BTF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgIi0tZGlzcGxheS1uYW1l
            IiwKICAgICAgICAgICAgIlJvb2sgUkdXIEFkbWluIE9wcyB1c2VyIiwKICAgICAgICAgICAgIi0t
            Y2FwcyIsCiAgICAgICAgICAgICJidWNrZXRzPSo7dXNlcnM9Kjt1c2FnZT1yZWFkO21ldGFkYXRh
            PXJlYWQ7em9uZT1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmVn
            cm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAog
            ICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyAiICIuam9pbihjbWQp
            KQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNz
            LkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGFscmVhZHkgZXhpc3RzLCB3ZSBqdXN0IHF1ZXJ5IGl0CiAgICAgICAgICAgIGlmIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSBlcnJuby5FRVhJU1Q6CiAgICAgICAgICAgICAgICBjbWQgPSBbCiAgICAg
            ICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgICAgICJ1c2Vy
            IiwKICAgICAgICAgICAgICAgICAgICAiaW5mbyIsCiAgICAgICAgICAgICAgICAgICAgIi0tdWlk
            IiwKICAgICAgICAgICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAi
            LS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgICAg
            ICAgICAgXQogICAgICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVj
            RXJyOgogICAgICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAg
            ICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5v
            dXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJu
            Y29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCiAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRv
            IGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAg
            ICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVj
            RXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2Us
            ICItMSIKCiAgICAgICAgIyBzZXBhcmF0ZWx5IGFkZCBpbmZvPXJlYWQgY2FwcyBmb3Igcmd3LWVu
            ZHBvaW50IGlwIHZhbGlkYXRpb24KICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQgPSBUcnVlCiAg
            ICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICJj
            YXBzIiwKICAgICAgICAgICAgImFkZCIsCiAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAg
            IHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICItLWNh
            cHMiLAogICAgICAgICAgICAiaW5mbz1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAg
            Ii0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVn
            cm91cF9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAg
            ICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNz
            LlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4
            ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIGNlcGggdmVyc2lvbiBub3Qgc3VwcG9ydGVkIGZv
            ciBhZGRpbmcgYGluZm89cmVhZGAgY2FwKHJnd192YWxpZGF0aW9uKQogICAgICAgICAgICBpZiAo
            CiAgICAgICAgICAgICAgICAiY291bGQgbm90IGFkZCBjYXBzOiB1bmFibGUgdG8gYWRkIGNhcHM6
            IGluZm89cmVhZFxuIgogICAgICAgICAgICAgICAgaW4gZXhlY0Vyci5zdGRlcnIuZGVjb2RlKCJ1
            dGYtOCIpCiAgICAgICAgICAgICAgICBhbmQgZXhlY0Vyci5yZXR1cm5jb2RlID09IDI0NAogICAg
            ICAgICAgICApOgogICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gRmFsc2UKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICBqc29ub3V0cHV0ID0ganNvbi5sb2Fkcyhv
            dXRwdXQpCiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJhY2Nlc3Nfa2V5Il0sCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsic2VjcmV0
            X2tleSJdLAogICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAgICIiLAog
            ICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX3Bvb2woc2VsZiwgcG9vbF9uYW1lKToKICAg
            ICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKHBvb2xfbmFtZSk6CiAgICAgICAg
            ICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRo
            ZSBwcm92aWRlZCBwb29sLCAne3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAg
            ICAgKQoKICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYsIHJiZF9wb29sX25hbWUpOgogICAgICAg
            IGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0
            dXJuCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1l
            KQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0
            KGlvY3R4LCBUcnVlKQoKICAgIGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAg
            ICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgog
            ICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFf
            ZWNfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5y
            YWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAg
            ICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScs
            ICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMgdXBwZXIgY2FzZSIKICAgICAgICAgICAgKQog
            ICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIu
            b3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9l
            eGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkgaXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRl
            ZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAg
            IGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAg
            ICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVt
            ZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSk6CiAgICAgICAgY21kID0gWwogICAgICAg
            ICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIsCiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91
            cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1f
            bmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgIF0KICAgICAgICB0cnk6
            CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJw
            cm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9y
            OgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            ICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgXQogICAgICAgICAgICB0
            cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRl
            cnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQ
            cm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vidm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3Jv
            dXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVkIgogICAgICAgICAgICAgICAgKQoKICAgIGRl
            ZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9zZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9
            IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAic3Vi
            dm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGluIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVz
            eXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICBwaW5f
            dHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcsCiAgICAgICAgXQogICAgICAgIHRyeToKICAg
            ICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nl
            c3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQg
            cGlubmVkIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3Vy
            bCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5f
            T1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAgc2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsi
            UkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXQogICAgICAgIHJnd19lbmRwb2ludCA9IHNl
            bGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAgICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCAr
            ICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1pbi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7
            ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5j
            b2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAgICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAog
            ICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAgICAgICAgICAgICAgICBhdXRoPVMzQXV0aChh
            Y2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5kcG9pbnQpLAogICAgICAgICAgICAgICAgdmVy
            aWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRp
            b25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgICAg
            ICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJsZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3
            X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAgICAgICAgICByZXR1cm4gIiIsICItMSIKICAg
            ICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYgcjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8i
            KSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAne3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiAoCiAg
            ICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAgICAgICItMSIsCiAgICAgICAgICAgICkKCiAg
            ICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3JhZ2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9p
            ZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3Vw
            cG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBv
            bmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAg
            ICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAg
            cmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVuZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBp
            cyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5k
            cG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUK
            ICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19z
            a2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAg
            ICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBi
            YXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBp
            cF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYgZXJyICE9ICIiOgogICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0
            aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3Vw
            cG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3VwcG9ydGVkOgogICAgICAgICAgICBmc2lkID0g
            c2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJnd19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3df
            ZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAg
            ICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAg
            ICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgICAgIGYiVGhl
            IHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0n
            LCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGluZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBh
            cGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZhbGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3ty
            Z3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBoIGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgICMgY2hl
            Y2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRl
            IGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQgZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2
            YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNlIGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAg
            IHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSIsCiAgICAgICAgICAgICAgICAiLnJndy5y
            b290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19w
            b29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3df
            cG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNfdG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAg
            ICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6
            CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tfcmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHJl
            dHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlz
            aXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmlnX25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmln
            KToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2NvbmZpZyAhPSAiIjoKICAgICAgICAgICAgY21k
            ID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgcmd3
            X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICAiZ2V0IiwKICAgICAgICAgICAgICAg
            ICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICByZ3dfbXVs
            dGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAgICBdCiAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9j
            ZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJv
            ciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAg
            ICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIu
            b3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29k
            ZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0dXJuICIt
            MSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYgY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9f
            YXJyYXkoc2VsZiwgdmFsdWUpOgogICAgICAgIHJldHVybiB2YWx1ZS5zcGxpdCgiLCIpCgogICAg
            ZGVmIHJhaXNlX2V4Y2VwdGlvbl9pZl9hbnlfdG9wb2xvZ3lfZmxhZ19pc19taXNzaW5nKHNlbGYp
            OgogICAgICAgIGlmICgKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9m
            YWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAg
            KQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9t
            YWluX3ZhbHVlcyA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIG9yICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVy
            ZV9kb21haW5fdmFsdWVzICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAgICAgICAgICAg
            ICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICApOgogICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgInBy
            b3ZpZGUgYWxsIHRoZSB0b3BvbG9neSBmbGFncyAtLXRvcG9sb2d5LXBvb2xzLCAtLXRvcG9sb2d5
            LWZhaWx1cmUtZG9tYWluLWxhYmVsLCAtLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIK
            ICAgICAgICAgICAgKQoKICAgIGRlZiB2YWxpZGF0ZV90b3BvbG9neV92YWx1ZXMoc2VsZiwgdG9w
            b2xvZ3lfcG9vbHMsIHRvcG9sb2d5X2ZkKToKICAgICAgICBpZiBsZW4odG9wb2xvZ3lfcG9vbHMp
            ICE9IGxlbih0b3BvbG9neV9mZCk6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCB0b3BvbG9neSBwb29scywg
            J3t0b3BvbG9neV9wb29sc30nLCBhbmQgIgogICAgICAgICAgICAgICAgZiJ0b3BvbG9neSBmYWls
            dXJlIGRvbWFpbiwgJ3t0b3BvbG9neV9mZH0nLCIKICAgICAgICAgICAgICAgIGYiYXJlIG9mIGRp
            ZmZlcmVudCBsZW5ndGgsICd7bGVuKHRvcG9sb2d5X3Bvb2xzKX0nIGFuZCAne2xlbih0b3BvbG9n
            eV9mZCl9JyByZXNwY3RpdmVseSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybgoKICAgIGRl
            ZiB2YWxpZGF0ZV90b3BvbG9neV9yYmRfcG9vbHMoc2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToK
            ICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9yYmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYu
            dmFsaWRhdGVfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgaW5pdF90b3BvbG9neV9yYmRfcG9vbHMo
            c2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToKICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9y
            YmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYuaW5pdF9yYmRfcG9vbChwb29sKQoKICAgICMgdGhp
            cyB3aWxsIHJldHVybiB0aGUgZmluYWwgYXJncyB0aGF0IHNjcmlwdCB1c2VzIHRvIHByb2Nlc3MK
            ICAgICMgdGhlIHByaW9yaXR5IHRvIHNldCBhIHBhcnRpY3VsYXIgdmFsdWUgaXMsCiAgICAjIGNv
            bW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmluaSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVz
            CiAgICBkZWYgZ2V0RmluYWxVc2VkQXJncyhzZWxmKToKICAgICAgICBhcmd1bWVudCA9IGYiW0Nv
            bmZpZ3VyYXRpb25zXVxuIgogICAgICAgIGZvciBhcmcgaW4gdmFycyhzZWxmLl9hcmdfcGFyc2Vy
            KToKICAgICAgICAgICAgaWYgc3RyKGdldGF0dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSk6CiAg
            ICAgICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxhZy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRl
            cm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxhZy1uYW1lLAogICAgICAgICAgICAgICAg
            IyBzbyB3ZSBjYW4gZ2V0IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAg
            ICAgICBhcmdWYWx1ZSA9IGFyZy5yZXBsYWNlKCJfIiwgIi0iKQogICAgICAgICAgICAgICAgIyBk
            byBub3QgYWRkIHRoZSBjb25maWctZmlsZSBmbGFnIGFuZCBhbHNvIG5vdCBhZGQgdGhlIGJvb2xl
            YW4gZmxhZ3Mgd2hpY2ggYXJlIHNldCB0byBGYWxzZQogICAgICAgICAgICAgICAgIyBiZWNhdXNl
            IGNvbmZpZy5pbmkgZmlsZSB0cmVhdHMgYm9vbGVhbiBmbGFncyBhcyBUcnVlIGFsd2F5cwogICAg
            ICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGFyZ1ZhbHVlICE9ICJjb25maWct
            ZmlsZSIKICAgICAgICAgICAgICAgICAgICBhbmQgZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBh
            cmcpICE9IEZhbHNlCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIGFyZ3Vt
            ZW50ICs9IGYie2FyZ1ZhbHVlfSA9IHtzdHIoZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcp
            KX1cbiIKICAgICAgICByZXR1cm4gYXJndW1lbnQKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAg
            IyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBj
            bHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRl
            X3JiZF9wb29sKHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNl
            bGYuaW5pdF9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVk
            ZWRfa2V5cy5hZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19k
            YXRhX3Bvb2xfZGV0YWlscygpCiAgICAgICAgIyBkb3VibGUgc3RyaW5nIG5lZWRlZCBmb3IgdXBz
            dHJlYW0gZXhwb3J0cyBvZiBmbGFncwogICAgICAgIHNlbGYub3V0X21hcFsiQVJHUyJdID0gZici
            e3NlbGYuZ2V0RmluYWxVc2VkQXJncygpfSInCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BB
            Q0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJL
            OFNfQ0xVU1RFUl9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgp
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1
            bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RB
            VEEiXSA9IHNlbGYuZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0
            X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktl
            eSgKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZGFzaGJvYXJkX2xpbmsoKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoImNsaWVudC5jc2ktcmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2Vs
            Zi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIi
            KQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19N
            RVRBREFUQV9QT09MX05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfbWV0YWRhdGFfcG9vbF9uYW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQ
            SEZTX0ZTX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAg
            ICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJn
            X3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9H
            Uk9VUCJdID0gc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBD
            ZXBoRlMgbm9kZSBhbmQgcHJvdmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFw
            WyJDRVBIRlNfUE9PTF9OQU1FIl06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNl
            bGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQog
            ICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJP
            VklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBI
            RlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAg
            ICAgICAgICAgIyBjcmVhdGUgdGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAg
            ICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAg
            ICJjc2kiLCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAg
            ICAgKQogICAgICAgICAgICAjIHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAog
            ICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9V
            UCJdOgogICAgICAgICAgICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUs
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAg
            ICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAog
            ICAgICAgICAgICAgICAgICAgICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEi
            LAogICAgICAgICAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0g
            PSAiIgogICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAg
            ICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lO
            VCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BP
            UlQiXSwKICAgICAgICAgICAgKSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkK
            ICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNf
            UE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2Vj
            X3Bvb2xfbmFtZSgpCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSA9IHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9t
            YXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMK
            ICAgICAgICApCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X3Bvb2xzICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5
            X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2Vs
            Zi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZ
            X1BPT0xTIl0pLAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90
            b19hcnJheSgKICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX1ZBTFVFUyJdCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAg
            c2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9M
            T0dZX1BPT0xTIl0pCiAgICAgICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5
            X3JiZF9wb29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRf
            dG9fYXJyYXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAg
            ICAgICAgZWxzZToKICAgICAgICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9s
            b2d5X2ZsYWdfaXNfbWlzc2luZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJF
            RklYIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0
            X21hcFsiUkdXX0VORFBPSU5UIl0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X2VuZHBvaW50OgogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAg
            ICAgICAgICAgICBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5f
            YXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAp
            OgogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwg
            InJlYWxtIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBz
            ZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0
            aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVf
            bmFtZSwgInpvbmUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAg
            IGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAg
            ICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAg
            ICBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAg
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9L
            RVkiXSwKICAgICAgICAgICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAgICAgICAgICAgICBlcnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxp
            ZGF0ZV9yZ3dfZW5kcG9pbnQoaW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVy
            cm9yLCBzZXQgdGhlIFJHV19FTkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAi
            LTEiOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRp
            c2l0ZSBwYXJhbWV0ZXJzIG9yIG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMu
            c3RkZXJyLndyaXRlKGVycikKCiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAg
            ICBmb3IgaywgdiBpbiBzZWxmLm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQg
            ayBub3QgaW4gc2VsZi5fZXhjbHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3Jp
            dGUoZiJleHBvcnQge2t9PXt2fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdl
            dHZhbHVlKCkKICAgICAgICBzaE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAg
            ICBkZWYgZ2VuX2pzb25fb3V0KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkK
            ICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAi
            IgogICAgICAgIGpzb25fb3V0ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJleHRlcm5hbC1jbHVzdGVyLXVzZXItY29tbWFuZCIsCiAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgImFyZ3MiOiBzZWxmLm91dF9tYXBbIkFSR1MiXSwKICAgICAgICAgICAgICAgIH0sCiAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2st
            Y2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZpZ01hcCIs
            CiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHNl
            bGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAgICAgICAg
            ICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5nIjogInt9
            IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAg
            ICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZnNp
            ZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAgICAgICAg
            ICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgt
            b3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91
            dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAgICAidXNl
            cktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYgJ01PTklU
            T1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1lbmRwb2lu
            dCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05J
            VE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9OSVRPUklO
            R19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9uaXRvcmlu
            Zy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0ZXIiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiTW9u
            aXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAg
            ICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVfU0VDUkVU
            JyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicgU2VjcmV0
            CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNS
            RVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JC
            RF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAog
            ICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgK
            ICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAg
            ICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ1c2Vy
            S2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAj
            IGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            ICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAg
            ICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMg
            aWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBq
            c29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5h
            bWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119
            IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAg
            ICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0
            X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRoZW4gb25s
            eSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9M
            SU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAg
            ICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            IHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAgICAgICAg
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0
            X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQgbm90IHNl
            bGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6IHNlbGYu
            b3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            Ijogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmImNlcGgt
            cmJkLXJhZG9zLW5hbWVzcGFjZS17c2VsZi5vdXRfbWFwWydSQURPU19OQU1FU1BBQ0UnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAg
            ICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRf
            bWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsn
            Q1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIs
            CiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAg
            ICAgICkKICAgICAgICBlbHNlOgogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJCRF9NRVRB
            REFUQV9FQ19QT09MX05BTUUiXToKICAgICAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAg
            ICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgt
            cmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICAiZGF0YVBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9P
            TF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlv
            L3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0K
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJuYW1lIjogImNlcGgtcmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3Rv
            cmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25l
            ci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICMgaWYgJ1RPUE9MT0dZX1BPT0xTJywgJ1RPUE9MT0dZX0ZBSUxVUkVf
            RE9NQUlOX0xBQkVMJywgJ1RPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUycgIGV4aXN0cywK
            ICAgICAgICAjIHRoZW4gb25seSBhZGQgJ3RvcG9sb2d5JyBTdG9yYWdlQ2xhc3MKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXQogICAgICAgICAg
            ICBhbmQgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXRvcG9sb2d5IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAg
            ICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidG9wb2xvZ3lGYWlsdXJlRG9tYWlu
            TGFiZWwiOiBzZWxmLm91dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xP
            R1lfRkFJTFVSRV9ET01BSU5fTEFCRUwiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5WYWx1ZXMiOiBzZWxmLm91
            dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xPR1lfRkFJTFVSRV9ET01B
            SU5fVkFMVUVTIgogICAgICAgICAgICAgICAgICAgICAgICBdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAidG9wb2xvZ3lQb29scyI6IHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNy
            ZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9j
            b250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lf
            UkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBI
            RlNfRlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZz
            TmFtZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAg
            ICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFt
            ZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQ
            SEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JH
            V19FTkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNz
            CiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3Mi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            ZW5kcG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAgICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAi
            a2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAg
            ICAgICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VT
            RVJfQUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2Vs
            Zi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAg
            ICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdS
            R1dfVExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0
            IiBzZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5k
            IjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJl
            dHVybiBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vy
            c19wZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlz
            aW9uZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNl
            bGYucnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2Vs
            Zi5ydW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2Vs
            Zi51cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25h
            bWVfZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAi
            Y2xpZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBw
            cm92aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0
            aGUgcmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2Fz
            IHVzZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1
            bHQnIHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0
            YSIKICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAg
            ICAgICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3Jh
            ZGUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lv
            bnMoc2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4
            aXN0cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVu
            dGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29u
            X291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAg
            ICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJp
            bnQoZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICBy
            ZXR1cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAg
            IHNlbGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5n
            X2NhcHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNl
            cikKICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAg
            ICBjYXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAg
            Y3VyX2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAg
            ICAgbmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            Y3VyX2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4g
            Y3VyX2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAg
            ICAgICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBm
            b3IgeCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAg
            ICBdCiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8g
            bWFpbnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3Qu
            ZXh0ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNh
            dGVzIHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1z
            IGluIHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5l
            d19jYXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4
            IGluIGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAg
            ICAgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAg
            ICAgICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFw
            cGVuZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tl
            YWNoQ2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRo
            IGNhcHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBj
            YXBzLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRf
            dmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pz
            b24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29t
            bWFuZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHBy
            aW50KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihz
            ZWxmKToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9u
            cygpCiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAg
            ICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRl
            ZF9vdXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3Vw
            cG9ydGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQog
            ICAgICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmls
            ZSBhbmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRw
            dXRfZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3Jp
            dGUoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJq
            T2JqID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2Vw
            dCBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1
            dGlvbiBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJv
            ciBhcyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQg
            T1NFcnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBv
            dXRwdXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0
            ZG93bigpCg==
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:62cd98508ab41f9526f10c55f6c8054475a1b029230f255fe7461fb750fd1def
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        version: 4.18.11-rhodf
      entries:
      - name: rook-ceph-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: rook-ceph-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: rook-ceph-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: rook-ceph-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: rook-ceph-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: rook-ceph-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: rook-ceph-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: rook-ceph-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: rook-ceph-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: rook-ceph-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: rook-ceph-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: rook-ceph-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/ipm-bundle@sha256:68c38ac59cf0d6cc145eb694e5b5a724da733caddb82c5585ef132e880dde544
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.connect.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:19b163f055da43946643057c07500784210d02db868d2aa10f25a6501d767d3d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.connect.redhat.com/cpopen/common-service-operator@sha256:13f5b6699ab505d02550b09e100f7e205c7fde3f2bc36317025bc795be216761
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - registry.connect.redhat.com/isovalent/clife-bundle@sha256:6d2ba514e40559fe3568dd0107f5e6648d43b543b3dbe4319e4114ca05fe20a6
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/vault-config-operator@sha256:696bfe3769605870286e228f463dc2582202d6aeb8f199d78d87f153c68d7401
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-emrcontainers-controller@sha256:cde2c51ca18d0969be0549e37ecd03cc430be3b4746d53b8744cb552a2c475df
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:29ec87ce725d2c277824312607915bb7de92e18015c388543173dbdadd5834a5
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:c906c822fbab57c1eb95880b0dfb9cd04100c86222a5d5cfe6342298e3bb1b9e
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:e26bc4b183f69a448552f9fe740292d4e803165d341d7d66047688d01fcd060c
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
        version: 4.17.13-rhodf
      entries:
      - name: ocs-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:a4b5197db1ccda9934f30bfd63e41b0bef430ea9bd224cd3dc022466742b82b5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:0fcc80541d8966f3ea3502da1c4fdda24de277d8d73d0752b46e7d69fbcf7038
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
        - registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: ocs-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:9fcdfa4615b91a5bb919aff71fc4df046365b2ee6d540b5f5f47f9ccaeb9877c
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-003
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:9f5b63888cbf4b1c90e3803d0017a01ad9c0c22693b978c9ee8299bfc5337832
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:212ab9249707ca8ff835186d88e7888af18118fcbe98d948a3156b9700307a79
        version: 7.6.12-opr-003
      entries:
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/community-operator-pipeline-prod/cluster-manager@sha256:a4c8a54148538b371e5cea3b2f2d4109f7b6bc4a7f2e5b52a840c9340fbf5e47
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:42102a2e462e27dd1de28ca3f7db02d227c44736f2d57c05037a83ecbff6dd9f
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:b405d6148aa680bf31bdc6185c1702916c3219983d8482f27d89ed2f93659aa7
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-recyclebin-controller@sha256:9d916ed0af37d60d90f9b4711b1544b9373bd247476278c9c4041c9dbce6acc9
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/openshift-community-operators/oadp-operator@sha256:f8c5c6e59f13db8847080ea6d964da6e50e56f7d373bcc4e4278b1a8a80395e4
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/openshift-community-operators/oadp-operator@sha256:b7b4d340157513da2ebef6bc7bc262967d0e825e2baa1f2bc2ab2e170457e015
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a6399dfd6fe0e8dc9f59fac4382320d23f9e1709e48ed6da7db69e76354b2646
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:0a86a1808d9906eb57886d3d3cd5c04cf9578a80860342132d36ecfc411bf807
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:b02bab3af52a2469ced96618f1d122464568ba21ebc0a3a945b78ec4e6bdc80d
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a614ec94c2468e34e129e9ecd4bc0d1895eb67b96e77eb5ad94cd8bd507860f5
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Net AI
      provider-url: https://netai.tech
    name: netai-demo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netai-demo-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deployments.netai.tech/v1alpha1",
                "kind": "AntennaDeployment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netai-demo-operator"
                  },
                  "name": "antennadeployment-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Networking
          com.redhat.openshift.versions: v4.18
          containerImage: ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
          createdAt: "2025-07-09T11:33:08Z"
          description: Net AI Operator that uses EnergAIze engine forecasts to scale
            Red Hat OpenShift resources.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Net AI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntennaDeployment is the Schema for the antennadeployments
              API
            displayName: Antenna Deployment
            kind: AntennaDeployment
            name: antennadeployments.deployments.netai.tech
            version: v1alpha1
          - description: Demo is the Schema for the Demo API
            displayName: Demo
            kind: Demo
            name: demos.deployments.netai.tech
            version: v1alpha1
        description: POC Operator that uses saved EnergAize forecasts to manage OpenShift
          Custom Resources
        displayName: netai-demo-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - energaize
        - scaling
        - scale
        - manage
        - forecast
        - cells
        - 5g
        - networking
        - net
        - ai
        - network
        - telecommunication
        - xapp
        links:
        - name: Netai Demo Operator
          url: https://netai-demo-operator.domain
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Net AI
          url: https://netai.tech
        relatedImages:
        - registry.access.redhat.com/ubi8/ubi@sha256:68fecea0d255ee253acbf0c860eaebb7017ef5ef007c25bee9eeffd29ce85b29
        - registry.connect.redhat.com/net-ai/demo-operator-bundle@sha256:b1a30de1e984da0487a7d042bd940876bedb5a2f43dd73c7895e684a3a35d81e
        - ghcr.io/net-ai-tech/netai-demo-operator-bundle@sha256:0783aba7b039a569b49de2a61ef6dc6ff1d1f301d93c26d623b7f652321871c5
        - ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
        version: 0.0.1
      entries:
      - name: netai-demo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: netai-demo-operator
    provider:
      name: Net AI
      url: https://netai.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-efs-controller@sha256:338a28600b08d1ead95ed409e3326e03f3c5e02ed840d3be56a950db6b1de061
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/ipfs-operator@sha256:03a61ef328feb626408d42904cdfb9d45f66e022d5a1e1088a15b74455487743
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:c60664e5da8eb23bbd918b1c13b83565b1f33a1aa8961b425232fa1ef4d1a6f1
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:82cafd48db8107dd141064ce7530468b11804c8c54235a317eb450a7606bac3e
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:5d9b56300a3b17178467954b533bee19622f8e96fe9d4ac9745ead2c59b9ec5f
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3e0fde49a8e8c5f780a76087bbc42507a42fea0e1fea572ae05ba66d27338365
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3b36dcea36ebf4b2e95fac6691f6fc8f1242f7e5c38195a9af28bfad514c28cb
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:36dc3710ac11d86e1cf0a3c6994ae0e6feb6c8ff1e8110997f27f4e573c259bc
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:64cadb49077eff7863dd678c4aec70f4101f17ee8e0bf7ca64ce7010c9bf4b9b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:4981d96081cc3d10cd26f20aab792e9c195b05bdcd609367011b7e9c7e6759b1
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-apigateway-controller@sha256:44d34788b43d103530115fc53b0252bfdbe6e07a692847a3d32aa21e45892ff9
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.connect.redhat.com/yugabytedb/yugabyte-platform-operator-bundle@sha256:4cee86e973206b5300e88af088aca700cde3b2e2e5e1c70b9a20289e9f592349
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - registry.connect.redhat.com/streamnative/function-mesh-bundle@sha256:e4377187c20a70d768a253bc0794f3152aa6fcee22ddbf2fc88f45756e9681fe
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - registry.connect.redhat.com/loadcore/loadcore@sha256:7fca6883b092ac7955bb42f893f163403960c81b83e360cfd1eac802e57ea839
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        - registry.connect.redhat.com/tigera/operator@sha256:4d0d814fcfeb628aae8a05cb7a0a416ae85395197950085c907b6cdcf2fb4317
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        - registry.connect.redhat.com/tigera/operator@sha256:b1a8b56ff5e1bb5326e1bde4931303358410da9768f0ebf9c08e2c2d8661d742
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        - registry.connect.redhat.com/tigera/operator@sha256:cb7f3367ff064cd01dc636a43fd89d5f6c4e6a7004cb7c94b1758b2398e36d71
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        - registry.connect.redhat.com/tigera/operator@sha256:de9724b84ca493d3b14c8f801c1049ac4713166b00d9bf3ea0b1986aac5040fa
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        - registry.connect.redhat.com/tigera/operator@sha256:8e0c026f369f80e8b36d0c94f1fd60f7f53656e57dc5be0b8b12189546874909
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        - registry.connect.redhat.com/tigera/operator@sha256:1fb1007f7aad744f0e1b0f6a32022c2571ee8558e9db53c85976bbd2a891a6b0
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - registry.connect.redhat.com/tigera/operator@sha256:42a4d0d0e24752dbb51daebcfca4d4850d04e79578bfb8b41cdc2a2a289df4a2
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        - registry.connect.redhat.com/tigera/operator@sha256:7894b770bb02b2271e1414d67f300218cc646534c9f2a6917e51dc8bff8ba731
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/openshift-community-operators/awss3-operator-registry@sha256:9ab3a2cd00b8704a8bae8e9b75f2a95d0a3e60fa3cb52589db8a4e7641b73027
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/openshift-community-operators/syndesis@sha256:9e8c08793aebc8857dcd451fa2fde2b3f05b9ff4f3d3cc604f9f752193f58d42
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/verticadb-operator@sha256:b7691da906474a8ab829966bb3da4a846eaf12909c97bd0d9cb6a1e5e7ff6c03
        - docker.io/vertica/verticadb-operator:1.11.2
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        - quay.io/community-operator-pipeline-prod/verticadb-operator@sha256:be706a582a4b8cae487835e920288b41229834cd3477b7389adb363db4b4f2ca
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              },
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "ServiceFunctionChain",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "servicefunctionchain-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "servicefunctionchain",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "servicefunctionchain-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-10-15T19:06:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-dpu-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
          - description: ServiceFunctionChain is the Schema for the servicefunctionchains
              API
            displayName: Service Function Chain
            kind: ServiceFunctionChain
            name: servicefunctionchains.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-dpu-operator-bundle@sha256:4e98c0f629e938a0c9d011e6a675d45589b044a02b3f9551fc0e8f7cd172c8a0
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:acb5ab931f4d7a46bb195c59d156441af5e9092b2739bcb0379b89d47ccf4e81
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:5dedc4733081fc21bcb77dbf4f4c032bf418f00bc4371c29c7ce028f51833c0b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: dpu-operator.v4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: dpu-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: dpu-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: dpu-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: dpu-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: dpu-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: dpu-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: dpu-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: dpu-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: dpu-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: dpu-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: dpu-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: dpu-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: dpu-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: dpu-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:4ab17b92016288f1ebfb2106c826a17e75d041f0dba7eaaf7545193c3305a408
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        - quay.io/openshift-community-operators/horreum-operator@sha256:a9d82ce368b015e45b93f917d4069978a0ba323f3dd684d470cdfd262e0c9f05
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.connect.redhat.com/cpopen/odlm@sha256:064b597625550ca75dbc9f3c8fd6b2de6e0b9cb7f42d604e871bfc668ad05ac1
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:eccacdf9ac2b0b1f6061aea457c705319f268bfa30631ed9919ae393f0ef7f71
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:cf330a9e405a9641c213a74baf28307a6e247c4f03cc0523d071dad8f5f9e2e9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:412403b096566965b2c92a0a55beed8ebd6d8113a7e1fdf5eb3334769929ca41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:c89d8be470c2fd052c5be7a7dbfb133c9e3528bfa5b3681ffb8b3223c4c6bac2
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:6c9e04df3f5ddc45788333a85f6aeff8953b008a4b165e9cc2d5a079c6c73ece
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:5611a2fd53859d55d5dfdfdd117e80a0f30be0a4489ad838d571e872d3b5158a
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        - registry.connect.redhat.com/loadcore/keysight-wap-stats-dashboards-bundle@sha256:f69f61eb818a1f0f4035ae79487402237b739c55eb104fc4643a7d3ef8d7d0c4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:6d9eabdb4f645aa1b8b6c08ca94f9ec89466cdc35b3cb02aaba22e160d8fd9ea
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:5484f8ecf079ee126d24c05e1bcb06467146b3b8e751e90e0fd68f3e933125f0
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:11ce1c35f9c3d9b5ef9539b4402eb8358df843b29ffd94af7a453e5d765e4767
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-wafv2-controller@sha256:cb9a426baec724225c0132973d5b90a83672f9b0b75451adc6d7fd858c79ab58
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-glue-controller@sha256:2e1e23de186f3c61be8c7a30f3b295fe3b8b101509cf30bd6b7bff055ab74c82
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.109.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
          createdAt: "2025-09-18T09:13:25Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/eclipse/che-dashboard@sha256:8e2285991a725e87d2a9131ea23e08ce97cd5bdf1118f53a55a140c2c97ca25d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:1df1b75f702952b650e21717b04de5f39c84ca8f55f4afa0f32644ef8638d03a
        - quay.io/che-incubator/che-code@sha256:3b5c7e1ae664c82ddb3eaedc8161eb72a3a1a22c6c5e3f342f4aad972a6282a5
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/community-operator-pipeline-prod/eclipse-che@sha256:545d98c512c99bc7bf4ccd02c1eb554bbc4bca7da8cc66e0e26934edbfd50c59
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/che-idea-dev-server@sha256:0344c9574a6544c4c0ef93a798a9705ca3a4624508a55c8f94c2bae7196f914a
        - quay.io/eclipse/che-plugin-registry@sha256:bd71aa6817eeda0449056e78a6e832b82dd37b65ec6cba2bac3ffb9defa4048e
        - quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/che-incubator/configbump@sha256:77bd0fc269ff3724583b43ec68e158fcd300e825e1b44e24e2b9e67ed28df6da
        - quay.io/devfile/universal-developer-image@sha256:f76ea662cfdb9e80b93bcb58a393aed8b52aa72e7da73762d30a3ad22c3f06d9
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/eclipse/che-server@sha256:e5a542da0484eb2223ac292e6005cf7e116bec15d377557f92949728e1f5c8bf
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        version: 7.109.0
      entries:
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/group-sync-operator@sha256:fe6fda2de67fc4996d3f388e11faa0130fee19453111c829ba0aaa5be5410d01
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:e6d5ef5583ae26a8312dc331bd6003fda6c7c248de15b6790a128d0cd03965dd
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:35e0ee6b3ec5cb654b709c5e4a76a1f08182d826b47f077031d9139d4c2cc308
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b876ec88e1e64c17572e956b1c7e4020f5a423c1275f7f57025823ff7f749c61
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b3a45189ae8fc5302fb8a9e15d5747bacce044537923621edf21e9d3b267a27e
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-bundle@sha256:6cd4cba02a9e5c22062108363f6a052734c398019a19f84afc18a7b05a71d578
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ram-controller@sha256:a90b07270f0cfa8929e01569895ab9a209ee44dcc452a3e1f6e0e9a52c5f13b9
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecs-controller@sha256:280d5ec57ad4deb45c145e702242468a4cc04d23112c33f4e72722a269c900e8
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-object-store-operator@sha256:c5af6adc5e7f9bb2d5bf6025b1b8c3e33ce5b9349a0246df1a26b7ec91b2ea2f
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lightbits Labs
      provider-url: https://www.lightbitslabs.com
    name: lightbits-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightbits-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "CSIDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "csidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "csidriver",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "csi.lightbitslabs.com",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "controller": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "tag": "v1.21.0-ubi9"
                  },
                  "node": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "sidecars": {
                      "lb-nvme-discovery-client": {
                        "registry": "docker.lightbitslabs.com",
                        "repository": "lightbits-operator/lb-nvme-discovery-client",
                        "tag": "v1.21.0-ubi9"
                      }
                    },
                    "tag": "v1.21.0-ubi9"
                  }
                }
              },
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "cluster1",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cluster",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "cluster1",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "addresses": "192.168.40.46,192.168.40.47,192.168.40.48",
                  "apiService": "192.168.40.46:443,192.168.40.47:443,192.168.40.48:443",
                  "clusterAdminSecretName": "lb-secret",
                  "clusterName": "cluster1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          com.redhat.openshift.versions: v4.17
          containerImage: docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
          createdAt: "2025-08-20T11:35:15Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/supported-versions: v4.17
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '[]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/suggested-namespace: lightbits-operator
          repository: https://github.com/LightBitsLabs/lightbits-operator
          support: Lightbits Labs <support@lightbitslabs.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster is the Schema for the clusters API
            displayName: Lightbits Cluster
            kind: Cluster
            name: clusters.storage.lightbitslabs.com
            version: v1alpha1
          - description: CSIDriver is the Schema for the csidrivers API
            displayName: Lightbits CSI Driver
            kind: CSIDriver
            name: csidrivers.storage.lightbitslabs.com
            version: v1alpha1
        description: |-
          Operator for deploying and managing Lightbits CSI driver.

          By using the operator users can quickly and easily deploy and upgrade Lightbits CSI Driver.

          ## Requirements

          - Kernel support for NVMe-TCP
          - Kernel support for NVMe Asymmetric Namespace Access (ANA)

          ## CSI driver Supported Features

          - Filesystem Volumes (xfs/ext4)
          - Raw Block Volumes
          - Dynamic/Static Provisioning
          - Volume Snapshots
          - Volume Clones
          - Multi-Tenancy & Credentials
          - Volume Expansion (Online/Offline)
          - Storage Capacity Tracking
          - Host-Side Encryption (LUKS)
          - Publish ReadOnly volumes

          ## How to install

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/install_lightbits_operator.html#installation)

          ## Documentation

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/)
        displayName: Lightbits Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - storage
        - lightbits
        links:
        - name: Lightbits Operator
          url: https://github.com/lightbitslabs/lightbits-operator
        maintainers:
        - email: k8s@lightbitslabs.com
          name: Lightbits Labs
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Lightbits Labs
          url: https://www.lightbitslabs.com
        relatedImages:
        - docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - registry.connect.redhat.com/lightbitslabs/lightbits-operator@sha256:53db496f7a254632a25acf8974aca4041cbddb109499871622ffc2182e7aaa83
        version: 1.2.0
      entries:
      - name: lightbits-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: lightbits-operator
    provider:
      name: Lightbits Labs
      url: https://www.lightbitslabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-bedrockagent-controller@sha256:6688b26f71500af0e2ffbd0ccc53926b0c41adaaf817a88f0fdfd6fb398075e7
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        - quay.io/community-operator-pipeline-prod/trident-operator:25.6.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-09-30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
          createdAt: "2025-09-30T03:27:25Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.41429998

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - quay.io/community-operator-pipeline-prod/sailoperator:1.28.0-nightly-2025-09-30
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-testing/proxyv2@sha256:28f058629e7969e8441b677b308a09bf5d4968dbd5024d54df8472d1f43e1a8e
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-testing/ztunnel@sha256:f4895a66a99ca0c87e54260d6d5784068a48fe7fbfa25144debad420ea7223fa
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-testing/pilot@sha256:42d64978872f6be8540027002a361933e4fe69590458353e39d2643b2459bb3c
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-testing/install-cni@sha256:a8c8f6fb3048461e08b86056000a3034f7945df067d5ebf43cca9bcb0ffa3916
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.28.0-nightly-2025-09-30
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      - name: sailoperator.v1.28.0-nightly-2025-09-21
        version: 1.28.0-nightly-2025-09-21
      - name: sailoperator.v1.28.0-nightly-2025-09-19
        version: 1.28.0-nightly-2025-09-19
      - name: sailoperator.v1.28.0-nightly-2025-09-18
        version: 1.28.0-nightly-2025-09-18
      - name: sailoperator.v1.28.0-nightly-2025-09-17
        version: 1.28.0-nightly-2025-09-17
      - name: sailoperator.v1.28.0-nightly-2025-09-16
        version: 1.28.0-nightly-2025-09-16
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/install-cni:1.23.0
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/pilot:1.23.0
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:a5698068f57fb27c6e4b8c385878c4651967ae5f17412c6ea6cf8bd34bfe76a1
        - docker.io/istio/proxyv2:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/pilot:1.24.1
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:5416c5fc62c7d9498349ed771bf93bfc0f712ad8da0b8ee109dc49d1e684fa56
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.3
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/proxyv2:1.24.2
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:9cc4d5a9f46e4d05f8adb5232459002a073c101cc813330325afb6682e2138ae
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.4
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - quay.io/community-operator-pipeline-prod/sailoperator:1.26.3
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-sagemaker-controller@sha256:c962ee90891a05197edc08b4cf015fce1e2431cced90bd274e89bb87708420f3
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:279f5d817f9dc90b202924fc2b2ab59e5303e9dddc1d78d71017d65cdd66d565
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
          createdAt: "2025-08-26T06:53:59Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/hpestorage/xspc-csi-operator-bundle@sha256:658dd12fe1b92bf38dd24601043cb84e27a27b0d987a36a0ec7d99b0514574e2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:4c289e5023fd3e59b2cf40335eba9eb53caf562472f16dd97b370773a0ba608b
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: xspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.32.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90",
                  "kernelVfDrivers": [
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
          description: This operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Intel device
            plugins for Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              GpuDevicePlugin is the Schema for the gpudeviceplugins API. It represents
              the GPU device plugin responsible for advertising Intel GPU hardware resources to
              the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              QatDevicePlugin is the Schema for the qatdeviceplugins API. It represents the QAT device
              plugin responsible for advertising Intel QuickAssist Technology hardware resources
              to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              SgxDevicePlugin is the Schema for the sgxdeviceplugins API. It represents
              the SGX device plugin responsible for advertising SGX device nodes to
              the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fpga
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4
        - registry.connect.redhat.com/intel/intel-device-plugins-operator-bundle@sha256:4a422ca897caeabb9fc3be3dfef3746819868de1674ea61c657cad257fbddf12
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
        version: 0.32.1
      entries:
      - name: intel-device-plugins-operator.v0.32.1
        version: 0.32.1
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        - quay.io/community-operator-pipeline-prod/flink-kubernetes-operator@sha256:7c919847b88a5969d0c0830e4154512bbca28015cf6fb7fa70cd20824a726685
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:13013220d7376d23df4780fe65c3bea325bdeb7b7a89d5704aa52e948ed7a2d5
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-operator-bundle@sha256:3fb13020c6a34b329825d268dd70c3c9310075780768312dc52cac1e7134a030
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:aa2267c1c871699db37763e938125f290ace4aff2aca6f668b2ce01f7df92b8d
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        version: 4.18.0-202509100344
      entries:
      - name: aws-efs-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: aws-efs-csi-driver-operator.v4.18.0-202508260754
        version: 4.18.0-202508260754
      - name: aws-efs-csi-driver-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: aws-efs-csi-driver-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: aws-efs-csi-driver-operator.v4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: aws-efs-csi-driver-operator.v4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: aws-efs-csi-driver-operator.v4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: aws-efs-csi-driver-operator.v4.18.0-202505280604
        version: 4.18.0-202505280604
      - name: aws-efs-csi-driver-operator.v4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: aws-efs-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: aws-efs-csi-driver-operator.v4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: aws-efs-csi-driver-operator.v4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: aws-efs-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: aws-efs-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: aws-efs-csi-driver-operator.v4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: aws-efs-csi-driver-operator.v4.18.0-202503190402
        version: 4.18.0-202503190402
      - name: aws-efs-csi-driver-operator.v4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: aws-efs-csi-driver-operator.v4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: aws-efs-csi-driver-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: aws-efs-csi-driver-operator.v4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-operator-bundle@sha256:7f3d41672fa8c60725304f0d45e468a4f01defcab7352501737ffddfd32edded
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-operator-bundle@sha256:75cb78f7a40d3daedbc11cbbb5505e029328b4120f0de11d4f6423de943e92d8
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        - quay.io/community-operator-pipeline-prod/node-discovery-operator@sha256:899baace103c58e7337ae5cd75e3764e6de26966aab7340037c612c26ca7cf87
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle@sha256:8cba51274389030d632c8b82255399217d485146f4797907134e3c4ee55cb4d2
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:3bc76999ece42fcb3acd913b32be3fa0e0da056832922819b81119dcfb5f1582
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:2aeba259864bbddaa8cdd46ab6d8c3495450a0081469a27b9acde53008d91bed
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:f58f60283f6cf9ea745f4a25378ef6690d98af8fda0eb63f663066631e71450b
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:48b37c9f9417ef442d896e79851b176505045b487bddf77bb4e6d91fd5ad3dd1
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:53777788253eaa1199e1854d14782619aeec0df9844d86b83742fe0005671cde
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:e3a13ea8b84b9e1abd158d394e11113e5f6c04d879547fa1d3e8638f1ac979d8
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:31258400bb37f4b8837ef294820f8a728ae83f8031d4e01ff0392690a315a30f
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        - quay.io/openshift-community-operators/move2kube-operator@sha256:cdf288be5ea779e03900af9b57c4d922b8760b88da4fcbf88437f8031d679aef
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:236a53d6fad949285b6abafcdd1a4541a7748f952c0c96fa2e71500079579ba4
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:38d03dacca046dda3ba08c39943ec2e7db33f1c7bb897e84fe1308a0982c63f1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - quay.io/openshift-community-operators/ocm-operator@sha256:ca6a3b24ca4ecfcab52ee0402e007d6f14b23fa7f3408d740923806989b170e4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/community-windows-machine-config-operator@sha256:2b79f0e6b8bc069fac7d6cb5c86783fdc98ccae30d279ca0492f1eca96b6ea3f
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        - registry.redhat.io/rhcl-1/limitador-operator-bundle@sha256:4c56cd8462531cf8ba39316a0c6f7493240bfa4a6ace77352fe7ac1597e2fd90
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        - registry.connect.redhat.com/netapp/trident-operator@sha256:53a4889e1036af9f858c9fdb1bd58f59b758a301bc773379572cf279f930f5c6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-eventbridge-controller@sha256:3a284fb7884e50ed9976368f312c75fc1f1cb2ec361297b38ee72d0ac6b66fc7
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/global-load-balancer-operator@sha256:50771c320c8b7faea144f765535e89979f9f81e61d34c8dd2e86139fa68e222e
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:1f4fd5ed8e398b713e733fb552e7294f22b3927d4e017f2ac15ea20024d3de32
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:759dc7d6836f8c8830ec145e3088bddc1b63bd40ff3047079fa8f0e986e74c30
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/openshift-community-operators/konveyor-operator@sha256:04e63481c5104bd034283d5d4b0567167f62480d2c61154b1d16dbd41c344dfb
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/openshift-community-operators/konveyor-operator@sha256:d4c590033b1a2bcd67375465298e5cb6d136fe091abb12b8d009aa683c23ef62
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:848db711d9e8a329b689b40003293fce19c43faedd31afd80d9ae9d5c11cda49
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:e29f175a799597c84f0663a66767bba24feb78f802516056b35b5beb14d9eb41
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:2d58f12044b0951c87c851527240196db148acb9797e6ac49f9474a6e6039d71
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:7b8bdecbe3b3def1ec810ad53f604c7a71daebcfa7433cbb3a0db937580d087c
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:16c4f165143b8bab7f3eb08c815c2cae0dbf7fffd9ce7a625357be5806c24c5c
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:c28522543b039454d2d97b48198d099ac0b8211cbc58f13468d5f70b9f69b037
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:18ba5c587dc613f9c1a8bc9e66081e52f3a37eedeea5e75881e628e9c35ef460
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - registry.connect.redhat.com/nvidia/nim-operator-bundle@sha256:3dbd307d32a39e37a02ac8fa9c9fb15a0c2fd6dc5aea12d55abb061e86365124
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-directory-operator@sha256:9d93e3aeef3cfbba0b655de7783ba9fd1f4cb4ee5de4389005b7c31d971d4947
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opentelemetry-operator@sha256:5dc347a5085c10823a122fab0bd42afd5d9b9aa19ba17e8992e9492000886629
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/aerospike-kubernetes-operator@sha256:bfa83ff729ef8b3a4ebca27b02e223327d3e1351d8bc78a06c18873d288dfb02
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        - quay.io/community-operator-pipeline-prod/aerospike-kubernetes-operator@sha256:63cdb7ea6cd6bda4c6ab2c1f3467489c2cef6c8a95836763b9c88b20c43e8168
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - quay.io/openshift-community-operators/sap-data-intelligence-observer-operator@sha256:5217360266f62fc8ba087fbc730074c14aacf62dbd124fdc3aa46650d51a243b
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sap-data-intelligence-observer-operator@sha256:79aa33cb7e4b570fc9c962fc6bd59708c7e28e81e2b188608780fa085db91442
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/kueue-operator
    name: kueue-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kueue-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "policy": "Disabled"
                    },
                    "integrations": {
                      "frameworks": [
                        "AppWrapper"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "FairSharing"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameOptional"
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob",
                        "Pod",
                        "Deployment",
                        "StatefulSet"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "byWorkload": "Parallel",
                      "policy": "ByWorkload"
                    },
                    "integrations": {
                      "frameworks": [
                        "RayJob",
                        "RayCluster",
                        "AppWrapper",
                        "PyTorchJob"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "Classical"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameRequired"
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-05-01T20:58:29Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.operator.openshift.io
            version: v1alpha1
        description: Kueue Operator description. TODO.
        displayName: Kueue Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:de644d640f355815f3aa40ced5456bb5b0e958aeba8bfbdc4ab360fcdb90f26d
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:0c4d9cd97f7579adbf2afd238c953ca5f170408cf27e3b2bf4aa16c108a76881
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:dd765c446564d4397570f4bc056b55989f6c0bae0d70514874e67765c9c77889
        version: 0.1.0
      entries:
      - name: kueue-operator.v0.1.0
        version: 0.1.0
      name: stable-v0.1
    - currentCSV: kueue-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-06-24T02:48:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:30659b42f5053dd756190211454c01f049bbc7689dddda3ecd8c406b157873a7
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:9a0b85b0752135694193befe00d10f7abdc0c99c6b9f295d59b237fa0fe3e6d2
        - registry.redhat.io/kueue-tech-preview/kueue-must-gather-rhel9@sha256:e5306c95bf6c5f55be6e791fcb66fe9b34f4f982536f4f90f8ed74acf3f6c31d
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:6f0b5f62f6e549c6f6769dd9a358a14668d7244f9c20753d1690d3e09948588c
        version: 0.2.1
      entries:
      - name: kueue-operator.v0.2.1
        version: 0.2.1
      - name: kueue-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: kueue-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-07-15T01:52:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue/kueue-rhel9-operator@sha256:5db0af8088cce6e9e7a88faf80855b6eaa1dd8679c6a56dd5ee98c50a3a8e2e4
        - registry.redhat.io/kueue/kueue-rhel9@sha256:16007bd38c480b607b63c82b3ee8721441cc374ee31c04b9e329426314590195
        - registry.redhat.io/kueue/kueue-must-gather-rhel9@sha256:f5f624ef8dbb3676ed4e37335e226e76e5f886c1990c6e0e8c0e4762740ed98d
        - registry.redhat.io/kueue/kueue-operator-bundle@sha256:1bafac48c00fb47972f512ccd5d846efbbccccd2259395de17537b39834d89df
        version: 1.0.1
      entries:
      - name: kueue-operator.v1.0.1
        version: 1.0.1
      - name: kueue-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: kueue-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/kueue-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-bundle-certified@sha256:a9b8102ab0ec81e5bea84332ef20b023d06660b410e7e2deadb00904d2098613
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:e39a0b57e73d92c5db330ac9b8a184eeb400397f14c05c1ef75c952d91e4f5b2
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:5024be9a3190288983e3c42e8a2dbb49aa44cd596237ad9ed0e72de5090dc9b9
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:ece58f1e360eccf7d970dca2f9da0c4d99a692648f165f677f11ff080649cf62
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:6bd423056b555d39337c0e3a9c31e7732e80f34d0c01f0766a6aadb2aff586f8
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle@sha256:261b77cfb315c608505e5659b2d01dc4e1ea96cd37fae25f417cd7df148f4ce4
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        - quay.io/openshift-community-operators/snyk-operator@sha256:97908144406700ae9c038316661dc7904f3e0f4e01b1bbdf68259b59dfb86d6c
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cadee58c1c1040fad5a46fc9be5fae4a1944845c5e736c50fa192192d5c17a36
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cf782e9e2a62a2fc71c81032e6bf40f9ba15d37c33feff2df551edb111c4e17c
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:60a9e4b05e7eb17ebfe58dd55db96ab2c68580d9282117f74f21869cb63e4a73
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:1461576c2e06c2ee8d84fd40836184566a346ed596b99cb81fe6a08d0ad99096
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        - quay.io/openshift-community-operators/camel-k@sha256:2b8369f729c18384f45a458e88c9f551eb00f26792b40dbf286e6ea2837e9400
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:6154ff0f72c9c49f1b7ea0edbe93e0a3b04f6e63abbd851d0e6706796050d672
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:21ce00e81ce495c142cbe9611f494cfb26b52b3c09a507bee0572c1e1e3333d7
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        - quay.io/openshift-community-operators/camel-k@sha256:26df9d918e942f7fc637212df87a114b5edd52570da18cc212b3499354cdb993
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:11dac0881f5ed66b5d42b2d69b13aba31cd56e6b32157cb3ce977be197189b79
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        - quay.io/openshift-community-operators/camel-k@sha256:cba8cde86a870b31f7dae9a554f4c496ed81731f91a8e64cd62b0caa88efaa38
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:06d31bb777b37f8f7c8981efc24ca26a2f9c56a1c579243d0c7bdddb34c76a98
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-dependencies
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-dependencies.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          createdAt: "2025-09-11T15:25:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          deploys dependencies.
        displayName: Data Foundation Dependencies
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-dependencies-operator-bundle@sha256:6c2223357b672b1a59d22eeab742d3c241d87d5524935974876c9371ad650afa
        version: 4.18.11-rhodf
      entries:
      - name: odf-dependencies.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-dependencies.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-dependencies.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-dependencies.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-dependencies.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-dependencies.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-dependencies.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-dependencies.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-dependencies.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-dependencies.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-dependencies.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-dependencies.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-dependencies
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "debug": false,
                  "ebpfProgramManagerMode": false,
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
          createdAt: "2024-09-13T12:46:10Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:577ab5d4780ba01e130ca5713fb7b3764244cdcb937b6b00e6bb17638f15fea2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ingress-node-firewall-operator-bundle@sha256:6c6a76287353a02c05164f9a0a3e07d27b75f0e5d403eb1b47f1e9dc7c9883d8
        version: 4.18.0-202509100344
      entries:
      - name: ingress-node-firewall.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ingress-node-firewall.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ingress-node-firewall.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ingress-node-firewall.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ingress-node-firewall.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: ingress-node-firewall.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: ingress-node-firewall.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ingress-node-firewall.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ingress-node-firewall.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: ingress-node-firewall.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ingress-node-firewall.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ingress-node-firewall.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: ingress-node-firewall.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ingress-node-firewall.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ingress-node-firewall.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        - registry.connect.redhat.com/loadcore/keysight-nats-operator-bundle@sha256:89bde716c2aa4dba0738d2481a2c4707c59a748f84349f9fb01cb430e69d7beb
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        - quay.io/community-operator-pipeline-prod/beegfs-csi-driver-operator@sha256:1a9f2aab026c2d1549be5b4624c56d65e813c619fdd5beefb062d5c9fb8981b8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        - quay.io/openshift-community-operators/nexus-operator-m88i@sha256:46cd06d477b4d09dbfca7ca4434ea58f881314138ed744c6390033bd6fb7dc0f
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-secretsmanager-controller@sha256:c92ea9b492b6585a816668bd2a77d65471e339e02e25e44c9558a34b352db593
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:151e7d9bc500b7252af46d0856e3e16559ff310a1d0efedcc5d914164bbdef55
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:309ba89dece3ee9901ed66c40bd28d71f20c7d91a7feff552e28ea0c772b9178
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:63bd0faf0b64e4672014205f9dcfeec6c436edd7dc2aaf83cd053d6e4569d90a
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:1b464c682cd29c8fb7bbc566a1575b9002ad61163c2b6e94138e0e4b0ce51140
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:8223bda16ac5ffac4b7407c849c4c30ee006ddafd34cad8a03d2af8314d56f58
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:d7333e377a5aace0d42e42ca936fa7a84d288b181eda4747f9367a9539f2bedc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-bundle@sha256:0064fcdc252a04010e605706be377417de4c9be4289358ec0f331d6a37befb0b
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        - quay.io/community-operator-pipeline-prod/percona-server-mongodb-operator@sha256:8659c9f25245e541609079970c8193fc6a58a7d323c249a81d743de696dc15d4
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-rhmp-bundle@sha256:b66e3cfad24cf28e0a738179cab062667b0103c52cf2e84abb88d5e018e12b2c
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:27c5d91d62371f3cf0dea4d50044d8a6e0a6015a453c4216a030683ea2474c64
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:9fe87e1b0c49dd00a6ab0c35b68ebf641f347ac5b405d31f3c5ace02a7a30540
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:35d2c7c2849f89474fd5648a0d2781dde28131303cc8a597ce243a3e314fe0af
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:dd0ab94059f782aa31696b15e1b25787cd58bab4905fe64faec7e49bbae3d289
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator-bundle@sha256:0f20db33b1eb069a767089862412846aacb723dc22b44adeb9793495ad20ea50
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle@sha256:506afdc804dc30e0a720e8e6f151b75a32148844de42d90d8c0648e9108d9ea7
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        - registry.connect.redhat.com/elastic/eck@sha256:063a1f405b87751c5102693451ae5c2e9ad203c6b8a31b062a5b403bb8b2febc
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/fence-agents-remediation@sha256:6ab21048f86979bcc34a84965dc2a97467a79507ab03469fe8c6deed224881f7
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/proactive-node-scaling-operator@sha256:0ae01e9adc67ff7341fa3faf51a67c02206c41342374cab0bb2e6079d84fef4a
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        - quay.io/openshift-community-operators/oci-ccm-operator@sha256:eaf63c6f9ee9e9c3562d6ced4cff091c1e401daa826ee3a92b321a5cc8f13da2
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:18:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:1a98d55c28a0679134e24b6aafb3c4dc9db0098094c8d2bcfaa41a63c33308e1
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: odr-cluster-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-cluster-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-cluster-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-cluster-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-cluster-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-cluster-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-cluster-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-cluster-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-cluster-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-cluster-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-cluster-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-cluster-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-cluster-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-cluster-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-cluster-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:16f820575acdca39c0b7e85edff7b910809f6d184a9a479de2bb2f687c94a6f2
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-cluster-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-cluster-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-cluster-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-cluster-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-cluster-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-cluster-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-cluster-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-cluster-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-cluster-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-cluster-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-cluster-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-cluster-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a93b382258a34e0b56ef6616a8490e5f44ef1c6aa5d29c0d08d2147ae5b440e6
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3170da029869d15a246ad31d0debd4b9c03ba13855fb877a754cfdffb2286b80
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a7ce92f7f4aadca2781c44839bc357615395587bc8b68a735642967cf8694362
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:093c6a66d4faa1d980e4319a048db8b2869eba8f13f40f02615d6696257e6719
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:5a19032dda9a0f584b41cbe9e729ed23e6003d90ebbb026eb9e7bb9769d4a4e4
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3b2cf972e1ced0a46d3559da4eb1a1b32c50457293ede4958a13538a84b45316
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-operator-bundle@sha256:1a950c41ce2ae204d6bcba959b2d1a3df2d994bf1fefce4c7095c4ca20a5fd6d
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-iam-operator@sha256:d7e2507481b280118997ec9fe8062a9d56ed0867265732ecb5c89ef6fccd98fc
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.connect.redhat.com/vastdataorg/csi-operator-bundle@sha256:925cc412650ec3a32ea4a55ffe65d16b76916d51f0b57ff62182b4ce12c5d2d8
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator-bundle@sha256:e0e7fc3a69575199a6fcf93985451dd66a5eb7d1ab8caecf059d7d52f3953dc4
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle-rhmp@sha256:d90f6bce87b783a80b1b81275cb61162c7c2a7898aa67ea85e3223deb1dfcb4b
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator-metadata@sha256:7fb2750941193d64f8cca278ff7d81d3ca5e29a2ad0568dff7ddae4396548bb3
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:57e796822bcd705974c0efcd6c190ce39eac2e70b9d6a0c27f69d85dfe966db2
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:1e7781ff09c5cfb87ede98b96d31e58b760bedbf7a23f41778654faba3507591
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:f234e8af232f2d3a9472307ae82834f96756841b8d60b8c42ac4f9ca3fd17545
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-lambda-controller@sha256:9b2943aa770ebd5638927b19f05f629575e1f8fef129f8520e01c2238771d7dc
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-operator-bundle@sha256:bcc3521d20a1662d4032901f46e3bd4744c9910b060f52bdf10c6a8dfea1276b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/sap-commerce-operator@sha256:9b6a3c9c7a3af1a91578fa06c202d128fc7bfff3ae21300a4382cffb11def17a
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        - quay.io/community-operator-pipeline-prod/cass-operator-community@sha256:428effade29fcb3e8981e84f7df945f9bc49fe1a9b56217573db8b387706c26b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - quay.io/openshift-community-operators/seldon-operator@sha256:864e87ef21e81624010a1cbcee6b17c8c816dcb0a699c66badd32eec3baf39fd
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.connect.redhat.com/ogaye/ogaye@sha256:f35a21fb3936b259aabe27b453612616fca4e6327c9042361b4d00ef332905c0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - registry.connect.redhat.com/hpestorage/csi-driver-operator-bundle@sha256:39fedc5650ff2fef6ed1d47a5745851101b7dd8f6ce1f8b72ac3a78bf3c1fd6a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso-operators/openstack-operator-bundle@sha256:945271aec89541d6d8a749c41af2fff3fe24e41fd50e935e3944451c940bf940
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
          createdAt: "2025-09-23 17:13:29"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.16-3,hyperconverged-cluster-operator-test-rhel9:v4.18.16-1,kubevirt-common-instancetypes-test-rhel9:v4.18.16-1,kubevirt-tekton-tasks-test-rhel9:v4.18.16-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:86d21c945452fb380a895aa96630726fdba73b4e10549aa8ae353c7cac37fe44
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:aa24c4a937c9c1e52c30f229af5b26ce413b684375801d0f5f1ef2a50e8502e2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:6542fcf51604d3ba95ec0d42aa7bc19bdef940e1959808efc98f170a68a7d01c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:82dc19865ed64c129b5214b413f5f8da16d90da532c40d3e02b28f767d72f3c5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:a67782645abd4a869ca8db132953f5f4da72c853fe16f9521c77f241a0703bfa
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:134cb74301a9f9f704d35dea38ffa73aa3ee0ecf43b7fb3e73a89c9b432af9d4
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:6cbb8785d06e58af9b8fc1a66d0d5989a1b490b04d0191252cbb16ace14a3cdd
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:424ea484cd63823b4585e630ffc4053c32309e3f7dc24a212582afc60064e503
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:01082e3cdbac6fa50bec206401660dedd832a71f1f9e52c27281587f1f0122ba
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:a4b7120550d915cb74f03dc3530082f3a0d6850809062469f59934d8518479b9
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:0db7542ab7d67866046b1dde2f3ec8a04c47caa9c622ea9832dc59d127b4e919
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:fb9321f06c842dcb628f9082094fb697146e8740130c889bee81ba6941177303
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:cfba9b634c5a8871511ae6c1644e63560fa51d8b2729eb0db207b89ea9a1b3e9
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:0384141b5c99df83759f69c88895bc23f946da385d2c6db0fc75f80c0853a77c
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:2cffcd87d0a42748ffdbc63707751eb8212ce34ac33d9225690f04ce696178cb
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:42e5ad70d114edd15c6e4d13d48399f572ce71ead1f45a9d94de2262a77754db
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:3da5b2baa67e468fe72efcdb4caeeb4f3384365719d5edb5fff6656edcd74141
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:521b8f630c112e9d8eb4d5686ec7a6907fff8cf491628b99d6aa4a10253b3d78
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:350bc99751e2cac633b98c633b6c83e60bfd55dde4301282ff8eb61f85ea8d49
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:e21e94b2778c786d2b6aa0c107df267055b1f7d36c589a5900d313b3fe3404c6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8ab2c1ea24981ca684505c2f9e050cd938efbcbdedabd74897e0c6c0d5c5a433
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:68b318dd7fed09f60f6df215c9812b38791707b4e9af1de27acf91365766808f
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:c3a694c846fc85709e756a617d2581c06062d482b0bba88983192f1efbf8cb21
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:7e60c4e1fb9a0e10b039887747287cc06ca156f813d88c7421f3950eaf07bf31
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:710d24725c6aa3b8282d32a310d3e107b6a2e57708425d5e3acde9e82e33591f
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:deaf31fb3ac137556fedec55c44abb71d01c6927aaa21e5d23627ce139f12107
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:8c3f276434dfb7462c556f7be3f93a0c857a463e2fa2befcaef3a7c0859c7368
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:e51dfca6233c67b1b5372c682a2fc6e0fe54f2beef86e4c9451aabc80017a5fb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e3af0269b0a3225b2b5d43cce57a40aef6788830e8af8f4330bd36e4aaf8ec97
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:1c4bcdb3e2074b3f99c4dd0c57a264330c7989b8564efcc50bfe0e086d487f5c
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:27d1ea63ca9dc4fe073a28793f574da3a991a4c36e8da55fc2c7889cede187ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:9f8ab3e7e76ccad5a794de310ed0ba39e070e00647e42cf2b9a1bcdd8c377f3f
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:e0d239237d8052c36aff1509d65c00909b8563ce14341b7eb7fb79e9b84deace
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:af0b3d4305dcc5ccf00569deddf122aabf444bda0d3f0c2a5c4fad7e0c4117e2
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:071b2bf8f6470a6f87c3804eac3a10a947376469e42a55e4f19d3b2fc98c7e95
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:8339ad1cbefafbb91be533f1980ffb0e22114d52fd5c11c1b06cc55d9554c9f4
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:f514eb6d468fdbc6f5bc2b60cafdaff19ea55cb9a9c60f22b9db466f5b8f9000
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:c839c0eef642ac854e04129b007bdd529f9b380840e052fc306ecc0d62e4d055
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:d661d52528a89d87764e16b5d1dcb6ed328075477453e69c6cb9814e2a2d0ea9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:21c02925a871875a8eb70efda72f79d4c287da5cb9c0a4fc19de9ebfcb06e496
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:ae92990aa7ee3cf0bb98549ba78882c032afc8ed323fa83d1e8de24951b09d11
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:ecbdb919859e032ba4f2fe3af4dc79e87a5ad28545942d9d75c734aa814b7b23
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:e6507ebc8d71561772e573691b160f60ee6819b3f6404d5f0db123f0e4bc17e4
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:250cb53bbe119c5609a99ad32e97fc23572af9c08490e0f5212cf0977f555827
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:7eb05621f7cfe8162b08bff8a7411d6b05959affd7dc0b7729b4ccf1d90bb005
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:55996948c6d66abfafc630d7f6fe4b67201af0b61613a39e1747218393fb5e01
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:7f61b66615c9b22ac0e7e6ded50fb5a4f0a71ca228d34340977d82c8eb8322d4
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:09abeaf5bd5a7ac9f272759932e8442828efe6935dcd3b365eb42b7655c1de18
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:d32a745d70478945bb25f04ca85c17dbc1fefa0a791522e42591a6876f085029
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:db86fa7625d770a7eb8a4af977499f4569023992d33caac0caf8e3b71871f574
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:f5ebfeb2d0bcb0e86999d4efdd08d03c34c90daa2523b1382e02f9d4930ded67
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:e3f9e0c1b53a54354246d9c934ea48c038a547845d6a75c7bfac5e82f780ee8a
        version: 4.18.16
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.16
        version: 4.18.16
      - name: kubevirt-hyperconverged-operator.v4.18.15
        version: 4.18.15
      - name: kubevirt-hyperconverged-operator.v4.18.14
        version: 4.18.14
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.12
        version: 4.18.12
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.10
        version: 4.18.10
      - name: kubevirt-hyperconverged-operator.v4.18.9
        version: 4.18.9
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.7
        version: 4.18.7
      - name: kubevirt-hyperconverged-operator.v4.18.6
        version: 4.18.6
      - name: kubevirt-hyperconverged-operator.v4.18.5
        version: 4.18.5
      - name: kubevirt-hyperconverged-operator.v4.18.4
        version: 4.18.4
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:a32214d166ee692a4cae917fe97e0535a4a696155e1c878ae5335c77dadef0df
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.13
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
          createdAt: "2025-08-17 16:56:34"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.13-6,hyperconverged-cluster-operator-test-rhel9:v4.18.13-4,kubevirt-common-instancetypes-test-rhel9:v4.18.13-3,kubevirt-tekton-tasks-test-rhel9:v4.18.13-2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:a4d0418e49929e186e37f20f64c46f8eca20e314f6da4deb55f0b5407686c2e2
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:9fbf91cd986fe4064b99d6eb4bde7a79f58966763765a5001d9ae75ed22655f5
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:710dd1bcf4e171943d5260bd16121abd75058bbeb80bc8c237d24aadca99585c
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:396eec2f40ad14e9348b229498313e478903125dbb8c6f7c5f3dc3dad359d18f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:36a46082ca4ce5fd4784a9613d49c30585883bccf7433dc2ecd8de787eace937
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c73b90403d587fa6993e736fd84c08ed39381ac506502c3711da99dd4ea39b4b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:47d3263ddc6d80e844f8240502df81c0b71c2ed85a29a0d4afac7b5c98379091
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:6c39b35d0142e12757091831e2f9900c5f8ebcb1498e0eb75b62c86ee2e7238c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:8a69cfd1bcf5745779a019499632141ecdbc4b2d9a4858c66d28df11087359f3
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:b7771fb216f69af28f49c615fa050bde00d27067acdb505578d104ce1337febd
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:21ad0c18f57b964b86ced3498ac0c36a6daf8a60bdeece77a0b41b168dbd6586
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:877fe3539fe1a4d9f8cf25bfe2c1bb847ff9413693848ad539c26442472c1d3e
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:6ceaacb96b29fbd5a07dac9e8aeb038b76d53f6dc30188b88068755f35da68a9
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:4f806c32d077c6021033733674ed5701bd45086f9fdbfe90985813e2efc27ee9
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:1e73a912388b8b3de05e669128d30dd3081d5747932dc10287d613e19f59a49b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ca5969eb64a112ae5c12ae3b6752b6858dfb999d78d61daffd61db5e07a7cdb9
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:b0a4d68244205f839c0bb2c5cd4f4f305537ce57b42d24d0b46fa4fe2fe5cbba
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:960ab26612067d50aae973b63ffe482649c90ff8f42c2dde698ef548422460cd
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9600fd87e8331989db167870e83618bfb5f273e34e28c2006fa72b49aceb86e0
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:d90907bbc0d2810d592316f609bc8fd0f551de4c37ac02c1c03ccc014cdef056
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:c78c9ee196047ada7ecb614a42cbbfe3dec934534942ee5cacf34cf030e50378
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:f54b97303ba8579a2ef38d971035117efbab46d0eece22a759bc2fba98d11511
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:1d984a34589472f1c554bb4a53b6df470104affacd341eccd5fa1e334a7d8854
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:28d40c603ff685d922149ae30dc77ba1d9bc9af0ce2ab6d9038ff8272940ca08
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:c399069fc6a0d3b3e7093ab6cc2d540d05a04a68213a24a7a500a83d084246df
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:3c2706973b2bca2fbbf88d5af1de0b3c53ef1c7e5029b9a3ef42f961795ad65c
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:725286e0c9265a66e3f4b2a86fcfa7cbdda4e963863af9fb2af9e1226e10f34f
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:3e546a40d6a3e2cce2567750f93260b3db0978477e8ce34f63cd5d195c58c195
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:3765263eb3b193d860699d3e348346b762eabca51ecf31f8f97db69f3ca5e4bd
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:8985f049a49cae810ffb59dd4c94afca632808d3c8ba40809c5447a382b63bad
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:7095f50cb41030a9f49af2ad9a99d81407b5d8491c421c34934f788f02eb7528
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:f501a890dd87d37eb24cdd86d5889eadfad1e454fcc782e8ca2454150e8dc494
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:04b135c11a1e4518febbad5ee56e118ca711ad3454db19f1824cda0d4d497742
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ebc17f70a654181527a8a1d39e069076af82343cab55531fd7a86eaa85b12fd2
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:a9788ba6756715f286ede2fa04e460fd021b24f2a0df068c2aa626c89814a7b5
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:5794ed8b4af6cc37d278cd91f08f0418b89550269d69c9196772fcca1a4933e7
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:0c2c54d18dddf8ae7f2993dd42b3d07217f63266218c329119dda13e0c431fb6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:f15c753ca6687605d956d5200ec9104067c9bfe66e987dce7c32bc45ed79c57d
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:ee7ae60e4ba4b1d13b0aecb053e3ccd22b62d2277d28413f394d32e6a3ffe139
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:b5a6c60e663b81cef29e9622af8f0eb62838be99ff51f576ca1b275c6e5a204a
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:4c58b6e5356866deb394d55e2f96e4d83489d24c988e8016ccddd8738b895b7b
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:88d6db3caa017e4c2d8114af1abd28df4d0751b01f291d42b84586378e0c9484
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:0838f54f7ab4fbc04346ef67501fbca0b6c2097558bbd9c6fec2632ba0ce7bc8
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:9e46d9a0ca2e1678a02a52d14e2d8f0eb007b21ae105100106532cd12f317a41
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:a0821002f32c485c28d43dcc3afe5f447ecec4a567ef1b5f07b162f96245cbe6
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:fd17e015ad32d54f8464bc337c874b9131e0edca750adb4165161f0d382b8c3a
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:45c7a05260790a6ea78c809316dc58ed98b32cd94a884e82f24650051a27a163
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:2e6b6d7d04690a616eb210fe4b142372807bb9c59b6af6e7d6fa196f8dadff3e
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:8e69000f409c115b6af634c057106dbe894395708a694c3a4cb030b06e77407b
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:498282a9a9afceec8587eef445214f5db6ed6c09c2c3362049c4097a210e5aa5
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:11031e69287b4fafa77e2d54dc40d8c7de64e7ed8528abfca7a798b5aa8970b9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:faf31a311d34ea8bad50c83918d5ac1dc32277b9d8e356a866d166f97209e40d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:61f96cac775c4f636f0685900c419a6771fe3f3b44648d93a176b4ab5efa9092
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:70219973c00c12519a1e5abd9fa3061669de8e7273ea1dcf9f72d40ae683d244
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:f0eba5e54eced4a4478d395b098b3dca201b7b374f9e6b7f4a5c1669953c0991
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:be1b6c7e71d25c45b0f4bc2ac8507a1c215e30e39f51bc9f907436f6bcc96b4b
        version: 4.18.13
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      - name: kubevirt-hyperconverged-operator.v4.18.2
        version: 4.18.2
      - name: kubevirt-hyperconverged-operator.v4.18.1
        version: 4.18.1
      - name: kubevirt-hyperconverged-operator.v4.18.0
        version: 4.18.0
      - name: kubevirt-hyperconverged-operator.v4.17.4
        version: 4.17.4
      - name: kubevirt-hyperconverged-operator.v4.17.3
        version: 4.17.3
      - name: kubevirt-hyperconverged-operator.v4.17.2
        version: 4.17.2
      - name: kubevirt-hyperconverged-operator.v4.17.1
        version: 4.17.1
      - name: kubevirt-hyperconverged-operator.v4.17.0
        version: 4.17.0
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        - registry.connect.redhat.com/fortanix/ccm-node-agent-operator-bundle@sha256:12746020eefd383963e08a86e3542608aed1999b4e520435e366a307a6a9eebc
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - registry.connect.redhat.com/microsafe/operador-aims-mobile-bundle@sha256:1572f715c13d852c2a824b726912c4e5cc49aea43d0d91b6fd673f7198b0d038
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - quay.io/openshift-community-operators/ibm-security-verify-access-operator@sha256:ec3970c7a6a5ee3ac863dfe6f16234fe19cd7ea4a541fd0cbae9a8183d0763a2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.18.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-05-14T20:50:15Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.17.0 <10.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.18/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.18/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.18 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.18/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.18/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.18/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:f391499545a2981a3c3492766bbaae6a72a82db07d88251e85fda5d672a108e0
        - registry.redhat.io/openshift4-wincw/windows-machine-config-operator-bundle@sha256:4ebf4ace7dcaf0a935b83314b2a514baffca704b2f36cec5cd61dcf13e459946
        version: 10.18.1
      entries:
      - name: windows-machine-config-operator.v10.18.1
        version: 10.18.1
      - name: windows-machine-config-operator.v10.18.0
        version: 10.18.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:6e58f2d6a1619a06be05807ff5407924087b7b27611628f9baf999733b67645b
        - registry.redhat.io/web-terminal/web-terminal-operator-bundle@sha256:fb969c8b8ab53ab6639795086efad6870dd212f04c94934b75910598efcf3ca3
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:0b133afa920b5180a3c3abe3dd5c73a9cfc252a71346978f94adcb659d683404
        version: 1.13.0
      entries:
      - name: web-terminal.v1.13.0
        version: 1.13.0
      - name: web-terminal.v1.12.1
        version: 1.12.1
      - name: web-terminal.v1.12.1-0.1745393748.p
        version: 1.12.1+0.1745393748.p
      - name: web-terminal.v1.12.1-0.1743657317.p
        version: 1.12.1+0.1743657317.p
      - name: web-terminal.v1.12.1-0.1742484227.p
        version: 1.12.1+0.1742484227.p
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/customized-user-remediation@sha256:dc517e2349fe993e3417b858865fb73b98b7b163c0b4231d4d7be1f11992bb78
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pubsubplus-eventbroker-operator@sha256:2761ebb89dace09a1c1ccad92c28c877f3a6f75a674191ffc3788be7fa549926
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/openshift-community-operators/ibm-spectrum-scale-csi-operator@sha256:01d24dfa0d6b6146230459f8f7e74f2749d7fff3c4d75c40de386e014f9367d6
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        - quay.io/openshift-community-operators/openebs@sha256:ff179bdd6e63b8e2cb1ce3b96d5e03d20c1a2410296fed6fb331bf5bc2e6b14d
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/ember-csi-community-operator@sha256:5dce9b3018cfb300cd651c7980353cedc6617383d5c525e6a4a39ffb8e105774
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator.2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v2",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "600m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "baseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c",
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "image": "registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733",
                    "resources": {
                      "limits": {
                        "cpu": "400m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "10Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "50Gi",
                    "analyticsvol": "50Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-02-15T09:21:28Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v2
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: ABot-Operator-v2.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-operator-bundle@sha256:35f90c08438e4e5aadd01de20bc82e59454c394c0e1d24ade6c56d6e02053bd4
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c
        - registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:e3b3c32d12223097a9d0eb466c3f49fbdab519a675123c7e69597b53ab4c79fe
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:50f1b4c65541b56b85f42bc9b3127bc7bc9faad1974067ce11481de7596e4c6e
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        version: 2.0.0
      entries:
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:4603bf04e53ce948bbf0ee5d239d5e265b40598044ef9e1f4539f4e466e59cb2
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:a178716e62ab8a752bcb9023ad9631af2dbc72357266cd13b43398e21f230828
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:c5288fea5b2ef8125f7b1ff7f7633eccd4e52d14ac9cfc5e473c2e317018ca3d
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:704925128f81d4ffadc4dcfaa69fc8011cf3d0f431d24222a8d7a03bad675331
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:8da482ee4c94b2546bf2b83345b1e7df6e70dcc63f68d8a2e839f10138951d2e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-operator-bundle@sha256:bd5bdc7085f3e9ce8494488f29b4d83a29796c47d536106cd6e14a7b67ab07ee
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/3scale-amp2/3scale-operator-bundle@sha256:934fd9123d1da778b9f8375854f2a61167f0efbc76371edfdc9962f7626e694a
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.11.8-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:9c2aba45971c924fd55b2aab50ee8e6861a9defd3808b38ff30316727af31f9a
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:4f807ee6e885e966853f70e1c40bbc179dc1f994b401ac868b7f2b921f9d17a0
        - registry.redhat.io/rhel8/mysql-80@sha256:7c05e61c0f4f5dad6aeda136e9a1af2f04f4de3c51004e997d4eacae3778da17
        - registry.redhat.io/3scale-mas/3scale-operator-bundle@sha256:a6c2eff926feb11da27eb9972cb86ee0ef9ca0b013afb6fe6af4f72439d557ef
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:419eb09d1b165363d5b90df01c1fa10f1f91f03649a8758620ffa49276685feb
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:a03a1ae7f01259192b8d7b4ecee6ddfcdc0cd9e2d24a8c2306accb93ee767135
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:7d07caf7b6c5a195ba4706e82f3d28c3793afab99b1a6d5a7020547c48534cdc
        - registry.redhat.io/3scale-mas/searchd-rhel7@sha256:1426e4bd4aa0eb28b479900ffc02621203886280d24997cd11818387ff30d9df
        - registry.redhat.io/3scale-mas/system-rhel7@sha256:b0fd85a4994502a3a2ae9ba6a7610c38af64b6cc897efac12feb4b115a484f77
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:22431acc2c991b6ee6b4de0219366dee6ea745122a49a0175d5fb58ed7d9205f
        version: 0.11.8-mas
      entries:
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:895fcf14b337398585f51e4b6e1f46a4a17c01a153ca889bd4374d691d89b9c8
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:dbf68e729484ba11aaafbcf1a590aca7f233de10e71b01d78c319dc5b5c6c278
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:e9824dd4ca39accfc674b88f7a2b76fba3006ff856e78402348c720f2aa8a6b7
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:767d55dc7343406637dd16f1b3a73bfaec353e0b058143fd68df2474271e3b76
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:43c90bbe3b20e95e1f352cdad097d63820a1282deedeb769352cb100da43ecc3
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:2f853ad4fece4510e847e96fe8085b53c56ee342e1f387677ff8cef345298e72
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:70c5237a251adb5f44a085d091ca6c24d5b313aa73f6545389a2b938740a9a20
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/socmmd@sha256:5adf9706daf49b35cfcf74b6fb07b056f87714b87047803b797a56f138c77486
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        - quay.io/openshift-community-operators/must-gather-operator@sha256:a71e5e25aaa92f03334776fee87316809bf2a46dfb34c497d63b228a194982f2
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: nbde-tang-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nbde-tang-server.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
          createdAt: "2025-04-24T18:06:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/nbde-tang-server
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tang Server allows definition of the NBDE Tang Server deployment
            displayName: Tang Server
            kind: TangServer
            name: nbde.openshift.io
            version: v1alpha1
        description: NBDE Tang Server operator provides Network-Bound Disk Encryption
          in OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NBDE
        - PBD
        - Installation
        - Network
        links:
        - name: NBDE Tang Server
          url: https://github.com/openshift/nbde-tang-server
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c15928ecc2f0ff370f567a2ca909b7a71e2becaff0af649275396e59232b7b8c
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-operator-bundle@sha256:79bf610b144cbdf4404a1153f1c395179ce1743274eec7102c862aae4f45890a
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
        version: 1.1.1
      entries:
      - name: nbde-tang-server.v1.1.1
        version: 1.1.1
      - name: nbde-tang-server.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: nbde-tang-server
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:6f0e07eb583201703f50ce9b3bf47e7aa63630074afcb588902caf05ff3ba2bf
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:35fae43ffc92770e13919f081195b35d8dd293f223fb9e577d75ae53f156305a
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:e135baa6b7eefcaf70780b6833acb689c5aef076b6280ff29c02949fbb33b06e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:6a447f455b8a3a3a8390787b36fe1d59a7d8bbf834f66a63fb0023f0557796e4
        version: 4.17.13-rhodf
      entries:
      - name: odf-prometheus-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-prometheus-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-prometheus-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-prometheus-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-prometheus-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-prometheus-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-prometheus-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-prometheus-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-prometheus-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-prometheus-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-prometheus-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-prometheus-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-prometheus-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-prometheus-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-prometheus-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:64c4cc3155e8ea6630bc7fecd392a41d71193cfb20f832b6334cd4795cf0735e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7916e81066baac027fb575a1edaab3c98a00505c73c4ac3c2530c111c11f0030
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:4352d0aecaac2efc3a15bf87925c08f7f85ec7baac9eebc29787f05af047514e
        version: 4.18.11-rhodf
      entries:
      - name: odf-prometheus-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-prometheus-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-prometheus-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-prometheus-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-prometheus-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-prometheus-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-prometheus-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-prometheus-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-prometheus-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-prometheus-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-prometheus-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-prometheus-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
          createdAt: "2025-08-26T06:53:32Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.1.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.1
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:498bcfc3c56bf642aaa70aec4db61b8610af0a22e9e378dfba02356770566198
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hitachi/hspc-operator-bundle@sha256:2d5e1ed3782a992034bec29d85d24befed31ad5eaf7a21ea44acdd775f90cb73
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: hspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        - registry.connect.redhat.com/opdev/operator-certification-operator-bundle@sha256:c390f32b40449fb09477e288fe5347319b691d3e89e020163e32007e6679c685
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator-bundle@sha256:4629bd129addb09478e48964a62acee88f5dd2746875801bd19d2d55cadd9da2
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        - registry.connect.redhat.com/loadcore/keysight-wap-test-core-bundle@sha256:16c938699ab0d71138693641ba7dcce2f4304da1d357a67a057c4effb4300401
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-cluster-operator@sha256:681e51da33e3cd9d6a4b617dca4338827237b70d1e8065322bc4135846039c56
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/cluster-aas-operator@sha256:aad6afc2f6f1f4e3b2c0d14b6c178d6f71ef318b0d3efc97faa5fdb736eaa721
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        - registry.redhat.io/openshift4/compliance-rhel8-operator-metadata@sha256:1c06868bce1f5d7145ef658bfcc284c44011aa24a17166aa124547b38f2259e4
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:8a2194d8d26dff03ec53ac0ab93c62c268e48675079a88db18bb14a977a391be
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:06e4abf4a94d44728b6bfa685b6ed027962b60fc2a0060a7256297857fd5f6fa
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:b1dc39979be1c347b123b2dd8070e4509179b7b499d4cf3fc7feb4552c6e2413
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:14fb2aa627ea7f38fba738cc4302cf1962660b10d3806c71161d8c7d64a8d060
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/openshift-community-operators/tagger@sha256:a528169d305dbc2bc69794437665e4f940ea708dfd957c58d69bc79316b68612
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/community-operator-pipeline-prod/mongodb-atlas-kubernetes@sha256:f12108323fc990f869394e09b4ff6d9751d0a07f3c152fbd45b0f738746a468b
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - registry.connect.redhat.com/amd/gpu-operator-v1@sha256:8a92a0d18a5ae6a837ca6590c2beb65d889cd744a098c22d3f01d2fffdc4dc57
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:4530ed52522a74051aaedabd62b99d5c5274a0ab72eb28d4e75207d9e9d3391b
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:02e6baf8619cc2b782cbb1f3072e554a7638eaf3fd804abe1ba5f43f692e511b
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:df59ef548b82e9d72fecd06aa6b9c569f96485693c8a4ecb8ec278ae92df436f
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:e68814481ca33a422ed5c2b55b8a369e8b48092fbc71b138cb938a119d39e689
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:8ce258f36d3baab3175b29d5ed6df89588f128d249285bb1e7cb45539933c511
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
          createdAt: "2025-04-23T07:03:43Z"
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - kind: NodeFeatureGroup
            name: nodefeaturegroups.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.18/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-nfd-operator-bundle@sha256:ffb2d53dfdc63bfc9320bc4e927d037244111ab0cdbe5721ad69ce06792d9753
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:a1a60b53db81b32de0127dbb63f8c3a9ef7f70c95968aa53ffe8b49eced7dc43
        version: 4.18.0-202509100344
      entries:
      - name: nfd.4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: nfd.4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: nfd.4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: nfd.4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: nfd.4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: nfd.4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: nfd.4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: nfd.4.18.0-202505052133
        version: 4.18.0-202505052133
      - name: nfd.4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: nfd.4.18.0-202504232132
        version: 4.18.0-202504232132
      - name: nfd.4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: nfd.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: nfd.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: nfd.4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: nfd.4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: nfd.4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: nfd.4.18.0-202502101733
        version: 4.18.0-202502101733
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:f6de1136ced78ddf1ff48fdadc259c251919dd3a70819fe0e9247411999d51f3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
          createdAt: "2025-05-15 07:58:52"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.20.1'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:0885c82cd09105a078d4308b4ec1047b96f3cbf1fd555a1c90a8b053befee27e
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5783a00c386858d1eb45f224431d07d40bb9be746de965e7b129ee30969d4014
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5e822b12b5aadaa08a2f95b879be911455607a9c04922a641625515f955dc9ca
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:2217d9128fe50d61d1dff48b588113b5267002f72d025b1b6fbf87a1c2e670d1
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:90505e5ffaee7af330306c4e045d3755e0c74d30b9f45e1e7739c205d945872d
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:2e18b858f09e7595a89c4dc6ea5a85f8c5952f581b769093bfaab8156bab0813
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:6b4d34e7a40b21cf464e2f42cef21ef72a69ae8a3f4e7600a7bd85aa689117d9
        version: 0.20.1
      entries:
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:e533406d832461cd8889f5e95569ea9f8dbbd57b51dcd08b56050ba1e3775ad0
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        - quay.io/openshift-community-operators/rhoas-operator@sha256:5299cd4c88012692b31d560ef981f80f994a0468552f43e03623adbf89b49ff6
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-networkfirewall-controller@sha256:96c35a47f58c48532fa3a5750e52a6d7bc2d530b54e807d57547f8780e9d7aea
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/korrel8r@sha256:f00f8a7e2011bcc97d87fc7f5bb40a19d6b5737d1f5d7f3b2322ec39009c0618
        - quay.io/korrel8r/operator:0.1.8
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53resolver-controller@sha256:542229539932458d76180d23f1d5c5795e2de89169d8054f7b95be6f3700881b
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:3a96ab2db995d24a9ca064b5566821828d6e73ea20df77c9932d8972cb8e5780
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:c2b996e8dac2f4c81a3e48e4203499f48e4e15d0692e98d61dc8c97192e0a776
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:68eb666168d3a41cc98f91a0f2d74656af73cf76ec2cbe2cc308ae2ffab84875
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        - registry.redhat.io/rhosdt/opentelemetry-operator-bundle@sha256:1844430afea89707b70bba22cb9a2db98161cfc62b53b1086b63f109058028c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - registry.connect.redhat.com/citrix/admagent-operator-bundle@sha256:2a32d9a8cc3bd67ba7ebfe3f96b59e42009121d630918d417717d8ac9f2e40f4
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:9d24507aabd06856c7c35d73417b6ce065164f9c966053b494d68f14f0e260b7
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:00e6945419be8ac83aa7c888fe243c2fb67ee4ae5030cae117c111fdf80f5fe4
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:20:34Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:85ba83a60fece2cc1c39ff245193712f918fbd55e1e283e940425fbb1b1f0e29
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: recipe.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: recipe.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: recipe.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: recipe.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: recipe.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: recipe.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: recipe.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: recipe.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: recipe.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: recipe.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: recipe.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: recipe.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: recipe.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: recipe.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: recipe.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:27:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:3c7033c68ff5e65533d2c3d0f2032cf34aecc153fd34f33ad3239331442d370e
        version: 4.18.11-rhodf
      entries:
      - name: recipe.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: recipe.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: recipe.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: recipe.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: recipe.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: recipe.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: recipe.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: recipe.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: recipe.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: recipe.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: recipe.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: recipe.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        - registry.connect.redhat.com/sonatype/nxiq-operator-bundle@sha256:48bccf9c8cc07127153af7974b949b8cd8429adf8714dcc866f026806e38fb08
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:afbe40e3dd6894210741fa65a3f0bb6e4a3d00ca261eb6bc3adac6a22251a74e
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.connect.redhat.com/vfunction/vfunction-server-operator-bundle@sha256:f1784afe63b2e834a6b0bf6da254339ffe3597d02dc2d389f9162d806bedbe74
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-organizations-controller@sha256:76ea247239dde4e52e5016944a41908de8fdcc4173cad8dfff04413a0e3cdf48
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - quay.io/openshift-community-operators/universal-crossplane@sha256:b3104adb471083de96d49fa53f663dcf0e3e7ff5e3573bb1422aa6045277bea0
        - upbound/crossplane:v1.5.1-up.1
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
          createdAt: "2024-06-12"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:87c29bad5a920ea676a5aa8f6e582492547e391ea166694f84cbb75044718a19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:a29fc3bfed3add2768499e94435e8993bfd704a8a46268e6e63a8d20d2c4ca6c
        - registry.redhat.io/openshift4/ose-ptp-operator-bundle@sha256:7202697ea75c3503d4165b1d970802ac776c2977755a8b240094046f74c162c0
        version: 4.18.0-202509100344
      entries:
      - name: ptp-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ptp-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ptp-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ptp-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ptp-operator.v4.18.0-202506260833
        version: 4.18.0-202506260833
      - name: ptp-operator.v4.18.0-202506121732
        version: 4.18.0-202506121732
      - name: ptp-operator.v4.18.0-202505291136
        version: 4.18.0-202505291136
      - name: ptp-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ptp-operator.v4.18.0-202505071704
        version: 4.18.0-202505071704
      - name: ptp-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ptp-operator.v4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: ptp-operator.v4.18.0-202504170406
        version: 4.18.0-202504170406
      - name: ptp-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ptp-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ptp-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: ptp-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ptp-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ptp-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator-bundle@sha256:b473fba287414d3ccb09aaabc64f463af2c912c322ca2c41723020b216d98d14
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
          createdAt: 2025/06/23
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.0.0 <5.1.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kube-descheduler-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * LongLifecycle
          * CompactAndScale
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `mode` - Configures the descheduler to either evict pods or to simulate the eviction.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
          * `evictionLimits` - Restricts the number of evictions during each descheduling run.
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:4c653bee02c0135c8625fea6a01e18a046013ee330c5253ff66a7cd705ba08c1
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-operator-bundle@sha256:37afe091e3b9656c107196d7120ce171d8b0bf895264a9791c4559e97c81a00a
        version: 5.1.3
      entries:
      - name: clusterkubedescheduleroperator.v5.1.3
        version: 5.1.3
      - name: clusterkubedescheduleroperator.v5.1.2
        version: 5.1.2
      - name: clusterkubedescheduleroperator.v5.1.1
        version: 5.1.1
      - name: clusterkubedescheduleroperator.v5.1.0
        version: 5.1.0
      - name: clusterkubedescheduleroperator.v5.0.1
        version: 5.0.1
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle@sha256:47ad4efd37ad85bccf915a6d9448d404b7ce3164d7afb4c944648eeba9f0b04b
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-bundle@sha256:3e7ab287f64398b2aa434d9cfd5d99e2e9cf2d40c2e9fa1b1418537807b7aeea
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator-bundle@sha256:b2bc3e161311f8f5d2f5fc967815259697ce30bf94e980214bc05aa0e1529d55
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/odf-node-recovery-operator@sha256:77b0ef6110a71da56216719ceed0f448257dec9633c98d7dfabc1038b23b6e29
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/hyperfoil-bundle@sha256:1d3625d12d95deb4f48660459fd0fc8b250b52b46d68136927d0f12465aaadf1
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        - quay.io/community-operator-pipeline-prod/wandb-operator@sha256:9bebaf0e72f0b3ce20dabf7ace62cb5ecd1015243438248e8101a238988cd60e
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:bf79fd57e1df791fcd45dd4745c923046dd7759ff5b3762cd7a329c5bf611316
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:b03f4e88ab1fed4744356a584d673fb2fdb15de7e1de7786bdd222b853690670
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:aa261c79b421bd49f7b46974a3bb5a790d5a1a73dd8dea4e1d8a4634fb68a51f
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:123488f0877c054a348f5b4135e68d77c17523bc4e5ea14dd73168372592f4ce
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:e66501276e3a10ae6d08dad3c4658fdb8b0b635e4b2fe72fa943b0e5300d7f2d
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/datatrucker-operator@sha256:4ab1d683d95d3eeba944b977594aff2e3b3e7203f56598b2fb2f7cff75354be3
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        - quay.io/community-operator-pipeline-prod/neuvector-community-operator@sha256:788f742c9f2def80d0b768eb96c5138f2cebdad767ea4ebb8cc26451771e23f5
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle@sha256:6c09391edf49d71b558f8ddbaf4e947dee17490202521539910d78b5ddcb3531
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        - quay.io/community-operator-pipeline-prod/ack-prometheusservice-controller@sha256:dd9c62f7493d43c6f28055e6f1aa7d586e25b2d6d359418782cf18de45da32e8
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - registry.connect.redhat.com/kedify/kedify-agent-helm-operator@sha256:d176136a44e790298158ea3c4232e2aa7d0ad88ff18b6a5b39238162ed41226a
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/community-operator-pipeline-prod/codeflare-operator@sha256:3dbd2b65d7841cc97c37db59ae8aff9991c4d0f0324c4391afd41c4de20c232a
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        - quay.io/community-operator-pipeline-prod/trustify-operator@sha256:49def1c9612621f90d2398cbe7385488c167cc8ccdcd951f098433eb42cb9ac5
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/susql-operator@sha256:4a47a8448610e54b8dc9700828c9250b84ee19126d6d4edbb62b35597e40bf37
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/grafana-operator@sha256:ae9aa1d05ab9dadcc9961ca4293d0878fd7656be79c869c7fdbee718f259ad86
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/orchestrator-operator-bundle@sha256:b82fd43d2dad2e525464177081869c923e239dcb09319458162f33a96370a6e6
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle-marketplace@sha256:b7f67aec4e911d89b3e6217e8718c349c20a859149b829a0390807c113e27196
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        - quay.io/openshift-community-operators/tidb-operator@sha256:4f73b39de6129b906b5eb7bb0b7485f3da406cfad8616f81e375cc8ffcb156ed
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:167b9063a2dec7f9a5f1c5f3c02daf481c45c5e6a2f1b25353102e85921b6fd7
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:83c5e966addca5197f25c5f9c81a52648db44c2fd9e4969bf30b82c4880f0a40
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:bfbb9d0d5556eb067543f4e139a935000319b439f700240adaf393c68bd8244b
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:e0b1cff317bd4e35334429d1cc7c548dbda54065716aea96ef30610f0b91901b
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:eef6e7cf8f76ccd5e9ed16a5e4bdeb4398859121a6b0e86284b9766408cc2b30
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/openshift-community-operators/dbaas-operator@sha256:b2474b569406fb0f47a7042042429bbc5cd2edbbd119e701c6f440a139e1b3ac
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.829
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.829'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/aquasec-operator-bundle@sha256:59e62e5cbd479eed9ddd083db881469056c35253d76aef6bda9cbdb578e05b3f
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/console@sha256:05bd172efdb4729027ea6a491c874a1a3368156a506b7b5d228837b12282ee1d
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/gateway@sha256:cc8a3e950ae45fb40866b7369aa9393e92d1ab91dc0795ee682217fc421e9ec2
        - registry.connect.redhat.com/aquasec/enforcer@sha256:b48f026d7876656c0b365a90bcaaca7f4544f38b85400c110cf20e57c67ac8f6
        - registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:ce827db4a2b01b054e33f1091ee7046a14757b8781444bc9868783fe1efbf641
        - registry.connect.redhat.com/aquasec/scanner@sha256:d0ab5e0dad4e042324874de02ee36585a93b53ad1c3e905416af2fcb98f15e2f
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.829
      entries:
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:ad6b8e10abfff1cdac1ea35b44c456e3d59b4dd1372cd47c2e4221ed93e1f71f
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - registry.connect.redhat.com/nginx/nginx-ingress-operator-bundle@sha256:4e032549f2c36c5a660994b5ec460ef214bb875a53f6f7e9b814d57049683a7a
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-bundle@sha256:31fd95c5232105d3354515102f099b9b662a8659f08da5d67d4e147d4f934fc7
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
          createdAt: "2025-09-09T14:18:23Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.18"}]'
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5189547a19ea81eb18142ad41e6412577eb9ac64d3987ba15b6665a1117a2e55
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:63d8b19d73e5cddcdce67c94e677666d3385ab8dfa1a534b9860fe1db33a2f91
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 4.17.13-rhodf
      entries:
      - name: odf-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
          createdAt: "2025-09-11T15:26:46Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:d45bf7fc74c58bc192a0c9d30157bd745f7d48b2a47820e3b768e8cf29b3c6a0
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:e76c59039b39b3caeabca73353a91501e8374e832c45eaf4005b1c08244e00b1
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odf-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:b4e4d262801b8ff616710075ff05689a8956e975e8ac563e50204a26d1725458
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:afc1e5bdade3ff0e3918fffb8fafb2700dfba7c473f99a2f347cfc7f889c8678
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: cluster-logging.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing, Observability
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
          createdAt: "2025-05-14T12:47:05Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.

              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:255a73d34b85f25b3655fb2b63954aa6b635b89080a8c249fc41568fd8365b10
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:2495de935f503808bf8700d9d66b8412d0f15fc754cf569d02abfbd74b5a6ee8
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:8c6decb25972dffb37917fb5f5a91ea05fb21b87823eb75807d8c601c12fc1bb
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
        version: 6.3.1
      entries:
      - name: cluster-logging.v6.3.1
        version: 6.3.1
      - name: cluster-logging.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        - quay.io/community-operator-pipeline-prod/apicurio-api-controller@sha256:c783cd6db1ae5a6152848f5e4258e89b37077a177199be905d0cfe451a48e5ef
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        - quay.io/openshift-community-operators/fossul-operator@sha256:95c043ddc818f44a303e27508e0acdcbbcb3d9b1329712c85a8e3fd4cb647e9a
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:17:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dd20cf6cd0b69404d2ddf5554e7f89870429c117740d2df51f3be5a8163b8db9
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:26c47d0498f8b8a61338cda3fffb10724c7e0f78d74d14537580763614130256
        version: 4.17.13-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-csi-addons-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-csi-addons-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-csi-addons-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-csi-addons-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-csi-addons-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-csi-addons-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-csi-addons-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-csi-addons-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-csi-addons-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-csi-addons-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-csi-addons-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-csi-addons-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-csi-addons-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-csi-addons-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:25:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationCronJob is the Schema for the encryptionkeyrotationcronjobs
              API
            displayName: EncryptionKeyRotation CronJob
            kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationJob is the Schema for the encryptionkeyrotationjobs
              API
            displayName: EncryptionKeyRotation Job
            kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFenceClass is the Schema for the networkfenceclasses
              API
            displayName: Network Fence Class
            kind: NetworkFenceClass
            name: networkfenceclasses.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationClass is the Schema for the volumegroupreplicationclasses
              API
            displayName: Volume Group Replication Class
            kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationContent is the Schema for the volumegroupreplicationcontents
              API
            displayName: Volume Group Replication Content
            kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dc3cbef1d890b1be5cd59d4e2fcb91f4f88e2aa53bf59c78a2cbe9af136ec3cd
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:c3e9d2deb1454fbd423ff9eeca34225df86f92b536c962300a75158505341678
        version: 4.18.11-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-csi-addons-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-csi-addons-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-csi-addons-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-csi-addons-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-csi-addons-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-csi-addons-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-csi-addons-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-csi-addons-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-csi-addons-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-csi-addons-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-csi-addons-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:a7b60f1c65a6fb4ee0109c7b2b0bd12fb9487e2f6edb059bd18cfd2337ea9116
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-elbv2-controller@sha256:f40f12efb734af5930c225925e99f31eb8db5fecece704a27ea50bd8661658c2
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - rasen708/cxl-operator:latest
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/ncn-community-operator@sha256:15edc1898fe9f8f56708da7d6de3548d06f1df3bdbb3f66a9cf4f7990fe9fad6
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        - quay.io/community-operator-pipeline-prod/victoriametrics-operator:0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        - registry.connect.redhat.com/flux/flux-operator@sha256:13c2f3b8092fc7b84548dd6383edbb8e41b7a8f3acefe63620f07f7a6454d8b9
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.connect.redhat.com/ibm/modelbuilder-bundle@sha256:4940d853ba477aeeec99550c35c132b69bab8bd7b6fbc1a2fe118c828c7ba088
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-access-operator@sha256:6e37f565564de371b5797895f3e96e0efd4f592d7d3e899ea8507438bdb60e78
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - quay.io/observatorium/opa-openshift:latest
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - docker.io/grafana/loki-operator:0.8.0
        - docker.io/grafana/loki:3.4.2
        - quay.io/community-operator-pipeline-prod/loki-operator@sha256:b7e52fc9b9a988eed8b8338a2b2eba34b020cb24a189c2fa4db63110118dbbc1
        - quay.io/observatorium/api:latest
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
          createdAt: "2025-09-19T19:13:23Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sns-controller@sha256:f535d7861f6ef1488ccf95d0c5629a924033acb133d9452dc86e4c80352e69d2
        version: 1.2.1
      entries:
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
          createdAt: 2025/06/24
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-run-once-duration-override-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-operator-bundle@sha256:8ff7f1c707b40448a719137f37f6a3e09860812f7f07c3e61b7f73d1216399dc
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:42f73634df2c06216fac7b3742b2ac21ef5f12ee98d8694c06f7c9073773acd0
        version: 1.2.2
      entries:
      - name: runoncedurationoverrideoperator.v1.2.2
        version: 1.2.2
      - name: runoncedurationoverrideoperator.v1.2.1
        version: 1.2.1
      - name: runoncedurationoverrideoperator.v1.2.0
        version: 1.2.0
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-paygo-bundle-certified@sha256:e0a8c4d0de20c91b73957d6ea09fb94005e680a092c54389e6e411edbeb59296
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        - registry.connect.redhat.com/loadcore/keysight-wap-db-bundle@sha256:3940b8cbf7579d9ce888f176f96c4d191a7cdfaec4be9251e8da4942e7897e97
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:2647a21dfb60364cfb7c4d1b81508adc13d859e97553bb2f710901e6ad59f4a3
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:dfaeb7ac2ccfed03042febea709f0ab66a4c57d62fa041b7ea887a0783d60a36
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:8d20352c0091985cf93e39cd18b2bb092bc15fc5e9900fe46ed5c6fc3c0af885
        version: 8.17.30
      entries:
      - name: kubeturbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        - quay.io/community-operator-pipeline-prod/ack-ec2-controller@sha256:58d3249744b82a7f75ea2b1461adb03259350930da791da62841f05d91e94d01
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/egressip-ipam-operator@sha256:1e2443488e0aad361cccece51689e2b8da882862cb62c1208a14ab1b5ec2922d
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.connect.redhat.com/intel/sriov-fec-operator-bundle@sha256:c30849cbf451eb016786bc367e3716389b1d45676ad298c4de6a492c7bb00f4f
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:2338eaab863304250b3b1d267e9f2261c48eef1f97c280b7fa0af0ebc3ba208b
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:40802ba8bf531637f19adbbaeba3631e0b0557168ecf61f566fb5f2da25d1d4f
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:504672b2db7fd25d1c85c9614f8242f98b4d2d04184a947dbaaada41ec205b45
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-codeartifact-controller@sha256:9bc95a0db7be356d2f657ae61c1c3098ff6e7e8bf1a675810dfb1b18805c6c8b
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - quay.io/openshift-community-operators/airflow-helm-operator@sha256:b3d05d2e2b4ff8ae64e4fc0338fcfa1ffdc345c6fdda8eaefed4e9108750fe58
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/kubernetes-imagepuller-operator@sha256:a72e3728dcf395667820c10c4c5ccb3f3d040ae5d04f50680e193398d3b72b88
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:5d4e4e62b03888bcbf6336934f2ddf5c5cb3d91bc8482a3aa1697e09625c6865
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-operator:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:df0939534c6aaa988714e64728ecdb387ff7dc4b32a5d97d56703ab143787171
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:ef65de2143df2c93aa73f099fa66d39e3239c30a9d547ed7bf258da0a19e8e35
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        - quay.io/openshift-community-operators/prometurbo@sha256:83444ab464bce4fb5021a2fca6d4d6ff72b40e90a6cfc9c5ac6c74357c775ad5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:032f053f21e273714e9f8a8f6854c83bc71cb84f5b32174704b490eca09eebc8
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:40b961a28ea9761c052a5cd17d2452fb72ea4786a92e13a0a1b589c7de31bf26
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:e4d902722d83f82372021d49c80f9880dfdb8bb312d4635e1d97980b60ca1fb6
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:961aa224525116fd575cff3a37a6465e9c652ee446927dd2bd03249b29469472
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:ad94f962d7aa064dc2b5b12ac2ea66ae5973909c82e6a96829165eaeff7ea952
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:c3913778f986ae4f3a7d94da2a9d0157233dbb676e3c1eecacbdbc02a864685a
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:e80f87f289ea944a78b5c38df600c0a22fa470a2ce90f0b1e9453cd63586d75d
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:bddfd4ebd973089f7446cb16ed9668d418f91c3907a5ede22052be4ccbea6d0c
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:b31ebfab28ec75c0d7110d6f01e3b26fd973c36013308e12f69584ff38ab3779
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:7edf59729c2ae8b4506d5361eda01a98f33243987ac082abf758719c76a4804a
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:94bade4574c3679a425628f68067f9aead57bdd78e60f635041f421ed30341de
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:f5dda99364d07df3a3680a2d7031c27df071c8e80746b1ececb5487240e9bbb0
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:a4704a3977fccceb145e8972d2d7de074455bc228f2e925e5b427ffe3dcc623d
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:4f8f9dae947e9d34e86dc4237aec9167a9a25f83cecfae4982dc952494502486
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:6c0899610a78d2737c2096552e5146d9d5602b5568fab24b51b409dbffcf1961
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:84c6c826473b844293bd2048a498c5e0e8a558224a09f8e2e0cc85a2f10fcacf
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/openshift-community-operators/nsm-operator@sha256:330a930f885dacf65537beb600fe7f428e02687e83778e47dcda3dc2df15ae24
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:30375ef2cab46b967f6d2cd46862377c83e58bc6489eb0e75b495a173ebfd84f
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:466b9c30c02a3dfcd9cc94fe8eb432dca7076b8c79eb97ebe5802d716fd963f8
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:d5121af03b32a4b2286ff4f39499931a1ec04e05af45dccc6599dcc1b49e27ad
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:a0739ee7420b64e6195ec53057b032c356faf0ea4b8c1943725052f925eaeb19
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-s3control-controller@sha256:114ea0669459ae148ab32f5366f239da9cfde23cd6bd6a67c14e5844856837a4
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acmpca-controller@sha256:c3929eb9a46af8feb2c830fd0b9225b6e3ad2559a9b5e04fda9239ef4cc07d9a
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        - quay.io/openshift-community-operators/machine-deletion-remediation@sha256:34d1e6cbfaf248a7b28c2a6e6a39f3fe46c183b9c38ac5fbb98eb3f421ba903c
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/machine-deletion-remediation@sha256:99b0b439d37ecdac8ec74433acf1a0ef787d0d6d1d7ef7cfe5f0a29f9dfe6071
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.connect.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:02d101a5929b60e612bf64a944e282ad9fd353c47d4ff28d2c9d5c0b84704ce5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        - registry.connect.redhat.com/coroot/coroot-operator@sha256:44aa4ea98d468f1d1ced671bd2e2bd5bf79ce467f8c572c2248abdce2b7c5acf
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - registry.connect.redhat.com/mongodb/mongodb-kubernetes-bundle@sha256:354b16ea43cfcb9bff704b6265fb10d94037beb231138e1fb44963f3b348bb40
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sqs-controller@sha256:a22009a5d2037aacf52d223e45b74e6aa08020d4cfd1836ff137254b73e19bb9
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:90f512ce90433a33b50c58613d9b7136e00f116c48e179c4a1be89d6a35f9abc
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:53e73c53c31e0e06c990d3d966c2b4053d559a58a17a781155a3ceff3efac831
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:246712ce23cdd590ff83d84e9800b4ed761886ed5bab9f9cda4125a5d930e7e5
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:68e766235bdcd8f3335e2ab6e09942d98b183b3e498abb65acab6481664d46ed
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:91afe3cccd5eabbe4933459e63fd64a6bd87e399ea1aa81ae30dc3e8341c83e4
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:56db2fd54dbe982dde6a4558b664b19a676a8ac99bfbc8082d097cd6d71a03ea
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:d17af97462de060dc21f997dac973d9f895ce91f80821f1a810d93dbcbda31e3
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:783380361431859301477dfe63b6915240e934ffd07a2af2651094579be47a15
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:46f8c0ea6e383b742ab9cdaed1c8f624c2cf405bb1ce39c0ea6977940d52de07
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:e4ea95a86e77aa64da0b1800f00c2e0e598776cde7cfc7bb9941a60bac207ed9
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:aca3d91b7b9f8c1e8a1e0d045c42ba5f27444e0c1c0bc4a87c697b428c2fff78
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:5af7a32da67d39e7a011672de333792e63604cbfee997c19a495fc759d2fd739
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:74f0cf91e2feb60fe794f71b9f56fa8afc60e53379a0aef104701dd6692272e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:72724358b360ab74cc0857e2f18749d3d5b4c76bb6e930f31c5b04c7716e4c2d
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.connect.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:61f935c5985994994eb2f454fb12d74312f7d7d1627fc2b55f3d6ec9bdeb4bc7
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        - quay.io/openshift-community-operators/eventing-kogito@sha256:abe8d5803d2d96ebbb9fe4cda0feaa0217e92def6a669404362bcf741d934690
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.21.0
          createdAt: "2025-09-15T11:39:08Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        - quay.io/community-operator-pipeline-prod/kepler-operator@sha256:37fe88e7234e5eb07c493a6412f27bb8df33896f7eab0a94b0e396fcdf870715
        - quay.io/sustainable_computing_io/kepler-operator:0.21.0
        - quay.io/sustainable_computing_io/kepler:v0.11.0
        version: 0.21.0
      entries:
      - name: kepler-operator.v0.21.0
        version: 0.21.0
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/namespace-configuration-operator@sha256:971991d2bf717001a98b175c50580cd7db9d230492e72670606c2ed50c53feb0
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ecr-controller@sha256:03ca7f898cfbf14f18a06cf3b510cddade3863275df934be891734b8c177c1e6
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-community-operators/openshift-nfd-operator@sha256:880517267f12e0ca4dd9621aa196c901eb1f754e5ec990a1459d0869a8c17451
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:e5e0c3b3a25c13ea916d70fa6f6ddc20f2c730bd9909aa2654672280c27afda2
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:8c65e65bbaf0ebb10dba03655c5854c69c82ddc0e6b2a29293eee029222e5efa
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
          createdAt: "2025-03-03 15:18:52"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virtio-container-disk@sha256:3cff6989c7d7f72d2f0ac27e512d99873768766ba420d65a6b7794676d0c1e88
        - quay.io/kubevirt/virt-operator@sha256:d5fce85dd4232745408daebfbf70c2bc879a06f8340d1f7448b141d5ba3f1821
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-uploadserver@sha256:2b59d3c0a7c154fb6915cf3439b42eb4aaf48f94025ddb75fa9b73b16148fc13
        - quay.io/kubevirt/ssp-operator@sha256:61ac0c182528a9b54d84755fc0931ad8d408ae74e7e4b644316a5e9e630ab360
        - quay.io/kubevirt/libguestfs-tools@sha256:d655b9abbe8cce45bce3934e11d01715e170ddda4da78d7c51f356e7ee1e3263
        - quay.io/kubevirt/bridge-marker@sha256:e492ca4a6d1234781928aedefb096941d95babee4116baaba4d2a3834813826a
        - quay.io/kubevirt/virt-exportserver@sha256:ec1693cc6e8a14ab83f2feea87427643a968d65832c44b25a1bb911482403ff2
        - ghcr.io/k8snetworkplumbingwg/ovs-cni-plugin@sha256:516791acf430bc747f01366e2b748ac76c0b5e39ad8592be2b3dcc809429fa1e
        - ghcr.io/kubevirt/kubesecondarydns@sha256:8273cdbc438e06864eaa8e47947bea18fa5118a97cdaddc41b5dfa6e13474c79
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:66c55f5ffa4b44f2e58990baa7c672ec9776612a5f1c57fee5a5ab1c080f17bb
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:902af624e05eca764c17137c27813b1618f388e4b1a53b27b54665e8735251a2
        - quay.io/kubevirt/aaq-operator@sha256:5e53b755adaf6c66f59812b590a2eae2d84554fa658e93621ac977316843d918
        - quay.io/kubevirt/hostpath-csi-driver@sha256:f8d44aabe0024058b4380e1af42a57a90b7f8c66c8627091bf425fc384c0880f
        - quay.io/kubevirt/virt-exportproxy@sha256:55e38274e59bfe5a2cb5671534b8b6e75d90e43107abecc95fd463f7be8f0026
        - quay.io/kubevirt/cdi-apiserver@sha256:1849bc14687449d36561eb1c768fdbf0cd14ababf9c254b33a4ddb14abfba01f
        - registry.k8s.io/sig-storage/livenessprobe@sha256:33692aed26aaf105b4d6e66280cceca9e0463f500c81b5d8c955428a75438f32
        - quay.io/kubevirt/network-passt-binding@sha256:c794cc1b9173267f1e3d34e666d9e2f782b88d93b47fb4449675346b467f802c
        - quay.io/kubevirt/virt-controller@sha256:68a4967ef5925d4c24fb5eab0aa5acf0a82946beb7f1139bd45c85efe3653cec
        - quay.io/kubevirt/cdi-controller@sha256:94d16a6bd0d9776eddedf64165f39e63e4dfe4f8a26e858bd099db5d09136047
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:92f2ac10f28a7e80699bd984dbc4de897011b1f68c2be95f3fcb90df410b119d
        - quay.io/kubevirt/virt-artifacts-server@sha256:789f68f4e27cf947cdd4f1c0732042176e67b7faa5e37e1b503b8dace3a46125
        - quay.io/kubevirt/hostpath-provisioner@sha256:156ae46a03e2ff5550967773100122f9abc82e25bd7acaa55dbdec22238dd2ca
        - quay.io/kubevirt/sidecar-shim@sha256:cb4025f7275f8de2891c2195c5faf128ac1b68595cdb3eff776a2e5360b8f034
        - quay.io/kubevirt/cdi-importer@sha256:f9fff83c8307bca61a22e8296cd103b32e63d7c8b2a170c6781c6cd962211d26
        - quay.io/kubevirt/pr-helper@sha256:fd434b4ea1ca5682fc5b05f23de01130c0c10f77a0a2d9d089d8f3cb5b545aa6
        - quay.io/kubevirt/kubevirt-template-validator@sha256:b227679f8ccc7c53469574ee5f58c9cc39eb6c54ec01377e4b3df3205a2f4ab4
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:89ddc40700d52cda23104978498fb4d74a643e807450b630beaa71ffdfa1fc57
        - ghcr.io/kubevirt/ipam-controller@sha256:fd7c38e129ea4ed8a86d1801c9e1ca387802376fc4abc0b0d228d4278f448ecd
        - quay.io/kubevirt/aaq-controller@sha256:b9fd0818e4019cccfb38f0923db058166eafdaea8bb35a720aa22b5ce55daf26
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:16030e8320088cf74dc5fbc7dccfea40169f09722dfad15765fa28bceb8de439
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:ac955c4ad03c40d74341623b22f99e8f796a6ed2ef5cc539fba3c1504d159955
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:8061bd1276ff022fe52a0b07bc6fa8d27e5f6f20cf3bf764e76d347d2e3c929b
        - quay.io/kubevirt/cdi-uploadproxy@sha256:6a98ef32c3778e46812c698970e14736f178fd2a11b1f74ccc36c34ce2e34466
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/kubevirt/passt-binding-cni@sha256:f3fea2ec66a26850489197edf7c85d074eb78bbac52916b10016ac0de8ef7c14
        - quay.io/kubevirt/cdi-cloner@sha256:feefb3933780bf352ba03b62b900907d02c64e73f7e74b6d336f89daf2752f62
        - quay.io/kubevirt/cdi-operator@sha256:8297771144f2f358d893771a9dde208a6bedf6bfd67ab2927b16c27424ca3f26
        - quay.io/kubevirt/kubemacpool@sha256:6cc8ca860960c003b36754c6dde249f26f78d32eb22159a25b5647e197dbc9f9
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/aaq-server@sha256:18c23eb4e32ed083047c8808b35fa81a22294a7e8bbabc74e2612fd2eba97dd7
        - quay.io/kubevirt/virt-launcher@sha256:18508d8767f8530df4d6aca4ffe467c40f9891fbe052548b3b2316f8bfb1bc38
        - quay.io/kubevirt/virt-api@sha256:d92c78b3048756fd11d3c6f1a48f4ce8555dd39e866039313dc23bfa70668443
        - quay.io/kubevirt/virt-handler@sha256:4ddd39539282e79cf3602700b791eb91a6d79991bf2e4dc4cba24e3da9db3d1c
        version: 1.14.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.14.0
        version: 1.14.0
      - name: kubevirt-hyperconverged-operator.v1.13.0
        version: 1.13.0
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - quay.io/openshift-community-operators/eclipse-amlen-operator@sha256:bb89e2a7cb1c9172b080c72ad3227017d135fc30076d2a68b4e215fa8fa55846
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/eclipse-amlen-operator@sha256:5fbfb0bd4f764643cdc1021e99e489514a1f035d85131b13e56b11a3045883c2
        - controller:latest
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/microcks@sha256:1313bb08037e57cde02d9d48c4d8c9663eed08651a3c29f872b5ef85ac764112
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-operator-bundle@sha256:44d82d5f8434fb6e88d2da406e87df0bf4274661d9d77554b80e09d19fce03de
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        - quay.io/community-operator-pipeline-prod/aqua@sha256:ca2b998fb0fa4e80d6f4c3be57041e8e55e43cf0b5f1fee579733fda9ea6a107
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ses-controller@sha256:0cdbea8336c2825bf1bc8dbd3f5235ff41d97fbed30106e7775c6e993f92d8b6
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        - registry.redhat.io/ansible-automation-platform/cloud-addons-operator-bundle@sha256:2cb3b1887e8eeae6708f5f46c9bfb304d7c3df1537ea433b0b23772acacf9f90
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
          createdAt: 2025/05/14
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.3.0 <1.4.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-operator-bundle@sha256:3804fcb77893d9198527b4801e2a778bf612596c9a0b125a0f0f872b23138434
        version: 1.4.1
      entries:
      - name: secondaryscheduleroperator.v1.4.1
        version: 1.4.1
      - name: secondaryscheduleroperator.v1.4.0
        version: 1.4.0
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.3
        version: 1.1.3
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      - name: secondaryscheduleroperator.v1.1.1
        version: 1.1.1
      - name: secondaryscheduleroperator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        - registry.connect.redhat.com/anchore/engine-operator-bundle@sha256:fbbe7e6c1d75c4de2f47e2c825c930568e85f1134545e9d890a0c9f3d9187a4d
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - cloudoperators/ibmcloud-operator:1.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/ibmcloud-operator@sha256:b22e89daf6b7f0fd55ba6db7eb7d7a48cbaef2844dfe20e558d7b821482e9480
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudfront-controller@sha256:235ad7b4ad07644d9a638909c681620b24a85148a8655b7c903faf6aa42c2242
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator-bundle@sha256:0192000a2994a62a904ce4aabfc1ff04caf2fa9d779d27c016922b9121877fac
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.connect.redhat.com/synaxg/synaxg-operator@sha256:ff3d9ab5506c4487e5b6697ffe960584aadfe9f16408e7976d345c3d93375171
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/community-operator-pipeline-prod/nfs-provisioner-operator@sha256:d7ef0fa3b5a1a6094b6fd60c6197a52cb14766865b3549806de100bb3522821a
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-kafka-controller@sha256:1a32352f30d0835299e14e8d0f0bac5585396dd6a10a63fe633e3d81c4d71b50
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/community-operator-pipeline-prod/windup-operator@sha256:4e2bafba54eb01b0ae77555d2ff3da2c34286c7c173756dd20a90de28be36535
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/layer7-operator:v1.2.1
        - quay.io/community-operator-pipeline-prod/layer7-operator:1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/mongodb-operator@sha256:1c06d3e936348ceeed773eb45144c9dd702e50ba293610421a93e3b3942ac9ef
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        - registry.connect.redhat.com/raftt/operator-bundle@sha256:7108686c168c35acd996e61a981bdd885b454837cf6823a48a28cdeeebbb0af9
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:d0bd55bad99bfd1ce09da5ee885a38229829570ebf80e1acc4341dff17a54e64
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:b9ffc18b2d88dfab92583af35b27e0ae36bceac7778cb448ab2ea2eae2ab3e94
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:60b1d25ebb3aada94985e359bb059d4491bf0148ac75d51b68a461670659e429
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:3128cd81b56975643c1c88c6ceff6d72e918a67ec2d6a5a606b6e04753c5b0ce
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:808529c14a8873324096fbd1622fb895c4dc7f60c1cf9e3ded646f9a81d7b496
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:708291528f26c70c561745967a1200edcb2f984687f749a49a2860fee580423f
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:d1c636cc17155c4764d331e050f9bf3ec9f39f6ec6ec14e972179b185ea9f5b8
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:5339c41013727fa5dd3ae5bc1d51eb90e02e5baa85d4039b1efdde30baeae1ae
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:1532fc1351ca44f863feaaf54ba8810be7a62396f98dd292ccaf975fde31ce5b
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:fa72cc3e0b09510d71a097b92ee101e057f1c11e1cb363172bd26b2d94272417
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:b897e25e03133513f9afdb5abe38cbf0f2ac0556c27518d9d80e792eb07ecf28
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-sfn-controller@sha256:01036536877e5b236dc83774eb3b6fe9493eaa593cf5d1cb5b103798d0727bf8
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        - quay.io/community-operator-pipeline-prod/postgresql@sha256:f9b7137598dcaf0774de71f4633c7e3b0359b0b40f313ad9c91dc8eca58592eb
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle-rhmp@sha256:875d81ba1b91a72d9bb4f1d86e2285992370a04ace3831b3c914127ab55a57e7
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:a657269d0212d952f5107514db63c3ea2b6c86ffed0a44f0f33eca96a242cf35
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:1ec1264dcc4ee289ada0747083ad9b072bad2e33544a3a879383694c2974a885
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:d7bd0e90aff6e8fc9fbc4119e5f1ec89eba93487b84356e3af95208c1250fe44
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:fc374491270b9566b0a5c80f3dd665b2227b1e8fb224eada7e5c98c82926ae08
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:20:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:36393d8945938325e4f330eee801ba28ab9b7604f58ea927004e17bb514f328f
        version: 4.17.13-rhodf
      entries:
      - name: odr-hub-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-hub-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-hub-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-hub-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-hub-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-hub-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-hub-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-hub-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-hub-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-hub-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-hub-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-hub-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-hub-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-hub-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-hub-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:bb352ffcfe7e79981f96fc5b5a4f2f557fc7fe39de27c36f195349bf4ad6248b
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-hub-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-hub-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-hub-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-hub-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-hub-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-hub-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-hub-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-hub-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-hub-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-hub-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-hub-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-hub-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - registry.connect.redhat.com/kong/gateway-operator-bundle@sha256:91b3d73bc75eb170035110e5c63747e4918f738841aa97b6df27d22e37b4879b
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        - registry.connect.redhat.com/loadcore/keysight-wap-core-bundle@sha256:2f8295f8f56ef2c56e8215d4125cf08facd2f1d07fad0ef3a01eb0a2ef5f0555
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.connect.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:623a93173e8f703849238806b40aac843cb2ed016c8763a4f8ea1a0b0990eeea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:411a47151609b7da29974ef3ae3b04f1bb1526f4328fd3e1d9b0165380018049
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:076c345de449353cc9837c603024ba9d790d26c5a8cebfc97db61ded1482ab78
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:f481b2e364b9fe417a1cbbe9eff70180a3b4153a14712a8935527e25f919ee0d
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:a85d4f1c8a686abc0b9032dddda75991216d44dfade214625723d7009be314d3
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:b71367012d1ed17d3967cd77fc28b38c69c76f75848fb20021dbb176dc382eed
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:32b442eb6344d3ab64d2885f53817e61f0e8b72b5c6e28845c8e056fc176b666
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:268aa90a9f6eebf6cdeb32dc47335e75e7483dcce4ea3331b1b736c2fd79bea4
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecrpublic-controller@sha256:6acdf23661d603339e998307e9d5130b21c45665f4b0c35019ea80aac3d7343e
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
          createdAt: "2024-09-09T20:32:09Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:6e80289c93ce0ef26c776beca59eec792832dc7f2f57cb76a7321eabc12405c8
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:8b8d39f64cef8a1ed37a9ee44ab3bd5b98575ff32fd3962f94cf7a9c79e623ae
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:6173ab8b48c178ea8bad0d6b98e73801aa28135fcb7455146449a66543891e26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:63ef9626a82b5d591abaff2e9fe76f52e2bac16b910da40af5e92920694a2a67
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:c88a193654c191820bee307d12ae5e721ef82740ec912de9f710481b386edde5
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:c3cbe6cf4158aec38c67582fb79358315d196de76e0da611b4b8d6183babeb58
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:33cefb8803cd212932fe2bbb6052e53562d04084403f0166c8e077957b92c58d
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:ad206c2f78dbcdec43755eca3cb92229ee77e737e741dc276efec23f6379ecd9
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:e90e8df1726698a79ec689358d323eea7a26a72290660db62443e3f0cd696a4e
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:66dfedd62d474aaed03b1fee142b91358271591b651a3766d8e0542d08da914f
        version: 1.0.17
      entries:
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:7317ed83242340597b5cdde8ee314606ddd563d0067ee85c12eb21037e2feb99
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:16b39335e97bcc7c970a8ba385c61d8333e9912faaee813ccf8520d434b7bb7a
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:51ac9aced282f62c8fb2d3696da3696b8496d8f877e9846fe3a4a743ccd63bea
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:a82288782b61ea46d068cc09ee10608d3201c136abfc767e4d37734775101390
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:828d31bf8a441091b8aed03da213a6b4ffae260070eff58c3319156f9bf04f5a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:f938cc534fc4b24bb9449f1e1e1e904dc588e6646c98474de4b8c81d083b4d96
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:095629572a40469eb651c9f81d6c10ff46254eff8b9813cadce54d6f0075c2fb
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:91456907c61952017daf07e376b565472ef86d67a2925f60e189c639ceb68595
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-03-25T14:27:42Z"
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.0 <4.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-operator-bundle@sha256:a372f2b3b61c0012618837e349d389f9f300ccd2ae0fa005e9cd7d379b8569a1
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:da231dc87dad63f4fe28e3b438a1ca9a96f04839389f7c3c19e8368bf4604283
        version: 4.18.1
      entries:
      - name: numaresources-operator.v4.18.1
        version: 4.18.1
      - name: numaresources-operator.v4.18.0
        version: 4.18.0
      name: "4.18"
    defaultChannel: "4.18"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/special-resource-operator@sha256:df87c400713311f2495d6d02769acf29e5251330194f992c11719063095fdaea
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.5.2
          createdAt: "2023-10-07T15:15:08Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.5.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/openshift-community-operators/kube-green@sha256:e0c9f5ecdef1f7172af098f20cce17afebbedf02c0e0bd516d4d6aae736d1b27
        version: 0.5.2
      entries:
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
          createdAt: "2025-09-19T19:13:34Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-iam-controller@sha256:2d904b6462679d578505409d9bc585795983df7a4f0dd6668253c724aaa5063b
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
        version: 1.5.1
      entries:
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        - quay.io/openshift-community-operators/edp-keycloak-operator@sha256:82ff141bb86599f12d4a516fd1b6c5252f9ff24ef2547be6c58e2f3d88219914
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        - quay.io/community-operator-pipeline-prod/edp-keycloak-operator@sha256:10151387a79d0fccb47114b56525f28f507faee96e4fe29babd2ec46c3dd9343
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/ibm-bamoe/bamoe-operator-bundle@sha256:54df7eba0292459c0491ba39c15bfcf793e1c0aef042a4b3f6c587b0d45f77af
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/cincinnati-operator-bundle@sha256:50e9c88ca31182c54f9c4bc6682642a5115fa280044e6c125f99626c69d999ad
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:8ad2282afef801c794e316c93a12834d688cfa44766132231f93013c3124944f
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:d4ec9966370c3e770209042e437c53b0780401150f5385069b8476e26c343919
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:43bf1b9f06a7ace06e0b87921ba9b62f430ee3c8b2357c713054151e01f9cd02
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:5ebd9fb9277844967f55b985fc44069c1c0209360735c5321eea3aaef8628528
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:e668de3c5a2eed2b2b89515860f735be8ac03c2646ab7b65ae3c1270262bc1ed
        version: 8.17.30
      entries:
      - name: prometurbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchserverless-controller@sha256:e31d29d7c3389932e42c82645fd7b697aa16e03e20b17be0f355ef6b301d5ab8
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle@sha256:10c0eed1e8ca3154f4c815241f6275f40113ae2ba0abdc940ad9a04682edafef
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-07-07T18:31:48Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:abe7b5029a0956943b5af20cd66192a6fa816998523c05b671dc1280a0219683
        - registry.connect.redhat.com/minio/volume-manager-operator@sha256:e3a69a9ef99d98d1cd8eb43d3acca48a6806ccae5b6747fff0ce79712a7d363d
        version: 5.0.0
      entries:
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator@sha256:6b7a1e34c33b16c464745f70658a10d6ea3d96099c1e2ab008e96589ef73e7eb
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle-rhmp@sha256:d54904777c861e00858ddd8f1218cc39db47fbfb1225b8d35902ed4e13e357b1
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - quay.io/openshift-community-operators/kuadrant-operator@sha256:398c9230878feb756ad393d7de5675a7370c883a18acf73483a660e9303d807c
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        - quay.io/community-operator-pipeline-prod/kuadrant-operator@sha256:f21456392af3b37ae48d77dc7f19fa23c6946269555bb1280978b29420bbf31e
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:fb7bc294096fa78f199d909f6ff8f7fcbcd771c3dc19f3f767d77dd13a2081a9
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:10e4bdca44a1a0cdef27b415765238249f81ae6ba53f94f992e37d350bc9f800
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:3e8aa02c71fc99e5eeabfce39c8fc24a9f9e8e826a58f9ff59c8b7c0e9c14e0d
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        - quay.io/community-operator-pipeline-prod/shipwright-operator@sha256:f4d3b97812c69b396c8e6d081d8fa9b412ebe1365432658288e6ab065c220fca
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/istio-workspace-operator@sha256:7a581f048aee64701b23f1a0e94d77f2e6932ee8229fd46b1635d8df6994e174
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        - quay.io/community-operator-pipeline-prod/kubecost@sha256:ff1e00f5a04e49fe9daf06db6b60c2ca14692efd098a38ecd622ef6bc5de8dd5
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        - registry.connect.redhat.com/loadcore/keysight-wap-ui-bundle@sha256:7ca0365e849fc98950021f86f3acb9048303fad3dcb6eac0ee09dd4bcdaf8fe0
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator-bundle@sha256:3715abae1ffeea2d0bebf665b1dfaddce4686bb8a61f0697e6fd201b157798b6
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/rhosdt/jaeger-operator-bundle@sha256:f250e39033d7cb1d786e5a7ec6798c25d4c5d8c6ecbcf6828915605fc4658da5
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - quay.io/openshift-community-operators/datadog-operator@sha256:b1977170e5991e1f08c49037d6f34873c8543d29d3a2c3692ebb1d66734f5b3e
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        - quay.io/community-operator-pipeline-prod/datadog-operator@sha256:0a5684e5bb442c5daad6737f1560198ca380507f2da583efd0a7d66e18f2dade
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:4e12e48d0b2b884de2a0d32682da044cff419aea53bb162915538e2179fcd7f3
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:9eca42cc297a93293200a1c46e8ed949a61770550366044ef54e36bed9d7a116
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:31eda4f1888b180feb023c42a6ec93478b2e956f13420d55987d9abf7c7797c3
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:7d5f9e02f3d6c3e2ac63f1fd33240dc696ee79563c214d21d2897ee773517e9c
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:5b061126dc3adf1fc9c8c22dfded18f4652bd62537aefca88ddcc585078b94ae
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:007a3d43633fe355202002737f766a7c636f537f3daeffa1f62356c0bbf969aa
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:87594d78977ead074e53b39a52e905a2d2861d24d69d2f2e33b066077cd279a4
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:ff8e563d97f0d0e42c690ca3e61d3625e920947264ea99c396392c440cd43201
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:12a02131f13e532a26572a524ae542766092ff75624113994d304824951161b4
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-operator-bundle@sha256:006edc597eec839f960bfd191b75217fc8c6bfaea9d1694e7ac9fa34266f9b4f
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/service-telemetry-operator-bundle@sha256:c4ec01d0d27eaf93cf820d8da1f2ad7ef3e557c5b4ed651c5a1d07a1fc116f79
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
          createdAt: "2025-05-30T01:26:40Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/saison-technology/hulft-bundle@sha256:b7d1348b16ae564bc1092dbdb2042bc2a1ee8a3b7fdc85b7d6fe18be9bb99c19
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
        version: 10.4.0
      entries:
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/annotationlab@sha256:b69fde0eb341114893aba097d58455c9e316a97cb829b41c0d080987490a83d3
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:2dded4ad8e80f711fc18ada34ea825e844045da44f0813550a586371dac84b52
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:a20986fafd0a475a5a77296b3f8a18c78edb84da5cc8246b06c805dd5f693511
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        - quay.io/openshift-community-operators/hawtio-operator@sha256:38b36745547a08afd10e1620f8243e6f1ce18a3eb8883ec78cd631577177c5ee
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: helloworld-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: helloworld-operator.v4.18.0-202506160933
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.example.com/v1alpha1",
                "kind": "HelloWorld",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "helloworld-operator"
                  },
                  "name": "helloworld-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-05-18T20:30:40Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202506160933'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HelloWorld is the Schema for the helloworlds API
            displayName: Hello World
            kind: HelloWorld
            name: helloworlds.apps.example.com
            version: v1alpha1
        description: helloworld-operator
        displayName: helloworld-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - helloworld-operator
        links:
        - name: Helloworld Operator
          url: https://helloworld-operator.domain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/helloworld-operator-bundle@sha256:056744dd517acc3c12908e5ce06bfffe2c634d427b837e244dd7185c70915c00
        - registry.redhat.io/openshift4/helloworld-operator-rhel9@sha256:7c9a290bef3368d01e44bbecafdd99d2e18376221557938ab82c483bf189f370
        version: 4.18.0-202506160933
      entries:
      - name: helloworld-operator.v4.18.0-202506160933
        version: 4.18.0-202506160933
      - name: helloworld-operator.v4.18.0-202506090359
        version: 4.18.0-202506090359
      - name: helloworld-operator.v4.18.0-202506050817
        version: 4.18.0-202506050817
      name: stable
    defaultChannel: stable
    packageName: helloworld-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:13caed923cfbf9d5ff215c8f87fc64ecd753d183f5573c31ea89c9bc5915c6e1
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:c47f099ae5ddd9b25fb7aa5ecfc253fdf24df2e7929318bbcf162583df91f886
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - quay.io/openshift-community-operators/azure-service-operator@sha256:2430ebe724ead60e84f12aa485c8dae4af82af10d14400ce76356392030c8c0b
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:e16accb01b2ca6bf4ebd60c6ce7faab73b784828578ee74305a3a97ff9db1e26
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:ea37c68775b14e7d590b6eeec3c9b59eba8e988860fc04d40775da3957d6d040
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:1673f08529a94671f881edec6fd29b73f43c2115dc86b643f9dd79d7e36be38e
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:5eabf561c0549d81862e521ddc1f0ab91a3f2c9d99dcd83ab5a2cf648a95dd19
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:758bfd5455d0ca96f5404229a73fdcd56686564fc4c315ca3468ebe8588dd9ca
        - registry.redhat.io/openshift4/ose-tools-rhel8@sha256:486b4d2dd0d10c5ef0212714c94334e04fe8a3d36cf619881986201a50f123c7
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:0765d248fd1c4b4f8cbcbee23f35506c08c9eeb96a8b0e8ff1511319be6c0ae6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:0ad1961d6dd1c8636f5f1bcb5ffdbea51ad29b88c6539df366305ef801c79877
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:06382df99dfff9f002365a31eee0855f8032e351bc9583e42e489ab7623fe287
        - registry.redhat.io/rhtas/client-server-cg-rhel9@sha256:f0c8f5129c0f611332f7b34dfc5f0dc39c73ed8a985a0845d1fe6ce31d2ba020
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:d4ea970447f3b4c18c309d2f0090a5d02260dd5257a0d41f87fefc4f014a9526
        - registry.redhat.io/rhtas/tuf-server-rhel9@sha256:6c26d2394b1d82995a6f3d3483d2c18befd6a4ae826af2cd25aac30dfabef1cc
        - registry.access.redhat.com/ubi9/httpd-24@sha256:7874b82335a80269dcf99e5983c2330876f5fe8bdc33dc6aa4374958a2ffaaee
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:496b29b00a9831278eb3a8316bd5fb233ad73cb9a6217b1ff4a5f72a91b87414
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:58140d6b7fee762b148336b342a6745b1910ab758a701b2e8077b1fca2af1180
        - registry.redhat.io/rhtas/client-server-re-rhel9@sha256:83b9f9a5ea40473c8525a16d9965b1ed1eb2d8f7c2ec1d17ae38e5f10ac1652e
        version: 1.0.2
      deprecation:
        message: channel stable-v1.0 is no longer supported.  Please switch to channel
          'stable' or channel 'stable-v1.1'.
      entries:
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:540556453c98db3919e5f305a99c699e767e2891a95fe7527ace578f3c67a308
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:6131053778ea04e437f3005f90d1138aa11ebc58e3a9295e2a8d8ef6713a52be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:09b8259ec9e982a33bde84c7c901ad70c79f3e5f6dd7880b1b1e573dafcbd56c
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:571c48ecb2658ce70199c06dc483ea48a16e032a764a6830388ad845f508d981
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        - registry.redhat.io/ubi9/httpd-24@sha256:366cc5bc3b96069fecaa0abeb9d3510c507e69a20de1d205d881da8ae8561ab4
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:dc994a95be22b0f4bab022fc362c4f44c6a7d1887a2eb0d04870d75654ec013b
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:5a752cefdaf28bfc53847185cdd5fef1ee47e3dcff8472f8a8bf7bbdc224ef57
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:75f1049431f9e92898a4217870309cbbb3b39c8362e929c0bad3b53cad4459db
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:603a2ffa54df0bd45ff42ac710dd16331560eda2be93442232b045c959773c4d
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:329a0dad7842d4ef60430b0e3f188231bea9e579427443ecd42eb62a5c282266
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1f5a30a285a16635a7234c3c1763dfb385c8bffd605fc862b782bdb5c6c61ea3
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:3b8f49c41df15022f8ffdf3a8f8605b14c14f4e10eae754a06a86b6585d158b3
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:775b261b8f258b369213246f2da5c91218678acb5e3b8bb4670b143f487767af
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:76c24a38ac89ed632d38e44049f37e4997abfa27fa8cadbb8afb42575031296f
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:796860a3e85712c60398c36983e0ff4d45325c7a4de869da2ebf1b6ba4b19825
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:7ce611aefdfedd8b2a633def482cf41f390c95b8f8c800b6163a585f117a9e2e
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/openshift-community-operators/prometheus@sha256:1de089becfcceb3747b3a3ce310c1612861b03446906cab5e981071db10709a9
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/openshift-community-operators/spark-helm-operator@sha256:9bd49276e02f2fb28bb1ff116112cd7ebe0aa12bdb9f1e46aae2501fd9e4e755
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-operator-bundle@sha256:c6bc928d220fa001e2a7b5e8661a3c8a604c21385dc1bb97b8dd121ca2347916
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-iam-operator@sha256:d7e2507481b280118997ec9fe8062a9d56ed0867265732ecb5c89ef6fccd98fc
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.connect.redhat.com/vastdataorg/csi-operator-bundle@sha256:925cc412650ec3a32ea4a55ffe65d16b76916d51f0b57ff62182b4ce12c5d2d8
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator-bundle@sha256:e0e7fc3a69575199a6fcf93985451dd66a5eb7d1ab8caecf059d7d52f3953dc4
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle-rhmp@sha256:d90f6bce87b783a80b1b81275cb61162c7c2a7898aa67ea85e3223deb1dfcb4b
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-operator-bundle@sha256:1a950c41ce2ae204d6bcba959b2d1a3df2d994bf1fefce4c7095c4ca20a5fd6d
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator-metadata@sha256:7fb2750941193d64f8cca278ff7d81d3ca5e29a2ad0568dff7ddae4396548bb3
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:57e796822bcd705974c0efcd6c190ce39eac2e70b9d6a0c27f69d85dfe966db2
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:1e7781ff09c5cfb87ede98b96d31e58b760bedbf7a23f41778654faba3507591
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:f234e8af232f2d3a9472307ae82834f96756841b8d60b8c42ac4f9ca3fd17545
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-lambda-controller@sha256:9b2943aa770ebd5638927b19f05f629575e1f8fef129f8520e01c2238771d7dc
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-operator-bundle@sha256:bcc3521d20a1662d4032901f46e3bd4744c9910b060f52bdf10c6a8dfea1276b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/sap-commerce-operator@sha256:9b6a3c9c7a3af1a91578fa06c202d128fc7bfff3ae21300a4382cffb11def17a
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - quay.io/openshift-community-operators/seldon-operator@sha256:864e87ef21e81624010a1cbcee6b17c8c816dcb0a699c66badd32eec3baf39fd
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.connect.redhat.com/ogaye/ogaye@sha256:f35a21fb3936b259aabe27b453612616fca4e6327c9042361b4d00ef332905c0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - registry.connect.redhat.com/hpestorage/csi-driver-operator-bundle@sha256:39fedc5650ff2fef6ed1d47a5745851101b7dd8f6ce1f8b72ac3a78bf3c1fd6a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        - quay.io/community-operator-pipeline-prod/cass-operator-community@sha256:428effade29fcb3e8981e84f7df945f9bc49fe1a9b56217573db8b387706c26b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso-operators/openstack-operator-bundle@sha256:945271aec89541d6d8a749c41af2fff3fe24e41fd50e935e3944451c940bf940
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
          createdAt: "2025-09-23 17:13:29"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.16-3,hyperconverged-cluster-operator-test-rhel9:v4.18.16-1,kubevirt-common-instancetypes-test-rhel9:v4.18.16-1,kubevirt-tekton-tasks-test-rhel9:v4.18.16-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:86d21c945452fb380a895aa96630726fdba73b4e10549aa8ae353c7cac37fe44
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:aa24c4a937c9c1e52c30f229af5b26ce413b684375801d0f5f1ef2a50e8502e2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:6542fcf51604d3ba95ec0d42aa7bc19bdef940e1959808efc98f170a68a7d01c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:82dc19865ed64c129b5214b413f5f8da16d90da532c40d3e02b28f767d72f3c5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:a67782645abd4a869ca8db132953f5f4da72c853fe16f9521c77f241a0703bfa
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:134cb74301a9f9f704d35dea38ffa73aa3ee0ecf43b7fb3e73a89c9b432af9d4
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:6cbb8785d06e58af9b8fc1a66d0d5989a1b490b04d0191252cbb16ace14a3cdd
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:424ea484cd63823b4585e630ffc4053c32309e3f7dc24a212582afc60064e503
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:01082e3cdbac6fa50bec206401660dedd832a71f1f9e52c27281587f1f0122ba
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:a4b7120550d915cb74f03dc3530082f3a0d6850809062469f59934d8518479b9
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:0db7542ab7d67866046b1dde2f3ec8a04c47caa9c622ea9832dc59d127b4e919
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:fb9321f06c842dcb628f9082094fb697146e8740130c889bee81ba6941177303
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:cfba9b634c5a8871511ae6c1644e63560fa51d8b2729eb0db207b89ea9a1b3e9
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:0384141b5c99df83759f69c88895bc23f946da385d2c6db0fc75f80c0853a77c
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:2cffcd87d0a42748ffdbc63707751eb8212ce34ac33d9225690f04ce696178cb
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:42e5ad70d114edd15c6e4d13d48399f572ce71ead1f45a9d94de2262a77754db
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:3da5b2baa67e468fe72efcdb4caeeb4f3384365719d5edb5fff6656edcd74141
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:521b8f630c112e9d8eb4d5686ec7a6907fff8cf491628b99d6aa4a10253b3d78
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:350bc99751e2cac633b98c633b6c83e60bfd55dde4301282ff8eb61f85ea8d49
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:e21e94b2778c786d2b6aa0c107df267055b1f7d36c589a5900d313b3fe3404c6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8ab2c1ea24981ca684505c2f9e050cd938efbcbdedabd74897e0c6c0d5c5a433
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:68b318dd7fed09f60f6df215c9812b38791707b4e9af1de27acf91365766808f
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:c3a694c846fc85709e756a617d2581c06062d482b0bba88983192f1efbf8cb21
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:7e60c4e1fb9a0e10b039887747287cc06ca156f813d88c7421f3950eaf07bf31
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:710d24725c6aa3b8282d32a310d3e107b6a2e57708425d5e3acde9e82e33591f
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:deaf31fb3ac137556fedec55c44abb71d01c6927aaa21e5d23627ce139f12107
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:8c3f276434dfb7462c556f7be3f93a0c857a463e2fa2befcaef3a7c0859c7368
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:e51dfca6233c67b1b5372c682a2fc6e0fe54f2beef86e4c9451aabc80017a5fb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e3af0269b0a3225b2b5d43cce57a40aef6788830e8af8f4330bd36e4aaf8ec97
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:1c4bcdb3e2074b3f99c4dd0c57a264330c7989b8564efcc50bfe0e086d487f5c
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:27d1ea63ca9dc4fe073a28793f574da3a991a4c36e8da55fc2c7889cede187ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:9f8ab3e7e76ccad5a794de310ed0ba39e070e00647e42cf2b9a1bcdd8c377f3f
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:e0d239237d8052c36aff1509d65c00909b8563ce14341b7eb7fb79e9b84deace
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:af0b3d4305dcc5ccf00569deddf122aabf444bda0d3f0c2a5c4fad7e0c4117e2
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:071b2bf8f6470a6f87c3804eac3a10a947376469e42a55e4f19d3b2fc98c7e95
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:8339ad1cbefafbb91be533f1980ffb0e22114d52fd5c11c1b06cc55d9554c9f4
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:f514eb6d468fdbc6f5bc2b60cafdaff19ea55cb9a9c60f22b9db466f5b8f9000
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:c839c0eef642ac854e04129b007bdd529f9b380840e052fc306ecc0d62e4d055
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:d661d52528a89d87764e16b5d1dcb6ed328075477453e69c6cb9814e2a2d0ea9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:21c02925a871875a8eb70efda72f79d4c287da5cb9c0a4fc19de9ebfcb06e496
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:ae92990aa7ee3cf0bb98549ba78882c032afc8ed323fa83d1e8de24951b09d11
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:ecbdb919859e032ba4f2fe3af4dc79e87a5ad28545942d9d75c734aa814b7b23
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:e6507ebc8d71561772e573691b160f60ee6819b3f6404d5f0db123f0e4bc17e4
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:250cb53bbe119c5609a99ad32e97fc23572af9c08490e0f5212cf0977f555827
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:7eb05621f7cfe8162b08bff8a7411d6b05959affd7dc0b7729b4ccf1d90bb005
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:55996948c6d66abfafc630d7f6fe4b67201af0b61613a39e1747218393fb5e01
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:7f61b66615c9b22ac0e7e6ded50fb5a4f0a71ca228d34340977d82c8eb8322d4
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:09abeaf5bd5a7ac9f272759932e8442828efe6935dcd3b365eb42b7655c1de18
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:d32a745d70478945bb25f04ca85c17dbc1fefa0a791522e42591a6876f085029
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:db86fa7625d770a7eb8a4af977499f4569023992d33caac0caf8e3b71871f574
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:f5ebfeb2d0bcb0e86999d4efdd08d03c34c90daa2523b1382e02f9d4930ded67
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:e3f9e0c1b53a54354246d9c934ea48c038a547845d6a75c7bfac5e82f780ee8a
        version: 4.18.16
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.16
        version: 4.18.16
      - name: kubevirt-hyperconverged-operator.v4.18.15
        version: 4.18.15
      - name: kubevirt-hyperconverged-operator.v4.18.14
        version: 4.18.14
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.12
        version: 4.18.12
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.10
        version: 4.18.10
      - name: kubevirt-hyperconverged-operator.v4.18.9
        version: 4.18.9
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.7
        version: 4.18.7
      - name: kubevirt-hyperconverged-operator.v4.18.6
        version: 4.18.6
      - name: kubevirt-hyperconverged-operator.v4.18.5
        version: 4.18.5
      - name: kubevirt-hyperconverged-operator.v4.18.4
        version: 4.18.4
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:a32214d166ee692a4cae917fe97e0535a4a696155e1c878ae5335c77dadef0df
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.13
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
          createdAt: "2025-08-17 16:56:34"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.13-6,hyperconverged-cluster-operator-test-rhel9:v4.18.13-4,kubevirt-common-instancetypes-test-rhel9:v4.18.13-3,kubevirt-tekton-tasks-test-rhel9:v4.18.13-2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:a4d0418e49929e186e37f20f64c46f8eca20e314f6da4deb55f0b5407686c2e2
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:9fbf91cd986fe4064b99d6eb4bde7a79f58966763765a5001d9ae75ed22655f5
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:710dd1bcf4e171943d5260bd16121abd75058bbeb80bc8c237d24aadca99585c
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:396eec2f40ad14e9348b229498313e478903125dbb8c6f7c5f3dc3dad359d18f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:36a46082ca4ce5fd4784a9613d49c30585883bccf7433dc2ecd8de787eace937
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c73b90403d587fa6993e736fd84c08ed39381ac506502c3711da99dd4ea39b4b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:47d3263ddc6d80e844f8240502df81c0b71c2ed85a29a0d4afac7b5c98379091
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:6c39b35d0142e12757091831e2f9900c5f8ebcb1498e0eb75b62c86ee2e7238c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:8a69cfd1bcf5745779a019499632141ecdbc4b2d9a4858c66d28df11087359f3
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:b7771fb216f69af28f49c615fa050bde00d27067acdb505578d104ce1337febd
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:21ad0c18f57b964b86ced3498ac0c36a6daf8a60bdeece77a0b41b168dbd6586
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:877fe3539fe1a4d9f8cf25bfe2c1bb847ff9413693848ad539c26442472c1d3e
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:6ceaacb96b29fbd5a07dac9e8aeb038b76d53f6dc30188b88068755f35da68a9
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:4f806c32d077c6021033733674ed5701bd45086f9fdbfe90985813e2efc27ee9
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:1e73a912388b8b3de05e669128d30dd3081d5747932dc10287d613e19f59a49b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ca5969eb64a112ae5c12ae3b6752b6858dfb999d78d61daffd61db5e07a7cdb9
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:b0a4d68244205f839c0bb2c5cd4f4f305537ce57b42d24d0b46fa4fe2fe5cbba
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:960ab26612067d50aae973b63ffe482649c90ff8f42c2dde698ef548422460cd
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9600fd87e8331989db167870e83618bfb5f273e34e28c2006fa72b49aceb86e0
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:d90907bbc0d2810d592316f609bc8fd0f551de4c37ac02c1c03ccc014cdef056
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:c78c9ee196047ada7ecb614a42cbbfe3dec934534942ee5cacf34cf030e50378
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:f54b97303ba8579a2ef38d971035117efbab46d0eece22a759bc2fba98d11511
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:1d984a34589472f1c554bb4a53b6df470104affacd341eccd5fa1e334a7d8854
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:28d40c603ff685d922149ae30dc77ba1d9bc9af0ce2ab6d9038ff8272940ca08
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:c399069fc6a0d3b3e7093ab6cc2d540d05a04a68213a24a7a500a83d084246df
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:3c2706973b2bca2fbbf88d5af1de0b3c53ef1c7e5029b9a3ef42f961795ad65c
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:725286e0c9265a66e3f4b2a86fcfa7cbdda4e963863af9fb2af9e1226e10f34f
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:3e546a40d6a3e2cce2567750f93260b3db0978477e8ce34f63cd5d195c58c195
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:3765263eb3b193d860699d3e348346b762eabca51ecf31f8f97db69f3ca5e4bd
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:8985f049a49cae810ffb59dd4c94afca632808d3c8ba40809c5447a382b63bad
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:7095f50cb41030a9f49af2ad9a99d81407b5d8491c421c34934f788f02eb7528
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:f501a890dd87d37eb24cdd86d5889eadfad1e454fcc782e8ca2454150e8dc494
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:04b135c11a1e4518febbad5ee56e118ca711ad3454db19f1824cda0d4d497742
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ebc17f70a654181527a8a1d39e069076af82343cab55531fd7a86eaa85b12fd2
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:a9788ba6756715f286ede2fa04e460fd021b24f2a0df068c2aa626c89814a7b5
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:5794ed8b4af6cc37d278cd91f08f0418b89550269d69c9196772fcca1a4933e7
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:0c2c54d18dddf8ae7f2993dd42b3d07217f63266218c329119dda13e0c431fb6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:f15c753ca6687605d956d5200ec9104067c9bfe66e987dce7c32bc45ed79c57d
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:ee7ae60e4ba4b1d13b0aecb053e3ccd22b62d2277d28413f394d32e6a3ffe139
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:b5a6c60e663b81cef29e9622af8f0eb62838be99ff51f576ca1b275c6e5a204a
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:4c58b6e5356866deb394d55e2f96e4d83489d24c988e8016ccddd8738b895b7b
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:88d6db3caa017e4c2d8114af1abd28df4d0751b01f291d42b84586378e0c9484
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:0838f54f7ab4fbc04346ef67501fbca0b6c2097558bbd9c6fec2632ba0ce7bc8
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:9e46d9a0ca2e1678a02a52d14e2d8f0eb007b21ae105100106532cd12f317a41
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:a0821002f32c485c28d43dcc3afe5f447ecec4a567ef1b5f07b162f96245cbe6
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:fd17e015ad32d54f8464bc337c874b9131e0edca750adb4165161f0d382b8c3a
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:45c7a05260790a6ea78c809316dc58ed98b32cd94a884e82f24650051a27a163
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:2e6b6d7d04690a616eb210fe4b142372807bb9c59b6af6e7d6fa196f8dadff3e
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:8e69000f409c115b6af634c057106dbe894395708a694c3a4cb030b06e77407b
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:498282a9a9afceec8587eef445214f5db6ed6c09c2c3362049c4097a210e5aa5
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:11031e69287b4fafa77e2d54dc40d8c7de64e7ed8528abfca7a798b5aa8970b9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:faf31a311d34ea8bad50c83918d5ac1dc32277b9d8e356a866d166f97209e40d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:61f96cac775c4f636f0685900c419a6771fe3f3b44648d93a176b4ab5efa9092
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:70219973c00c12519a1e5abd9fa3061669de8e7273ea1dcf9f72d40ae683d244
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:f0eba5e54eced4a4478d395b098b3dca201b7b374f9e6b7f4a5c1669953c0991
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:be1b6c7e71d25c45b0f4bc2ac8507a1c215e30e39f51bc9f907436f6bcc96b4b
        version: 4.18.13
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      - name: kubevirt-hyperconverged-operator.v4.18.2
        version: 4.18.2
      - name: kubevirt-hyperconverged-operator.v4.18.1
        version: 4.18.1
      - name: kubevirt-hyperconverged-operator.v4.18.0
        version: 4.18.0
      - name: kubevirt-hyperconverged-operator.v4.17.4
        version: 4.17.4
      - name: kubevirt-hyperconverged-operator.v4.17.3
        version: 4.17.3
      - name: kubevirt-hyperconverged-operator.v4.17.2
        version: 4.17.2
      - name: kubevirt-hyperconverged-operator.v4.17.1
        version: 4.17.1
      - name: kubevirt-hyperconverged-operator.v4.17.0
        version: 4.17.0
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        - registry.connect.redhat.com/fortanix/ccm-node-agent-operator-bundle@sha256:12746020eefd383963e08a86e3542608aed1999b4e520435e366a307a6a9eebc
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - registry.connect.redhat.com/microsafe/operador-aims-mobile-bundle@sha256:1572f715c13d852c2a824b726912c4e5cc49aea43d0d91b6fd673f7198b0d038
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - quay.io/openshift-community-operators/ibm-security-verify-access-operator@sha256:ec3970c7a6a5ee3ac863dfe6f16234fe19cd7ea4a541fd0cbae9a8183d0763a2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.18.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-05-14T20:50:15Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.17.0 <10.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.18/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.18/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.18 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.18/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.18/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.18/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:f391499545a2981a3c3492766bbaae6a72a82db07d88251e85fda5d672a108e0
        - registry.redhat.io/openshift4-wincw/windows-machine-config-operator-bundle@sha256:4ebf4ace7dcaf0a935b83314b2a514baffca704b2f36cec5cd61dcf13e459946
        version: 10.18.1
      entries:
      - name: windows-machine-config-operator.v10.18.1
        version: 10.18.1
      - name: windows-machine-config-operator.v10.18.0
        version: 10.18.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:6e58f2d6a1619a06be05807ff5407924087b7b27611628f9baf999733b67645b
        - registry.redhat.io/web-terminal/web-terminal-operator-bundle@sha256:fb969c8b8ab53ab6639795086efad6870dd212f04c94934b75910598efcf3ca3
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:0b133afa920b5180a3c3abe3dd5c73a9cfc252a71346978f94adcb659d683404
        version: 1.13.0
      entries:
      - name: web-terminal.v1.13.0
        version: 1.13.0
      - name: web-terminal.v1.12.1
        version: 1.12.1
      - name: web-terminal.v1.12.1-0.1745393748.p
        version: 1.12.1+0.1745393748.p
      - name: web-terminal.v1.12.1-0.1743657317.p
        version: 1.12.1+0.1743657317.p
      - name: web-terminal.v1.12.1-0.1742484227.p
        version: 1.12.1+0.1742484227.p
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/customized-user-remediation@sha256:dc517e2349fe993e3417b858865fb73b98b7b163c0b4231d4d7be1f11992bb78
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pubsubplus-eventbroker-operator@sha256:2761ebb89dace09a1c1ccad92c28c877f3a6f75a674191ffc3788be7fa549926
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        - quay.io/openshift-community-operators/openebs@sha256:ff179bdd6e63b8e2cb1ce3b96d5e03d20c1a2410296fed6fb331bf5bc2e6b14d
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/ember-csi-community-operator@sha256:5dce9b3018cfb300cd651c7980353cedc6617383d5c525e6a4a39ffb8e105774
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator.2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v2",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "600m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "baseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c",
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "image": "registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733",
                    "resources": {
                      "limits": {
                        "cpu": "400m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "10Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "50Gi",
                    "analyticsvol": "50Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-02-15T09:21:28Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v2
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: ABot-Operator-v2.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-operator-bundle@sha256:35f90c08438e4e5aadd01de20bc82e59454c394c0e1d24ade6c56d6e02053bd4
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c
        - registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:e3b3c32d12223097a9d0eb466c3f49fbdab519a675123c7e69597b53ab4c79fe
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:50f1b4c65541b56b85f42bc9b3127bc7bc9faad1974067ce11481de7596e4c6e
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        version: 2.0.0
      entries:
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:4603bf04e53ce948bbf0ee5d239d5e265b40598044ef9e1f4539f4e466e59cb2
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:a178716e62ab8a752bcb9023ad9631af2dbc72357266cd13b43398e21f230828
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:c5288fea5b2ef8125f7b1ff7f7633eccd4e52d14ac9cfc5e473c2e317018ca3d
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:704925128f81d4ffadc4dcfaa69fc8011cf3d0f431d24222a8d7a03bad675331
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:8da482ee4c94b2546bf2b83345b1e7df6e70dcc63f68d8a2e839f10138951d2e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/openshift-community-operators/ibm-spectrum-scale-csi-operator@sha256:01d24dfa0d6b6146230459f8f7e74f2749d7fff3c4d75c40de386e014f9367d6
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/3scale-amp2/3scale-operator-bundle@sha256:934fd9123d1da778b9f8375854f2a61167f0efbc76371edfdc9962f7626e694a
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.11.8-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:9c2aba45971c924fd55b2aab50ee8e6861a9defd3808b38ff30316727af31f9a
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:4f807ee6e885e966853f70e1c40bbc179dc1f994b401ac868b7f2b921f9d17a0
        - registry.redhat.io/rhel8/mysql-80@sha256:7c05e61c0f4f5dad6aeda136e9a1af2f04f4de3c51004e997d4eacae3778da17
        - registry.redhat.io/3scale-mas/3scale-operator-bundle@sha256:a6c2eff926feb11da27eb9972cb86ee0ef9ca0b013afb6fe6af4f72439d557ef
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:419eb09d1b165363d5b90df01c1fa10f1f91f03649a8758620ffa49276685feb
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:a03a1ae7f01259192b8d7b4ecee6ddfcdc0cd9e2d24a8c2306accb93ee767135
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:7d07caf7b6c5a195ba4706e82f3d28c3793afab99b1a6d5a7020547c48534cdc
        - registry.redhat.io/3scale-mas/searchd-rhel7@sha256:1426e4bd4aa0eb28b479900ffc02621203886280d24997cd11818387ff30d9df
        - registry.redhat.io/3scale-mas/system-rhel7@sha256:b0fd85a4994502a3a2ae9ba6a7610c38af64b6cc897efac12feb4b115a484f77
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:22431acc2c991b6ee6b4de0219366dee6ea745122a49a0175d5fb58ed7d9205f
        version: 0.11.8-mas
      entries:
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:895fcf14b337398585f51e4b6e1f46a4a17c01a153ca889bd4374d691d89b9c8
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:dbf68e729484ba11aaafbcf1a590aca7f233de10e71b01d78c319dc5b5c6c278
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:e9824dd4ca39accfc674b88f7a2b76fba3006ff856e78402348c720f2aa8a6b7
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:767d55dc7343406637dd16f1b3a73bfaec353e0b058143fd68df2474271e3b76
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:43c90bbe3b20e95e1f352cdad097d63820a1282deedeb769352cb100da43ecc3
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:2f853ad4fece4510e847e96fe8085b53c56ee342e1f387677ff8cef345298e72
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:70c5237a251adb5f44a085d091ca6c24d5b313aa73f6545389a2b938740a9a20
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/socmmd@sha256:5adf9706daf49b35cfcf74b6fb07b056f87714b87047803b797a56f138c77486
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        - quay.io/openshift-community-operators/must-gather-operator@sha256:a71e5e25aaa92f03334776fee87316809bf2a46dfb34c497d63b228a194982f2
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: nbde-tang-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nbde-tang-server.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
          createdAt: "2025-04-24T18:06:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/nbde-tang-server
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tang Server allows definition of the NBDE Tang Server deployment
            displayName: Tang Server
            kind: TangServer
            name: nbde.openshift.io
            version: v1alpha1
        description: NBDE Tang Server operator provides Network-Bound Disk Encryption
          in OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NBDE
        - PBD
        - Installation
        - Network
        links:
        - name: NBDE Tang Server
          url: https://github.com/openshift/nbde-tang-server
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c15928ecc2f0ff370f567a2ca909b7a71e2becaff0af649275396e59232b7b8c
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-operator-bundle@sha256:79bf610b144cbdf4404a1153f1c395179ce1743274eec7102c862aae4f45890a
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
        version: 1.1.1
      entries:
      - name: nbde-tang-server.v1.1.1
        version: 1.1.1
      - name: nbde-tang-server.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: nbde-tang-server
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:6f0e07eb583201703f50ce9b3bf47e7aa63630074afcb588902caf05ff3ba2bf
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:35fae43ffc92770e13919f081195b35d8dd293f223fb9e577d75ae53f156305a
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:e135baa6b7eefcaf70780b6833acb689c5aef076b6280ff29c02949fbb33b06e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:6a447f455b8a3a3a8390787b36fe1d59a7d8bbf834f66a63fb0023f0557796e4
        version: 4.17.13-rhodf
      entries:
      - name: odf-prometheus-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-prometheus-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-prometheus-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-prometheus-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-prometheus-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-prometheus-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-prometheus-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-prometheus-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-prometheus-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-prometheus-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-prometheus-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-prometheus-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-prometheus-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-prometheus-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-prometheus-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:64c4cc3155e8ea6630bc7fecd392a41d71193cfb20f832b6334cd4795cf0735e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7916e81066baac027fb575a1edaab3c98a00505c73c4ac3c2530c111c11f0030
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:4352d0aecaac2efc3a15bf87925c08f7f85ec7baac9eebc29787f05af047514e
        version: 4.18.11-rhodf
      entries:
      - name: odf-prometheus-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-prometheus-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-prometheus-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-prometheus-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-prometheus-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-prometheus-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-prometheus-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-prometheus-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-prometheus-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-prometheus-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-prometheus-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-prometheus-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-operator-bundle@sha256:bd5bdc7085f3e9ce8494488f29b4d83a29796c47d536106cd6e14a7b67ab07ee
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        - registry.connect.redhat.com/opdev/operator-certification-operator-bundle@sha256:c390f32b40449fb09477e288fe5347319b691d3e89e020163e32007e6679c685
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator-bundle@sha256:4629bd129addb09478e48964a62acee88f5dd2746875801bd19d2d55cadd9da2
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        - registry.connect.redhat.com/loadcore/keysight-wap-test-core-bundle@sha256:16c938699ab0d71138693641ba7dcce2f4304da1d357a67a057c4effb4300401
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-cluster-operator@sha256:681e51da33e3cd9d6a4b617dca4338827237b70d1e8065322bc4135846039c56
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/cluster-aas-operator@sha256:aad6afc2f6f1f4e3b2c0d14b6c178d6f71ef318b0d3efc97faa5fdb736eaa721
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        - registry.redhat.io/openshift4/compliance-rhel8-operator-metadata@sha256:1c06868bce1f5d7145ef658bfcc284c44011aa24a17166aa124547b38f2259e4
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:8a2194d8d26dff03ec53ac0ab93c62c268e48675079a88db18bb14a977a391be
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:06e4abf4a94d44728b6bfa685b6ed027962b60fc2a0060a7256297857fd5f6fa
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:b1dc39979be1c347b123b2dd8070e4509179b7b499d4cf3fc7feb4552c6e2413
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:14fb2aa627ea7f38fba738cc4302cf1962660b10d3806c71161d8c7d64a8d060
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
          createdAt: "2025-08-26T06:53:32Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.1.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.1
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:498bcfc3c56bf642aaa70aec4db61b8610af0a22e9e378dfba02356770566198
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hitachi/hspc-operator-bundle@sha256:2d5e1ed3782a992034bec29d85d24befed31ad5eaf7a21ea44acdd775f90cb73
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: hspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/openshift-community-operators/tagger@sha256:a528169d305dbc2bc69794437665e4f940ea708dfd957c58d69bc79316b68612
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - registry.connect.redhat.com/amd/gpu-operator-v1@sha256:8a92a0d18a5ae6a837ca6590c2beb65d889cd744a098c22d3f01d2fffdc4dc57
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/community-operator-pipeline-prod/mongodb-atlas-kubernetes@sha256:f12108323fc990f869394e09b4ff6d9751d0a07f3c152fbd45b0f738746a468b
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
          createdAt: "2025-04-23T07:03:43Z"
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - kind: NodeFeatureGroup
            name: nodefeaturegroups.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.18/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-nfd-operator-bundle@sha256:ffb2d53dfdc63bfc9320bc4e927d037244111ab0cdbe5721ad69ce06792d9753
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:a1a60b53db81b32de0127dbb63f8c3a9ef7f70c95968aa53ffe8b49eced7dc43
        version: 4.18.0-202509100344
      entries:
      - name: nfd.4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: nfd.4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: nfd.4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: nfd.4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: nfd.4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: nfd.4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: nfd.4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: nfd.4.18.0-202505052133
        version: 4.18.0-202505052133
      - name: nfd.4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: nfd.4.18.0-202504232132
        version: 4.18.0-202504232132
      - name: nfd.4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: nfd.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: nfd.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: nfd.4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: nfd.4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: nfd.4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: nfd.4.18.0-202502101733
        version: 4.18.0-202502101733
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:f6de1136ced78ddf1ff48fdadc259c251919dd3a70819fe0e9247411999d51f3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
          createdAt: "2025-05-15 07:58:52"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.20.1'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:0885c82cd09105a078d4308b4ec1047b96f3cbf1fd555a1c90a8b053befee27e
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5783a00c386858d1eb45f224431d07d40bb9be746de965e7b129ee30969d4014
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5e822b12b5aadaa08a2f95b879be911455607a9c04922a641625515f955dc9ca
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:2217d9128fe50d61d1dff48b588113b5267002f72d025b1b6fbf87a1c2e670d1
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:90505e5ffaee7af330306c4e045d3755e0c74d30b9f45e1e7739c205d945872d
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:2e18b858f09e7595a89c4dc6ea5a85f8c5952f581b769093bfaab8156bab0813
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:6b4d34e7a40b21cf464e2f42cef21ef72a69ae8a3f4e7600a7bd85aa689117d9
        version: 0.20.1
      entries:
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:e533406d832461cd8889f5e95569ea9f8dbbd57b51dcd08b56050ba1e3775ad0
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:4530ed52522a74051aaedabd62b99d5c5274a0ab72eb28d4e75207d9e9d3391b
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:02e6baf8619cc2b782cbb1f3072e554a7638eaf3fd804abe1ba5f43f692e511b
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:df59ef548b82e9d72fecd06aa6b9c569f96485693c8a4ecb8ec278ae92df436f
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:e68814481ca33a422ed5c2b55b8a369e8b48092fbc71b138cb938a119d39e689
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:8ce258f36d3baab3175b29d5ed6df89588f128d249285bb1e7cb45539933c511
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-networkfirewall-controller@sha256:96c35a47f58c48532fa3a5750e52a6d7bc2d530b54e807d57547f8780e9d7aea
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/korrel8r@sha256:f00f8a7e2011bcc97d87fc7f5bb40a19d6b5737d1f5d7f3b2322ec39009c0618
        - quay.io/korrel8r/operator:0.1.8
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53resolver-controller@sha256:542229539932458d76180d23f1d5c5795e2de89169d8054f7b95be6f3700881b
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:3a96ab2db995d24a9ca064b5566821828d6e73ea20df77c9932d8972cb8e5780
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:c2b996e8dac2f4c81a3e48e4203499f48e4e15d0692e98d61dc8c97192e0a776
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:68eb666168d3a41cc98f91a0f2d74656af73cf76ec2cbe2cc308ae2ffab84875
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        - registry.redhat.io/rhosdt/opentelemetry-operator-bundle@sha256:1844430afea89707b70bba22cb9a2db98161cfc62b53b1086b63f109058028c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - registry.connect.redhat.com/citrix/admagent-operator-bundle@sha256:2a32d9a8cc3bd67ba7ebfe3f96b59e42009121d630918d417717d8ac9f2e40f4
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        - quay.io/openshift-community-operators/rhoas-operator@sha256:5299cd4c88012692b31d560ef981f80f994a0468552f43e03623adbf89b49ff6
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:20:34Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:85ba83a60fece2cc1c39ff245193712f918fbd55e1e283e940425fbb1b1f0e29
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: recipe.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: recipe.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: recipe.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: recipe.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: recipe.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: recipe.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: recipe.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: recipe.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: recipe.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: recipe.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: recipe.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: recipe.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: recipe.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: recipe.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: recipe.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:27:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:3c7033c68ff5e65533d2c3d0f2032cf34aecc153fd34f33ad3239331442d370e
        version: 4.18.11-rhodf
      entries:
      - name: recipe.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: recipe.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: recipe.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: recipe.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: recipe.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: recipe.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: recipe.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: recipe.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: recipe.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: recipe.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: recipe.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: recipe.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        - registry.connect.redhat.com/sonatype/nxiq-operator-bundle@sha256:48bccf9c8cc07127153af7974b949b8cd8429adf8714dcc866f026806e38fb08
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:9d24507aabd06856c7c35d73417b6ce065164f9c966053b494d68f14f0e260b7
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:00e6945419be8ac83aa7c888fe243c2fb67ee4ae5030cae117c111fdf80f5fe4
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.connect.redhat.com/vfunction/vfunction-server-operator-bundle@sha256:f1784afe63b2e834a6b0bf6da254339ffe3597d02dc2d389f9162d806bedbe74
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-organizations-controller@sha256:76ea247239dde4e52e5016944a41908de8fdcc4173cad8dfff04413a0e3cdf48
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - quay.io/openshift-community-operators/universal-crossplane@sha256:b3104adb471083de96d49fa53f663dcf0e3e7ff5e3573bb1422aa6045277bea0
        - upbound/crossplane:v1.5.1-up.1
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
          createdAt: "2024-06-12"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:87c29bad5a920ea676a5aa8f6e582492547e391ea166694f84cbb75044718a19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:a29fc3bfed3add2768499e94435e8993bfd704a8a46268e6e63a8d20d2c4ca6c
        - registry.redhat.io/openshift4/ose-ptp-operator-bundle@sha256:7202697ea75c3503d4165b1d970802ac776c2977755a8b240094046f74c162c0
        version: 4.18.0-202509100344
      entries:
      - name: ptp-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ptp-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ptp-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ptp-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ptp-operator.v4.18.0-202506260833
        version: 4.18.0-202506260833
      - name: ptp-operator.v4.18.0-202506121732
        version: 4.18.0-202506121732
      - name: ptp-operator.v4.18.0-202505291136
        version: 4.18.0-202505291136
      - name: ptp-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ptp-operator.v4.18.0-202505071704
        version: 4.18.0-202505071704
      - name: ptp-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ptp-operator.v4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: ptp-operator.v4.18.0-202504170406
        version: 4.18.0-202504170406
      - name: ptp-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ptp-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ptp-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: ptp-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ptp-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ptp-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator-bundle@sha256:b473fba287414d3ccb09aaabc64f463af2c912c322ca2c41723020b216d98d14
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
          createdAt: 2025/06/23
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.0.0 <5.1.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kube-descheduler-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * LongLifecycle
          * CompactAndScale
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `mode` - Configures the descheduler to either evict pods or to simulate the eviction.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
          * `evictionLimits` - Restricts the number of evictions during each descheduling run.
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:4c653bee02c0135c8625fea6a01e18a046013ee330c5253ff66a7cd705ba08c1
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-operator-bundle@sha256:37afe091e3b9656c107196d7120ce171d8b0bf895264a9791c4559e97c81a00a
        version: 5.1.3
      entries:
      - name: clusterkubedescheduleroperator.v5.1.3
        version: 5.1.3
      - name: clusterkubedescheduleroperator.v5.1.2
        version: 5.1.2
      - name: clusterkubedescheduleroperator.v5.1.1
        version: 5.1.1
      - name: clusterkubedescheduleroperator.v5.1.0
        version: 5.1.0
      - name: clusterkubedescheduleroperator.v5.0.1
        version: 5.0.1
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:afbe40e3dd6894210741fa65a3f0bb6e4a3d00ca261eb6bc3adac6a22251a74e
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-bundle@sha256:3e7ab287f64398b2aa434d9cfd5d99e2e9cf2d40c2e9fa1b1418537807b7aeea
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator-bundle@sha256:b2bc3e161311f8f5d2f5fc967815259697ce30bf94e980214bc05aa0e1529d55
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/odf-node-recovery-operator@sha256:77b0ef6110a71da56216719ceed0f448257dec9633c98d7dfabc1038b23b6e29
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/hyperfoil-bundle@sha256:1d3625d12d95deb4f48660459fd0fc8b250b52b46d68136927d0f12465aaadf1
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        - quay.io/community-operator-pipeline-prod/wandb-operator@sha256:9bebaf0e72f0b3ce20dabf7ace62cb5ecd1015243438248e8101a238988cd60e
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:bf79fd57e1df791fcd45dd4745c923046dd7759ff5b3762cd7a329c5bf611316
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:b03f4e88ab1fed4744356a584d673fb2fdb15de7e1de7786bdd222b853690670
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:aa261c79b421bd49f7b46974a3bb5a790d5a1a73dd8dea4e1d8a4634fb68a51f
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:123488f0877c054a348f5b4135e68d77c17523bc4e5ea14dd73168372592f4ce
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:e66501276e3a10ae6d08dad3c4658fdb8b0b635e4b2fe72fa943b0e5300d7f2d
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle@sha256:47ad4efd37ad85bccf915a6d9448d404b7ce3164d7afb4c944648eeba9f0b04b
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        - quay.io/community-operator-pipeline-prod/neuvector-community-operator@sha256:788f742c9f2def80d0b768eb96c5138f2cebdad767ea4ebb8cc26451771e23f5
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle@sha256:6c09391edf49d71b558f8ddbaf4e947dee17490202521539910d78b5ddcb3531
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        - quay.io/community-operator-pipeline-prod/ack-prometheusservice-controller@sha256:dd9c62f7493d43c6f28055e6f1aa7d586e25b2d6d359418782cf18de45da32e8
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/datatrucker-operator@sha256:4ab1d683d95d3eeba944b977594aff2e3b3e7203f56598b2fb2f7cff75354be3
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - registry.connect.redhat.com/kedify/kedify-agent-helm-operator@sha256:d176136a44e790298158ea3c4232e2aa7d0ad88ff18b6a5b39238162ed41226a
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/community-operator-pipeline-prod/codeflare-operator@sha256:3dbd2b65d7841cc97c37db59ae8aff9991c4d0f0324c4391afd41c4de20c232a
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        - quay.io/community-operator-pipeline-prod/trustify-operator@sha256:49def1c9612621f90d2398cbe7385488c167cc8ccdcd951f098433eb42cb9ac5
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/susql-operator@sha256:4a47a8448610e54b8dc9700828c9250b84ee19126d6d4edbb62b35597e40bf37
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/grafana-operator@sha256:ae9aa1d05ab9dadcc9961ca4293d0878fd7656be79c869c7fdbee718f259ad86
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/orchestrator-operator-bundle@sha256:b82fd43d2dad2e525464177081869c923e239dcb09319458162f33a96370a6e6
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle-marketplace@sha256:b7f67aec4e911d89b3e6217e8718c349c20a859149b829a0390807c113e27196
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        - quay.io/openshift-community-operators/tidb-operator@sha256:4f73b39de6129b906b5eb7bb0b7485f3da406cfad8616f81e375cc8ffcb156ed
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/openshift-community-operators/dbaas-operator@sha256:b2474b569406fb0f47a7042042429bbc5cd2edbbd119e701c6f440a139e1b3ac
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.829
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.829'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/aquasec-operator-bundle@sha256:59e62e5cbd479eed9ddd083db881469056c35253d76aef6bda9cbdb578e05b3f
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/console@sha256:05bd172efdb4729027ea6a491c874a1a3368156a506b7b5d228837b12282ee1d
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/gateway@sha256:cc8a3e950ae45fb40866b7369aa9393e92d1ab91dc0795ee682217fc421e9ec2
        - registry.connect.redhat.com/aquasec/enforcer@sha256:b48f026d7876656c0b365a90bcaaca7f4544f38b85400c110cf20e57c67ac8f6
        - registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:ce827db4a2b01b054e33f1091ee7046a14757b8781444bc9868783fe1efbf641
        - registry.connect.redhat.com/aquasec/scanner@sha256:d0ab5e0dad4e042324874de02ee36585a93b53ad1c3e905416af2fcb98f15e2f
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.829
      entries:
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:ad6b8e10abfff1cdac1ea35b44c456e3d59b4dd1372cd47c2e4221ed93e1f71f
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:167b9063a2dec7f9a5f1c5f3c02daf481c45c5e6a2f1b25353102e85921b6fd7
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:83c5e966addca5197f25c5f9c81a52648db44c2fd9e4969bf30b82c4880f0a40
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:bfbb9d0d5556eb067543f4e139a935000319b439f700240adaf393c68bd8244b
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:e0b1cff317bd4e35334429d1cc7c548dbda54065716aea96ef30610f0b91901b
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:eef6e7cf8f76ccd5e9ed16a5e4bdeb4398859121a6b0e86284b9766408cc2b30
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-bundle@sha256:31fd95c5232105d3354515102f099b9b662a8659f08da5d67d4e147d4f934fc7
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
          createdAt: "2025-09-09T14:18:23Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.18"}]'
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5189547a19ea81eb18142ad41e6412577eb9ac64d3987ba15b6665a1117a2e55
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:63d8b19d73e5cddcdce67c94e677666d3385ab8dfa1a534b9860fe1db33a2f91
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 4.17.13-rhodf
      entries:
      - name: odf-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
          createdAt: "2025-09-11T15:26:46Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:d45bf7fc74c58bc192a0c9d30157bd745f7d48b2a47820e3b768e8cf29b3c6a0
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:e76c59039b39b3caeabca73353a91501e8374e832c45eaf4005b1c08244e00b1
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odf-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:b4e4d262801b8ff616710075ff05689a8956e975e8ac563e50204a26d1725458
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:afc1e5bdade3ff0e3918fffb8fafb2700dfba7c473f99a2f347cfc7f889c8678
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: cluster-logging.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing, Observability
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
          createdAt: "2025-05-14T12:47:05Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.

              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:255a73d34b85f25b3655fb2b63954aa6b635b89080a8c249fc41568fd8365b10
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:2495de935f503808bf8700d9d66b8412d0f15fc754cf569d02abfbd74b5a6ee8
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:8c6decb25972dffb37917fb5f5a91ea05fb21b87823eb75807d8c601c12fc1bb
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
        version: 6.3.1
      entries:
      - name: cluster-logging.v6.3.1
        version: 6.3.1
      - name: cluster-logging.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - registry.connect.redhat.com/nginx/nginx-ingress-operator-bundle@sha256:4e032549f2c36c5a660994b5ec460ef214bb875a53f6f7e9b814d57049683a7a
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        - quay.io/openshift-community-operators/fossul-operator@sha256:95c043ddc818f44a303e27508e0acdcbbcb3d9b1329712c85a8e3fd4cb647e9a
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:17:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dd20cf6cd0b69404d2ddf5554e7f89870429c117740d2df51f3be5a8163b8db9
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:26c47d0498f8b8a61338cda3fffb10724c7e0f78d74d14537580763614130256
        version: 4.17.13-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-csi-addons-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-csi-addons-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-csi-addons-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-csi-addons-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-csi-addons-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-csi-addons-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-csi-addons-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-csi-addons-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-csi-addons-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-csi-addons-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-csi-addons-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-csi-addons-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-csi-addons-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-csi-addons-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:25:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationCronJob is the Schema for the encryptionkeyrotationcronjobs
              API
            displayName: EncryptionKeyRotation CronJob
            kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationJob is the Schema for the encryptionkeyrotationjobs
              API
            displayName: EncryptionKeyRotation Job
            kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFenceClass is the Schema for the networkfenceclasses
              API
            displayName: Network Fence Class
            kind: NetworkFenceClass
            name: networkfenceclasses.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationClass is the Schema for the volumegroupreplicationclasses
              API
            displayName: Volume Group Replication Class
            kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationContent is the Schema for the volumegroupreplicationcontents
              API
            displayName: Volume Group Replication Content
            kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dc3cbef1d890b1be5cd59d4e2fcb91f4f88e2aa53bf59c78a2cbe9af136ec3cd
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:c3e9d2deb1454fbd423ff9eeca34225df86f92b536c962300a75158505341678
        version: 4.18.11-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-csi-addons-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-csi-addons-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-csi-addons-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-csi-addons-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-csi-addons-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-csi-addons-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-csi-addons-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-csi-addons-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-csi-addons-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-csi-addons-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-csi-addons-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:a7b60f1c65a6fb4ee0109c7b2b0bd12fb9487e2f6edb059bd18cfd2337ea9116
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        - quay.io/community-operator-pipeline-prod/apicurio-api-controller@sha256:c783cd6db1ae5a6152848f5e4258e89b37077a177199be905d0cfe451a48e5ef
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - rasen708/cxl-operator:latest
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/ncn-community-operator@sha256:15edc1898fe9f8f56708da7d6de3548d06f1df3bdbb3f66a9cf4f7990fe9fad6
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        - quay.io/community-operator-pipeline-prod/victoriametrics-operator:0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        - registry.connect.redhat.com/flux/flux-operator@sha256:13c2f3b8092fc7b84548dd6383edbb8e41b7a8f3acefe63620f07f7a6454d8b9
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.connect.redhat.com/ibm/modelbuilder-bundle@sha256:4940d853ba477aeeec99550c35c132b69bab8bd7b6fbc1a2fe118c828c7ba088
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-elbv2-controller@sha256:f40f12efb734af5930c225925e99f31eb8db5fecece704a27ea50bd8661658c2
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - quay.io/observatorium/opa-openshift:latest
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - docker.io/grafana/loki-operator:0.8.0
        - docker.io/grafana/loki:3.4.2
        - quay.io/community-operator-pipeline-prod/loki-operator@sha256:b7e52fc9b9a988eed8b8338a2b2eba34b020cb24a189c2fa4db63110118dbbc1
        - quay.io/observatorium/api:latest
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
          createdAt: "2025-09-19T19:13:23Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sns-controller@sha256:f535d7861f6ef1488ccf95d0c5629a924033acb133d9452dc86e4c80352e69d2
        version: 1.2.1
      entries:
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
          createdAt: 2025/06/24
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-run-once-duration-override-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-operator-bundle@sha256:8ff7f1c707b40448a719137f37f6a3e09860812f7f07c3e61b7f73d1216399dc
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:42f73634df2c06216fac7b3742b2ac21ef5f12ee98d8694c06f7c9073773acd0
        version: 1.2.2
      entries:
      - name: runoncedurationoverrideoperator.v1.2.2
        version: 1.2.2
      - name: runoncedurationoverrideoperator.v1.2.1
        version: 1.2.1
      - name: runoncedurationoverrideoperator.v1.2.0
        version: 1.2.0
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-paygo-bundle-certified@sha256:e0a8c4d0de20c91b73957d6ea09fb94005e680a092c54389e6e411edbeb59296
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-access-operator@sha256:6e37f565564de371b5797895f3e96e0efd4f592d7d3e899ea8507438bdb60e78
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:2647a21dfb60364cfb7c4d1b81508adc13d859e97553bb2f710901e6ad59f4a3
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:dfaeb7ac2ccfed03042febea709f0ab66a4c57d62fa041b7ea887a0783d60a36
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:8d20352c0091985cf93e39cd18b2bb092bc15fc5e9900fe46ed5c6fc3c0af885
        version: 8.17.30
      entries:
      - name: kubeturbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        - quay.io/community-operator-pipeline-prod/ack-ec2-controller@sha256:58d3249744b82a7f75ea2b1461adb03259350930da791da62841f05d91e94d01
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/egressip-ipam-operator@sha256:1e2443488e0aad361cccece51689e2b8da882862cb62c1208a14ab1b5ec2922d
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.connect.redhat.com/intel/sriov-fec-operator-bundle@sha256:c30849cbf451eb016786bc367e3716389b1d45676ad298c4de6a492c7bb00f4f
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:2338eaab863304250b3b1d267e9f2261c48eef1f97c280b7fa0af0ebc3ba208b
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:40802ba8bf531637f19adbbaeba3631e0b0557168ecf61f566fb5f2da25d1d4f
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        - registry.connect.redhat.com/loadcore/keysight-wap-db-bundle@sha256:3940b8cbf7579d9ce888f176f96c4d191a7cdfaec4be9251e8da4942e7897e97
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-codeartifact-controller@sha256:9bc95a0db7be356d2f657ae61c1c3098ff6e7e8bf1a675810dfb1b18805c6c8b
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - quay.io/openshift-community-operators/airflow-helm-operator@sha256:b3d05d2e2b4ff8ae64e4fc0338fcfa1ffdc345c6fdda8eaefed4e9108750fe58
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/kubernetes-imagepuller-operator@sha256:a72e3728dcf395667820c10c4c5ccb3f3d040ae5d04f50680e193398d3b72b88
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:5d4e4e62b03888bcbf6336934f2ddf5c5cb3d91bc8482a3aa1697e09625c6865
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-operator:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:df0939534c6aaa988714e64728ecdb387ff7dc4b32a5d97d56703ab143787171
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:ef65de2143df2c93aa73f099fa66d39e3239c30a9d547ed7bf258da0a19e8e35
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        - quay.io/openshift-community-operators/prometurbo@sha256:83444ab464bce4fb5021a2fca6d4d6ff72b40e90a6cfc9c5ac6c74357c775ad5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:032f053f21e273714e9f8a8f6854c83bc71cb84f5b32174704b490eca09eebc8
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:40b961a28ea9761c052a5cd17d2452fb72ea4786a92e13a0a1b589c7de31bf26
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:e4d902722d83f82372021d49c80f9880dfdb8bb312d4635e1d97980b60ca1fb6
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:961aa224525116fd575cff3a37a6465e9c652ee446927dd2bd03249b29469472
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:ad94f962d7aa064dc2b5b12ac2ea66ae5973909c82e6a96829165eaeff7ea952
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:504672b2db7fd25d1c85c9614f8242f98b4d2d04184a947dbaaada41ec205b45
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:c3913778f986ae4f3a7d94da2a9d0157233dbb676e3c1eecacbdbc02a864685a
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:e80f87f289ea944a78b5c38df600c0a22fa470a2ce90f0b1e9453cd63586d75d
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:bddfd4ebd973089f7446cb16ed9668d418f91c3907a5ede22052be4ccbea6d0c
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:b31ebfab28ec75c0d7110d6f01e3b26fd973c36013308e12f69584ff38ab3779
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:7edf59729c2ae8b4506d5361eda01a98f33243987ac082abf758719c76a4804a
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:94bade4574c3679a425628f68067f9aead57bdd78e60f635041f421ed30341de
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:f5dda99364d07df3a3680a2d7031c27df071c8e80746b1ececb5487240e9bbb0
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:a4704a3977fccceb145e8972d2d7de074455bc228f2e925e5b427ffe3dcc623d
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:4f8f9dae947e9d34e86dc4237aec9167a9a25f83cecfae4982dc952494502486
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:6c0899610a78d2737c2096552e5146d9d5602b5568fab24b51b409dbffcf1961
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:84c6c826473b844293bd2048a498c5e0e8a558224a09f8e2e0cc85a2f10fcacf
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:30375ef2cab46b967f6d2cd46862377c83e58bc6489eb0e75b495a173ebfd84f
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:466b9c30c02a3dfcd9cc94fe8eb432dca7076b8c79eb97ebe5802d716fd963f8
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:d5121af03b32a4b2286ff4f39499931a1ec04e05af45dccc6599dcc1b49e27ad
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:a0739ee7420b64e6195ec53057b032c356faf0ea4b8c1943725052f925eaeb19
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-s3control-controller@sha256:114ea0669459ae148ab32f5366f239da9cfde23cd6bd6a67c14e5844856837a4
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/openshift-community-operators/nsm-operator@sha256:330a930f885dacf65537beb600fe7f428e02687e83778e47dcda3dc2df15ae24
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        - quay.io/openshift-community-operators/machine-deletion-remediation@sha256:34d1e6cbfaf248a7b28c2a6e6a39f3fe46c183b9c38ac5fbb98eb3f421ba903c
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/machine-deletion-remediation@sha256:99b0b439d37ecdac8ec74433acf1a0ef787d0d6d1d7ef7cfe5f0a29f9dfe6071
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.connect.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:02d101a5929b60e612bf64a944e282ad9fd353c47d4ff28d2c9d5c0b84704ce5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        - registry.connect.redhat.com/coroot/coroot-operator@sha256:44aa4ea98d468f1d1ced671bd2e2bd5bf79ce467f8c572c2248abdce2b7c5acf
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acmpca-controller@sha256:c3929eb9a46af8feb2c830fd0b9225b6e3ad2559a9b5e04fda9239ef4cc07d9a
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sqs-controller@sha256:a22009a5d2037aacf52d223e45b74e6aa08020d4cfd1836ff137254b73e19bb9
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:90f512ce90433a33b50c58613d9b7136e00f116c48e179c4a1be89d6a35f9abc
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:53e73c53c31e0e06c990d3d966c2b4053d559a58a17a781155a3ceff3efac831
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:246712ce23cdd590ff83d84e9800b4ed761886ed5bab9f9cda4125a5d930e7e5
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:68e766235bdcd8f3335e2ab6e09942d98b183b3e498abb65acab6481664d46ed
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:91afe3cccd5eabbe4933459e63fd64a6bd87e399ea1aa81ae30dc3e8341c83e4
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:56db2fd54dbe982dde6a4558b664b19a676a8ac99bfbc8082d097cd6d71a03ea
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:d17af97462de060dc21f997dac973d9f895ce91f80821f1a810d93dbcbda31e3
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:783380361431859301477dfe63b6915240e934ffd07a2af2651094579be47a15
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:46f8c0ea6e383b742ab9cdaed1c8f624c2cf405bb1ce39c0ea6977940d52de07
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:e4ea95a86e77aa64da0b1800f00c2e0e598776cde7cfc7bb9941a60bac207ed9
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:aca3d91b7b9f8c1e8a1e0d045c42ba5f27444e0c1c0bc4a87c697b428c2fff78
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:5af7a32da67d39e7a011672de333792e63604cbfee997c19a495fc759d2fd739
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:74f0cf91e2feb60fe794f71b9f56fa8afc60e53379a0aef104701dd6692272e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:72724358b360ab74cc0857e2f18749d3d5b4c76bb6e930f31c5b04c7716e4c2d
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.connect.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:61f935c5985994994eb2f454fb12d74312f7d7d1627fc2b55f3d6ec9bdeb4bc7
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - registry.connect.redhat.com/mongodb/mongodb-kubernetes-bundle@sha256:354b16ea43cfcb9bff704b6265fb10d94037beb231138e1fb44963f3b348bb40
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.21.0
          createdAt: "2025-09-15T11:39:08Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        - quay.io/community-operator-pipeline-prod/kepler-operator@sha256:37fe88e7234e5eb07c493a6412f27bb8df33896f7eab0a94b0e396fcdf870715
        - quay.io/sustainable_computing_io/kepler-operator:0.21.0
        - quay.io/sustainable_computing_io/kepler:v0.11.0
        version: 0.21.0
      entries:
      - name: kepler-operator.v0.21.0
        version: 0.21.0
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/namespace-configuration-operator@sha256:971991d2bf717001a98b175c50580cd7db9d230492e72670606c2ed50c53feb0
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ecr-controller@sha256:03ca7f898cfbf14f18a06cf3b510cddade3863275df934be891734b8c177c1e6
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        - quay.io/openshift-community-operators/eventing-kogito@sha256:abe8d5803d2d96ebbb9fe4cda0feaa0217e92def6a669404362bcf741d934690
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:e5e0c3b3a25c13ea916d70fa6f6ddc20f2c730bd9909aa2654672280c27afda2
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:8c65e65bbaf0ebb10dba03655c5854c69c82ddc0e6b2a29293eee029222e5efa
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
          createdAt: "2025-03-03 15:18:52"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virtio-container-disk@sha256:3cff6989c7d7f72d2f0ac27e512d99873768766ba420d65a6b7794676d0c1e88
        - quay.io/kubevirt/virt-operator@sha256:d5fce85dd4232745408daebfbf70c2bc879a06f8340d1f7448b141d5ba3f1821
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-uploadserver@sha256:2b59d3c0a7c154fb6915cf3439b42eb4aaf48f94025ddb75fa9b73b16148fc13
        - quay.io/kubevirt/ssp-operator@sha256:61ac0c182528a9b54d84755fc0931ad8d408ae74e7e4b644316a5e9e630ab360
        - quay.io/kubevirt/libguestfs-tools@sha256:d655b9abbe8cce45bce3934e11d01715e170ddda4da78d7c51f356e7ee1e3263
        - quay.io/kubevirt/bridge-marker@sha256:e492ca4a6d1234781928aedefb096941d95babee4116baaba4d2a3834813826a
        - quay.io/kubevirt/virt-exportserver@sha256:ec1693cc6e8a14ab83f2feea87427643a968d65832c44b25a1bb911482403ff2
        - ghcr.io/k8snetworkplumbingwg/ovs-cni-plugin@sha256:516791acf430bc747f01366e2b748ac76c0b5e39ad8592be2b3dcc809429fa1e
        - ghcr.io/kubevirt/kubesecondarydns@sha256:8273cdbc438e06864eaa8e47947bea18fa5118a97cdaddc41b5dfa6e13474c79
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:66c55f5ffa4b44f2e58990baa7c672ec9776612a5f1c57fee5a5ab1c080f17bb
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:902af624e05eca764c17137c27813b1618f388e4b1a53b27b54665e8735251a2
        - quay.io/kubevirt/aaq-operator@sha256:5e53b755adaf6c66f59812b590a2eae2d84554fa658e93621ac977316843d918
        - quay.io/kubevirt/hostpath-csi-driver@sha256:f8d44aabe0024058b4380e1af42a57a90b7f8c66c8627091bf425fc384c0880f
        - quay.io/kubevirt/virt-exportproxy@sha256:55e38274e59bfe5a2cb5671534b8b6e75d90e43107abecc95fd463f7be8f0026
        - quay.io/kubevirt/cdi-apiserver@sha256:1849bc14687449d36561eb1c768fdbf0cd14ababf9c254b33a4ddb14abfba01f
        - registry.k8s.io/sig-storage/livenessprobe@sha256:33692aed26aaf105b4d6e66280cceca9e0463f500c81b5d8c955428a75438f32
        - quay.io/kubevirt/network-passt-binding@sha256:c794cc1b9173267f1e3d34e666d9e2f782b88d93b47fb4449675346b467f802c
        - quay.io/kubevirt/virt-controller@sha256:68a4967ef5925d4c24fb5eab0aa5acf0a82946beb7f1139bd45c85efe3653cec
        - quay.io/kubevirt/cdi-controller@sha256:94d16a6bd0d9776eddedf64165f39e63e4dfe4f8a26e858bd099db5d09136047
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:92f2ac10f28a7e80699bd984dbc4de897011b1f68c2be95f3fcb90df410b119d
        - quay.io/kubevirt/virt-artifacts-server@sha256:789f68f4e27cf947cdd4f1c0732042176e67b7faa5e37e1b503b8dace3a46125
        - quay.io/kubevirt/hostpath-provisioner@sha256:156ae46a03e2ff5550967773100122f9abc82e25bd7acaa55dbdec22238dd2ca
        - quay.io/kubevirt/sidecar-shim@sha256:cb4025f7275f8de2891c2195c5faf128ac1b68595cdb3eff776a2e5360b8f034
        - quay.io/kubevirt/cdi-importer@sha256:f9fff83c8307bca61a22e8296cd103b32e63d7c8b2a170c6781c6cd962211d26
        - quay.io/kubevirt/pr-helper@sha256:fd434b4ea1ca5682fc5b05f23de01130c0c10f77a0a2d9d089d8f3cb5b545aa6
        - quay.io/kubevirt/kubevirt-template-validator@sha256:b227679f8ccc7c53469574ee5f58c9cc39eb6c54ec01377e4b3df3205a2f4ab4
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:89ddc40700d52cda23104978498fb4d74a643e807450b630beaa71ffdfa1fc57
        - ghcr.io/kubevirt/ipam-controller@sha256:fd7c38e129ea4ed8a86d1801c9e1ca387802376fc4abc0b0d228d4278f448ecd
        - quay.io/kubevirt/aaq-controller@sha256:b9fd0818e4019cccfb38f0923db058166eafdaea8bb35a720aa22b5ce55daf26
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:16030e8320088cf74dc5fbc7dccfea40169f09722dfad15765fa28bceb8de439
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:ac955c4ad03c40d74341623b22f99e8f796a6ed2ef5cc539fba3c1504d159955
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:8061bd1276ff022fe52a0b07bc6fa8d27e5f6f20cf3bf764e76d347d2e3c929b
        - quay.io/kubevirt/cdi-uploadproxy@sha256:6a98ef32c3778e46812c698970e14736f178fd2a11b1f74ccc36c34ce2e34466
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/kubevirt/passt-binding-cni@sha256:f3fea2ec66a26850489197edf7c85d074eb78bbac52916b10016ac0de8ef7c14
        - quay.io/kubevirt/cdi-cloner@sha256:feefb3933780bf352ba03b62b900907d02c64e73f7e74b6d336f89daf2752f62
        - quay.io/kubevirt/cdi-operator@sha256:8297771144f2f358d893771a9dde208a6bedf6bfd67ab2927b16c27424ca3f26
        - quay.io/kubevirt/kubemacpool@sha256:6cc8ca860960c003b36754c6dde249f26f78d32eb22159a25b5647e197dbc9f9
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/aaq-server@sha256:18c23eb4e32ed083047c8808b35fa81a22294a7e8bbabc74e2612fd2eba97dd7
        - quay.io/kubevirt/virt-launcher@sha256:18508d8767f8530df4d6aca4ffe467c40f9891fbe052548b3b2316f8bfb1bc38
        - quay.io/kubevirt/virt-api@sha256:d92c78b3048756fd11d3c6f1a48f4ce8555dd39e866039313dc23bfa70668443
        - quay.io/kubevirt/virt-handler@sha256:4ddd39539282e79cf3602700b791eb91a6d79991bf2e4dc4cba24e3da9db3d1c
        version: 1.14.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.14.0
        version: 1.14.0
      - name: kubevirt-hyperconverged-operator.v1.13.0
        version: 1.13.0
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - quay.io/openshift-community-operators/eclipse-amlen-operator@sha256:bb89e2a7cb1c9172b080c72ad3227017d135fc30076d2a68b4e215fa8fa55846
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/eclipse-amlen-operator@sha256:5fbfb0bd4f764643cdc1021e99e489514a1f035d85131b13e56b11a3045883c2
        - controller:latest
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/microcks@sha256:1313bb08037e57cde02d9d48c4d8c9663eed08651a3c29f872b5ef85ac764112
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-community-operators/openshift-nfd-operator@sha256:880517267f12e0ca4dd9621aa196c901eb1f754e5ec990a1459d0869a8c17451
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        - quay.io/community-operator-pipeline-prod/aqua@sha256:ca2b998fb0fa4e80d6f4c3be57041e8e55e43cf0b5f1fee579733fda9ea6a107
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ses-controller@sha256:0cdbea8336c2825bf1bc8dbd3f5235ff41d97fbed30106e7775c6e993f92d8b6
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        - registry.redhat.io/ansible-automation-platform/cloud-addons-operator-bundle@sha256:2cb3b1887e8eeae6708f5f46c9bfb304d7c3df1537ea433b0b23772acacf9f90
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
          createdAt: 2025/05/14
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.3.0 <1.4.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-operator-bundle@sha256:3804fcb77893d9198527b4801e2a778bf612596c9a0b125a0f0f872b23138434
        version: 1.4.1
      entries:
      - name: secondaryscheduleroperator.v1.4.1
        version: 1.4.1
      - name: secondaryscheduleroperator.v1.4.0
        version: 1.4.0
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.3
        version: 1.1.3
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      - name: secondaryscheduleroperator.v1.1.1
        version: 1.1.1
      - name: secondaryscheduleroperator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-operator-bundle@sha256:44d82d5f8434fb6e88d2da406e87df0bf4274661d9d77554b80e09d19fce03de
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - cloudoperators/ibmcloud-operator:1.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/ibmcloud-operator@sha256:b22e89daf6b7f0fd55ba6db7eb7d7a48cbaef2844dfe20e558d7b821482e9480
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudfront-controller@sha256:235ad7b4ad07644d9a638909c681620b24a85148a8655b7c903faf6aa42c2242
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator-bundle@sha256:0192000a2994a62a904ce4aabfc1ff04caf2fa9d779d27c016922b9121877fac
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.connect.redhat.com/synaxg/synaxg-operator@sha256:ff3d9ab5506c4487e5b6697ffe960584aadfe9f16408e7976d345c3d93375171
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        - registry.connect.redhat.com/anchore/engine-operator-bundle@sha256:fbbe7e6c1d75c4de2f47e2c825c930568e85f1134545e9d890a0c9f3d9187a4d
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-kafka-controller@sha256:1a32352f30d0835299e14e8d0f0bac5585396dd6a10a63fe633e3d81c4d71b50
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/community-operator-pipeline-prod/windup-operator@sha256:4e2bafba54eb01b0ae77555d2ff3da2c34286c7c173756dd20a90de28be36535
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/layer7-operator:v1.2.1
        - quay.io/community-operator-pipeline-prod/layer7-operator:1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/mongodb-operator@sha256:1c06d3e936348ceeed773eb45144c9dd702e50ba293610421a93e3b3942ac9ef
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        - registry.connect.redhat.com/raftt/operator-bundle@sha256:7108686c168c35acd996e61a981bdd885b454837cf6823a48a28cdeeebbb0af9
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:d0bd55bad99bfd1ce09da5ee885a38229829570ebf80e1acc4341dff17a54e64
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:b9ffc18b2d88dfab92583af35b27e0ae36bceac7778cb448ab2ea2eae2ab3e94
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/community-operator-pipeline-prod/nfs-provisioner-operator@sha256:d7ef0fa3b5a1a6094b6fd60c6197a52cb14766865b3549806de100bb3522821a
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-sfn-controller@sha256:01036536877e5b236dc83774eb3b6fe9493eaa593cf5d1cb5b103798d0727bf8
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        - quay.io/community-operator-pipeline-prod/postgresql@sha256:f9b7137598dcaf0774de71f4633c7e3b0359b0b40f313ad9c91dc8eca58592eb
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle-rhmp@sha256:875d81ba1b91a72d9bb4f1d86e2285992370a04ace3831b3c914127ab55a57e7
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:a657269d0212d952f5107514db63c3ea2b6c86ffed0a44f0f33eca96a242cf35
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:1ec1264dcc4ee289ada0747083ad9b072bad2e33544a3a879383694c2974a885
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:60b1d25ebb3aada94985e359bb059d4491bf0148ac75d51b68a461670659e429
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:3128cd81b56975643c1c88c6ceff6d72e918a67ec2d6a5a606b6e04753c5b0ce
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:808529c14a8873324096fbd1622fb895c4dc7f60c1cf9e3ded646f9a81d7b496
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:708291528f26c70c561745967a1200edcb2f984687f749a49a2860fee580423f
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:d1c636cc17155c4764d331e050f9bf3ec9f39f6ec6ec14e972179b185ea9f5b8
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:5339c41013727fa5dd3ae5bc1d51eb90e02e5baa85d4039b1efdde30baeae1ae
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:1532fc1351ca44f863feaaf54ba8810be7a62396f98dd292ccaf975fde31ce5b
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:fa72cc3e0b09510d71a097b92ee101e057f1c11e1cb363172bd26b2d94272417
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:b897e25e03133513f9afdb5abe38cbf0f2ac0556c27518d9d80e792eb07ecf28
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:d7bd0e90aff6e8fc9fbc4119e5f1ec89eba93487b84356e3af95208c1250fe44
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:fc374491270b9566b0a5c80f3dd665b2227b1e8fb224eada7e5c98c82926ae08
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:20:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:36393d8945938325e4f330eee801ba28ab9b7604f58ea927004e17bb514f328f
        version: 4.17.13-rhodf
      entries:
      - name: odr-hub-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-hub-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-hub-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-hub-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-hub-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-hub-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-hub-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-hub-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-hub-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-hub-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-hub-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-hub-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-hub-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-hub-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-hub-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:bb352ffcfe7e79981f96fc5b5a4f2f557fc7fe39de27c36f195349bf4ad6248b
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-hub-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-hub-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-hub-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-hub-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-hub-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-hub-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-hub-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-hub-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-hub-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-hub-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-hub-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-hub-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - registry.connect.redhat.com/kong/gateway-operator-bundle@sha256:91b3d73bc75eb170035110e5c63747e4918f738841aa97b6df27d22e37b4879b
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.connect.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:623a93173e8f703849238806b40aac843cb2ed016c8763a4f8ea1a0b0990eeea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:411a47151609b7da29974ef3ae3b04f1bb1526f4328fd3e1d9b0165380018049
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:076c345de449353cc9837c603024ba9d790d26c5a8cebfc97db61ded1482ab78
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:f481b2e364b9fe417a1cbbe9eff70180a3b4153a14712a8935527e25f919ee0d
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:a85d4f1c8a686abc0b9032dddda75991216d44dfade214625723d7009be314d3
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:b71367012d1ed17d3967cd77fc28b38c69c76f75848fb20021dbb176dc382eed
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:32b442eb6344d3ab64d2885f53817e61f0e8b72b5c6e28845c8e056fc176b666
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:268aa90a9f6eebf6cdeb32dc47335e75e7483dcce4ea3331b1b736c2fd79bea4
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecrpublic-controller@sha256:6acdf23661d603339e998307e9d5130b21c45665f4b0c35019ea80aac3d7343e
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
          createdAt: "2024-09-09T20:32:09Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:6e80289c93ce0ef26c776beca59eec792832dc7f2f57cb76a7321eabc12405c8
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:8b8d39f64cef8a1ed37a9ee44ab3bd5b98575ff32fd3962f94cf7a9c79e623ae
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:6173ab8b48c178ea8bad0d6b98e73801aa28135fcb7455146449a66543891e26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:63ef9626a82b5d591abaff2e9fe76f52e2bac16b910da40af5e92920694a2a67
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:c88a193654c191820bee307d12ae5e721ef82740ec912de9f710481b386edde5
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:c3cbe6cf4158aec38c67582fb79358315d196de76e0da611b4b8d6183babeb58
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:33cefb8803cd212932fe2bbb6052e53562d04084403f0166c8e077957b92c58d
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:ad206c2f78dbcdec43755eca3cb92229ee77e737e741dc276efec23f6379ecd9
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:e90e8df1726698a79ec689358d323eea7a26a72290660db62443e3f0cd696a4e
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:66dfedd62d474aaed03b1fee142b91358271591b651a3766d8e0542d08da914f
        version: 1.0.17
      entries:
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:7317ed83242340597b5cdde8ee314606ddd563d0067ee85c12eb21037e2feb99
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:16b39335e97bcc7c970a8ba385c61d8333e9912faaee813ccf8520d434b7bb7a
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:51ac9aced282f62c8fb2d3696da3696b8496d8f877e9846fe3a4a743ccd63bea
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:a82288782b61ea46d068cc09ee10608d3201c136abfc767e4d37734775101390
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:828d31bf8a441091b8aed03da213a6b4ffae260070eff58c3319156f9bf04f5a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:f938cc534fc4b24bb9449f1e1e1e904dc588e6646c98474de4b8c81d083b4d96
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:095629572a40469eb651c9f81d6c10ff46254eff8b9813cadce54d6f0075c2fb
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:91456907c61952017daf07e376b565472ef86d67a2925f60e189c639ceb68595
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        - registry.connect.redhat.com/loadcore/keysight-wap-core-bundle@sha256:2f8295f8f56ef2c56e8215d4125cf08facd2f1d07fad0ef3a01eb0a2ef5f0555
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/special-resource-operator@sha256:df87c400713311f2495d6d02769acf29e5251330194f992c11719063095fdaea
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.5.2
          createdAt: "2023-10-07T15:15:08Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.5.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/openshift-community-operators/kube-green@sha256:e0c9f5ecdef1f7172af098f20cce17afebbedf02c0e0bd516d4d6aae736d1b27
        version: 0.5.2
      entries:
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
          createdAt: "2025-09-19T19:13:34Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-iam-controller@sha256:2d904b6462679d578505409d9bc585795983df7a4f0dd6668253c724aaa5063b
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
        version: 1.5.1
      entries:
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        - quay.io/openshift-community-operators/edp-keycloak-operator@sha256:82ff141bb86599f12d4a516fd1b6c5252f9ff24ef2547be6c58e2f3d88219914
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        - quay.io/community-operator-pipeline-prod/edp-keycloak-operator@sha256:10151387a79d0fccb47114b56525f28f507faee96e4fe29babd2ec46c3dd9343
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/ibm-bamoe/bamoe-operator-bundle@sha256:54df7eba0292459c0491ba39c15bfcf793e1c0aef042a4b3f6c587b0d45f77af
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/cincinnati-operator-bundle@sha256:50e9c88ca31182c54f9c4bc6682642a5115fa280044e6c125f99626c69d999ad
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-03-25T14:27:42Z"
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.0 <4.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-operator-bundle@sha256:a372f2b3b61c0012618837e349d389f9f300ccd2ae0fa005e9cd7d379b8569a1
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:da231dc87dad63f4fe28e3b438a1ca9a96f04839389f7c3c19e8368bf4604283
        version: 4.18.1
      entries:
      - name: numaresources-operator.v4.18.1
        version: 4.18.1
      - name: numaresources-operator.v4.18.0
        version: 4.18.0
      name: "4.18"
    defaultChannel: "4.18"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:8ad2282afef801c794e316c93a12834d688cfa44766132231f93013c3124944f
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:d4ec9966370c3e770209042e437c53b0780401150f5385069b8476e26c343919
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchserverless-controller@sha256:e31d29d7c3389932e42c82645fd7b697aa16e03e20b17be0f355ef6b301d5ab8
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle@sha256:10c0eed1e8ca3154f4c815241f6275f40113ae2ba0abdc940ad9a04682edafef
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-07-07T18:31:48Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:abe7b5029a0956943b5af20cd66192a6fa816998523c05b671dc1280a0219683
        - registry.connect.redhat.com/minio/volume-manager-operator@sha256:e3a69a9ef99d98d1cd8eb43d3acca48a6806ccae5b6747fff0ce79712a7d363d
        version: 5.0.0
      entries:
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:43bf1b9f06a7ace06e0b87921ba9b62f430ee3c8b2357c713054151e01f9cd02
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:5ebd9fb9277844967f55b985fc44069c1c0209360735c5321eea3aaef8628528
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:e668de3c5a2eed2b2b89515860f735be8ac03c2646ab7b65ae3c1270262bc1ed
        version: 8.17.30
      entries:
      - name: prometurbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle-rhmp@sha256:d54904777c861e00858ddd8f1218cc39db47fbfb1225b8d35902ed4e13e357b1
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - quay.io/openshift-community-operators/kuadrant-operator@sha256:398c9230878feb756ad393d7de5675a7370c883a18acf73483a660e9303d807c
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        - quay.io/community-operator-pipeline-prod/kuadrant-operator@sha256:f21456392af3b37ae48d77dc7f19fa23c6946269555bb1280978b29420bbf31e
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator@sha256:6b7a1e34c33b16c464745f70658a10d6ea3d96099c1e2ab008e96589ef73e7eb
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:fb7bc294096fa78f199d909f6ff8f7fcbcd771c3dc19f3f767d77dd13a2081a9
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:10e4bdca44a1a0cdef27b415765238249f81ae6ba53f94f992e37d350bc9f800
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:3e8aa02c71fc99e5eeabfce39c8fc24a9f9e8e826a58f9ff59c8b7c0e9c14e0d
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        - quay.io/community-operator-pipeline-prod/shipwright-operator@sha256:f4d3b97812c69b396c8e6d081d8fa9b412ebe1365432658288e6ab065c220fca
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/istio-workspace-operator@sha256:7a581f048aee64701b23f1a0e94d77f2e6932ee8229fd46b1635d8df6994e174
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        - quay.io/community-operator-pipeline-prod/kubecost@sha256:ff1e00f5a04e49fe9daf06db6b60c2ca14692efd098a38ecd622ef6bc5de8dd5
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        - registry.connect.redhat.com/loadcore/keysight-wap-ui-bundle@sha256:7ca0365e849fc98950021f86f3acb9048303fad3dcb6eac0ee09dd4bcdaf8fe0
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator-bundle@sha256:3715abae1ffeea2d0bebf665b1dfaddce4686bb8a61f0697e6fd201b157798b6
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - quay.io/openshift-community-operators/datadog-operator@sha256:b1977170e5991e1f08c49037d6f34873c8543d29d3a2c3692ebb1d66734f5b3e
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        - quay.io/community-operator-pipeline-prod/datadog-operator@sha256:0a5684e5bb442c5daad6737f1560198ca380507f2da583efd0a7d66e18f2dade
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:4e12e48d0b2b884de2a0d32682da044cff419aea53bb162915538e2179fcd7f3
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:9eca42cc297a93293200a1c46e8ed949a61770550366044ef54e36bed9d7a116
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:31eda4f1888b180feb023c42a6ec93478b2e956f13420d55987d9abf7c7797c3
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:7d5f9e02f3d6c3e2ac63f1fd33240dc696ee79563c214d21d2897ee773517e9c
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:5b061126dc3adf1fc9c8c22dfded18f4652bd62537aefca88ddcc585078b94ae
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:007a3d43633fe355202002737f766a7c636f537f3daeffa1f62356c0bbf969aa
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:87594d78977ead074e53b39a52e905a2d2861d24d69d2f2e33b066077cd279a4
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:ff8e563d97f0d0e42c690ca3e61d3625e920947264ea99c396392c440cd43201
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:12a02131f13e532a26572a524ae542766092ff75624113994d304824951161b4
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-operator-bundle@sha256:006edc597eec839f960bfd191b75217fc8c6bfaea9d1694e7ac9fa34266f9b4f
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/service-telemetry-operator-bundle@sha256:c4ec01d0d27eaf93cf820d8da1f2ad7ef3e557c5b4ed651c5a1d07a1fc116f79
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/rhosdt/jaeger-operator-bundle@sha256:f250e39033d7cb1d786e5a7ec6798c25d4c5d8c6ecbcf6828915605fc4658da5
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
          createdAt: "2025-05-30T01:26:40Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/saison-technology/hulft-bundle@sha256:b7d1348b16ae564bc1092dbdb2042bc2a1ee8a3b7fdc85b7d6fe18be9bb99c19
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
        version: 10.4.0
      entries:
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/annotationlab@sha256:b69fde0eb341114893aba097d58455c9e316a97cb829b41c0d080987490a83d3
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:2dded4ad8e80f711fc18ada34ea825e844045da44f0813550a586371dac84b52
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:a20986fafd0a475a5a77296b3f8a18c78edb84da5cc8246b06c805dd5f693511
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        - quay.io/openshift-community-operators/hawtio-operator@sha256:38b36745547a08afd10e1620f8243e6f1ce18a3eb8883ec78cd631577177c5ee
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: helloworld-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: helloworld-operator.v4.18.0-202506160933
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.example.com/v1alpha1",
                "kind": "HelloWorld",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "helloworld-operator"
                  },
                  "name": "helloworld-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-05-18T20:30:40Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202506160933'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HelloWorld is the Schema for the helloworlds API
            displayName: Hello World
            kind: HelloWorld
            name: helloworlds.apps.example.com
            version: v1alpha1
        description: helloworld-operator
        displayName: helloworld-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - helloworld-operator
        links:
        - name: Helloworld Operator
          url: https://helloworld-operator.domain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/helloworld-operator-bundle@sha256:056744dd517acc3c12908e5ce06bfffe2c634d427b837e244dd7185c70915c00
        - registry.redhat.io/openshift4/helloworld-operator-rhel9@sha256:7c9a290bef3368d01e44bbecafdd99d2e18376221557938ab82c483bf189f370
        version: 4.18.0-202506160933
      entries:
      - name: helloworld-operator.v4.18.0-202506160933
        version: 4.18.0-202506160933
      - name: helloworld-operator.v4.18.0-202506090359
        version: 4.18.0-202506090359
      - name: helloworld-operator.v4.18.0-202506050817
        version: 4.18.0-202506050817
      name: stable
    defaultChannel: stable
    packageName: helloworld-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:13caed923cfbf9d5ff215c8f87fc64ecd753d183f5573c31ea89c9bc5915c6e1
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:c47f099ae5ddd9b25fb7aa5ecfc253fdf24df2e7929318bbcf162583df91f886
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:e16accb01b2ca6bf4ebd60c6ce7faab73b784828578ee74305a3a97ff9db1e26
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:ea37c68775b14e7d590b6eeec3c9b59eba8e988860fc04d40775da3957d6d040
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:1673f08529a94671f881edec6fd29b73f43c2115dc86b643f9dd79d7e36be38e
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:5eabf561c0549d81862e521ddc1f0ab91a3f2c9d99dcd83ab5a2cf648a95dd19
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:758bfd5455d0ca96f5404229a73fdcd56686564fc4c315ca3468ebe8588dd9ca
        - registry.redhat.io/openshift4/ose-tools-rhel8@sha256:486b4d2dd0d10c5ef0212714c94334e04fe8a3d36cf619881986201a50f123c7
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:0765d248fd1c4b4f8cbcbee23f35506c08c9eeb96a8b0e8ff1511319be6c0ae6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:0ad1961d6dd1c8636f5f1bcb5ffdbea51ad29b88c6539df366305ef801c79877
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:06382df99dfff9f002365a31eee0855f8032e351bc9583e42e489ab7623fe287
        - registry.redhat.io/rhtas/client-server-cg-rhel9@sha256:f0c8f5129c0f611332f7b34dfc5f0dc39c73ed8a985a0845d1fe6ce31d2ba020
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:d4ea970447f3b4c18c309d2f0090a5d02260dd5257a0d41f87fefc4f014a9526
        - registry.redhat.io/rhtas/tuf-server-rhel9@sha256:6c26d2394b1d82995a6f3d3483d2c18befd6a4ae826af2cd25aac30dfabef1cc
        - registry.access.redhat.com/ubi9/httpd-24@sha256:7874b82335a80269dcf99e5983c2330876f5fe8bdc33dc6aa4374958a2ffaaee
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:496b29b00a9831278eb3a8316bd5fb233ad73cb9a6217b1ff4a5f72a91b87414
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:58140d6b7fee762b148336b342a6745b1910ab758a701b2e8077b1fca2af1180
        - registry.redhat.io/rhtas/client-server-re-rhel9@sha256:83b9f9a5ea40473c8525a16d9965b1ed1eb2d8f7c2ec1d17ae38e5f10ac1652e
        version: 1.0.2
      deprecation:
        message: channel stable-v1.0 is no longer supported.  Please switch to channel
          'stable' or channel 'stable-v1.1'.
      entries:
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:540556453c98db3919e5f305a99c699e767e2891a95fe7527ace578f3c67a308
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:6131053778ea04e437f3005f90d1138aa11ebc58e3a9295e2a8d8ef6713a52be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:09b8259ec9e982a33bde84c7c901ad70c79f3e5f6dd7880b1b1e573dafcbd56c
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:571c48ecb2658ce70199c06dc483ea48a16e032a764a6830388ad845f508d981
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        - registry.redhat.io/ubi9/httpd-24@sha256:366cc5bc3b96069fecaa0abeb9d3510c507e69a20de1d205d881da8ae8561ab4
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:dc994a95be22b0f4bab022fc362c4f44c6a7d1887a2eb0d04870d75654ec013b
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:5a752cefdaf28bfc53847185cdd5fef1ee47e3dcff8472f8a8bf7bbdc224ef57
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:75f1049431f9e92898a4217870309cbbb3b39c8362e929c0bad3b53cad4459db
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:603a2ffa54df0bd45ff42ac710dd16331560eda2be93442232b045c959773c4d
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:329a0dad7842d4ef60430b0e3f188231bea9e579427443ecd42eb62a5c282266
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1f5a30a285a16635a7234c3c1763dfb385c8bffd605fc862b782bdb5c6c61ea3
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:3b8f49c41df15022f8ffdf3a8f8605b14c14f4e10eae754a06a86b6585d158b3
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:775b261b8f258b369213246f2da5c91218678acb5e3b8bb4670b143f487767af
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:76c24a38ac89ed632d38e44049f37e4997abfa27fa8cadbb8afb42575031296f
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:796860a3e85712c60398c36983e0ff4d45325c7a4de869da2ebf1b6ba4b19825
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:7ce611aefdfedd8b2a633def482cf41f390c95b8f8c800b6163a585f117a9e2e
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/openshift-community-operators/prometheus@sha256:1de089becfcceb3747b3a3ce310c1612861b03446906cab5e981071db10709a9
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - quay.io/openshift-community-operators/azure-service-operator@sha256:2430ebe724ead60e84f12aa485c8dae4af82af10d14400ce76356392030c8c0b
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-operator-bundle@sha256:c6bc928d220fa001e2a7b5e8661a3c8a604c21385dc1bb97b8dd121ca2347916
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/openshift-community-operators/spark-helm-operator@sha256:9bd49276e02f2fb28bb1ff116112cd7ebe0aa12bdb9f1e46aae2501fd9e4e755
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        - registry.connect.redhat.com/qc-qranium/aic-operator-bundle@sha256:581b21f44d6b993888681ed702ed10b9d1bb8e52dcca53a91e856febed2613a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patterns-operator@sha256:99ac398b06fa9d9a7626b99cb47fb380d0cf4a4307fa11da8e486df35966debb
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        - quay.io/openshift-community-operators/varnish-operator@sha256:46de72c5abad80cef3d6426be967131c51c2c15029a6a38744975578c805a44e
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:3f7fc52c79f3e4ef7ddd6f5efc8eec1ee4836bbf2076e7dd5f68bb3b7c6e9d8f
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:34d4f950bce7863f15d5f1e7d6555a3ccf97ec31dc3ee140f68442890c4f0f96
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:96b77663961daea3a336b97e34e9412281bfb2727854dc970a69edd029b9af85
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:5170255ecd45b9bb7a70abec0f6eeb0518bbc2871de759e468dd61a80261c0ad
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:9a3f5b6c403fb84c9cb93c9d7cf902c80990196983607c8ca568d4b558938e17
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/kasten-rhmp-bundle@sha256:c6b58e5f6c9b2a1f5c4bb770d1fa1db6eaaee8ae7d9751a48338623d0d812438
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.connect.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:99e845965f7eca4b5a3361c494a385292d655e7c1906185197826c773cf9aaf6
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/openshift-community-operators/infinispan@sha256:3ba9d20e83825f3c041a7dfd195b0800f74fa0f5cd1e6218fbd3547fedd85d64
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:f165693aa550903d7a26b429ed4d093f60194a7a04148e65e61c52d67b527fad
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:5e1a1fdb5a78a34155abe77754c774c65335ea8cbfdec2a222ee22a062c2db46
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:e43b0581415af648c0b82830377e354a951814d3eb7825bbe31b9f3e633917c5
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/devspaces-operator-bundle@sha256:a48dd94d5616183843ba939146b7e4d985210cdd902038ff80ff6df3df3d74ca
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:18452ba29f4e0f52b765e8f9359e4143ebfe1203160e6f64ce46ec9d4f095a62
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:9716402247ed874f19574a0d5125d1035fd198f36081d9129c1d4d4ad2f714a6
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:397feef6fc1dffb7e2c3798bbe3bfebb86d6173c225eb61473029cc4500b0cbd
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a70fceaad3b6a23e6f247b4dff41fa64349c4bbf841b66adbd0f8309c6acce74
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a7a4d20b5acc59dbb8c99a37c19b2749f15ed0be7f841490a829796865c5a956
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        - quay.io/openshift-community-operators/loki-helm-operator@sha256:fa3662ec16e6cb3f866ad46982a50d8fdefb43c68f3caa13291239013e03af13
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        - registry.connect.redhat.com/appian/appian-operator-bundle@sha256:2cfb2d2e43fa207c8c06df6de8d3e90123c4cd52cdb6a4c6d618ec576b79c2eb
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:4ec16d88828e580fd9223125043ca49569324c4f06b7a65bbaa7d76923de2bad
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:7f959a4076c36d76f8f1489a92d6984d1cfd2ae5c0533a78e8b16010eb6d4d3c
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:a3132caf9cddf6d6186b3e982a6f6232a208fc727b433ee6d200a3b72f50ca88
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:d89f4d647534569f401790d6e10bc42433ae791be86558b100a4027a74b92cd3
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
          createdAt: "2025-09-30T14:21:42Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/flux-operator@sha256:557ae37fa6a6942e3a9d06966df67721011ac7905e2241867c975948092f2936
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
        version: 0.30.0
      entries:
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-operator-bundle@sha256:84d65eed99300b0ba87369cbf9a9fc17a37781f8b6f9de2819d0a159f9071a6c
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:e140415e0361cabeab192ba6d22634283cd47749dd25e9cee4a775d3c409cd97
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:4b7796235a0a545db83570add7a675fdd9d82b17f931eb81d0cafbedb1fbf22c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:9f9dadba5a35c8ab98ddacee22593a247a31e96a4b6e1b4f1a946c2e34006bb9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:01ca6a65a041a47efd611673a720e69464f02424d21feadc496c356065ff2f65
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:ac244e06ecd2f8a5e4de7b121efb8e82e3e7d23194878cb8327ff0b2affb5f04
        - registry.redhat.io/openshift4/ose-smb-csi-driver-operator-bundle@sha256:91d04cbe0eee6c4fe0230d0d179366942e6c54afb65bb77a43a681e1a2411caf
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        version: 4.18.0-202509100344
      entries:
      - name: smb-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: smb-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: smb-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: smb-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: smb-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: smb-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: smb-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: smb-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: smb-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: smb-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: smb-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: smb-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: smb-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: smb-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: smb-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        - registry.connect.redhat.com/enterprisedb/pg4k-pgd@sha256:fb206b95690e9604c41715369c69125a906b866dc35cfc5c63638ce7a4337fa2
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        - registry.connect.redhat.com/amd/amd-t2-january-bundle@sha256:7c87354aa07e783a2cbf4ab29877a8ba56ea4269bc1f41bfbe08c558b3666841
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.connect.redhat.com/loc-devops/operator-bundle@sha256:b9f7a678ceec2d367d55499a48a15a9dd729bc34b0e707989be23c97fc3c5f4e
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "aes_key_rotation_interval": "2160h",
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_location_id": "devspatcher",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "enable_server_session": "true",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "session_duration_seconds": "86400",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          },
                          {
                            "email": "admin@mycompany.com",
                            "hash": "$2y$10/yYwtXRrOagYO2jNtwVzGQPPEfDvX0QllV8NXbku",
                            "userID": "b7e2c2e2-2e7a-4c7b-9e2a-7e6b8c1f2a3d",
                            "username": "Admin MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
          createdAt: "2025-09-24T08:25:29Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/edb-hcp-operator@sha256:87833c7edfa714b433fb066533bd415049b4ae96588ea8457a85a62326801d5c
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
        version: 1.3.0
      entries:
      - name: edb-hcp-operator.v1.3.0
        version: 1.3.0
      - name: edb-hcp-operator.v1.2.3
        version: 1.2.3
      - name: edb-hcp-operator.v1.2.2
        version: 1.2.2
      - name: edb-hcp-operator.v1.2.1
        version: 1.2.1
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator-bundle@sha256:a26586b5f2d748f56dc29fbe3871d84de422e4103ce13c93dbe477718c582487
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        - quay.io/openshift-community-operators/yaks@sha256:0031de9a68502cb65923c4e17453195663f59e46b1cda37619a857add503c016
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:91632b89b4001e2ffa0f1aef56207eb48f0275e20c635e7f1ff7c6de57a59ec2
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:eed8048d085eef7eefaa643a09627521fc2596ebd26280e05c79dddaa0b01afe
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        version: 4.17.13-rhodf
      entries:
      - name: mcg-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: mcg-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: mcg-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: mcg-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: mcg-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: mcg-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: mcg-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: mcg-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: mcg-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: mcg-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: mcg-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: mcg-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: mcg-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: mcg-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: mcg-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{},"bucketNotifications":{"enabled":false}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:ed5c7c8cbbe455b6d31b1c55871585a0679a91800f53743d81d4cc36c6aac7cc
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:88387ba7a9e85c1bcf57966558db57c099fd301ea80a68e43afc2bca3dfe94a7
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.18.11-rhodf
      entries:
      - name: mcg-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: mcg-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: mcg-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: mcg-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: mcg-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: mcg-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: mcg-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: mcg-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: mcg-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: mcg-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: mcg-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: mcg-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-operator-bundle@sha256:b637a02d23bd8364cc8914421049eb60169c163ac70bff2f33591df1a1193002
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e9612ec99daa171c403b5ddf37788526e5375b83987e5de9b1ebe519198607b4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:34d177abb56b948eaf9ce6fa62e37427ddee9fcecca48e752d09f72c69847a46
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:932e83fe9c9b54934767a7c5af683b2ab08bd2b4f0524c191f1c766dc64d9c3d
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e1c19bbcd4c010ef2bd4f760eb24b2c96cdb82b4e025361a433e6a241eb23cf5
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:a6e02e1876fd5382ba4921dd65650b2d747666feb96c839f34e99d2972dfc323
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cb711fcb32b5ca0f34c4e3c67624bc4e035de4247617d988a9e841fd58c1c911
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:26af4e950a33755114298386a70229672a09ce45b6203025856f460a55c4c8a0
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/flux@sha256:cd94037ff3055ab6ce40b0670f900248a5fd233cc67a2a4f92c1559cf9d2f3ad
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/deployment-validation-operator@sha256:2803298c7cf3eeaffe9da3153142058d94d65ff1dae90efd8819a4cd23e0518f
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        - quay.io/community-operator-pipeline-prod/eginnovations-operator@sha256:19b12c75f7a2dfab027881ed9f0b09d25e3baa936d21f7be3ec1e0177ed9fdf2
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        - quay.io/openshift-community-operators/limitador-operator@sha256:7aa0a4fe55e5e099e2e6e4c2cd44b9bfeba69d8cfade10500ccba3092fb60dd2
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        - quay.io/community-operator-pipeline-prod/limitador-operator@sha256:450f9bf43b57b83e9975a1865981af22e28c725193573354d1180f406887664a
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-framework-bundle@sha256:53d056ac37177263d79098cda775deff86a327c65106ea5a40109b08a856cf84
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/keycloak-permissions-operator@sha256:4de19e52e84cdba0963bdd852a63912f097280800d65add8d474c0953da478a0
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle@sha256:a90a748aa70ee5f1b8fd375a7877db1292792ba46f7b49f629533c82a25c43a0
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.connect.redhat.com/cpopen/ibm-zen-operator@sha256:5a9120d5977e934ca3801eab44ed48ecd4a9fee10c0b4980594d33cb02e64855
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        - registry.connect.redhat.com/airlock/microgateway@sha256:80d49a299a97bed33c28941f93312e86b949ce945e843c93da6bca49865c81d0
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        - registry.connect.redhat.com/airlock/microgateway@sha256:087d4d726f8080111d2b18cd0f9ab290f4560333fef308e553991484b6210c61
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/openshift-community-operators/nlp-server@sha256:6f80f44822301252540c6b808a20cea327f0eedeadc7643d31fc743de5c5e696
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:fdc0c228ca38db30f89812e41c09ec96d9276ed844b0801013375bbffd6c1a16
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:ea606dcfa4cbe4df7b27c10d3fc45f464845518c4896845079ca9194600ceb6f
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - registry.connect.redhat.com/loadcore/keysight-kcos-ingress-bundle@sha256:f04b8f15b069c06eac5d9a97f81a2f069704237eea998d35727b89854c793746
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        - registry.connect.redhat.com/loadcore/keysight-wap-msg-broker-bundle@sha256:a7dfca98b4e09c20317cbb486e7893de8d88faf5db2ee26549ba81332ec526cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - quay.io/community-operator-pipeline-prod/dell-csm-operator@sha256:6e2e813b7ca426a793da717d42f3620edbc3eaa84be710600f2a5262924378b4
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:0739d5035590cece6d2a573aa1e799b0b0cab8b548a41cc2488cf465486cf4a5
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:3028579f4909be6b8dcd970e778bc88d0576c76e55b98c886738b8f601cdf4c1
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:e50343879b4c3c4002c5e7ff67f4664df441cf5f239f59fa98b7fddd75516d41
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        - quay.io/openshift-community-operators/alvearie-imaging-ingestion@sha256:f65cc3cb83b1c75385970ce70c6a84b15b143e682ad0d1784148f1643df0991e
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:add5d14653f57ac1e3254c2d1c70fc9232959f76ee6286fe5cb572696a5a4fd3
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:7fc4d97b5ee63085672fc9a562ebdad3d4c00d1b555e06679d1dd9ac945f1bb9
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:05e12cc4caaebb7f185f651f8ca876c442c02d79794e561b743beba0d189fa04
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:84796005b6e71efa5d783c55b1f72bcc3762944bf32740cb75498bf8b908bffb
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:424e138a17e2af91603e834ff65d529400d6793771ddaab92d3988f4e6d2f2e3
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:9d505bf561650eb9785811ba50c920480d85679eac6a611a1c5794544e175934
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3abd8c3f97c9bba127b6e8c7efcc078c90c9d7cc35d74cfcd9bc0d303f62d2ed
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:21c17448932f053f1d4845111288a317786b190154bae9da7947c315ba12f796
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:023b9cc1e4ec2542b8d68b636b21a2e8c4eb143424581c62d16598fddb47da6a
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3843fd48c57be102576db93e8483a30f8efe8cd46b5e09c704ac0bc5f18553a3
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:55bb4133abc945ce7b75e6c01c6beaf63c19c129fc8fbfb48f1b53b8e28aae97
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:43b959fcdcf107b70e4224d80b63d1c123560c3af055109bbcc7d68476546584
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:2845843f32b895dee0242e31380b44625e3f8381caee7ecb3258648311c302f7
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:c7cd3ffa88eae188e94abcc0af56d328aff43463b98a3bc3c0b574f29373944b
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: openshift
      provider-url: ""
    name: pf-status-relay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pf-status-relay-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pfstatusrelay.openshift.io/v1alpha1",
                "kind": "PFLACPMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "pf-status-relay-operator"
                  },
                  "name": "pflacpmonitor-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.19.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PFLACPMonitor is the Schema for the pflacpmonitors API
            displayName: PFLACPMonitor
            kind: PFLACPMonitor
            name: pflacpmonitors.pfstatusrelay.openshift.io
            version: v1alpha1
        description: It deploys the pf-status-relay application.
        displayName: pf-status-relay-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pf-status-relay
        - sriov
        - lacp
        links:
        - name: Pf Status Relay Operator
          url: https://pf-status-relay-operator.domain
        maintainers:
        - email: marguerr@redhat.com
          name: Marcelo Guerrero
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: openshift
        relatedImages:
        - registry.redhat.io/openshift4/pf-status-relay-rhel9@sha256:8cdc0714d02144f5b32e26c97f7678d0377e0c3826c475a784e76a42f781d3e3
        - registry.redhat.io/openshift4/pf-status-relay-operator-bundle@sha256:fb9635a614b6676fbe6a532d07b3206f12ee525b7947378785e79951d43940b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/pf-status-relay-rhel9-operator@sha256:e6719af6fdc99c652238f3c31a014782811eb46da91e65f93c083a7a001e2496
        version: 4.18.0-202509100344
      entries:
      - name: pf-status-relay-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: pf-status-relay-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: pf-status-relay-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: pf-status-relay-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: pf-status-relay-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: pf-status-relay-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: pf-status-relay-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: pf-status-relay-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: pf-status-relay-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: pf-status-relay-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: pf-status-relay-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: pf-status-relay-operator
    provider:
      name: openshift
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/rhpam-7/rhpam-operator-bundle@sha256:46e7d899d64ea8cbfc35221d05535db1023c41c1d3a7c580cb03e7e7d769144b
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-rhel9-operator-bundle@sha256:896a615487c0ffd5335b869f0dc0c0ec1c9200dbcafa52f2493003fecbc030bc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-operator-bundle@sha256:f596dc9323da89f5e45e71192f0fa698b7bb29b8205eaa0fc1659d68fa7401f4
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:8e223a537f487bc3cef10769adc6fb619085dafe3ea797cfd52c20e3066f694f
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle-rhmp@sha256:81476f897726c85293e64e6188ed314e3e1e637c568c6b0ff8f475a876bee29c
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:4ae986761a69d25697b9b6a3a99de88d8587beba136c3de29b90624e2ea6738f
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:adbfebc73bc5a9cd12d17e9e1fbdf87c2176aae7a7217bcefcec3068e51bc904
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    defaultChannel: stable-v0.2
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.connect.redhat.com/nec/sanstoragecsi-operator-bundle@sha256:1957586c13e3bf80bdc688f519be440ec1dcc4ffe63343f526018dcb93e21cea
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/gitlab/gitlab-runner-operator-bundle@sha256:68b22c412a44d5fa09b45e204d7c8e7b6ca6afa4ed6879de608dbcb0b1929676
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/openshift-community-operators/pcc-operator@sha256:e99a466ddb52ad44c435ccf952754d685744fb5b7b7e4cf77dd542e530e4f968
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53-controller@sha256:227a306cc90998489bfb681cb29b416238b404bc2db28a09b8949d4cca2d87a4
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-operator-bundle@sha256:e83707d0b396c38cf55046d3159f1f427188404aa11dc47d7b1c1f25605182de
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.3-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:64a623d5d536c790dee04c1604ec2f4429f8b8f1973b4232b543e8cbe23d6590
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:7c9b4327d8677f48c48888230fc38eec8c20c7f646dde568dbd96b5e4c8f7967
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:8797a3be94ce6f4e4da15e4efefe41ea5b6ffb4cdc6d698638c80976bb22491c
        version: 4.17.6
      entries:
      - name: lvms-operator.v4.17.6
        version: 4.17.6
      - name: lvms-operator.v4.17.5
        version: 4.17.5
      - name: lvms-operator.v4.17.4
        version: 4.17.4
      - name: lvms-operator.v4.17.3
        version: 4.17.3
      - name: lvms-operator.v4.17.2
        version: 4.17.2
      - name: lvms-operator.v4.17.1
        version: 4.17.1
      - name: lvms-operator.v4.17.0
        version: 4.17.0
      name: stable-4.17
    - currentCSV: lvms-operator.v4.18.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:5d0534e08790e91d5d355eac2e5c187a217c1ec6eb59464ea9a771490ac040e9
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:6e809f8393e9c3004a9b8d80eb6ea708c0ab1e124083c481b48c01a359684588
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:b3b1c4836a82a7ec6c9b02e4a86c352f4ae8c45a3fe027ae7200869afe3ab028
        version: 4.18.3
      entries:
      - name: lvms-operator.v4.18.3
        version: 4.18.3
      - name: lvms-operator.v4.18.2
        version: 4.18.2
      - name: lvms-operator.v4.18.1
        version: 4.18.1
      - name: lvms-operator.v4.18.0
        version: 4.18.0
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-rhmp@sha256:1502ccda3c067d74689d0341008f05362bc23904c8e35b20e020fe0dca82aef7
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        - registry.connect.redhat.com/netapp/netapp-supported-apache-spark-on-openshift-operator@sha256:4f15a0fc375156fafdf2d504f9980a8eb4451808b5554357877345cacf8f3594
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-key-manager-operator@sha256:7a998ea17ef537ec5e998e1627b065ae04ab01b7988196da1cdd1763e05a9924
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:6ef2956b297795a9d0c83119e71309549975df3ae5652e6a98cef035ace5d51a
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:8831faee6979acfb5bd031b64cbc09fb90e9ea4f48bbff15e9d39ccdf6bf3d95
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        - quay.io/openshift-community-operators/prometheus-exporter-operator@sha256:e62276f4c7e583d1fb24997fa970662f1a8d510de0baea159ba796b48cff709b
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        - quay.io/community-operator-pipeline-prod/ack-dynamodb-controller@sha256:547bdd3d1037dbc3b189f4c2f8e1d49379e1d547cd9063391d6c745e1a103b80
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:59ff4966f4f653e49d42d97bfed6351d3730128db74f666ee4ecacff11790baf
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:01f2ca529d2486f113bcefc9fedce6a6fd07bcb48aaf534394b5b04c353f8853
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.connect.redhat.com/cpopen/ds-operator@sha256:6ec2387e45d9557cea2df749831e136120945794773ed0d48e0b04f0b3d9e115
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:ca70ebce43d451d040aedcec036413f986e620234fd9a56d45983805396db0b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:4aec6aa2d2b9a24167bb6bc92db7c2c323847b7a2f3ad7f998c96bf1d2903366
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-athena-controller@sha256:8cff9258583f8acbdd26d263a61ffa11792cf9f4c1d65410bb4c71f9f7aaf178
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - quay.io/openshift-community-operators/devopsinabox@sha256:1c2ad76102151a56ecd753d9b86fd9f155d64b912067738d1858d33a02557928
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        - quay.io/community-operator-pipeline-prod/activemq-artemis-operator@sha256:1e3c6ff2d2612c3e85d40b1974b5a3c4f603b77353752389fa588728bd4263c4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - registry.connect.redhat.com/mariadb/mariadb-enterprise-operator-bundle@sha256:73e7f243a6676d559068bf8ab80cc0ffe0db58152675603ca8bd360334439f26
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator-bundle@sha256:df0dcd31867bb1c3c4321db4412a744fd1b3ba8faddb509d191fab92407b6677
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        - quay.io/community-operator-pipeline-prod/apicurio-registry@sha256:7087ec1377c2612edf3b8a76f6e115d66c3ec2b4b20f5d8374638d6bf0037e93
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-pipes-controller@sha256:e46b540219ca42c3b6c08ff99afabf09daea2dca5a335cda6013343df2f35878
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        - registry.redhat.io/openshift-serverless-1/logic-operator-bundle@sha256:5fff2717f7b08df2c90a2be7bfb36c27e13be188d23546497ed9ce266f1c03f4
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:53d8cf8bf5c1103860ffc4110e2a422371761e76ee79edfef8680e947407c5ce
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        version: 1.3.2
      deprecation:
        message: '''builds-1.3'' has reached the end of maintenance support. Please
          upgrade to ''builds-1.5''.'
      entries:
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:aadba00eadb3ac46d321a14498fc3dffd5be9afbf698d80adddde0a7a1681fc4
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        - registry.connect.redhat.com/smilecdr/smilecdr@sha256:c44f5661a598aa18c4c10372918a45edb0f26d7d2772c47a55c66e4ff874cb5f
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
          createdAt: "2025-09-09T14:14:54Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:ca10a842900d6515f714a17916a32dfd5a2909543e4d339acefc7e3e41b8d7ac
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:31d145c7074ad6cd85684b5191a3033533bf2149124c7ead5367871da5ed0dd8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        version: 4.17.13-rhodf
      entries:
      - name: ocs-client-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-client-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-client-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-client-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-client-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-client-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-client-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-client-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-client-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-client-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-client-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-client-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-client-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-client-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-client-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
          createdAt: "2025-09-11T15:23:25Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:60a91d6e4fe7e0ab999eab2224ef09fd0681e74920b23b306c7a9893bbb29cb8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:725eabb38876b93ab7693d75013d0292756a2efc00b8608f5716f31f8c324a53
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        version: 4.18.11-rhodf
      entries:
      - name: ocs-client-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-client-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-client-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-client-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-client-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-client-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-client-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-client-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-client-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-client-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-client-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-client-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-sso-bundle@sha256:f2c5a8abd1a9ec7665edf1e6206a6212cb752082b09532e4c218bc0990e48287
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:50b7f0a5f3fdd98cfb0294d664cc01ffd1e5359be66a3da14f5552e9cb3bae2c
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:29d46db99d97fe28758c0efd2acf27203ce9958734b82cf37c3a9063b5159506
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: nexus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        - quay.io/openshift-community-operators/nexus-operator@sha256:277b94985c0dc982abb76b5dd8c93b37305ccca32c000d67552d540d4918cbba
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:b83c3263e77904ce1c0f95b7bfab969605c97297a5598fd794d64daf508c52c0
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:ea88a3090291f3742ccd4641785c19a80f4d3bd485a4f9c362a7645f73896db3
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:3dbb71004362a70683a5ee9f0916bca75c0d1c351d144677fd9d85bcb2ca8bf7
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:89d480ae76480829660d4cccfede74717faeb7e62a4a48b932750e41a945783d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a71b6fe33997db924669253992436e1e02d88769bb97376c0871848c2d330fed
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:1074f45a625df6ea412a9ffef3d4f4e53f3107ca45bb1de9bb030c47ce6518d9
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:0543e615d8a48ba6aaf90a8de0b0a79d37689c64bfe71f91c340455d9a359426
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a93fdee9e8d05ebe65e9dd8682fa7bc99c54f9d9f26d2cc8519f529255805da8
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:9d825d79fffd39a1149b262d48d08616795e5cfdb4c4bc40f5f192242f71042d
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kinesis-controller@sha256:de760dbe730e38cd9fadebe5fdb69299c04512e6a70f567ce678fc0bedc53e9c
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/jaeger@sha256:907dfe1d01f73c8a385da3661ddbe03db31e0b9ad6f5ab2336460402d0d915a0
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        - registry.connect.redhat.com/ibm/ibm-application-gateway-operator-bundle@sha256:e4ece7113fa7056980507408ac29d60676cf51c20d71a058a1accf765a6ea7e3
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchservice-controller@sha256:699f483cf96822576e3b79b690f256e2dbf4ca8d4426daf4298c980aefdeba4f
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/percona-xtradb-cluster-operator@sha256:d0c9d03130774d961f62032f684c494313787354186d028e9709e2463f69431c
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:a5645cb55c3e6c22655c6c8606b9bd51a49db44aecc3fe52e0eb80cca83d17c4
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:19e133c41b8f947aa83afba0417d8db3570cc2e1368813996fa5d135fdca4e04
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.102
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-09-03T09:23:19Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.102
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator-bundle@sha256:fdef8f42c5425b9607915439daf3e5e2b77705cd220a083a00b42ee2e122d894
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:80c6d45a2ca70f55809620a82c309abb1e8272a0515f9c1a2905b3302c1467a1
        version: 0.1.102
      entries:
      - name: hibernation-operator.v0.1.102
        version: 0.1.102
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-commonui-operator@sha256:87788afbc24851deb2c9c3f14b773bab174b2a2f02ac2fa0dd656a97433ffca2
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dns-operator@sha256:d5eadbf006aade69ac18156c891614da899d0779ba9e8fbaf755a256c46f055c
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/openshift-community-operators/gitops-primer@sha256:97af48469275f1b3a669f8e24a97a80ce9897fc5b12643a4f298b82dbe26e20b
        - quay.io/openshift/origin-oauth-proxy:4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso-operators/watcher-operator-bundle@sha256:01ab69b9c33a91cb194ad644bb64884b358af78467e2bb766ef373e4c36e795d
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4478b8a78bdde23e0d40476f8ad9e4bf0628fba7603a390e8de7e6a319abc1ec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:6c2ab763239f6e5a2a72890120bf1f20f6277850cec59a7e11cea2b6a06f12e8
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:8f30d4c7c491e9a804a8abf80570d1e1924cd0d787c6abb727331541c435d5ff
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:2b6b1a0b56cfb348ed4035b13d31501ff5757c101ed21027b330a7852576b891
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4c4c98b6cecd51c6644eb79b85789b7ebbcdc662edbb66ce435ea3b5e58d9273
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:b8aa74c5b2635a6d3413eeac6fc0f6eb9a2e459089aaabbf626ea3ef92a67ce0
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:73d91324ccc995e9a6db525f194bee340c17dd624ce520c767ea8fcf2346f786
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:d29abe14abc67d8836cc6f7656556a0a7fa8f4a8eea3f78e40265b339e7430f5
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:f7c4ed28a549fe7414312bd22c2ae140f1f28d30656fb9fb0c8ade05a8f82088
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c59956d70416fab0c6d7b3ecd8c9cfa3cf9415f75cfdfe2214f44b0b4aced594
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:7cace73b03bb02202a04b6026eb9f7323610fd0ccd5ba18a6d00fd66577a96b4
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c173a99026499700c08092f8418c79f39527d9b914690b199b41ed781c279de1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        - registry.connect.redhat.com/loadcore/keysight-wap-logging-bundle@sha256:2aeeefa7493eed8960004904567894244418982209e1b3a40488b759463edcdb
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/resource-locker-operator@sha256:9cc08639be2af4b72a9b816420d87dfa290a168fb83edc947a04b3e06cdb675b
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/lib-bucket-provisioner@sha256:132ecc88db79fb8b117a7a7974ca8f704600bac489923d918b78fd00378d5a5f
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-operator-bundle@sha256:aeae57b2fc706b782a3e3905e5c1825fe831f96f5b1c9478e1f2b217307a892b
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataDirect Networks
      provider-url: https://www.ddn.com
    name: vmstore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vmstore-csi-driver-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.tintri.com/v1alpha1",
                "kind": "VmstoreCsi",
                "metadata": {
                  "name": "vmstorecsi-sample"
                },
                "spec": {
                  "config": {
                    "debug": true,
                    "vmstore_map": {
                      "vmstore2": {
                        "dataIp": "\u003cdataIp\u003e",
                        "managementAddress": "\u003cmanagementAddress\u003e",
                        "mountPoint": "/tintri-csi",
                        "password": "tintri99",
                        "username": "admin",
                        "vmstoreFS": "/tintri/csi"
                      }
                    }
                  },
                  "csi_snapshotter": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                    "resources": null,
                    "tag": "v6.3.2",
                    "timeout": "120m"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "tintri/vmstore-openshift-file-driver",
                    "tag": "v1.1.0"
                  },
                  "priorityClassName": "system-cluster-critical",
                  "provisioner": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                    "resources": null,
                    "tag": "v4.0.1",
                    "timeout": "120m"
                  },
                  "registrar": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                    "resources": null,
                    "tag": "v2.10.1"
                  },
                  "resizer": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-resizer",
                    "resources": null,
                    "tag": "v1.12.0"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "512m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "64Mi"
                    }
                  },
                  "secretName": "vmstore-csi-file-driver-config",
                  "volumeNamePrefix": "pvc-vmstore"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-13T21:26:54Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: VmstoreCsi
            name: vmstorecsis.charts.tintri.com
            version: v1alpha1
        description: "A Container Storage Interface (CSI) Driver for Tintri Vmstore
          filesystems. This operator eases the installation\nand management of the
          Vmstore CSI Driver in an OpenShift or other Operator Lifecycle Manager enabled
          cluster.\n\n\n## Prerequisites\nMake sure that the top level folder where
          the CSI volumes will be created exists on VMStore and is shared over NFS.\nFor
          example `tintri/csi`.\n## Usage\n\n### Dynamically provisioned volumes\n\nFor
          dynamic volume provisioning, the administrator needs to set up a _StorageClass_
          in the PV yaml (examples/nginx-dynamic-volume.yaml for this example) pointing
          to the driver.\nFor dynamically provisioned volumes, Kubernetes generates
          volume name automatically (for example `pvc-vmstore-cfc67950-fe3c-11e8-a3ca-005056b857f8`).\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\n```\n\n#### Example\n\nRun
          Nginx pod with dynamically provisioned volume:\n\n```bash\noc apply -f examples/nginx-dynamic-volume.yaml\n\n#
          to delete this pod:\noc delete -f examples/nginx-dynamic-volume.yaml\n```\n\n###
          Static (pre-provisioned) volumes\n\nThe driver can use already existing
          Vmstore filesystem,\nin this case, _StorageClass_, _PersistentVolume_ and
          _PersistentVolumeClaim_ should be configured.\n\n#### _StorageClass_ configuration\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc-nginx-persistent\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\nallowedTopologies:\n-
          matchLabelExpressions:\n  - key: topology.vmstore.csi.tintri.com/zone\n
          \   values:\n    - zone-1\nmountOptions:                         # list
          of options for `mount -o ...` command\n#  - noatime                          #\n```\n\n####
          _PersistentVolume_ configuration\n\n```yaml\napiVersion: v1\nkind: PersistentVolume\nmetadata:\n
          \ name: vmstore-csi-file-driver-pv-nginx-persistent\n  labels:\n    name:
          vmstore-csi-file-driver-pv-nginx-persistent\nspec:\n  storageClassName:
          vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n    - ReadWriteOnce\n
          \ capacity:\n    storage: 1Gi\n  csi:\n    driver: vmstore.csi.tintri.com\n
          \   volumeHandle: vmstore1/172.30.230.128:/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img\n```\n\n####
          _PersistentVolumeClaim_ (pointing to created _PersistentVolume_)\n\n```yaml\napiVersion:
          v1\nkind: PersistentVolumeClaim\nmetadata:\n  name: vmstore-csi-file-driver-pvc-nginx-persistent\nspec:\n
          \ storageClassName: vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n
          \   - ReadWriteOnce\n  resources:\n    requests:\n      storage: 1Gi\n  selector:\n
          \   matchLabels:\n      # to create 1-1 relationship for pod - persistent
          volume use unique labels\n      name: vmstore-csi-file-driver-pv-nginx-persistent\n```\n\n###
          Topology configuration\nIn order to configure CSI driver with kubernetes
          topology, use the `zone` parameter in driver config or storageClass parameters.
          Example config file with zones:\n  ```bash\n  vmstore_map:\n    vmstore1:\n
          \     mountPoint: /tintri-csi                            # mountpoint on
          the host where the tintri fs will be mounted\n      dataIp: 172.30.200.37\n
          \     managementAddress: tafzq23.tintri.com\n      vmstoreFS: /tintri/fs01\n
          \     zone: zone-1\n      username: admin\n      password: password\n\n
          \   vmstore2:\n      mountPoint: /tintri-csi                            #
          mountpoint on the host where the tintri fs will be mounted\n      dataIp:
          10.136.40.71\n      managementAddress: tafbm05.tintri.com\n      vmstoreFS:
          /tintri/csi\n      zone: zone-2\n      username: admin\n      password:
          password\n  ```\n\nThis will assign volumes to be created on Vmstore that
          correspond with the zones requested by allowedTopologies values.\n\n\n###
          Defaults/Configuration/Parameter options\n\n| Default  | Config               |
          \ Parameter    |  Desc                             |\n|----------|----------------------|---------------|-----------------------------------|\n|
          \  -      | dataIp               |   dataIp      | Data address used to
          mount Vmstore filesystem |\n| -    | mountPoint       |    -      | Mount
          point inside of the driver container where Vmstore filesystem will be mounted|\n|
          \  -      | username             |      -        | Vmstore API username
          |\n|   -      | password             |      -        | Vmstore API password
          |\n|   -      | zone                 |      -        | Zone to match topology.vmstore.csi.tintri.com/zone
          \           |\n|   -      | vmstoreFS            |   vmstoreFS   | Path
          to filesystem on Vmstore |\n|   -      | managementAddress    |      -        |
          FQDN for Vmstore management operations |\n|   -      | defaultMountOptions
          \ |   mountOptions| Mount options |\n|   0777   | mountPointPermissions|
          \  mountPointPermissions | File permissions to be set on mount point |\n\n####
          Example\n\nRun nginx server using PersistentVolume.\n\n**Note:** Pre-configured
          volume file should exist on the Vmstore:\n`/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img`.\n\n```bash\noc
          apply -f nginx-persistent-volume.yaml\n\n# to delete this pod and volume:\noc
          delete -f nginx-persistent-volume.yaml\n```\n\n## Snapshots\nTo use CSI
          snapshots, the snapshot CRDs along with the csi-snapshotter must be installed.\n```bash\noc
          apply -f deploy/kubernetes/snapshots/\n```\n\nAfter that the snapshot class
          for VMStore CSI must be created\n\n```bash\noc apply -f examples/snapshot-class.yaml\n```\n\nNow
          a snapshot can be created, for example\n```bash\noc apply -f examples/snapshot-from-dynamic.yaml\n```\n\nWe
          can create a volume using the snapshot\n```bash\noc apply -f examples/nginx-from-snapshot.yaml\n```\n\n##
          Troubleshooting\n### Driver logs\nTo collect all driver related logs, you
          can use the `oc logs` command.\nAll in on command:\n```bash\nmkdir vmstore-csi-logs\nfor
          name in $(oc get pod -owide | grep vmstore | awk '{print $1}'); do oc logs
          $name --all-containers > vmstore-csi-logs/$name; done\n```\n\nTo get logs
          from all containers of a single pod \n```bash\noc logs <pod_name> --all-containers\n```\n\nLogs
          from a single container of a pod\n```bash\noc logs <pod_name> -c driver\n```\n\n####
          Driver secret\n```bash\noc get secret vmstore-csi-file-driver-config -o
          json | jq '.data | map_values(@base64d)' > vmstore-csi-logs/vmstore-csi-file-driver-config\n```\n\n####
          PV/PVC/Pod data\n```bash\noc get pvc > vmstore-csi-logs/pvcs\noc get pv
          > vmstore-csi-logs/pvs\noc get pod > vmstore-csi-logs/pods\n```\n\n####
          Extended info about a PV/PVC/Pod:\n```bash\noc describe pvc <pvc_name> >
          vmstore-csi-logs/<pvc_name>\noc describe pv <pv_name> > vmstore-csi-logs/<pv_name>\noc
          describe pod <pod_name> > vmstore-csi-logs/<pod_name>\n```"
        displayName: Vmstore CSI file driver
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - CSI
        - Vmstore
        - storage
        - Tintri
        - DDN
        links:
        - name: DDN vmstore Filesystem
          url: https://www.ddn.com/products/lustre-file-system-vmstore/
        maintainers:
        - email: akhodos@tintri.com
          name: Alexey Khodos
        - email: rrethinam@tintri.com
          name: Ranjith Rethinam
        maturity: alpha
        provider:
          name: DataDirect Networks
          url: https://www.ddn.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/ddn/vmstore-csi-file-driver-operator@sha256:64df1c504a9d70bd2b8b34520c8d431a92bb54702d601a57a2a989c6dd982462
        - tintri/vmstore-csi-driver-operator-controller@sha256:749ceff2bd46c71d81338510c1cc5866d39c66345f971d3093f986ea0417311d
        version: 1.1.0
      entries:
      - name: vmstore-csi-driver-operator.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vmstore-csi-driver-operator
    provider:
      name: DataDirect Networks
      url: https://www.ddn.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:042623e1faacc9fc22222a1c3e546a561abf7fc05d41ae9c915a69795f6bec7e
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:67b2b8783a04460019057cf826a74c18747c8c64ee7fe8ce83c8d0b7cd87cb36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle-marketplace@sha256:7864f9f1cf3ad1d551b698b2e803302d207704b448b15f6c0013cbd9e854593c
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator-bundle@sha256:e284d0839795f763ca53402c2b050fa9ffedebccd0f3832fe7259368c3ad32f2
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-s3-controller@sha256:058a37a9a2fe6a42bbdddabb3fc7495450f1ebbd7f46194866c1491c40ce1a9f
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ClusterPulse
      provider-url: ""
    name: clusterpulse
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterpulse.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.clusterpulse.io/v1alpha1",
                "kind": "ClusterPulse",
                "metadata": {
                  "name": "clusterpulse-prod"
                },
                "spec": {
                  "api": {
                    "enabled": true,
                    "imagePullPolicy": "Always",
                    "logging": {
                      "json": true,
                      "level": "INFO"
                    },
                    "oauth": {
                      "enabled": true,
                      "headerEmail": "X-Forwarded-Email",
                      "headerUser": "X-Forwarded-User"
                    },
                    "rbac": {
                      "auditEnabled": true,
                      "cacheTTL": 0,
                      "defaultDeny": true,
                      "enabled": true,
                      "groupSyncInterval": 300
                    },
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "security": {
                      "allowedOrigins": [
                        "*"
                      ],
                      "sessionExpireMinutes": 1440,
                      "trustedHosts": [
                        "*"
                      ]
                    },
                    "server": {
                      "host": "0.0.0.0",
                      "port": 8080,
                      "reload": false,
                      "workers": 1
                    }
                  },
                  "clusterEngine": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "timing": {
                      "cacheTTL": "600",
                      "fullScanInterval": "300",
                      "reconciliationInterval": "30"
                    }
                  },
                  "frontend": {
                    "config": {
                      "features": {
                        "devTools": false,
                        "mockData": false,
                        "nodeDetails": true,
                        "operatorDetails": true,
                        "registryStatus": true
                      },
                      "refreshInterval": 30000,
                      "theme": {
                        "defaultTheme": "auto"
                      }
                    },
                    "enabled": true,
                    "oauth": {
                      "enabled": true,
                      "image": "quay.io/openshift/origin-oauth-proxy:4.17"
                    },
                    "replicas": 3,
                    "route": {
                      "enabled": true,
                      "tls": {
                        "termination": "reencrypt"
                      }
                    }
                  },
                  "policyEngine": {
                    "batch": {
                      "cacheClearBatchSize": 500,
                      "redisPipelineBatchSize": 1000,
                      "redisScanBatchSize": 100
                    },
                    "cache": {
                      "groupCacheTTL": 300,
                      "policyCacheTTL": 300
                    },
                    "enabled": true,
                    "limits": {
                      "maxPoliciesPerUser": 100,
                      "policyValidationInterval": 300
                    },
                    "logging": {
                      "level": "INFO"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "enabled": true,
                    "volumePermissions": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "ClusterConnection",
                "metadata": {
                  "name": "example-cluster"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "cluster-credentials"
                  },
                  "displayName": "Example Production Cluster",
                  "endpoint": "https://api.example.openshift.com:6443",
                  "monitoring": {
                    "interval": 30,
                    "timeout": 10
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "MonitorAccessPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "access": {
                    "effect": "Allow",
                    "enabled": true
                  },
                  "identity": {
                    "priority": 100,
                    "subjects": {
                      "groups": [
                        "default-group"
                      ],
                      "users": [
                        "default-user"
                      ]
                    }
                  },
                  "operations": {
                    "audit": {
                      "logAccess": false,
                      "requireReason": false
                    }
                  },
                  "scope": {
                    "clusters": {
                      "default": "none",
                      "rules": [
                        {
                          "permissions": {
                            "view": true,
                            "viewAuditInfo": true,
                            "viewCosts": true,
                            "viewMetadata": true,
                            "viewMetrics": true,
                            "viewSecrets": true,
                            "viewSensitive": true
                          },
                          "resources": {
                            "namespaces": {
                              "visibility": "all"
                            },
                            "nodes": {
                              "filters": {
                                "hideMasters": false
                              },
                              "visibility": "all"
                            },
                            "operators": {
                              "visibility": "none"
                            },
                            "pods": {
                              "visibility": "all"
                            }
                          },
                          "selector": {
                            "matchPattern": ".*"
                          }
                        }
                      ]
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "RegistryConnection",
                "metadata": {
                  "name": "dockerhub",
                  "namespace": "clusterpulse"
                },
                "spec": {
                  "displayName": "Docker Hub",
                  "endpoint": "https://registry-1.docker.io",
                  "healthCheckPaths": [
                    "/v2/"
                  ],
                  "insecure": false,
                  "labels": {
                    "environment": "production",
                    "provider": "dockerhub"
                  },
                  "monitoring": {
                    "checkCatalog": false,
                    "interval": 60,
                    "maxCatalogEntries": 100,
                    "timeout": 10
                  },
                  "skipTLSVerify": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: docker.io/clusterpulse/clusterpulse-operator:0.1.0
          createdAt: "2025-09-26T21:09:27Z"
          description: Provides OpenShift cluster health monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/ClusterPulse/operator
          support: None
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a connection to a monitored Kubernetes/OpenShift
              cluster
            displayName: Cluster Connection
            kind: ClusterConnection
            name: clusterconnections.clusterpulse.io
            version: v1alpha1
          - description: Manages the deployment and configuration of ClusterPulse
              monitoring platform
            displayName: ClusterPulse Instance
            kind: ClusterPulse
            name: clusterpulses.charts.clusterpulse.io
            version: v1alpha1
          - description: Defines fine-grained access control policies for cluster
              monitoring
            displayName: Monitor Access Policy
            kind: MonitorAccessPolicy
            name: monitoraccesspolicies.clusterpulse.io
            version: v1alpha1
          - description: Defines a connection to a container registry for monitoring
            displayName: Registry Connection
            kind: RegistryConnection
            name: registryconnections.clusterpulse.io
            version: v1alpha1
        description: |
          ## About ClusterPulse
          ClusterPulse is a comprehensive multi-cluster monitoring and management platform designed for OpenShift environments. It provides centralized visibility, policy-based access control, and real-time monitoring across multiple clusters through a unified dashboard. For more information, visit [https://github.com/ClusterPulse/operator](https://github.com/ClusterPulse/operator).

          ## Features and Capabilities
          - **Multi-Cluster Monitoring**: Connect and monitor multiple OpenShift/Kubernetes clusters from a single interface
          - **Policy-Based Access Control**: Fine-grained access policies using MonitorAccessPolicy CRDs
          - **Real-Time Metrics**: Live cluster health, node status, and resource utilization tracking
          - **Registry Monitoring**: Monitor connected container registries
          - **OAuth Integration**: Seamless OpenShift OAuth integration for authentication
          - **Redis-Backed Storage**: Scalable data storage using Redis for metrics and state
        displayName: ClusterPulse
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - clusterpulse
        links:
        - name: ClusterPulse
          url: https://github.com/ClusterPulse/operator
        maintainers:
        - email: asmoore@redhat.com
          name: Asa Moore
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: ClusterPulse
        relatedImages:
        - docker.io/clusterpulse/clusterpulse-operator:0.1.0
        - quay.io/community-operator-pipeline-prod/clusterpulse:0.1.0
        version: 0.1.0
      entries:
      - name: clusterpulse.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: clusterpulse
    provider:
      name: ClusterPulse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:61685ecb519d17160149a5dc16ae28d6fd96264652a6f19fef861addfa4ec6f4
        - quay.io/rasen708/ncn-operator:v0.2.2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:a99a5ca46c5bd3065905588f4d0e3cdc6409d435abdc02793be88e77c1a8f0f5
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:7d2ad8e800339c15efa5d5954106e01151d34135b25e9a23e30f0027ee23f193
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:eaa77ae0f1a1f3d32e269e73eb29a73d11012d5e920af1011d1842bd904d7293
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:4d9d003842e3ffc386f4e07ce7a81cfa85e676031fb6fdc96aefa9654244974a
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:a7641fc5f4c7aed4b73d7d98ccb58ef64cf14ab16d72378f56dba3ad4b8714c6
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:7dffd170b9b7d021b1615e6d746cbdfcbd819a1306ecdbaf042bcaf2202f7daa
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:111d1f4ab8f006386102d1bec4cd6a41b446d9ae401a43c821b2f318c8ba74c8
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:d7a4f9b8e5eb9bf6d11e00898eb168e86636f0633959eeb2b36b2d6e5e505f71
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle-rhmp@sha256:1f3d005224eada1b71d04f21d07a7c4fde159725c7a44d967d996152c802bede
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator-bundle@sha256:85e75ec797f70c5807ec1ea25208cf0bb1990daa05eca9fbc82bc38aa4a840ab
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
          createdAt: "2025-07-16T09:03:58Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <2.8.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk@sha256:99aec4a74ea4c9910b50cc035b4c0c85ed4711a4c9b456028690d0f0c7af2657
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/splunk/splunk-operator-bundle@sha256:433bec6334c4771e140a24ea8753e7bcbf358c4504c942658fd3364915ac86ac
        - splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
        version: 2.8.1
      entries:
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
          createdAt: "2025-09-19T19:24:03Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-mq-controller@sha256:702ffd931cb4baafb1eac49f9fda5588d8b02f3b8c696e477dd626b0efad9f54
        version: 1.1.1
      entries:
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ssm-controller@sha256:922fd6daa660d2250bad8a3c54733cceae845624a14651cb20e049e29bf90e1c
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        - quay.io/openshift-community-operators/community-trivy-operator@sha256:e3652e68e1e569d7f40432fd8cbb7bfdc6738a116ecc1011c71a36f7ca5663c6
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:3b538de0bbaa43d2368402ea1c0eaf767c8b9e5a28e112a309148a630dd3c052
        - registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:5b61484c6a2bbc377435093bd9abb31df980ece470325cae92f3106963b6482c
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:af8ccc7a075f11bf6f31310830704e3a27b23587094f56e518b1f264c92b0fe4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: kubernetes-nmstate-operator.4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: kubernetes-nmstate-operator.4.18.0-202508271223
        version: 4.18.0-202508271223
      - name: kubernetes-nmstate-operator.4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: kubernetes-nmstate-operator.4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: kubernetes-nmstate-operator.4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: kubernetes-nmstate-operator.4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: kubernetes-nmstate-operator.4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: kubernetes-nmstate-operator.4.18.0-202505280403
        version: 4.18.0-202505280403
      - name: kubernetes-nmstate-operator.4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: kubernetes-nmstate-operator.4.18.0-202505061007
        version: 4.18.0-202505061007
      - name: kubernetes-nmstate-operator.4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: kubernetes-nmstate-operator.4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: kubernetes-nmstate-operator.4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: kubernetes-nmstate-operator.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: kubernetes-nmstate-operator.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: kubernetes-nmstate-operator.4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: kubernetes-nmstate-operator.4.18.0-202503190933
        version: 4.18.0-202503190933
      - name: kubernetes-nmstate-operator.4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: kubernetes-nmstate-operator.4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: kubernetes-nmstate-operator.4.18.0-202502250802
        version: 4.18.0-202502250802
      - name: kubernetes-nmstate-operator.4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/cluster-observability-operator-bundle@sha256:8c27ac0dfae9e507601dc0a33ea19c8f757e744350a41f41b39c1cb8d60867b2
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        - registry.connect.redhat.com/loadcore/keysight-wap-diagnostics-bundle@sha256:4e8400a38a04c61dc20ece1ee7dac1ad8a56ac309621f3225e7389e84e20c870
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:f5f025c7f56643fab9148bb5de8ee34a84d02b78a6abb4b5592306cf54296654
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        - quay.io/community-operator-pipeline-prod/percona-postgresql-operator@sha256:44e687b38a48f68bbbb919e6df0d3b87b75492aca7bd41e508a2aa0e80d024a9
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        - quay.io/community-operator-pipeline-prod/ack-elasticache-controller@sha256:a489c815e9d3777d775ccd65bbd0ecd5fe7c577de36c504b63cf1620531d4ccb
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:301c3b793d8068f9e50c855196da91b6c07111ef500b02a91fbb354c9fc1316b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:8edd2f89d7adda6c1a5aae7414f50e0d786ec8443c1db13834ea4924009b6419
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:227159c5a93966a43fa3a84cb69536ff0051cc261da068561241261d3837998a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:60635156d6b4e54529195af3bdd07329dcbe6239757db86e536dbe561aa77247
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:e4d49cf6a2fb8b3585ecfe3b3da3e9de090fb99de3dee2da0601b1f6b071173c
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:095e93f8dd31a0985d6fef6a6168da456630febe89ab9e0cb9480170f1036541
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:524d315efa226f97ff26b8668f9297e755c6f0c815b8f3385b317667553f0e62
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-bundle@sha256:80ca5ad4634305808ac5a3e8ff656ce6ff69b3c16dccc5e03f42763fa89bc762
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:857c1bdff46d1a1d5abb51f75a80b7944bb461687b179b6f1622103947b8c3b1
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:d101e2b9ab4d8ca7c87c86ea4e3630dcc82a9c5a78885179c8b16773c8aac587
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:ecdd5c5fc609128e0af751f87b797b71e4ab0200a9abaaca835ff9a3f293fb10
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:be099e23851cb4b6d15e9c44c4da8fb6cba74aab005597d305757d1dae8b0122
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:96119f61731cec7eb037c3a4a7c2c8bfdf9ef11fd5b08abc25e3d24ddf342b06
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:2ea267aac67763b4f160c4e6876c0226d2fa135ee00614814a59b4de1d91f4c4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:218905f7902e6ac80366856d6c5ae37cd6d167c81e9521df7c3c3d6619f6c425
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:09fe022370abfcefcf3abfe8b6debc2f201a0ac8b8dacaf1278b8a6bee2b5273
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:5b865231214649649a2a9fb06934f931503f68ed740e150be1c7c8fcbb6f1317
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:16245de1e88228c231ab2edb4e9380753eff982bb72febadd00cdaad286a8eb3
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:491e7aea221eb0e1d2992310e53d50db287cbf32f62d090bd842236b603d38bb
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:7de09e5394985cdec87f5afab65b3343b8a37c0eb7f778774b21933ae69d691c
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle-marketplace@sha256:28f388286bc4b4981182ff1ab9e927f59744dcb755392907fe65cd3b96a44519
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        - registry.connect.redhat.com/loadcore/keysight-rest-api-browser-v1-bundle@sha256:89ef7cf0dc8b27e58a63a82a701894a7cc4d0643bf0c8c8e20c9bb59d1a1b1bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/openshift-qiskit-operator@sha256:633ce4190b02034d70b52d7e624fc9d417ef997607f9710a36e73bbc0963fe4d
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/community-operator-pipeline-prod/openshift-node-upgrade-mutex-operator@sha256:5ac79c9193934440464a595434fe4241e038505d32650843afcad5f5bc2ce8a7
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        - quay.io/openshift-community-operators/mercury-operator@sha256:0c253fd002cb389db40b2438d4d1c9e953a84125dbf0f4b6f5e3a6efd87cf3a6
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:9db455e082454734d2a17811dcfec53d3d60c28df6e7fe8bc979a45b28a35061
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:8650430afd813fd749fcd80ad5f6fd0aeaab3df048e8fe34e6aa72a07ab801fc
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:ed063f507fa8e2f378d0adbde823c110484d0212d124cc72d67efa987e86444e
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.8.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:68b310ed3cfd65db893ba015ef1d5442365201c0ced006c1915e90edb99933ea
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:bcc5ed9560667c7347f3044c7fa6148e3867c17e93e83595a62794339a5a6b0d
        version: 0.8.1
      entries:
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/3scale-apicast-operator-bundle@sha256:29fe669a56a69adb537798c91b9359fa3b051da5ffa9f27789a2067a3a72330a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.7.9-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:ea2229948fea841343fcaa337bb7e2854acf5f563d3ffb0af8bda1c21ec2a91f
        - registry.redhat.io/3scale-mas/apicast-operator-bundle@sha256:d3a608b37aa089ea46e803696ba4957585b875f8a810f517e891f385ad105937
        - registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
        version: 0.7.9-mas
      entries:
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        - registry.connect.redhat.com/loadcore/keysight-wap-storage-bundle@sha256:0955dbda03bd7307364a1a458ab8d6ac5a8c9feb6150706dcede9676a6513b0d
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:591fc4b5867dff26df1952653423b33889c0f56ba3f756057694d4c0e396eeb0
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:3225c49cf30a146784533496ce478466034d4549ff9721752958bac342fdec77
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:b0fa6b0356d83626509289b33dfa669bd7d32542aba986a18d296b85ce63389a
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:fe759127885502dc66423530ff03e37973355c5566a1c27e988b59d71528dd2b
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:c0336e5079f255ab6cd38e4e2e61d0cb4c6a35a976528fb44409bfde57e70dde
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a77d39ded03271f2deb8b438abd4328a1846ee0d4886cce14d869ba61f84d600
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a6f3a3a8aa52de4441ef3249a76dcff1b6b6defba417a9fdbba37eb458adee04
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:6208f64c7fe75f39ab7511e2b480c493d2ecc8d1e76a1ac498a93a6a24e8beb5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:11d64a921a98b7ec1a2126a03e446a055a5ae1705dc2005bb8b73cbbcf7ecf99
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        - quay.io/openshift-community-operators/hedvig-operator@sha256:81d956c12057f18bf920044a4e39c27c99b2c23afdccc0ff9468f0484e1961d7
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/coherence-operator:3.5.4
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-operator-bundle@sha256:f40e7955ac649c383134ab6e77bdd521b75b21ecd9b476436405d0674c7fc855
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-operator-bundle@sha256:d04033c2cd2d236f6319a3e35880b89c8f0c60d53cd32f6f097d07d6d27e5f2c
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-ati-updates-bundle@sha256:3d06a58b2a277a67339c22f3fe10c1daac1c8eabaac8540b01177f6f5688784c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatchlogs-controller@sha256:94145cae2b8cef1410acbb4f1c1eb571e71408341716af1742f27ddf4dc350b6
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/apimatic-kubernetes-operator@sha256:a5f68205291caa9bb7903180afc974fe86252ba87e671978c8b49e0206976d06
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
        - quay.io/community-operator-pipeline-prod/ndb-operator@sha256:5f33c2a8181ea8e91844afd1f4042affbd9eb878266b6b2bcc6906f0829ed00b
        version: 1.6.0
      entries:
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - quay.io/openshift-community-operators/keda@sha256:50816f675c76d36dbc62846579150477018ef87df325861751f3fc02a70c8781
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/community-operator-pipeline-prod/alloydb-omni-operator@sha256:08fafb811f709780c0ccb3c22e3c6417cb14a70a9305e46a7b50457647a82102
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:ed530db9b4a29fb591ea208cd0c6b3b83b49537ff5a0a4f501b54ce667f9b303
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:31bdff225bd825d02ce0f468b1edc77541ce85b7b9111f14a583c418c984232f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:80198f65f7e3f9022a34e0342e9423807b855dcb7bda810172eebc9a27afb5c2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:f04d565eb5162fd5a9c78cd8443af6a3b184a660277537f1d6d575caa2e85441
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:773592dfb9964970f444871a8812e8b74db3f205faaa0bf148defa1a2b87b649
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:0b1771f697a700263c5b827dbde56a5c9bab610e9d7df55ef93885a4e54cc0e9
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8441322a0fce02df407573675b5b5d92d56de97c8aec72541b33d2dadc4050d5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:be93be0341649215fb001725aac0c3c5925343adbf1ad00c979b8c3d489512d6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b22190de3162d3e03481cae8904ce563912756eea7ac7e7feec0acbbdae5e6f6
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:c2f39be75173b4cc57498c5845b14c464bf9af92b49716df9661c8135ac2cf7a
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b2d58f4dc20eb7e17e09af0e8c5de55fa04f9f4f697e10c6be81a7a6a08381b2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:27efa0ab1d9c69047a924bc1dcd2b56108f5f390a6da56afb6ab7a37cb4578e2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/volume-expander-operator@sha256:b775df99f891e2cee9c2169e9c948a8d61d65b4cc90f36752b46a56e2d4be602
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
          createdAt: "2025-09-19T22:01:50Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
        - quay.io/community-operator-pipeline-prod/ack-eks-controller@sha256:120566a02e487aaa8823323423fccfaa377135037e2e1cb9f3ff14313cf3bfeb
        version: 1.9.2
      entries:
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: sonar-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        - quay.io/openshift-community-operators/sonar-operator@sha256:5e7a87f52955162f090c1d7fc69f6d0917444b553355a11c061ab60f289a17d7
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    defaultChannel: alpha
    packageName: sonar-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-documentdb-controller@sha256:8a991c76f281ec30258b6317bd623894b78a4f39d9790a1cbb32e6f1b7ff63f5
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        - quay.io/community-operator-pipeline-prod/dynatrace-operator@sha256:c475f9cc0f6caa2bb316ff42a35a40b1d4bf8ef7c8a3384d9c78eb205bda7076
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        - quay.io/community-operator-pipeline-prod/kaoto-operator@sha256:4b6cc43eba8efbe17ffec746431d6cf6b045886ff865673a08a0820abecebdcd
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:5ea2dd102074798d1ee005dc03fbbd24f4324df831f279355dc2dfc262c9d880
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:e4dec1f39f8e24043cf16703efcac2cc3b953d3f4c37b3956fce0c65ef829764
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:ed8230e90f0ab0586a743d5e760a7469345b1136971aad17ef675c1beccf1982
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:08995782514e496c24d888352e9d1e6b788dd87adb761e0c2e745be2975ffcaf
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:7b324e897fbcc881ececa529822d7b738f82cd4c3adf4cc54b66080b888f5d9c
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:e2a329fc74ed5f17e395367126c7e96197178e8f886d475b657a7910a4a14152
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:0218d68f97e89b48bbdd9c7cd750d4a666cf33af1cb8a76507c067a332a7b6bd
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:2bc56e149c14920662b549a5367f98daf9afad0df29b4746b195f6afc0c9c597
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        - registry.connect.redhat.com/loadcore/keysight-nimbusmosaic-dusim-bundle@sha256:97d5f906ef83272a8d7a2a249f38f8b9b186f8e9b64729a049a7eac0dbda8b1a
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        - quay.io/community-operator-pipeline-prod/pelorus-operator@sha256:f54c42a30675897a971cac278d5d8ea24edf3121570264c07cde04f11cc4e7a8
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:57af9e32c31dad769bda449ebe35e09ea135e8a7d06cddb42be4d1a09005f518
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ab3cb65b88c34fd58524a2a51cd510c611924e5860834965bad0b4418a8b6dd3
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
        - registry.redhat.io/openshift4/ose-local-storage-operator-bundle@sha256:d303fc63f5b849f073e34a7791ea92cdc3cff269cd1b593e27480ae96cac1a15
        version: 4.18.0-202509100344
      entries:
      - name: local-storage-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: local-storage-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: local-storage-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: local-storage-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: local-storage-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: local-storage-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: local-storage-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: local-storage-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: local-storage-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: local-storage-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: local-storage-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: local-storage-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: local-storage-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: local-storage-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: local-storage-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: local-storage-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
          createdAt: "2024-09-04T17:53:42Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:07cafb83ba90378c1fa60338cf951a9d7145ec245710637e90fdb508875dcc45
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:1d86075469312e84f3d16379efcd6130ec9e8b23a1343e3945f071d82beff6e3
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:dc10544e1bb24883f38373a0445944745282e5a69ed1744f428f98871a7a39c9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:31b3658dd79c18111c21e99aea52cc198f8e95b458a1a4f8d2bfc4d21e9bdacb
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:1b8008a16a0f53c09ecc5a219964194689a1aeb01acee8313aa49a9c26259429
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:a539064234088a4ee13b8a9666ca9507208fe38189ee9b8b673a38a58aa75599
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:b137e1345038c9e566767f24cce60c03ec278f5a2efc28fa08e0bd8bfa016486
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:c001dc96a1478a58f30461ab018e40bede13a78f3cdce869a70dcb8b9bba22b1
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:92a3162c705537147005a065cdddbb410e1863d80c08d8b0f27b7ff5588500cd
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:8fbfffb0bc3adfb4d72961dad381e1daba414bf4a7084f03b5208f6be4191ceb
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:035855c3d4b32a29232846d73dc398e81f941fd90a03e33241ca06f6e80b2204
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:f86e96edb0418811df328debdee93b4a51cf7f651134ee35e9c50b9dbfc29fd8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:084f1e54ef0d7f457d5a7cfc57b567ad8d85237c12f707c82f7a30a54e3652ba
        version: 1.7.17
      entries:
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:9f5479ac795eb50fa5ea597828b2d50ddaa4878faa1a07b3de0e611cb4af5fc6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/operator-bundle@sha256:bc8a47d5f8ffe3a2fe48819f08c82525677131cfb4f3c0facd82470c0f10acee
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle-marketplace@sha256:4145f7a04d7dbdda226b75af4e4cabca63a803f290b1581ea4ae0dc223eca7df
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        - registry.connect.redhat.com/loadcore/keysight-wap-appsec-bundle@sha256:e5049593a52fb26c124de7005b45919f26fa0430293f18283a20c8cde981b025
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cli-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager/cli-manager-rhel9-operator@sha256:be0fc230fef111cff36a8b846bbac9aaa98aae94ad9200416e841aa487121de4
        - registry.redhat.io/cli-manager/cli-manager-rhel9@sha256:a6b7857c1acd5173c770e6ba1d03947eb13804d2512f7f7cd6bb96e5e5cdaa17
        - registry.redhat.io/cli-manager/cli-manager-operator-bundle@sha256:74c88f1cfef4cc31b794fc5ed8dcc10d3b0e6869c228d808f1d98819e69554b6
        version: 0.1.1
      entries:
      - name: cli-manager-operator.v0.1.1
        version: 0.1.1
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:a0e4c432590e65818e7ebaab127d483d68fdc5ccae9a13820665a8aef122fde7
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:242d754d5fe1f2a07091ae93da284b8912e0f4b4c973200cf6d1241e6abb88a9
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:437a786a301ea720cdb5651dff878a9f51470bab46ba28daae4928c7074c9157
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
          createdAt: "2024-11-21T23:48:37Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVSNetwork is the Schema for the ovsnetworks API
            displayName: OVSNetwork
            kind: OVSNetwork
            name: ovsnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-operator-bundle@sha256:4c8271d0bf6d94061fe8df8110b041cb0e1a840fb6263315f552166fc5053e73
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:2e71d4125a35f6e7bd2a04cdd8cb6aae058d745eda372ae3d3fd37783778f486
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:57a57f4f3a0adb8b7234683f2968352988ec9a414824f4f444cca096825a5df5
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:45835aba95af72aa8e73b630d1d2b4c1079217531f025f4715a2eb8e8856205f
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:8b412541a797aaff1b49aa229e310957a8d7da03a3d38e6bbfd65f2ca36c2bdd
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:6eb244a301ae5395b69b1bfcd0baac37c094f4c38a1a330c0a70fb81e71f6157
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:3b87d49b200ac4fdeac97f0e97a46001e91c8f4680e2975e90bd94ec37864403
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:0f18db6a5bbc952bee8f3e4ac9e93e807d1e85cd132affe94b1f4005c1e128ce
        - registry.redhat.io/openshift4/ose-sriov-rdma-cni-rhel9@sha256:65864bc235689ff38b63d2a7b2c72874ee155241a493fa9429fbe87d5100cfcf
        version: 4.18.0-202509100344
      entries:
      - name: sriov-network-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: sriov-network-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: sriov-network-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: sriov-network-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: sriov-network-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: sriov-network-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: sriov-network-operator.v4.18.0-202505260733
        version: 4.18.0-202505260733
      - name: sriov-network-operator.v4.18.0-202505121333
        version: 4.18.0-202505121333
      - name: sriov-network-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: sriov-network-operator.v4.18.0-202504221235
        version: 4.18.0-202504221235
      - name: sriov-network-operator.v4.18.0-202504160002
        version: 4.18.0-202504160002
      - name: sriov-network-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: sriov-network-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: sriov-network-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: sriov-network-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: sriov-network-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: sriov-network-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:72a54c9457e311001b25c1b414b7001c6d7f9a34b89efb1b4198084586386f31
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        - registry.connect.redhat.com/questlabs/alloy-operator@sha256:fe4e83fc4a50e1aa04e6a9ba2a57acc737077b86681c1d7c597646fad337f221
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - registry.connect.redhat.com/kove/kove-operator@sha256:04bf970760c4dfe910465b0ef6a18e9c17fa70f6388131f12ae0aa0cd646f663
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vertical-pod-autoscaler.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "autoscaling.k8s.io/v1",
                "kind": "VerticalPodAutoscaler",
                "metadata": {
                  "name": "myapp-vpa"
                },
                "spec": {
                  "resourcePolicy": {
                    "containerPolicies": [
                      {
                        "containerName": "*",
                        "controlledResources": [
                          "cpu",
                          "memory"
                        ],
                        "maxAllowed": {
                          "cpu": 1,
                          "memory": "500Mi"
                        },
                        "minAllowed": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    ]
                  },
                  "targetRef": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "myapp-deployment"
                  }
                }
              },
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "minReplicas": 2,
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "safetyMarginFraction": 0.15
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
          createdAt: "2024-12-05T23:41:17Z"
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1beta2
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1beta2
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: Vertical Pod Autoscaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - verticalpodautoscaler
        - vpa
        - autoscaling
        - scaling
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-operator-bundle@sha256:f5df5d878ef0a0b494c9bafe99c35f3152d30f6c426be789fcd636dd13eff6e6
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:5da084c4dc81251a707db1bd9347c13b39a1e01660120aa2e6a47f9d1d981fe9
        version: 4.18.0-202509100344
      entries:
      - name: vertical-pod-autoscaler.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: vertical-pod-autoscaler.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: vertical-pod-autoscaler.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: vertical-pod-autoscaler.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: vertical-pod-autoscaler.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: vertical-pod-autoscaler.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: vertical-pod-autoscaler.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: vertical-pod-autoscaler.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: vertical-pod-autoscaler.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: vertical-pod-autoscaler.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: vertical-pod-autoscaler.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: vertical-pod-autoscaler.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: vertical-pod-autoscaler.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: vertical-pod-autoscaler.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: vertical-pod-autoscaler.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator-metadata@sha256:d6d0e71f5451e8255edfff87ef28ba7743fe7057efd53b25bbbd01bf9e1d0307
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/lws-operator
    name: leader-worker-set
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: leader-worker-set.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "LeaderWorkerSetOperator",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-04-14T08:12:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lws-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/lws-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LeaderWorkerSetOperator
            name: leaderworkersetoperators.operator.openshift.io
            version: v1
        description: "Leader Worker Set Operator is based on the [LWS](https://lws.sigs.k8s.io/docs/)
          open source project.\nLeader Worker Set: An API for deploying a group of
          pods as a unit of replication. \nIt aims to address common deployment patterns
          of AI/ML inference workloads, \nespecially multi-host inference workloads
          where the LLM will be sharded and run across multiple devices on multiple
          nodes.\n"
        displayName: Leader Worker Set Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - leader-worker-set
        links:
        - name: Leader Worker Set
          url: https://lws.sigs.k8s.io/docs/overview/
        - name: Operand Source Code
          url: https://github.com/openshift/kubernetes-sigs-lws
        - name: Operator Source Code
          url: https://github.com/openshift/lws-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/lws-operator
        relatedImages:
        - registry.redhat.io/leader-worker-set/lws-rhel9@sha256:affb303b1173c273231bb50fef07310b0e220d2f08bfc0aa5912d0825e3e0d4f
        - registry.redhat.io/leader-worker-set/lws-operator-bundle@sha256:f641cf3b2d6659e26eb0ae49c67a3416152a2fbc2f26f490889a2bd169b9ea06
        - registry.redhat.io/leader-worker-set/lws-rhel9-operator@sha256:c202bfa15626262ff22682b64ac57539d28dd35f5960c490f5afea75cef34309
        version: 1.0.0
      entries:
      - name: leader-worker-set.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: leader-worker-set
    provider:
      name: Red Hat
      url: https://github.com/openshift/lws-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:5f10cfc1bc8571f44a99d88755adade0743106219673bbdd2e4d3a04ee714778
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:9f90cd7ab8c0ab5da6eb12f7f0dbd177df62f07c38fb759a97989899cfcbe75d
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:71ff93b137b6569f8fdefbd446cf12cd95477e4294bdea98fb5318e5d7d1cdd6
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:7a04dfd687dfe1646bd59a39c9e88781ffb815085bbd15f6305f6eac5f59bc49
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:d1a9246def81537e3417a6d48d1e818459583db658b54a87ca5af6911c8a3f53
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:e88676c0346c245b315f1a3bf7f85b10515bae3620c786fedda0533b39d2a400
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:66bd3eba0b68c317fe128c6713488cb6374f9647095768d3f7eed8120bf02e77
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        - quay.io/openshift-community-operators/authorino-operator@sha256:8fe34f94f87c567401cc2bbbbcc3a3e719364b80a2b17851496f86c85af5c09a
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino:v0.20.0
        - quay.io/community-operator-pipeline-prod/authorino-operator@sha256:92ad93f27c34f1406fc05a248d04709b0109adf37f2f9424f48fc6a73d9d359e
        - quay.io/kuadrant/authorino-operator:v0.16.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/logging-operator@sha256:d3ce3322f7f25c7b8bfabdd6e9fd32b833239fc39592240e0ccff3a13757fb17
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:f252d6f89a7cdcc520c3af461ca57d0dfabf736b9c38277d77477c4c880e182c
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:97d03647567ddb6f04edea39244bc0814e8f7e557251060d9b241a5fabe04dfc
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:f3497acc3216d820014cce7332e9d0683f4ca6438782ace02210192f787ca097
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:23f2f66e9f6d79d05e2be5a10c394834f093f4677ab06d8935ed33980408e8d3
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:c6d6c3e0408b851cdc9589cc0ee642d77397cdfaa40f50cf42557c5aaf9c0690
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:ff314ab983699455c32ab0ef27bb552d00825974a2d7088df2000f09afd8a2e2
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - k8s.gcr.io/redis:v1
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - quay.io/openshift-community-operators/ripsaw@sha256:ff8d796d5bd6c8cb0e155b3ced631122397758eb472470223e30352eeddba480
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle@sha256:2a28b64f7321df23928c8226b3eb3161cf520d38cc57d875b3cbfc828c164edc
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        - quay.io/community-operator-pipeline-prod/hazelcast-platform-operator@sha256:1209876fc9103eb2d69c34e97e4e2e0b384a9a778ae7e6a53222b8552fbf2f26
        version: 5.13.0
      entries:
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kms-controller@sha256:a2c60484a6556949318aa1d730c601b5bb05b1795b945a607209ef3434cce72f
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:68cdc37990bc7da8a6bbe3a4908378c8026e5ff980bf8c44c0277f689006fc51
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:97762fcd8bfeec64fe4d3d43a6e60bc705ff2c8773354357e545adf31c8f85c5
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:4fa7a30eb3f679a495aeb96794056ffd003237686b2b60e693c8cabf6bc15bae
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6dce519730a6eb73bc18bef8372cdde65c29af3cb52eb78afc808d8582526fd3
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:9f2c01e88b29154ba7a0f70a1c69a49c86a567808cf057cb5d115159278cf2cf
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:edec37e045212f74d160724ffc7b2c8ccf4e60eb95629aae80360494d78d5433
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:465f0bef1f167b1d16ba9ae5f2d9a1133d11786b6809698a84c5c962387dbf6f
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8be098242f971aefa61a4a24d2c6630ecaa7856ece842166c38887a66bce1bda
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6afbc7ffc98918e3a557b0b3d9b2c3aa80caa47b9a73205fac9e6572c66f4df7
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:eefdd05485167d490f0011d9e5d2bb6112db1daea25940c79e1f55b3af254548
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:05e66583757da895fd23c3af9830ae1ac6a84d107107e5b6bbaece57e4d79c9a
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:02f53d3395b2b98f92a96b466c83ca15c79a75d82451e2ed8ba921375e727c7f
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8abe0994052e2a328e54c9d30c920dc331e9b816a4606234a3023b9ef65ec4ed
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:ef3f85e1227e886a131d3ddec3fb431aef6d3daed46115b288246b9ace0e69e2
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8b9b06ead5e014b920c82034ded615413dca53cf36f804005408726c9bbda66f
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:51db956f1472ae0498c57b9488448c34cf7e24afab6b9997eb3689b001d73277
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:392f9299c75a3ca4f593eaebdfe515145696f12db9f990bb004cfa58dd6bf0ef
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:df4e4dce45f87e684e7e1044a7edb2456160141afb6ba57b9af28088ffbac407
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:fecad1a330421341e6946252f95735ed177a690eab6250be097413ea29ce9473
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:9dcf24fecf3ee0d999601c276ac15e8634c275590f2c42b8a6bcb83984bf1445
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a9950264bfefc536b4e0a0adc2348f92aa41da55dd50a2324ddad0253291b67c
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:4d59e60cc75de82bd3f3a063237cef8a5598d315f6d53e803d37589a94929b0e
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:40d17d4d0632c272779da152f710fb7c2de50acc3791eb0b9fe445a13c0cb102
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:8e321dc1bf74ba331b62d21bff4d8993956ab935cf5be32f01e7f7aa6897bf22
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:16fbdaaf303ab40202d49438737f3307926594a5259081eeeb39595b7ea41f8e
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-zero-trust-workload-identity-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zero-trust-workload-identity-manager.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "ZeroTrustWorkloadIdentityManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/part-of": "zero-trust-workload-identity-manager"
                  },
                  "name": "zerotrustworkloadidentitymanager-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
          createdAt: 2025-06-16T11:15:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: zero-trust-workload-identity-manager
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/zero-trust-workload-identity-manager
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterFederatedTrustDomain
            name: clusterfederatedtrustdomains.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterSPIFFEID
            name: clusterspiffeids.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterStaticEntry
            name: clusterstaticentries.spire.spiffe.io
            version: v1alpha1
          - kind: SpiffeCSIDriver
            name: spiffecsidrivers.operator.openshift.io
            version: v1alpha1
          - kind: SpireAgent
            name: spireagents.operator.openshift.io
            version: v1alpha1
          - kind: SpireOIDCDiscoveryProvider
            name: spireoidcdiscoveryproviders.operator.openshift.io
            version: v1alpha1
          - kind: SpireServer
            name: spireservers.operator.openshift.io
            version: v1alpha1
          - kind: ZeroTrustWorkloadIdentityManager
            name: zerotrustworkloadidentitymanagers.operator.openshift.io
            version: v1alpha1
        description: The Zero Trust Workload Identity Manager is an Openshift Day-2
          Operator designed to automate the setup and management of SPIFFE/SPIRE components
          (like spire-server, spire-agent, spiffe-csi-driver, and oidc-discovery-provider)
          within OpenShift clusters. It enables zero-trust security by dynamically
          issuing and rotating workload identities, enforcing strict identity-based
          authentication across workloads. This manager abstracts complex SPIRE configurations,
          streamlining workload onboarding with secure identities and improving the
          cluster's overall security posture.
        displayName: Zero Trust Workload Identity Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zero-trust-workload-identity-manager
        - ztwim
        - spire
        - spiffe-spire
        - security
        links:
        - name: Documentation
          url: https://github.com/openshift/zero-trust-workload-identity-manager/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-agent-rhel9@sha256:0281d6ab233a685ba0f739629584aacd9254e0b13a697fe8708efa41289e746e
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-operator-bundle@sha256:1974903a3e2b692eedab2602395fdc4a6e2ffbc79ed4e633705c05ce7771e1b7
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-helper-rhel9@sha256:a39c80c6e1495ea166a858c8ed03304c6455420c077d937c99bf8732a6829790
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-controller-manager-rhel9@sha256:67f0e6670195c84515e7d78139a2951e463db11445740fe1d9a449a7f338556c
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-oidc-discovery-provider-rhel9@sha256:5ba7e94b7ea06852d32002b055df556ff2a2fdc9a2e8f793b837c4856d773782
        - registry.access.redhat.com/ubi9@sha256:861e833044a903f689ecfa404424494a7e387ab39cf7949c54843285d13a9774
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:7d170ba9b3de2e88dc0d0a6b0c13630ca3b23eed96ea5e3c725516046d7c7fc7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-csi-driver-rhel9@sha256:a16ac0a4363f7cffada39a2f17aa42f5de3faf78b3688ae5ebf2ecf3a4759cb7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-server-rhel9@sha256:70f97c613feeaadfed99d37aabd62e1cb2247825292ad9be806bc2fd065c9166
        version: 0.1.0
      entries:
      - name: zero-trust-workload-identity-manager.v0.1.0
        version: 0.1.0
      name: tech-preview-v0.1
    defaultChannel: tech-preview-v0.1
    packageName: openshift-zero-trust-workload-identity-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-operator-bundle-rhel8@sha256:0fbbb0b32adadc5746251103d25bba6bd365b2ff60e289e4694061fc92c3ce6e
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        - registry.connect.redhat.com/paloalto/pcc-operator@sha256:c7200a8c652ee544b0e0bb4d9427eab3a3b2dee6ca61ee88abb8052316382673
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-messaging-topology-operator@sha256:4c62fb6d686a20a28437905d4e1a9b93ac05853b4dbf29fcce34421a43268724
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/sap-hana-express-operator@sha256:0fa4f29bd7669690be87a08405bed95d3ead2293a125b28149de2b10b02d2cd6
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/community-operator-pipeline-prod/falcon-operator@sha256:3916adf5083c83838564b1ed039decbadb2efa4ec288699ef0137a66983a3cc4
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.connect.redhat.com/hitachi/hnas-csi-operator-bundle@sha256:79e6ba3db78aa55a9f3839a172843d284d91e9cdaae9cb2c8e7e705f3d03115d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:60b293c23587072b3376f83afee95b0207228c5c7b80483b025adda572c3551d
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        - quay.io/community-operator-pipeline-prod/ack-rds-controller@sha256:80b0d6e33fc94471d05e05f6d370be3a84fcb5c2044486e007d57360eb69d9d7
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:b10c121395f5bc46d61fdeddd35a0c8d9ae4e67a9a77d8ada90e0ea53666323e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle-marketplace@sha256:9d84d46dd1eb74afa18466c73766402b46c2885ef483600aff83065f8065d8d1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        - quay.io/community-operator-pipeline-prod/argocd-operator@sha256:483854621b95dcefa03d18a970600b271726422ae3362479f73dcfca9085c92d
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        - registry.redhat.io/jboss-eap/eap-operator-bundle@sha256:0db70367917668964e2f4f5efdcef864ccc3f065848023e52caf4abdec728d9d
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:096946dd896b6d9d937eb6529c3a23a84ccb129649c5a0389d0ab83989c28b53
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/jhipster-online-operator@sha256:5f2436297c026832e5c15241d60090da581daa5502e352ea33d77f22ef3e7a52
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/openshift-community-operators/sigstore-helm-operator@sha256:ff4e548a8a003bd23e5076362bafb2e22f77b8445c5e3585d99cd5ce26fcaf56
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/mtr/mtr-operator-bundle@sha256:190a6fa2a6a9b6b182cdf9cab18814d05c203b54ca18dc533e553fa5dcca779e
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:2808a0397495982b4ea0001ede078803a043d5c9b0285662b08044fe4c11f243
        version: 1.0.2
      deprecation:
        message: |
          The preview-1.0 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:a91cee14f47824ce49759628d06bf4e48276e67dae00b50123d3233d78531720
        version: 1.1.0
      deprecation:
        message: |
          The preview-1.1 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:4477810692429ac498b6ec1b295e6ead0f8fea25e91f6ddf17255d7e349b82e1
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:ff2bb666c2696fed365df55de78141a02e372044647b8031e6d06e7583478af4
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        version: 1.2.2
      deprecation:
        message: |
          The stable-1.2 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:8da62ba1c19c905bc1b87a6233ead475b047a766dc2acb7569149ac5cfe7f0f1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        version: 1.3.3
      deprecation:
        message: |
          The stable-1.3 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-load-core-bundle@sha256:f62a7f8eb74d0d5309fede6c9c0201d1f0c3736f93f1aeec7446089d64fce28f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        - quay.io/openshift-community-operators/camel-karavan-operator@sha256:6dc78ecdf6e1936872626c02d41986d6553c1500002d85692c97e86fa77c129d
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:5a5e2f775760aa68c85d0f69b700d23ef2a6be7bb415bd96401d03eb6077f90c
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:79a9191f0cd5d11af004ead87ba9eb5359c67fd658c346db4c3cea3b16b23987
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:54ea11add0a1d988e0c965968f9a65e011978190f76fa2f826f1378474bac1d4
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
          createdAt: "2024-10-11T14:05:47Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8337924489a1747d4bbca6dd6ac42210731cf721b549b5830846b271621e2ac1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:6d0aeffbccaf9df83b42c530bf82707aa7c1cc16f72a5ef30b66fbd562c01ab3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8c05a303be5c536f4051ef291475b578d4d38ca0946fb4b4dd8bca376ef432aa
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:e91c2e5473f3e24c185dd25e9843f333e6d7419d9c5f4d4396669717b5969683
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:c543dbbb66fcac02345e1d3c12333780721bf049dca0f0a27d4c55b15a5d3e14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:a65780c5680a6fbbd0e3d0eed423b14a9c9b3d7d17f801c1eb6eb6b777384ca5
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fd6fb83bbeb9010f7e91f0afc91beac0bf38cb79c2de8f533a2bd07b1b238262
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:936fd68d7659683de7789fcf34d2e5873bd4bd1534c01c2d80501647f7c3c442
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1dd8f0f08ff1323d910fea8adfa13cef64c2fd2f01968e4d9d55f71cf42835d3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:365115407964d182de7e7564438c6fd089e2c0bcc0f94ec48621ca0cc2bf0fe4
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4bea07e101740afc36a1617363a79b1c474efcacd584f6e475a53b3f2395a079
        version: 2.7.2
      entries:
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:3d0409c1a0b17c51d3dbd5aa0371edd597e3c5ba194aad987fe867dd804d1923
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    - currentCSV: mtv-operator.v2.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
          createdAt: "2025-09-09T18:06:51Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:41cc528bf93000095b4eb56aaff9a6dfb124d1711a2f31f0745163ff934ac96e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:0c8ebfdd52053dbcc221a3df7d603163a821dd9e0ac6f54977e081adc6cb4504
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a299acb18821691cddba52904140cef8bf12632fbf2e752b33118a574a5ff394
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:ad3b548405e8c4c075ef7d4c0eb0d0430684066ed7eab4563a1f8378eaf3d939
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4b37b1aecbbf087257be5ba2bb559f2de005aacb4cd371192cdde1a9891739cf
        - registry.redhat.io/migration-toolkit-virtualization/mtv-vsphere-xcopy-volume-populator-rhel9@sha256:c0a4a232ff8adfb7d57e0761251ad0d79a0f02d27c4bd813f62f6b943deb6585
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f90f4eaaaa02b289704b19dd171282f3d11f370ed6cea992e55964729ad5116e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:258efbf073f78f8b5b48fdda22ef055e2315046cfecc9644da51959289c836d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d57b39674260f3bb1ae4c164022c2c7be3cc9645f1c46a83e43f40965633451a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:aeaa9ffcb0ffbb0f69b2712a3031f7e2dfc0824273b5c9c0f6349bc701a0e89e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:e438c5fccc957ad26ecfc3633561342d3b6f2a6d5a4c7839109564008de62222
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:83a523392a9c715972bba7044c7c3d3944594cc89778449e1f59fbd6a2659baf
        version: 2.9.4
      entries:
      - name: mtv-operator.v2.9.4
        version: 2.9.4
      - name: mtv-operator.v2.9.3
        version: 2.9.3
      - name: mtv-operator.v2.9.2
        version: 2.9.2
      - name: mtv-operator.v2.9.1
        version: 2.9.1
      - name: mtv-operator.v2.9.0
        version: 2.9.0
      name: release-v2.9
    defaultChannel: release-v2.9
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:0ee0bd2449f183277af845e312257b4a1ced4de865831b15e1758a397d8d29eb
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:2cf058c819b40512094f84f4d374457455370ec4994d34cee6e39b9d87841f05
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:f182078b0077b6df6d47ddcab0204e963c1475486bd61b8eccb73b46eb8b2995
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c8b3505ed491c1bf6a913f161715c8ac483d8d9adf1fd29916ea4edad987cd67
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:af5a13351f960c0938c68604b772f24743c0e997701838a7784c53406471e72a
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:0d9f5293d69eebd91fffd83ae371141e640cab8f9e2d559a83714ce0f94fee08
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:d6675617c19b4ac7c1913078eaf9ffbe3d03761b570dffa002d9de11f337f19f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:2a35a0ef852cb4d76e963df50449712a06ff13391c30675aaf931e21cb561b8a
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:00f7ce0148988e93679d4b53b45ae8d2ad54f308b908940e7b28504318f718c6
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:c352f5633beef2eaaa324dda66b9b484e5ea79efbbc3b781079f2e9d10c880a9
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-bundle@sha256:47a7bafff2715fbef6130ca0b84560b6c2358275821b72ae54ad46020e02bee5
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:30264b9f49912cc0e6bed59001869e3168caace20b292c9dad4cc2d4947d289d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:d7cc6d2575eb252fb3e2468106e801f192da7844b0b153dc3c6a12cf7cc0a8fe
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:60e99bb91309e448c652898dac070a6d920fda164f27879d94632c88a9e61ffa
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hive-operator:1.2.4953-c248a51
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:be43412a2e4bad2cd9d15e4fbabe0292585c5bbbd65da12d519f30f17d2fe594
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:822fd8e65c241cebedd835bfebf67645199561cfdd6148d5afffb8b35e499ae3
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        - quay.io/openshift-community-operators/hive-operator@sha256:9a3cb682afc66ad662045b3d8ff214203cc00be60f66f74564820f6dcf99cd69
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        - registry.connect.redhat.com/fortanix/ccm-node-agent-dcap-operator-bundle@sha256:545724c273dfd6a1fa923d468b2d054f39ec3186d01586a8b3fe494c0f22f8dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/ndmspc-operator@sha256:b1f47ce0cd6930d902dc3538a1a87140dbe9257dd50773a518fc2350185f90cc
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        - quay.io/community-operator-pipeline-prod/klusterlet@sha256:9efef9a034c9f12505a0d81bc9697c4a3ab7e3d05d1e7aaa4180490d70ddf6a2
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:21f4648e0e71b33ef24b4921ab34b9e2b5e90eac25ade735d51331ac4f7c4cf3
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:2ed6bfdf37e36c63cde3633443553dee410201200ad351e6d91a5651f8e20ecc
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:11edce56874ff1ee666752ec5fa521f3f6ae78c52cf1f289ceb97a86e6c083aa
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:5e4c319ce871a0dcbab7fff23325e6233c9dbcb2592fdeb1a9ba7c57ac2a63b2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:d86bb21fdeec35b1fc9789023a558750da76e036e0790dcfce97dcd137dcb6fb
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:7d94521a5cbdd34f8307653a3e1b596ec86415b2165fabdbabc8ed8bed35473d
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:3d5ad015a953708cc53ac2ea92b1b1d99c870a801b26e790d41458adc0a958e5
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:fea5b977249013181616f60b95c5e6d224cd0f248717171d946ff0e4d6eed6c0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-memorydb-controller@sha256:7f2e65d817479fcc5658cac745cef325b2bfd6256eeed8b819f1713afd59fe68
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patch-operator@sha256:437f5b7c661d3e9bc2c8ac13b0c8f3565f685b9e5ef52791312bc64d55e2e476
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator-certified.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-operator-bundle@sha256:4512f786fed0d9b2285b5a2827ec3b597452f90c217e27398f2f0f552377d844
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - quay.io/helmoperators/cockroachdb:v5.0.4
        - quay.io/openshift-community-operators/cockroachdb@sha256:f42337e7b85a46d83c94694638e2312e10ca16a03542399a65ba783c94a32b63
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        - quay.io/openshift-community-operators/cockroachdb@sha256:d3016b1507515fc7712f9c47fd9082baf9ccb070aaab58ed0ef6e5abdedde8ba
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/netobserv-operator@sha256:73af4040707ce79baa3115deed1e9148436b248ef3570f1175a6d31fdf9596c4
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:394d122efcf0da16bb68bfef1292395cda1bdb21909ae55f8fcc54a16814f1cf
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:62f93c5989d1d8217a3bac076addac05c0906a1a1f1eeaacb87bcf256812fdf9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:ea8c060444aa8e89dfbd90d61d6ec3900ab4d6634f6aa8067face1e6974b9065
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        - registry.connect.redhat.com/loadcore/keysight-wap-notifications-bundle@sha256:9c2ebb98476bf3374b71a838c394017e7d31aaa045bd52b3a66924b6f6b7b067
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.connect.redhat.com/cpopen/ibm-licensing-operator@sha256:a9f68d204b6acb50ac5023e5f2314ad0d06e06483f87290381d1e92476a79579
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - registry.connect.redhat.com/noiro/aci-operator-bundle@sha256:80e25268ab5aa6ea418deae5ccc197fbffe2b9cdf4f27894f5beeb9315f6b123
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:48b37460dd7245cf661009267c3353962f0f195efd6a3007c9cc70b51b91a9e3
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:9949a581df3960a770117b6cf6b9fee1c63ea12ca09ac2d677f1cbe0a83fbd93
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - quay.io/akaris/sosreport-operator:0.0.2
        - quay.io/openshift-community-operators/sosreport-operator@sha256:f0d28a3e2ac2e1cbd0c307797b728ad89ac33375135d72054a7b765e274e4bea
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/node-maintenance-operator@sha256:80dfaab6f7a5c34cebb45c874ce134bf60e3697e14a95b69f1972dd98ecc172a
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=4.11.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:75eaf7b8b71e1a15e002f6cc5430e5e50039699d7b52c3150f5f71b1595f00c5
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-operator-bundle@sha256:41ece1ea773fd610cf8a57483a4d2846e0bb3ce81db701021a3e0dd91e13b7c8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:f7159e880ec55da49ac774db0c454800055e7cd3f37ba94fbcb629d08aac1146
        version: 4.18.0-202509100344
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-operator-bundle@sha256:4f8908ba97517826fda14cab3624ed42a6348fb96f0be4abf9cd2a7b68f7d3e9
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-wap-migration-bundle@sha256:3deff8b48ab1a6f0e5fd0f8315bc1e191da55525cb3cf0f315f2f7afdb47a6c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:20a4f2c56cb186fff4e8d31021d84b690512cc65c3c9629ffe623a0213ca9b0a
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        - quay.io/community-operator-pipeline-prod/mariadb-operator@sha256:3cbbff7ae75fb4f4629c7b8b58ca85ce67b23d5da558241301266d010e8b1188
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle@sha256:333a6f476d03adcbfaa4ab38802859719a7dd8857783f8487c99b008d06815c5
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acm-controller@sha256:4d63542b39f41c458dd1eec28e44edce3a919f4dd9137ff64a9452cac1a1f954
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - docker.io/absaoss/k8gb:v0.11.4
        - quay.io/openshift-community-operators/k8gb@sha256:9743f957b00e5ad344fed0c062d616d8222990139ac4afbf14871fd9813f0ef0
        - absaoss/k8s_crd:v0.0.11
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/ucs-ci-solutions-operator@sha256:e284253263183eb2cd313f6fb009bb830b9a79a4d4e2481468821c30a9cdce03
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-operator-bundle@sha256:67723df19ab7d7871a6b639da3975c230d95dd250fa2995a9f44bdbb4b141454
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:439aeb806c3264948a15e325f7b1e71037099a085c8d971646188ace77fcf5d3
        version: 4.18.0-202509100344
      entries:
      - name: secrets-store-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: secrets-store-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: secrets-store-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: secrets-store-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: secrets-store-csi-driver-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: secrets-store-csi-driver-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: secrets-store-csi-driver-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: secrets-store-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: secrets-store-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: secrets-store-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: secrets-store-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: secrets-store-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: secrets-store-csi-driver-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: secrets-store-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: secrets-store-csi-driver-operator.v4.18.0-202502280802
        version: 4.18.0-202502280802
      - name: secrets-store-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: secrets-store-csi-driver-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-09T14:17:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:c5bd4f10120088f285a3b65e85bba825f9f55964758901f3b5827ac31516675b
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:26935de5f21a7865aba484599136df6a251c009e8fa9781e1040da3b19e7dd33
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:e05074d46e062268461d2689f7048c1a5dc4ed89a039d93e382f5ebc0726a269
        version: 4.17.13-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-multicluster-orchestrator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-multicluster-orchestrator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-multicluster-orchestrator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-multicluster-orchestrator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-multicluster-orchestrator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-multicluster-orchestrator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-multicluster-orchestrator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-multicluster-orchestrator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-multicluster-orchestrator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-multicluster-orchestrator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-multicluster-orchestrator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-multicluster-orchestrator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-multicluster-orchestrator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-multicluster-orchestrator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-11T15:26:27Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:4b78c8df4f4aa4be805a442b059b5e2822464bb029826f2f0090cf8b6b747768
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:47c754d0e2782dd8d885758e68803173c32f1fe6319ac09c786dd24e5124df25
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:957805157e7690deedaae63856c9ce71e5c299610bca401524397a40e2820537
        version: 4.18.11-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-multicluster-orchestrator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-multicluster-orchestrator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-multicluster-orchestrator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-multicluster-orchestrator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-multicluster-orchestrator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-multicluster-orchestrator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-multicluster-orchestrator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-multicluster-orchestrator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-multicluster-orchestrator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-multicluster-orchestrator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-multicluster-orchestrator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:3b683338f5849b887e1bc81ea078a9b799d272f824abd4247ff87e70ca7acbf2
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1a156d67fd5beb9ffddecab724490acb7143f8141ff3b392c6f519e76c3ad374
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:645fcb18684884aef8ca4d5299f0d053d496c3a5d3f37cadd66c51d38514a680
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:5a73ce3a8cc8422aeda37fd9ef97f5aecdbf8c69fb6ff488df26342e989b5c57
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1dfe9180a9fca4f773c1282dfd0006daa0bb48dfbac87e352661e051107f9033
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2cc940711539edd93a23355dcb06365821d140346e1df248386a77667c0f47ea
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2418015d00846dd0d7a8aca11927f1e89b4d8d525e6ae936360e3e3b3bd9e22f
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - registry.connect.redhat.com/a5f1f0cc7ac3db90370970f9b/tawon@sha256:3691c0948c8eb266ada348de504622227af6516f3b3d331eb09fd258dd0609a5
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    defaultChannel: alpha
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
          createdAt: "2025-09-19T13:48:41Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dapr-kubernetes-operator@sha256:8f2087df1987437f1c5e62aa9453fea220a41012077fec67b6ba8368ec267421
        - docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
        version: 0.0.11
      entries:
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-apigatewayv2-controller@sha256:1cf12140fcd67b0f3e8c5791a91c5b705bb3af8b2e03ea4ba2d4ce6f9f7a4809
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - registry.connect.redhat.com/mongodb/enterprise-operator-bundle@sha256:23c3c8387eb9377b52327d7193f59eeea5831f4050248b3cbb54cbcd9437fc71
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-websocket-service-bundle@sha256:da100dd9c5b542ba46de7166e783e28a0f7e675f9ac896f211405e5a9e2082ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "rawDeploymentServiceConfig": "Headed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.34.0
          createdAt: "2025-09-03T09:59:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Argo
          Workflows\n* Model Mesh - ModelMesh Serving is the Controller for managing
          ModelMesh, a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks\n* Feast - Feature
          Store is the fastest path to manage existing infrastructure to productionize
          analytic data for model training and online inference.\n* Llama Stack –
          unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opendatahub-operator:2.34.0
        - quay.io/opendatahub/opendatahub-operator:v2.34.0
        version: 2.34.0
      entries:
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        - quay.io/community-operator-pipeline-prod/opendatahub-operator@sha256:d3e3d5ae02b834522481fa9046bf2ad3ea06d706c5cb06a32d32e61e229e4a6c
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:2992841660e8b1664fa2ce06da894aa828d4ea1be3eb5b27476efdd73731486a
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:bf85a9da8a100a90deb3745497bc1505852910ef23dd12c624fa17de51bd8377
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:329d6eaeff787bcde5db0ba4f9403eee533a601b45dc2d5d349b318705452591
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:2cac5fdb55664c08928a7d924eacb18cc4c03c609261e3b6d812590e4d166375
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:b35da8aa4e6746952e5cfdbbac72950a48f7ad32a3d762a470ce62118c78643a
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:e5a0cb8af7879bf13d940df278ccdb8a5c400e5d7108f23b4abef94c964e40d6
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:91aae0500cd3cfc9a4a536241feb4fe50288e2a68f69ab5c602c02169e5f5b2e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:40c7d896355dfe67ff3616f1e1df86548bfcd0524f7e56d4d72ab4f5f1739f6a
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:3f69a124f9c31ebc0c78f249872a69b1289622bf4f5aad4c739f6e8caf9cb85b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:33a2e21a069ff24af4bf0ffc51e83ddf398df5d9735d0123120768e9dc2c27f9
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:826354c9c06a06f998e522a88d6a81239e003e1d6a2633e9b7b27b8a61d5fe38
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-operator-bundle@sha256:a59435a22c2f470b149ff721240f14144d2e56200b141e59a4f1011267a6c925
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:1346dcee82f2120bd2af22be9f041e057bdc32f9ad1c3ef717d599ba6519faeb
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:44a5c4cc3086e4eac42dc78a72b7492d13df0098017288278d9006e767a480a1
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:b9dbbce235b7a4f45f42184cf0b908e91e17094e971bc29a8710f35127a6b06f
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:026418bae8b2d425d45b2b3481f65c241d89e25f455e3fbfcf2f79536d978bca
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        - registry.connect.redhat.com/nec/webotx-operator-bundle@sha256:7e655933a7fab3ac3aa88eb5e4ca9ac3a71b98875ef1ef5bc49e87284f160ff8
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-licensing-bundle@sha256:2d3abd0263183bc92718316581675b92073f37c3786784fae24b3587a9441def
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/community-operator-pipeline-prod/cert-utils-operator@sha256:92a0c0b893b56330593e0a255b8df61031e92a61d172aecc58be5805d847798a
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/community-operator-pipeline-prod/gitwebhook-operator@sha256:2d33e4f4f4b607ee2006481efc7ad9341049085ef988777d23e586f7406a5109
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - quay.io/bpfman/bpfman-operator:v0.5.6
        - quay.io/community-operator-pipeline-prod/bpfman-operator@sha256:6b3603c7c1917c67b5aa57202d1d2fdc0d9ac6b3a9373afa7172f322b3b3fa9b
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:74775c7851ac98ba2903435c71b9d02cc504ca8ee98b65ad7282d671392756a6
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.connect.redhat.com/cpopen/ibm-elasticsearch-operator@sha256:0bedcf95b31e28d41b124780f3b97c68a87db5626ecf1d5e81bb06b8b3fcd91a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/federatorai-operator-bundle@sha256:d0ac978d749654ddf4cc20823805ba7d33ee62e9a7ec98fdd53d9926b697c347
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle-marketplace@sha256:19c9d44cb9a0c7f52501fcb589750f3e85e270d066423bcadac593dd79e48ff4
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        - quay.io/openshift-community-operators/github-arc-operator@sha256:d8d97d19cea466c4c240097b79bfb7de90312d937ad4db63f9e90901686e54da
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cognitoidentityprovider-controller@sha256:093670fcaee6ef20e471c0df45c9ea7f40e5f0c43208c9447623324fdb39dbe0
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/self-node-remediation@sha256:549585ca18d331cfbd5911fa3e79912428a0be3308588be2438ece3247746bd8
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |2

          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler,
          **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: kaliu@redhat.com
          name: KaiYiLiu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-operator-bundle@sha256:5ebbc06088e99b937327cd457324461cafdda9649b2dd2db1bcb892be8556eed
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:141b069ff9925390607315b0cf675003abbdac9555b29700e3b999fadd6c89f2
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:971174692a1aae68ac07ffa132254eb3c018a52507eb32621b6dc286fd25ca0d
        version: 0.5.0
      entries:
      - name: power-monitoring-operator.v0.5.0
        version: 0.5.0
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.2-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
          createdAt: "2025-09-11T18:04:40Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.2'
          olm.substitutesFor: cryostat-operator.v4.0.2
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:b6a6e401042daf1780ae6be19aedc551ed176d1a13bf1f0f48359552c638a156
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:8fd30d86f5468227899560071a28e92197c5790ea820e918c47559f0695cfb8a
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:d2788c8732fec62c7a27e7bd1497fea1ab3ff605e7e913bbe17d8ef97567329b
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:550de94fb20462c5cc5c52166820319544e774b8d45fc7beed7a095f5c1300aa
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:bbee66a44c2e3a5525c2fa81db72e8528bccf40583d0bd665841644473308310
        - registry.redhat.io/ubi9/nginx-124@sha256:126903359331eb9f2c87950f952442db456750190956b9301d00acfcadbc19ed
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:67e313e50c6ef100d6938dc1a734728661f328ab1f9757667c7a06de11689967
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:d9720fefa3e34d05a5536fe1a66e23b30fe0845ae2b96b9c9be4ee705be9633b
        - registry.redhat.io/cryostat/cryostat-operator-bundle@sha256:455ecde240a152987fcfe0013abee992014510de9e89c014559a5d0665a56512
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:9ba156d802c9c3eafc948afef88e0eaca6b5e73bf32ba80ea7b0651ba7ce923d
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:2a5e1d2d39a7d9d2bc74316110e1a65e9e18f3a9fb7f0a498c7457dd9e6684ee
        version: 4.0.2+4
      entries:
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
          createdAt: "2024-11-27T00:00:00Z"
          description: The deployment validation operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        - registry.redhat.io/dvo/deployment-validation-operator-bundle@sha256:c04bae05ee0c24fe5bf94257b0dd2cb020f5c9b5f173ceffffdd041b6dff4ee9
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.11
        version: 0.7.11
      - name: deployment-validation-operator.0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.0.7.5
        version: 0.7.5
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CephCSI Community
      provider-url: ""
    name: cephcsi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cephcsi-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:12:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:b023d95637934b16ec8342994edd4e3198579517b68817ce6b4cc14958746b9b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:10cfe59ee2d625ed67f3804ba7e471f10fa1e12bc4f96de5e2f453e5cdefcf89
        version: 4.17.13-rhodf
      entries:
      - name: cephcsi-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: cephcsi-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: cephcsi-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: cephcsi-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: cephcsi-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      name: stable-4.17
    - currentCSV: cephcsi-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:20:52Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:903e47fa88119b5507a0d773d58318236bbce53201b5e6a18a16311f704b1120
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:562703a8b197f03b29d0564df329ecfda192a94b7818c6fd5f919329a5cc9bd7
        version: 4.18.11-rhodf
      entries:
      - name: cephcsi-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: cephcsi-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: cephcsi-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: cephcsi-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: cephcsi-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: cephcsi-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: cephcsi-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: cephcsi-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: cephcsi-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: cephcsi-operator
    provider:
      name: CephCSI Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/openshift-community-operators/pulp-operator@sha256:1926e96b185443ff0c7a2861cefc2e3aded0c0493e0d7f7d78e45447e06750d4
        - quay.io/pulp/pulp-minimal:stable
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:1e75649538f13cf3bb2e6131160f038a70689d6f92a110bdcec348e865040c4a
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    - currentCSV: pulp-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 1
                  },
                  "custom_pulp_settings": "settings",
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "ingress_type": "nodeport",
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "example-pulp",
                  "postgres_configuration_secret": "example-pulp-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "example-pulp"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.1.0
          createdAt: "2025-09-18T18:00:01Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          metadata.annotations.support: Community
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://pulpproject.org/](https://pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://pulpproject.org/pulp-oci-images/docs/admin/reference/available-images/single-process-images/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://pulpproject.org/pulp_operator/](https://pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:92a3e95cb3b1e50100933ea08cb44074fc505f6c0e7fa24d028ec46fa9e95590
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.1.0
        - quay.io/pulp/pulp-web:stable
        version: 1.1.0
      entries:
      - name: pulp-operator.v1.1.0
        version: 1.1.0
      - name: pulp-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudtrail-controller@sha256:20fe3c3ba7688906de40d7fd36c00552be280595708935ff800ad5228829f781
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:fbf847c41e1736fb6efab7a6acdc03acc562fa7e667306773880b277883bc4e5
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-operator-bundle@sha256:7c808f6e4ea469d88703300bdcc6bdce160634833cbd10bd110dc69d05be588c
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
        version: 4.18.0-202509100344
      entries:
      - name: clusterresourceoverride-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: clusterresourceoverride-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: clusterresourceoverride-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: clusterresourceoverride-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: clusterresourceoverride-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: clusterresourceoverride-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: clusterresourceoverride-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: clusterresourceoverride-operator.v4.18.0-202505070803
        version: 4.18.0-202505070803
      - name: clusterresourceoverride-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: clusterresourceoverride-operator.v4.18.0-202504251302
        version: 4.18.0-202504251302
      - name: clusterresourceoverride-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: clusterresourceoverride-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: clusterresourceoverride-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: clusterresourceoverride-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: clusterresourceoverride-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: clusterresourceoverride-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: clusterresourceoverride-operator.v4.18.0-202501231202
        version: 4.18.0-202501231202
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator-metadata@sha256:a2f0513cd8b96d71855464384cad7f1cd8c39f8d445be1ff1c1b91f22008d16d
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-operator-bundle@sha256:c2229c2cfad9bbbca9179cb6617bce0033f5be40731641a69b41f9a90e034c8b
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        - registry.connect.redhat.com/hazelcast/hazelcast-platform-operator-bundle@sha256:0aabd3e7f5f993c1288450678d55e8f7b36c8b30894e5d844521a474aa6b2714
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/smart-gateway-operator-bundle@sha256:f12b2f1dd34c73b9c8a5ad461e771c0ce56c9ea4eb88f03225be1462a42c5a78
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:0cda4be8f74da7524a72c961c8762927c66e145d82e59af92073a54179bc9936
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:4bab095c5dd6f8c95a17a8ff2bf70ff3f53faea48b56eea8baa9cf9a42a5227a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/keycloak-operator@sha256:a4da3bcbb4e99d60b8217d9e62742eeb83eef163dfefb2f7efda131f312e46e3
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        - quay.io/openshift-community-operators/keycloak-operator@sha256:65fe00f310289b70538a6dde897ef663731d0da44a7bb1d0f55a3001a50e53fa
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/keycloak-operator@sha256:70bbd44f6ecf86aa8e9710175a96dd3c1d813aec40667a09800241ce30afd93f
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:9009198860c7a1e0de4f429261d442b75b5b2893a943bb77914b14d643db8829
        - registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
          createdAt: "2025-06-19T11:07:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-zookeeper@sha256:c8487e18318a92a64e84a5ab29f844a0973f3b8b2aa291723014731471459176
        - docker.io/confluentinc/cp-server@sha256:072719eef75ee61294027ba77f7b9cd821437158761495a1644b5687af9305a8
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:7a271565db9c500f8c5755de397e8093251673d2fc6b8a10974f927f382b4704
        - confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2520a514b848b53338c411f0e0dfa68491596761f8bbbb090d652f71eda010af
        - docker.io/confluentinc/cp-kafka-rest@sha256:9c19e7900669e555810d26676393737a7d573b39c31e284c0ad78807bfe05ce3
        - docker.io/confluentinc/cp-server-connect@sha256:05e77fe8e19747521e1f1d5630bbf384b0fe0ec9efe8dc798de6e4a3eb0bfa2d
        - docker.io/confluentinc/cp-schema-registry@sha256:5b6bf5b971dd8fe2c821e227391858f58f6ae2128f1ae1faa988c3ea59b86801
        - docker.io/confluentinc/confluent-init-container@sha256:550a6b92fc5aca81ad8e0d3d47c0063af9da877e45c1ecda1a9cc44a70e5d87a
        - docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df1f6a99ecfe8d8acc0af222848aad13349f06071fc2bf1360d6d20f99769394
        version: 2.10.2
      entries:
      - name: confluent-for-kubernetes.v2.10.2
        version: 2.10.2
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.2",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
          createdAt: "2025-06-19T11:14:20Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:b7cde5736b5275abafafbb81f4c917890abfe8f400f436c3a1f712cf200832f8
        - docker.io/confluentinc/cp-schema-registry@sha256:4f6cabeea11ae532cb47905eadaa3192d7113ea6d0a0da310cdfd5eb7142fc45
        - docker.io/confluentinc/confluent-init-container@sha256:37161a90b89dfad625bf61ba7d299707d7b2a96271e9667044c3eaba3ffc5f66
        - docker.io/confluentinc/cp-server@sha256:dfed43a1618de1fea6d202921ca4cc2168f7455c3d37fe046f3e64febf042a87
        - docker.io/confluentinc/cp-zookeeper@sha256:f1e6d3d698a6e2757e3c225b5c6f8f007251838226c46f652a4cbb4314b50277
        - confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:2baa0cb5df8017d9d14c2f0c7e02ce371ba92462d8d1e93a3ca1bc548c20a74e
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:507e1538a64c4232ffda86504594def45cb20bbcb6b13fc6196c1f41e71f93bf
        - docker.io/confluentinc/cp-ksqldb-server@sha256:744f5d844ac223c857116f3994c5be2105898ea3896d53026eb62e76c7240c4f
        - docker.io/confluentinc/cp-server-connect@sha256:73822ffff10d386b98c9ca2bb295b4d079deb210116c8e4283cdc26ce304526f
        version: 2.11.2
      entries:
      - name: confluent-for-kubernetes.v2.11.2
        version: 2.11.2
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.0",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
          createdAt: "2025-06-13T13:02:11Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:ba49d46654770994c302d586a609afdfc58a7cc05f204aa090b643f0680a5014
        - docker.io/confluentinc/cp-server-connect@sha256:e4cbc1836402497ea42b2c961c5d6b4420c25b3c6a5f6f71be3e99fcc0f09702
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:a283f2a72b6346febaf711e71dee1e0e3666d7265320f3f69307fc2c0b95d5ef
        - docker.io/confluentinc/cp-schema-registry@sha256:3dcb9652a4c83d4566a7334041c1ac3ec359665a9450a70c5d7b41adaea55c90
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:35f2799f71c308d87e6a7dada00263abb9e08e895b10f10c815e7a9edf19dd49
        - docker.io/confluentinc/confluent-init-container@sha256:82277228368a783ba84367c920a9fa15fad6844723df1e16b07c9dcc829a54ad
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-server@sha256:ffd5961412da42c81c9b20f35e334aa8bdf351dcb452610e84156ed511116ef5
        version: 3.0.0
      entries:
      - name: confluent-for-kubernetes.v3.0.0
        version: 3.0.0
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator-bundle@sha256:21b69b5eb896a37691a8e086f3a51c0f68086a0ff9370fd66667acc6c935a850
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/mta/mta-operator-bundle@sha256:65c0932bfda32b1e62be9edd0f96bb4d6e41e302a2e8080fc6cc794ee5c52276
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-operator-bundle@sha256:f3d585e24d65905b2e9f594dcdb67721aa95057ab7587ae531c3487aa2f94e53
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/mta/mta-operator-bundle@sha256:a13643117c2867351718a872f7f1b2350c67855ca73a727a1dc140754ffe6589
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/mta/mta-operator-bundle@sha256:9eeb43af2bcab84f5261d1575f7c897903a7696dba011d256abffa1fe850eba2
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-operator-bundle@sha256:6aef520983fe01351886e108d32b998b20e74116fc0280c754a914bfad03d0bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ed10e35f45a140e21d826bee71b3ab96e67f1b4dc8c2f7f1676380fa51b67fe
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        - registry.redhat.io/mta/mta-operator-bundle@sha256:68953a2a721309fd103e09f3171d0c9d50020413631b575f1a44602a5d2ac855
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ec977b0c9c26551db110dcf2d5069647b1a5495f869ff48a0319cb496b84cdc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - registry.connect.redhat.com/eginnovations/universal-agent-operator@sha256:65f851074e0e1db57a62b559c0f10f9d30543011727aec3b832f9a35468139e9
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator-bundle@sha256:8d130a5f2407931dbc73e45a53610b95ceb04cb507b8f6519f630cada1c0da45
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        - quay.io/openshift-community-operators/tf-controller@sha256:cdaeb470d9b331a3fa08167e5c904befb00be983e1354651670e21e13560bf9b
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/openshift-community-operators/sosivio@sha256:110d7f4fb993d65ff497eca57c42206faab07dc5f5b1f43b1b682387e7cd3a10
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        - quay.io/openshift-community-operators/l5-operator@sha256:e4cb13801a80753c99a47f009ca0a29416b3e04afe5314d07d97c2464a6f0d75
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:712d31214423b326ab8f107435c85431ce8d99c2e52fd8d615005a4274e0ff98
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/operator@sha256:2cb5ec186af3264cba42333245616dd01f0f9868a849129b1a31afbc51ee1410
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - registry.connect.redhat.com/harness/gitops-operator@sha256:f97cdc2245b7133a04235c55540d3a87165e080b1d5b9572a663731f6ccbe597
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.18.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
          - description: L2ServiceStatus is the Schema for the l2servicestatus API
            kind: ServiceL2Status
            name: servicel2statuses.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-metallb-operator-bundle@sha256:c80c3467cf52fc4b2a55e444fba9fdc8cf3b1efedad5d1276a889d07aed44e68
        - registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:6f23dd225f2bccfe39d05428a51de990a4a9338c49da079384eb30fb470c04f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509100344
      entries:
      - name: metallb-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: metallb-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: metallb-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: metallb-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: metallb-operator.v4.18.0-202506251003
        version: 4.18.0-202506251003
      - name: metallb-operator.v4.18.0-202506070804
        version: 4.18.0-202506070804
      - name: metallb-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: metallb-operator.v4.18.0-202505081435
        version: 4.18.0-202505081435
      - name: metallb-operator.v4.18.0-202504291304
        version: 4.18.0-202504291304
      - name: metallb-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: metallb-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: metallb-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: metallb-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: metallb-operator.v4.18.0-202503181802
        version: 4.18.0-202503181802
      - name: metallb-operator.v4.18.0-202503110933
        version: 4.18.0-202503110933
      - name: metallb-operator.v4.18.0-202502261233
        version: 4.18.0-202502261233
      - name: metallb-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: metallb-operator.v4.18.0-202502041632
        version: 4.18.0-202502041632
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:latest"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:latest
          createdAt: "2025-01-03T19:50:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The eBPF Manager Operator is a Kubernetes Operator for deploying
          [bpfman](https://bpfman.netlify.app/), a system daemon\nfor managing eBPF
          programs. It deploys bpfman itself along with CRDs to make deploying\neBPF
          programs in Kubernetes much easier.\n\n## Quick Start\n\nTo get bpfman up
          and running quickly simply click 'install' to deploy the bpfman-operator
          in the bpfman namespace via operator-hub.\n## Configuration\n\nThe `bpfman-config`
          configmap is automatically created in the `bpfman` namespace and used to
          configure the bpfman deployment.\n\nTo edit the config simply run\n\n```bash\nkubectl
          edit cm bpfman-config\n```\n\nThe following fields are adjustable\n\n- `bpfman.agent.image`:
          The image used for the bpfman-agent`\n- `bpfman.image`: The image used for
          bpfman`\n- `bpfman.log.level`: the log level for bpfman, currently supports
          `debug`, `info`, `warn`, `error`, and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`:
          the log level for the bpfman-agent currently supports `info`, `debug`, and
          `trace` \n\nThe bpfman operator deploys eBPF programs via CRDs. The following
          CRDs are currently available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n
          - KprobeProgram\n- TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n-
          XdpProgram\n - BpfNsApplication\n- TcProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n\n ## More information\n\nPlease checkout the [bpfman community
          website](https://bpfman.io/) for more information."
        displayName: eBPF Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - registry.redhat.io/bpfman/bpfman-operator-bundle@sha256:b6d43f1fec6ba7336c23d1fbceab15ebb2ed9e9e2eebb08c78d7541ee47b4204
        - registry.redhat.io/bpfman/bpfman-rhel9-operator@sha256:c3d878d77091bed03d99f72943d9b78e210e2f6194ea4b5cd2aa23f8d48f3f61
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      name: latest
    defaultChannel: latest
    packageName: bpfman-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-applicationautoscaling-controller@sha256:4f1b92eac30d72a66d8fe8c2c44339c7a5fd6af65d48c0d8697459c7af290741
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Agents",
                "metadata": {
                  "name": "loadcore-agents-sample",
                  "namespace": "keysight-loadcore-agents"
                },
                "spec": {
                  "agents": [
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent1",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "replicas": 2,
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    },
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent2",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    }
                  ],
                  "image": {
                    "registry": "prod.harbor.keysight.digital",
                    "repository": "loadcore/loadcore-agent",
                    "tag": "UBI8-0.1-7d990d7254"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "regcred"
                    }
                  ],
                  "networks": [
                    {
                      "config": "{ \"cniVersion\": \"0.3.1\", \"type\": \"ipvlan\", \"master\": \"ens192\", \"mode\": \"l2\", \"ipam\": { \"type\": \"whereabouts\", \"range\": \"192.168.1.0/24\", \"routes\": [ { \"dst\": \"0.0.0.0/0\" } ] } }",
                      "name": "loadcore-agent-network"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:53:55Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:f1d1e9f2aca11505407785bc33905744b667bc8aea49df73fcc50db379c0b375
        - registry.connect.redhat.com/loadcore/keysight-loadcore-agents-bundle@sha256:f16fcea424cb7b1eb647f3e9d3c08bd93c8a1b8bb7eab71a4e756da72a7a56a0
        version: 0.1.21
      entries:
      - name: keysight-loadcore-agents.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:a58c9c8454ac337c5a359dfe5b7c6b06983b7050ad62a0d11054f06fc17e92c1
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.89.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.89.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:f6c37a0808c137f226b6afb6cac1e2ce05184d6a01b9234969ce63056571e991
        version: 42.89.0
      entries:
      - name: t8c-operator.v42.89.0
        version: 42.89.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - quay.io/community-operator-pipeline-prod/gitlab-runner-operator@sha256:3e0314a2695c071fe0a34f3dc7684b43b5a31b6705f42982b4dbaacabe360d2b
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/docling-operator@sha256:9a3717ee195572b29d1ee31b96bb95561f5521b032dbaa09a000413b227b5f12
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-operator-bundle@sha256:2bab9fa63b128a3ba0a6031f70f0342f95d17dd53a3ec2930b19940b5b2e6f47
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - registry.connect.redhat.com/eweb/cloudnfv@sha256:0bb427954fb8df5de197353908a5189922f514e121ae8cdb8242d125b1b3807d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - quay.io/openshift-community-operators/odoo-operator@sha256:616e1067ebf8deb66636d57cae2ac1f9981b83419d8e7f023bc7df1c1b745536
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        - quay.io/community-operator-pipeline-prod/apicurio-registry-3@sha256:e207a4f7f53486b78f95ad958f7e0053a8a8561c3dcb1463a7176f27eb9ea082
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
          createdAt: "2025-09-19T21:59:41Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatch-controller@sha256:254df5f0e6821a65277bc3d029e7af87b10800b192f9724997996224525a9943
        version: 1.2.1
      entries:
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        - registry.connect.redhat.com/scaleops-sh/scaleops-operator-bundle@sha256:efd483d9c826e25b0d1cd79deff4263772260c72da79bad149ef7985cc00a44b
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:54bb383ac6c5f57ae77dcd96cdc33cea5c4c1aff9f9d356c1721591aa820e82d
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:948ad5eb47b7809ba5dd78cb6bb30b1bc64163617f229dc33cec251c1d5bb20e
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        - quay.io/community-operator-pipeline-prod/lbconfig-operator@sha256:b0bb724a5d9905633962a3e8267e75d371f754bce866a29cba3ac0d0c671ebff
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:610a5e66cf1b6c7f5451c9a0f5e0a6cc9435f2548335fbd50f44b8656adadd99
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:bc0d8b62d5814407e08817ece44626400cc3d9d29b14aa35dfe711b907fddf21
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.connect.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:15e8cbce994b783fa6dac737ac28f1b713d1c75891dbce4987d2c09ff295376b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:12d6ae3f139daf33dbf104499c195d9915470b957f401b5f622cfde92350a6c1
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:965ec7875eccc6a5fc56c3b8f0da668d826d34580117ba1438a607a9d0d13b73
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:f8a8cfcebba4b17fa6cfa99b3ef430af272c5237476dba4665803a90adfd5d2a
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:36561e5888cd34f8fb83ceff2910141fceaa6ea290bd246125930fe183e2ebc3
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:9e8bcbff46b97f3bc1071146b8db233abd1630016e6fc0efb679df8243c6fd03
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
          createdAt: "2024-07-10 00:59:51"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:518691d157b3e3b5031a2b1e4cc3f415839536079bb8570ae8690ac92d302ba0
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:a64f67f29a4eb8ce426ed842f1a5a5fcd83fd9d91acebb064cf65c89c9d8bfb6
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        version: 3.74.9
      deprecation:
        message: |
          The `latest` channel is no longer supported.  Please switch to the `stable` channel.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:7fd7595e6a61352088f9a3a345be03a6c0b9caa0bbc5ddd8c61ba1d38b2c3b8e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        version: 3.62.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:5187b04f8010a274d6bc017fc82f1b00d886124e520c1b7de8b4da8cd2f37bc0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        version: 3.64.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:d3bf33899ca8b199ca8782978618ca343164e43fd0ae2ac6a0cfbf697efa6428
        version: 3.65.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6536ee795e8aaeab9321ee4820e55933195983a768ba24fab40508d03e50a6e4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        version: 3.66.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:ad88773ab141f45fe24c61e83ec736aab27f652fb82ca2a85fee6f885568c5b9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        version: 3.67.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6a10a4859527946f61f497ff84b80f973cb7be5982d428423836b5aca81a1f5d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:297e7f863106c6b41f882bbd5b691a12c797c24d832473a04e604d4639e4a68e
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        version: 3.69.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:9b35ae6df2c19935783735d213483f3fcecead2f49f873a4f7844ef9fb5c7f1a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        version: 3.70.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e001b9f844a6c298dd74f1ab683f3dbbf0ad66044cec419c373ef4f97d47be73
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        version: 3.71.3
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:878ed77d07d3f40f049a594af44376797305786a2a264dae3d0c50401643bfa5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        version: 3.72.4
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:459a8d91beff91638872d377a8b82c279ad2905ce3f4ad4bfc00f6822418ac57
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        version: 3.73.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        version: 3.74.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:a0d9cd31d953aaf7e84f5a39f405c5aa2ea62e68d6c728a936ca1ffc4dc1b93e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        version: 4.0.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:b26133843687cc50e2c52d6067d8b3369b5cb28dfbfd15dd4b6062096673195a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        version: 4.1.6
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:08aebe6e8275d3314b6756d522ec568ccdfb9c52df98c9822f187f6c5a3af19c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        version: 4.2.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:30f9384244819da013358325cec364bb210b3b33b4cf9680698531fe50562aac
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        version: 4.3.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:123c0108bd7a190dd2accfd21786600f787147f040e57fb5f840861e2caa441d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        version: 4.4.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:41351e88688d23a443ad4f141caf0433b2c2789e6a1a85b4f5b36abeb5656bb1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        version: 4.5.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6d6b2696a12d49cf3403e338a0e028187473b6f43ef2b0bd72c7701200ac8e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:c7f894ab773d1d389656d13ed88fbf575faddd95c5b8396c6593ab3f993f331e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-bundle-rhm@sha256:5d14e0c44ffb543b7cf59817fe587c8e819542d8edfa261a44e58aea0ecdb79b
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        - registry.connect.redhat.com/randoli/app-insights-operator@sha256:21ed3d9b8b8466665c1cb784d0bc244279d0730f6319625caf46af111c25d333
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
          createdAt: "2025-05-09T05:49:47Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-operator-bundle@sha256:f49b7a2c60e79edaba1a5faabc3983f87f9123b6d84121a438d179ccbb8102e3
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
        version: 1.1.1
      entries:
      - name: multiarch-tuning-operator.v1.1.1
        version: 1.1.1
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.9-r3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
          createdAt: "2025-08-05"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:327a9899adfe89af9aebdcfac651a9eddec9ecf8a345caee1cef1b4a1fb279d5
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:df0493f7edbdedf754c70e4efc256fa4d246a24d6c8882c562e56ef3844278c1
        - registry.redhat.io/apicurio/apicurio-registry-3-operator-bundle@sha256:728354aee3729a1fba233c2aa92cacb21d27efd4869706ddd214a2d3688e33b2
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
        version: 3.0.9-r3
      entries:
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:d9930900b0d2f92ae813246fc322290234ad8bbd165b10e36671b53852228718
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:87448055a162e215a4d4982a8494fec6e7cc59fe6704494ddfdcfea076a2100c
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:a7c6f1c9260c79ebbdfb25ea996516ba1e2bcbedf25401e9e6848b955a2d3d78
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-operator-bundle@sha256:9ab3d5eb70c654270fef9e8e9d4e5b68e1ac3c86bd0e0dcd36cc6f8741be1603
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/cluster-impairment-operator@sha256:6fd114eb8f525f670e505373c2999a4e66b53beda988e99b267204d2fe0e1b4f
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/keepalived-operator@sha256:5a46fcb72f02a1f0d9cbc72f44948ef4b09149a4b3fe2e45fd31df90d20b3c56
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-keyspaces-controller@sha256:6e56400162867a4f55a70c340aeeb074726fd3dca13b6da60c395d6f35a48b48
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/automotive-infra@sha256:6d4180ff6899d4de8ba2cd4a06ece234bbd8f0ec1bd0b4b0ae4b7a0c047121da
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        - quay.io/openshift-community-operators/aws-efs-operator@sha256:6372847df81faddc23c938928157c3ac280e537a4eec9f7edd599480260bc566
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - quay.io/openshift-community-operators/reportportal-operator@sha256:67a7ffaf6a303b92541170f266d7b2871f5c2fe73f6b6c9a2b7b33dbc4555fa5
        - quay.io/waynesun09/rp5-operator:v0.0.8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:8197e4a96ac8949d43cda62ac2257a9b1e71f0b54f42283ee4925f88c369f873
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:728074151b245f5e0d65bd8956149f05ba22a1b6438120ebca2d85b0d7d0c837
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:4109698147632b858fe2787964772a1339cdebee72b75b42d9b1c5dbbbecb45c
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:088d0eebf60c8f6961ca72c268f61d12648d575aa41552ae6398fefc6d9700cb
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d6d444e642bfd567eb40c7e084f5df743c23b47935e52e35ef541c67fcc00433
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:b3469cd953ef2015fad2f7f96d960cee5ecaa6c19df631a9c882154ab5b3de98
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:f49d3eddbfcec1018fd3716cd0ec2addef9961d2d308099636541868566979b0
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        - quay.io/openshift-community-operators/t8c@sha256:403b29a6262ba0e06de117f788aea7224268b798868c3e5b738b05e96df20456
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4bfaf97ab09a1aff1316536b99957b1f82763b009dde6d08fc9eea1fdd290fbf
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        - quay.io/openshift-community-operators/t8c@sha256:61f1ea9b3851d7272703f5bc21203ff03bb35fd52ed9780d0cf996379c78a00b
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:48a5a9fbfa92b4b87b3653613467a3fcceacadbbabd274e367b40f6871039cc8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d65834d8f9946c2764d45666c90c50f428b74afe7b066d666c18e9545994c9a8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4964196aba839fde569ce58ab7e4eef438c9638ae4057194f2d7a58eaa5a98d9
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        - quay.io/openshift-community-operators/t8c@sha256:46c815fa635d58820757f41b1d8f04333638c39471f4d6c901a6500379680b57
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d8a575a84dd86d241acaaec53af1e164bc1ff1b95b5eb710851ae698bbed6bf5
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:dfdc8fbdfc0ee76b2d5c62f6fd5deac2c50756661871d8b4357e25043e5ce66a
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:081d7f4f20332599fdbbd07f22cc18eca6c5625b619efeae6e861239cde4b12a
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:cf7ccc1a60d6787ebcd1d76415cca4cc55c2c3ee18d52c0511e66ca6bf9310f8
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: loki-operator.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
          createdAt: "2025-06-11T13:21:40Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b311677919183c6c91c6da40b0ddb2529f92ed52b2c46a07e399a4bad5dea280
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:fce287241be64b7b2c20a971c4f2d468c079ae81556aaab756572ebad27a17f5
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:ff98532bd8f21ecf375062d465f2aee4cdcd70b647d258c1e09f7e2286f998c9
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:ea2d995744cd4a4bd6e7467c2d6927bfc89a47d551ef0d6b048c0c57652f81be
        version: 6.3.1
      entries:
      - name: loki-operator.v6.3.1
        version: 6.3.1
      - name: loki-operator.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b793aef88881aa75d47a22377d8932efc0b497bb01b2f3f376d64891d1dace54
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-operator-bundle@sha256:722cf3c05eb5663a737cb1385b971b6d1f0c8b2595d5ac425f7b100c25e2fd26
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-operator-bundle@sha256:ed97b11fa0f0f7aadf193810a5a599bf26015a0af8ee6fd90061110dcb0d50f4
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f19ae051041c46fe7587f903a9c6221bcbe8be2c4cc0235739f7592eb93ae164
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:28e37ec35de1307405264cd4f79aaa6ce4d4ecbd6cbc27f4fef5db415c678d64
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        - registry.redhat.io/quay/quay-operator-bundle@sha256:4a004584fa6c7187481e8c8480f17e0e4edd18c32f52a2e271c68d02f1a9655f
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f16c0a135dd7380e2a489f963b38bd76a63a973f482d06ece8d3656aff223703
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b499eebed32c0261db414e307f96f8aef9e743035544c80c048526e7ba4fbdc2
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-operator-bundle@sha256:55a5e6297109aea029bd40fac09a9fbc10c28a8c5558f9afc9b9b86c5dc7a4d4
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-operator-bundle@sha256:e3f63ea56de7ca5388a20c17ac63e16f07935ab09bcdcd38a86639c9d59c8a7d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-operator-bundle@sha256:53fcddadaa31a5f7d74c38f0bbd0d80b4c09796a580ef8494f60ff82fb6fe6c1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-bundle@sha256:9d0ed6d279800a189c1466e0b32f80af2f8bad96a65b7dc0322c64337dddf39c
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:9d21b738b0d357ef064690eae66bd4b3810664bcde57237822f1e90e9fd6ad03
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/application-services-metering-operator@sha256:c3600cc8cd789df1f3c8f48bb3b1bc4c9cbf0285b6f6ce06f1d563c52f50915e
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/openshift-community-operators/etcd@sha256:e42581eacbc8349f40c6a4ed6d7d82d6168518e15b1a24cba36817205bd8688f
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:1d308347e5e290ac4bb7dd7aab0a1c0659d661153541d81dd30733b52d1fbf83
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:f1757deff8c4048ae42cb033f3ddd7686dec2968e5a4cd61ff8f74fda69a8a09
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - docker.io/grafana/tempo-query:2.8.2
        - docker.io/grafana/tempo:2.8.2
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - quay.io/community-operator-pipeline-prod/tempo-operator@sha256:14282a0ca95401ab3e4c11dc2b611a6f57c00c461254f52ddf03ef74c725767b
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/jboss-webserver-5/jws5-operator-bundle@sha256:ce4060138117dbc8fc21abb60f0b8ef1bfa5ebc0dd8eb803c0e79d7447341964
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9y
            dCgzMzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQ
            LmFkZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAg
            ICAgICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24i
            LAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJh
            c2gpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3Zp
            ZGVkIGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAog
            ICAgICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmls
            ZXN5c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IG1ldGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBk
            ZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2Vw
            aGZzIGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1l
            bnQoCiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92
            aWRlcyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9v
            bC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRh
            IHBvb2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBp
            biB0aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAg
            KQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1l
            bmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZh
            bHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+
            OjxQT1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCki
            LAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAg
            ICAiLS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAg
            ICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBv
            aW50IFRMUyBjZXJ0aWZpY2F0ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRf
            YXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1za2lwLXRscyIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJJ
            Z25vcmUgVExTIGNlcnRpZmljYXRpb24gdmFsaWRhdGlvbiB3aGVuIGEgc2VsZi1zaWduZWQgY2Vy
            dGlmaWNhdGUgaXMgcHJvdmlkZWQgKE5PVCBSRUNPTU1FTkRFRCIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW1vbml0b3JpbmctZW5k
            cG9pbnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgZW5k
            cG9pbnRzIChjb21tYSBzZXBhcmF0ZWQgbGlzdCBvZiAoZm9ybWF0IGA8SVB2ND5gIG9yIGA8W0lQ
            djZdPmAgb3IgYDxGUUROPmApIGVudHJpZXMgb2YgYWN0aXZlIGFuZCBzdGFuZGJ5IG1ncnMpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tbW9uaXRvcmluZy1lbmRwb2ludC1wb3J0IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAg
            ICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkNlcGggTWFuYWdlciBw
            cm9tZXRoZXVzIGV4cG9ydGVyIHBvcnQiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1za2lwLW1vbml0b3JpbmctZW5kcG9pbnQiLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBhY3Rpb249InN0b3JlX3RydWUi
            LAogICAgICAgICAgICBoZWxwPSJEbyBub3QgY2hlY2sgZm9yIGEgbW9uaXRvcmluZyBlbmRwb2lu
            dCBmb3IgdGhlIENlcGggY2x1c3RlciIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgZXJhc3VyZSBjb2RlZCBSQkQgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWRyeS1ydW4iLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAg
            ICAgICBhY3Rpb249InN0b3JlX3RydWUiLAogICAgICAgICAgICBoZWxwPSJEcnkgcnVuIHByaW50
            cyB0aGUgZXhlY3V0ZWQgY29tbWFuZHMgd2l0aG91dCBydW5uaW5nIHRoZW0iLAogICAgICAgICkK
            ICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yYWRvcy1u
            YW1lc3BhY2UiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1G
            YWxzZSwKICAgICAgICAgICAgaGVscD0iRGl2aWRlcyBhIHBvb2wgaW50byBzZXBhcmF0ZSBsb2dp
            Y2FsIG5hbWVzcGFjZXMsIHVzZWQgZm9yIGNyZWF0aW5nIFJCRCBQVkMgaW4gYSBDZXBoQmxvY2tQ
            b29sUmFkb3NOYW1lc3BhY2UgKHNob3VsZCBiZSBsb3dlciBjYXNlKSIsCiAgICAgICAgKQogICAg
            ICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXN1YnZvbHVtZS1n
            cm91cCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgc3Vidm9sdW1lIGdy
            b3VwIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tcmd3LXJlYWxtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAg
            ICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhlIG5hbWUg
            b2YgdGhlIHJndy1yZWFsbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXJndy16b25lLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlk
            ZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cC1uYW1lIiwKICAg
            ICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAg
            ICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSByZ3ctem9uZWdyb3VwIiwKICAgICAg
            ICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9w
            b2xvZ3ktcG9vbHMiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJl
            ZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdG9wb2xv
            Z3ktY29uc3RyYWluZWQgcmJkIHBvb2xzIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3Vw
            LmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9wb2xvZ3ktZmFpbHVyZS1kb21haW4tbGFi
            ZWwiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwK
            ICAgICAgICAgICAgaGVscD0iazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gbGFiZWwgKGV4YW1w
            bGU6IHpvbmUsIHJhY2ssIG9yIGhvc3QpIGZvciB0aGUgdG9wb2xvZ3ktcG9vbHMgdGhhdCBtYXRj
            aCB0aGUgY2VwaCBkb21haW4iLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdGhlIGs4cyBjbHVzdGVyIGZhaWx1
            cmUgZG9tYWluIHZhbHVlcyBjb3JyZXNwb25kaW5nIHRvIGVhY2ggb2YgdGhlIHBvb2xzIGluIHRo
            ZSBgdG9wb2xvZ3ktcG9vbHNgIGxpc3QiLAogICAgICAgICkKCiAgICAgICAgdXBncmFkZV9ncm91
            cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikKICAgICAgICB1cGdyYWRlX2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFkZSIsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAg
            IGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9yIGV4YW1wbGU6IGNsaWVudC5j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0aGNoZWNrZXIgY2VwaCB1c2Vy
            cyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBuZXcgY2x1c3RlciB2ZXJzaW9u
            IGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBwbGllZC4iCiAgICAgICAgICAg
            ICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rl
            ci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVwZ3JhZGUgYWxsIHRoZSBkZWZh
            dWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAgICAgICsgIkZvciByZXN0cmlj
            dGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVu
            c2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2luZyAtLXJlc3RyaWN0ZWQtYXV0
            aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9yeSBmbGFncyIKICAgICAgICAg
            ICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29sLW5hbWUsIC0tazhzLWNsdXN0
            ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUgdXBncmFkaW5nIgogICAgICAg
            ICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3UgaGF2ZSBwYXNzZWQgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcgdXNlciB0aGVuIHdoaWxlIHVw
            Z3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAgICAgICAgICsgIlNhbXBsZSBy
            dW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMu
            cHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxpY2Fwb29sIC0tazhzLWNsdXN0
            ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGllbnQuY3NpLXJiZC1ub2RlLXJv
            b2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAiUFM6IEFuIGV4aXN0aW5nIG5v
            bi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0byBhIHJlc3RyaWN0ZWQgdXNl
            ciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRlIGZsYWcgc2hvdWxkIG9ubHkg
            YmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVzZXJzLCBpdCBzaG91bGRuJ3Qg
            YmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxpZWQgcGVybWlzc2lvbiwgZm9y
            IGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2ggcG9vbCB1c2VyIGhhcyBhY2Nl
            c3MiLAogICAgICAgICkKCiAgICAgICAgIyBBZGQgY29tbWFuZC1saW5lIGFyZ3VtZW50cwogICAg
            ICAgIGNvbmZpZ19ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb25maWciKQogICAg
            ICAgIGNvbmZpZ19ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWNvbmZpZy1maWxl
            IiwKICAgICAgICAgICAgdHlwZT1zdHIsCiAgICAgICAgICAgIGhlbHA9IlBhdGggdG8gdGhlIGNv
            bmZpZ3VyYXRpb24gZmlsZSwgUHJpb3JpdHk6IGNvbW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmlu
            aSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgKQoKICAgICAgICBpZiBhcmdzX3Rv
            X3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAogICAgICAgICAgICAgICAgdHlwZShhcmdzX3Rv
            X3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICksICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJz
            ZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgYXJnc190b19w
            YXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJldHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190
            b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6
            CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCgogICAgICAgICAgICBpZiByYmRfcG9v
            bF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBpZiByYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0
            YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAg
            KQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQi
            OiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            IG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCdd
            fSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21z
            ZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZh
            bHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtleSBpbiBqc29uX291dFsicG9vbHMiXToKICAg
            ICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wg
            bmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBvb2wKICAgICAgICAgICAgICAgIGlmICgKICAg
            ICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAg
            ICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bv
            b2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxl
            IGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAg
            ICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29s
            X25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgICAgIHBvb2xfZXhpc3Qg
            PSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAg
            ICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9wb29sIG5hbWUsIgogICAgICAgICAgICAgICAg
            ICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBwb29sX2V4aXN0OgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            IlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwge3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBl
            eGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihzZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIg
            KyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoOgogICAgICAg
            ICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rp
            bmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAgICAgY29udGVudHMgPSBmLnJlYWQoKQogICAg
            ICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJzdHJpcCgpCgogICAgZGVmIF9jaGVja19jb25m
            bGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJF
            aXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRhdGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBz
            aG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAgICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBv
            aW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAgICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdl
            dHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRlbGltaXRlcgogICAgICAgIHRyeToKICAgICAg
            ICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEp
            CiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikK
            CiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgog
            ICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4o
            ZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAgICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAg
            ICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNzKGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRB
            ZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAgICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJv
            cjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUUROIgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGln
            aXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQg
            bm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGludFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBp
            ZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioqMTYgLSAxOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9y
            dH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoKICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYs
            IGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMg
            aWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0
            IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBvaW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShz
            ZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gIiIsICIiLCAiIgog
            ICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJz
            cGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            Ik5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSAhPSAiWyI6CiAgICAgICAgICAgICAg
            ICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRwb2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAg
            ICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2VuZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3Ry
            X3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBbImh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJl
            c3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZvciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlcCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRf
            c3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9IE5vbmUKICAgICAgICAgICAgICAgICMgSWYg
            dmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBkaXJlY3RvcnksCiAgICAgICAgICAgICAgICAj
            IHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4gcHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFz
            aCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNTTC4KICAgICAgICAgICAgICAgIGlmIHByZWZp
            eCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RsczoKICAgICAgICAg
            ICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1GYWxzZSkKICAgICAgICAgICAgICAg
            IGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNlcnQ6CiAgICAgICAgICAgICAgICAgICAgdmVy
            aWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1l
            b3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAgICAgICAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10aW1lb3V0KQogICAgICAg
            ICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAyMDA6CiAgICAgICAgICAgICAgICAgICAgcmV0
            dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAgICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVy
            cjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vycm9yID0gZXJyCiAgICAgICAgICAgICAgICBj
            b250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRv
            IGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNw
            b25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICIiLAog
            ICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCItMSIpLAogICAgICAgICkKCiAgICBkZWYgcGFy
            c2VfY29uZmlnX2ZpbGUoc2VsZiwgY29uZmlnX2ZpbGUpOgogICAgICAgIGNvbmZpZyA9IGNvbmZp
            Z3BhcnNlci5Db25maWdQYXJzZXIoKQogICAgICAgIGNvbmZpZy5yZWFkKGNvbmZpZ19maWxlKQog
            ICAgICAgIGZvciBhcmcgaW4gbGlzdCh2YXJzKHNlbGYuX2FyZ19wYXJzZXIpKToKICAgICAgICAg
            ICAgIyBweXRob24gdHJlYXRzIGZsYWctbmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28g
            Y29udmVydGluZyBiYWNrIHRvIGZsYWctbmFtZSwKICAgICAgICAgICAgIyBzbyB3ZSBjYW4gZ2V0
            IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAgIGFyZ3VtZW50ID0gYXJn
            LnJlcGxhY2UoIl8iLCAiLSIpCiAgICAgICAgICAgIGFyZ3VtZW50VmFsdWUgPSBzdHIoZ2V0YXR0
            cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcpKQogICAgICAgICAgICBjb25maWdfdmFsdWUgPSBjb25m
            aWcuZ2V0KCJDb25maWd1cmF0aW9ucyIsIGFyZ3VtZW50LCBmYWxsYmFjaz1Ob25lKQogICAgICAg
            ICAgICAjIGdpdmUgcHJpb3JpdHkgdG8gY29tbWFuZCBsaW5lIGFyZ3VtZW50LCBpZiBjb21tYW5k
            IGxpbmUgYXJndW1lbnQgaXMgbm90IHByZXNlbnQgdXNlIGNvbmZpZy5pbmkgdmFsdWUsCiAgICAg
            ICAgICAgIGlmIChzdHIoc3lzLmFyZ3YpLmZpbmQoYXJndW1lbnQpID09IC0xKSBhbmQgKAogICAg
            ICAgICAgICAgICAgKGNvbmZpZ192YWx1ZSAhPSBOb25lKSBhbmQgKGNvbmZpZ192YWx1ZSAhPSAi
            IikKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuX19zZXRh
            dHRyX18oYXJnLCBjb25maWdfdmFsdWUpCgogICAgICAgIHJldHVybiBjb25maWcKCiAgICBkZWYg
            X19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAgICAgICAgc2VsZi5vdXRfbWFwID0ge30K
            ICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0KCkKICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3RvX3BhcnNlPWFyZ19saXN0KQogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGU6CiAgICAgICAgICAgIHNlbGYuY29uZmln
            ID0gc2VsZi5wYXJzZV9jb25maWdfZmlsZShzZWxmLl9hcmdfcGFyc2VyLmNvbmZpZ19maWxlKQog
            ICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoKQogICAgICAgIHNlbGYucnVu
            X2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRw
            dXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAgICAgc2VsZi5jZXBoX2NvbmYg
            PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNlbGYuY2VwaF9rZXlyaW5nID0g
            c2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1c2VyIG5vdCBwcm92aWRlZCwg
            Z2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxmLnJ1bl9hc191c2VyIGFuZCBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLnJ1bl9hc191c2Vy
            ID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGU6CiAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gc2VsZi5ERUZBVUxUX1JH
            V19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25mOgogICAgICAgICAgICBrd2Fy
            Z3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmluZzoKICAgICAgICAgICAgICAg
            IGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBoX2tleXJpbmd9CiAgICAgICAg
            ICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxlPXNlbGYuY2VwaF9jb25mLCAq
            Kmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIgPSByYWRvcy5S
            YWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3JlYWRfZmlsZSgpCiAgICAgICAg
            c2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToKICAgICAgICBp
            ZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAgICAgICAgICAgIHNlbGYuY2x1
            c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2Vw
            aCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rlci5nZXRfZnNpZCgpKQoKICAg
            IGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNvbik6CiAgICAgICAgY21kID0g
            anNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAgICAgICAgcmV0X3ZhbCwgY21k
            X291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFuZChjbWQsIGIiIikKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAgICAgIHByaW50KGYiQ29tbWFu
            ZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgIGYiUmV0
            dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2NtZF9vdXR9XG4iCiAgICAgICAg
            ICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0tLS0tLS0tLVxuIgogICAgICAg
            ICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJy
            b3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQnIGlzIG5vdCBlbXB0eQogICAg
            ICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24gb3V0cHV0CiAgICAgICAgaWYg
            cmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBqc29uX291dCA9IGpzb24ubG9h
            ZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cKCiAg
            ICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6CiAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9y
            dW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            IyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9ydW1fc3RhdHVzJyBjb21tYW5k
            IGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwg
            IT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICApCiAgICAgICAg
            cV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVyX25hbWUiXQogICAgICAgIHFf
            bGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21hdGNoaW5nX2xpc3QgPSBbCiAg
            ICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJdWyJtb25zIl0gaWYgbFsibmFt
            ZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQiKQogICAgICAgIHFfbGVhZGVy
            X2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAgICAgICAgIyBnZXQgdGhlIGFk
            ZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAgICAgcV9sZWFkZXJfYWRkcnZl
            YyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMiLCB7fSkuZ2V0KCJhZGRydmVj
            IiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9kZXRhaWxzWyJwdWJsaWNfYWRk
            ciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudjJfcG9ydF9l
            bmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIi
            OgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bImFkZHIiXQog
            ICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+IDEgYW5kIHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVh
            ZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAg
            c3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUg
            bm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAg
            ICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9h
            ZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBw
            b3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15
            IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRo
            b2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNp
            bnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYg
            aXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91
            bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUg
            aG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlw
            CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1h
            dCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlz
            IGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBm
            YWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3Zh
            bCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBh
            ZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28g
            dmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRw
            b2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAg
            ICAgICAgaXAgPSBbXQogICAgICAgICAgICAjIGV4YW1wbGUgb3V0cHV0IFsoPEFkZHJlc3NGYW1p
            bHkuQUZfSU5FVDogMj4sIDxTb2NrZXRLaW5kLlNPQ0tfU1RSRUFNOiAxPiwgNiwgJycsICgnOTMu
            MTg0LjIxNi4zNCcsIDgwKSksIC4uLl0KICAgICAgICAgICAgIyB3ZSBuZWVkIHRvIGdldCA5My4x
            ODQuMjE2LjM0IHNvIGl0IHdvdWxkIGJlIGlwWzBdWzRdWzBdCiAgICAgICAgICAgIGlmIGhvc3Rf
            aXBfdHlwZSA9PSAiSVB2NiI6CiAgICAgICAgICAgICAgICBpcCA9IHNvY2tldC5nZXRhZGRyaW5m
            bygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZhbWlseT1zb2NrZXQuQUZf
            SU5FVDYsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbGlmIGhvc3RfaXBfdHlwZSA9PSAiSVB2NCI6CiAgICAgICAgICAgICAgICBpcCA9IHNv
            Y2tldC5nZXRhZGRyaW5mbygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZh
            bWlseT1zb2NrZXQuQUZfSU5FVCwgcHJvdG89c29ja2V0LklQUFJPVE9fVENQCiAgICAgICAgICAg
            ICAgICApCiAgICAgICAgICAgIGRlbCBzb2NrZXQKICAgICAgICAgICAgcmV0dXJuIGlwWzBdWzRd
            WzBdCiAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQoKICAgIGRlZiBnZXRfYWN0aXZlX2FuZF9zdGFu
            ZGJ5X21ncnMoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAg
            ICAgICAgICByZXR1cm4gIiIsIHNlbGYuZHJ5X3J1bigiY2VwaCBzdGF0dXMiKQogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gc2VsZi5fYXJn
            X3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50CiAgICAgICAgc3RhbmRieV9tZ3JzID0gW10KICAg
            ICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0OgogICAgICAgICAgICBjbWRf
            anNvbiA9IHsicHJlZml4IjogInN0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1w
            dCwKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAg
            ICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICAgICAgIidtZ3Igc2VydmljZXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAg
            ICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlf
            T1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtb25pdG9yaW5nX2Vu
            ZHBvaW50ID0gKAogICAgICAgICAgICAgICAganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0
            KCJzZXJ2aWNlcyIsIHt9KS5nZXQoInByb21ldGhldXMiLCAiIikKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgImNhbid0IGZp
            bmQgbW9uaXRvcmluZ19lbmRwb2ludCwgcHJvbWV0aGV1cyBtb2R1bGUgbWlnaHQgbm90IGJlIGVu
            YWJsZWQsICIKICAgICAgICAgICAgICAgICAgICAiZW5hYmxlIHRoZSBtb2R1bGUgYnkgcnVubmlu
            ZyAnY2VwaCBtZ3IgbW9kdWxlIGVuYWJsZSBwcm9tZXRoZXVzJyIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgIyBub3cgY2hlY2sgdGhlIHN0YW5kLWJ5IG1nci1zCiAgICAgICAgICAgIHN0
            YW5kYnlfYXJyID0ganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0KCJzdGFuZGJ5cyIsIFtd
            KQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5IGluIHN0YW5kYnlfYXJyOgogICAgICAgICAg
            ICAgICAgaWYgIm5hbWUiIGluIGVhY2hfc3RhbmRieS5rZXlzKCk6CiAgICAgICAgICAgICAgICAg
            ICAgc3RhbmRieV9tZ3JzLmFwcGVuZChlYWNoX3N0YW5kYnlbIm5hbWUiXSkKICAgICAgICAgICAg
            dHJ5OgogICAgICAgICAgICAgICAgcGFyc2VkX2VuZHBvaW50ID0gdXJscGFyc2UobW9uaXRvcmlu
            Z19lbmRwb2ludCkKICAgICAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYi
            aW52YWxpZCBlbmRwb2ludDoge21vbml0b3JpbmdfZW5kcG9pbnR9IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBwYXJzZWRfZW5kcG9p
            bnQuaG9zdG5hbWUKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoK
            ICAgICAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHN0cihwYXJzZWRfZW5k
            cG9pbnQucG9ydCkKCiAgICAgICAgIyBpZiBtb25pdG9yaW5nIGVuZHBvaW50IHBvcnQgaXMgbm90
            IHNldCwgcHV0IGEgZGVmYXVsdCBtb24gcG9ydAogICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQ6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYu
            REVGQVVMVF9NT05JVE9SSU5HX0VORFBPSU5UX1BPUlQKCiAgICAgICAgIyB1c2VyIGNvdWxkIGdp
            dmUgY29tbWEgYW5kIHNwYWNlIHNlcGFyYXRlZCBpbnB1dHMgKGxpa2UgLS1tb25pdG9yaW5nLWVu
            ZHBvaW50PSI8aXAxPiwgPGlwMj4iKQogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlz
            dCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdC5yZXBsYWNlKCIsIiwgIiAiKQogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCA9IG1vbml0b3JpbmdfZW5kcG9pbnRf
            aXBfbGlzdC5zcGxpdCgpCiAgICAgICAgIyBpZiBtb25pdG9yaW5nLWVuZHBvaW50IGNvdWxkIG5v
            dCBiZSBmb3VuZCwgcmFpc2UgYW4gZXJyb3IKICAgICAgICBpZiBsZW4obW9uaXRvcmluZ19lbmRw
            b2ludF9pcF9saXN0X3NwbGl0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKCJObyAnbW9uaXRvcmluZy1lbmRwb2ludCcgZm91bmQiKQogICAgICAgICMg
            Zmlyc3QgaXAgaXMgdHJlYXRlZCBhcyB0aGUgbWFpbiBtb25pdG9yaW5nLWVuZHBvaW50CiAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9z
            cGxpdFswXQogICAgICAgICMgcmVzdCBvZiB0aGUgaXAtcyBhcmUgYWRkZWQgdG8gdGhlICdzdGFu
            ZGJ5X21ncnMnIGxpc3QKICAgICAgICBzdGFuZGJ5X21ncnMuZXh0ZW5kKG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXBfbGlzdF9zcGxpdFsxOl0pCiAgICAgICAgZmFpbGVkX2lwID0gbW9uaXRvcmluZ19l
            bmRwb2ludF9pcAoKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50ID0gIjoiLmpvaW4oCiAgICAg
            ICAgICAgIFttb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnRd
            CiAgICAgICAgKQogICAgICAgIGlwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KG1vbml0
            b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBv
            aW50X2lwID0gc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQog
            ICAgICAgICAgICApCiAgICAgICAgICAgICMgY29sbGVjdCBhbGwgdGhlICdzdGFuZC1ieScgbWdy
            IGlwcwogICAgICAgICAgICBtZ3JfaXBzID0gW10KICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRi
            eV9tZ3IgaW4gc3RhbmRieV9tZ3JzOgogICAgICAgICAgICAgICAgZmFpbGVkX2lwID0gZWFjaF9z
            dGFuZGJ5X21ncgogICAgICAgICAgICAgICAgbWdyX2lwcy5hcHBlbmQoCiAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgICAgICAg
            ICAgZWFjaF9zdGFuZGJ5X21nciwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAg
            ICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiQ29udmVyc2lvbiBvZiBob3N0OiB7ZmFpbGVkX2lwfSB0byBJUCBmYWlsZWQuICIKICAgICAg
            ICAgICAgICAgICJQbGVhc2UgZW50ZXIgdGhlIElQIGFkZHJlc3NlcyBvZiBhbGwgdGhlIGNlcGgt
            bWdycyB3aXRoIHRoZSAnLS1tb25pdG9yaW5nLWVuZHBvaW50JyBmbGFnIgogICAgICAgICAgICAp
            CgogICAgICAgIF8sIF8sIGVyciA9IHNlbGYuZW5kcG9pbnRfZGlhbChtb25pdG9yaW5nX2VuZHBv
            aW50LCBpcF90eXBlKQogICAgICAgIGlmIGVyciA9PSAiLTEiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGVycikKICAgICAgICAjIGFkZCB0aGUgdmFsaWRhdGVk
            IGFjdGl2ZSBtZ3IgSVAgaW50byB0aGUgZmlyc3QgaW5kZXgKICAgICAgICBtZ3JfaXBzLmluc2Vy
            dCgwLCBtb25pdG9yaW5nX2VuZHBvaW50X2lwKQogICAgICAgIGFsbF9tZ3JfaXBzX3N0ciA9ICIs
            Ii5qb2luKG1ncl9pcHMpCiAgICAgICAgcmV0dXJuIGFsbF9tZ3JfaXBzX3N0ciwgbW9uaXRvcmlu
            Z19lbmRwb2ludF9wb3J0CgogICAgZGVmIGNoZWNrX3VzZXJfZXhpc3Qoc2VsZiwgdXNlcik6CiAg
            ICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJhdXRoIGdldCIsICJlbnRpdHkiOiBmInt1c2Vy
            fSIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgIGlmIHJldF92YWwgIT0gMCBv
            ciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIHJldHVy
            biBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaGZzX3Byb3Zpc2lvbmVy
            X2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAi
            YWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRh
            dGE9KiIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFk
            YXRhPXtjZXBoZnNfZmlsZXN5c3RlbX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAg
            ICBkZWYgZ2V0X2NlcGhmc19ub2RlX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRp
            dHkgPSAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIKICAgICAgICBjYXBzID0gewogICAgICAgICAg
            ICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAg
            ICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAiYWxsb3cgcncgdGFn
            IGNlcGhmcyAqPSoiLAogICAgICAgICAgICAibWRzIjogImFsbG93IHJ3IiwKICAgICAgICB9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAi
            azhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVy
            LW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lz
            dGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAg
            IGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7
            ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVz
            eXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgKj17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5Cgog
            ICAgZGVmIGdldF9lbnRpdHkoCiAgICAgICAgc2VsZiwKICAgICAgICBlbnRpdHksCiAgICAgICAg
            cmJkX3Bvb2xfbmFtZSwKICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgIGs4c19j
            bHVzdGVyX25hbWUsCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgKToKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgIG9y
            IHJiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgIG9yIGFsaWFzX3JiZF9w
            b29sX25hbWUgIT0gIiIKICAgICAgICAgICAgIyBjaGVja2luZyBhbGlhc19yYmRfcG9vbF9uYW1l
            IGlzIG5vdCBlbXB0eSBhcyB0aGVyZSBtYXliZSBhIHNwZWNpYWwgY2hhcmFjdGVyIHVzZWQgb3Ro
            ZXIgdGhhbiAuIG9yIF8KICAgICAgICApOgogICAgICAgICAgICBpZiBhbGlhc19yYmRfcG9vbF9u
            YW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICAgICAicGxlYXNlIHNldCB0aGUgJy0tYWxpYXMtcmJkLWRhdGEt
            cG9vbC1uYW1lJyBmbGFnIGFzIHRoZSByYmQgZGF0YSBwb29sIG5hbWUgY29udGFpbnMgJy4nIG9y
            ICdfJyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAg
            YWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgICAgIG9yIGFs
            aWFzX3JiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgICk6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICInLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgdmFsdWUgc2hvdWxkIG5vdCBj
            b250YWluICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9
            IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXthbGlhc19yYmRfcG9vbF9uYW1lfSIKICAg
            ICAgICBlbHNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfS17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZToKICAgICAg
            ICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICByZXR1
            cm4gZW50aXR5CgogICAgZGVmIGdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNf
            Y2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJz
            ZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJi
            ZC1kYXRhLXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYg
            azhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAg
            ICAgICAgZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAg
            ICAgIGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1l
            LAogICAgICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmIHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFt
            ZX0gbmFtZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBv
            b2w9e3JiZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVm
            IGdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktcmJkLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            b3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9
            IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xfbmFtZQog
            ICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rl
            cl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFk
            b3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgazhzX2Ns
            dXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBmb3VuZCwg
            cGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAgICAgICAg
            ZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAg
            ICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAgICAgIGlm
            IHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gKAog
            ICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0gbmFt
            ZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBvb2w9e3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9k
            ZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gc2VsZi5y
            dW5fYXNfdXNlcgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLAogICAg
            ICAgICAgICAibWdyIjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwKICAgICAgICAgICAgIm9zZCI6
            IGYicHJvZmlsZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD17c2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5yb290LCBhbGxv
            dyBydyBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2ws
            IGFsbG93IHJ4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubG9n
            LCBhbGxvdyB4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuYnVj
            a2V0cy5pbmRleCIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQoKICAg
            ICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jYXBzX2FuZF9lbnRpdHkoc2Vs
            ZiwgdXNlcl9uYW1lKToKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIi
            IGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lv
            bmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3Ry
            aWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRf
            Y2VwaGZzX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBz
            ZWxmLmdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            Y3NpLXJiZC1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50
            LmNzaS1yYmQtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KCkKICAgICAg
            ICBpZiAiY2xpZW50LmNzaS1yYmQtbm9kZSIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAi
            Y2xpZW50LmNzaS1yYmQtbm9kZSIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAg
            cmV0dXJuIHNlbGYuZ2V0X3JiZF9ub2RlX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2VudGl0eSgpCgogICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgIGYibm8gdXNlciBmb3VuZCB3aXRo
            IHVzZXJfbmFtZToge3VzZXJfbmFtZX0sICIKICAgICAgICAgICAgImdldF9jYXBzX2FuZF9lbnRp
            dHkgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcihzZWxmLCB1c2VyKToKICAgICAgICAiIiIKICAgICAgICBjb21tYW5kOiBjZXBo
            IGF1dGggZ2V0LW9yLWNyZWF0ZSBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciBtb24gJ2Fs
            bG93IHInIG1nciAnYWxsb3cgcncnIG9zZCAnYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            JwogICAgICAgICIiIgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2Vu
            dGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1
            dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAg
            ICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBj
            YXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0K
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4g
            KAogICAgICAgICAgICAgICAgc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgICAgICJjZXBo
            ICIKICAgICAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAg
            ICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50aXR5Il0KICAg
            ICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21k
            X2pzb25bImNhcHMiXSkKICAgICAgICAgICAgICAgICksCiAgICAgICAgICAgICAgICAiIiwKICAg
            ICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAg
            dXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJf
            a2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXksIGYie2VudGl0eS5zcGxpdCgn
            LicsIDEpWzFdfSIKICAgICAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4nLDEpWzFdIHRvIHJlbmFt
            ZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1ub2RlCgogICAgICAgIHJl
            dF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRf
            anNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAg
            ICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGdl
            dC1vci1jcmVhdGUge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYi
            RXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJ
            U1R9IgogICAgICAgICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0p
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4n
            LDEpWzFdIHRvIHJlbmFtZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1u
            b2RlCgogICAgZGVmIGdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoc2VsZik6CiAgICAgICAg
            Y21kX2pzb24gPSB7InByZWZpeCI6ICJmcyBscyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            aWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlf
            cnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsIHJlcG9ydCBhbiBlcnJvcgog
            ICAgICAgIGlmIHJldF92YWwgIT0gMDoKICAgICAgICAgICAgIyBpZiBmcyBhbmQgZGF0YV9wb29s
            IGFyZ3VtZW50cyBhcmUgbm90IHNldCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSA9
            PSAiIgogICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9v
            bF9uYW1lID09ICIiCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICByZXR1cm4KICAgICAg
            ICAgICAgIyBpZiB1c2VyIGhhcyBwcm92aWRlZCBhbnkgb2YgdGhlCiAgICAgICAgICAgICMgJy0t
            Y2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgb3IgJy0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lJyBhcmd1
            bWVudHMsCiAgICAgICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIGFzIHdlIGFyZSB1bmFibGUg
            dG8gdmVyaWZ5IHRoZSBhcmdzCiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIidmcyBscycgY2VwaCBjYWxsIGZhaWxlZCB3aXRoIGVy
            cm9yOiB7ZXJyX21zZ30iCiAgICAgICAgICAgICkKCiAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQg
            PSB7fQogICAgICAgICMgaWYgJy0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgYXJndW1lbnQgaXMg
            cHJvdmlkZWQsCiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBwcm92aWRlZCBmaWxlc3lzdGVt
            LW5hbWUgZXhpc3RzIG9yIG5vdAogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZToKICAgICAgICAgICAgIyBnZXQgdGhlIG1hdGNoaW5nIGxpc3QKICAgICAg
            ICAgICAgbWF0Y2hpbmdfanNvbl9vdXRfbGlzdCA9IFsKICAgICAgICAgICAgICAgIG1hdGNoZWQK
            ICAgICAgICAgICAgICAgIGZvciBtYXRjaGVkIGluIGpzb25fb3V0CiAgICAgICAgICAgICAgICBp
            ZiBzdHIobWF0Y2hlZFsibmFtZSJdKSA9PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWUKICAgICAgICAgICAgXQogICAgICAgICAgICAjIHVuYWJsZSB0byBmaW5kIGEgbWF0
            Y2hpbmcgZnMtbmFtZSwgcmFpc2UgYW4gZXJyb3IKICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5n
            X2pzb25fb3V0X2xpc3QpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiRmlsZXN5c3RlbSBwcm92aWRlZCwg
            J3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWV9JywgIgogICAgICAgICAg
            ICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBmcy1saXN0OiB7W3N0cih4WyduYW1lJ10p
            IGZvciB4IGluIGpzb25fb3V0XX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1hdGNo
            aW5nX2pzb25fb3V0ID0gbWF0Y2hpbmdfanNvbl9vdXRfbGlzdFswXQogICAgICAgICMgaWYgY2Vw
            aGZzIGZpbGVzeXN0ZW0gbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgIyB0cnkgdG8gZ2V0
            IGEgZGVmYXVsdCBmcyBuYW1lIGJ5IGRvaW5nIHRoZSBmb2xsb3dpbmcKICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAjIGEuIGNoZWNrIGlmIHRoZXJlIGlzIG9ubHkgb25lIGZpbGVzeXN0ZW0gaXMg
            cHJlc2VudAogICAgICAgICAgICBpZiBsZW4oanNvbl9vdXQpID09IDE6CiAgICAgICAgICAgICAg
            ICBtYXRjaGluZ19qc29uX291dCA9IGpzb25fb3V0WzBdCiAgICAgICAgICAgICMgYi4gb3IgZWxz
            ZSwgY2hlY2sgaWYgZGF0YV9wb29sIG5hbWUgaXMgcHJvdmlkZWQKICAgICAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICMg
            YW5kIGlmIHByZXNlbnQsIGNoZWNrIHdoZXRoZXIgdGhlcmUgZXhpc3RzIGEgZnMgd2hpY2ggaGFz
            IHRoZSBkYXRhX3Bvb2wKICAgICAgICAgICAgICAgIGZvciBlYWNoSiBpbiBqc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSBpbiBlYWNoSlsiZGF0YV9wb29scyJdOgogICAgICAgICAgICAgICAgICAgICAgICBtYXRjaGlu
            Z19qc29uX291dCA9IGVhY2hKCiAgICAgICAgICAgICAgICAgICAgICAgIGJyZWFrCiAgICAgICAg
            ICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIG1hdGNoaW5nIGZzIGV4aXN0cywgdGhhdCBtZWFucyBw
            cm92aWRlZCBkYXRhX3Bvb2wgbmFtZSBpcyBpbnZhbGlkCiAgICAgICAgICAgICAgICBpZiBub3Qg
            bWF0Y2hpbmdfanNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhX3Bv
            b2wgbmFtZSwge3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfSwiCiAgICAg
            ICAgICAgICAgICAgICAgICAgICIgZG9lcyBub3QgZXhpc3RzIgogICAgICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgIyBjLiBpZiBub3RoaW5nIGlzIHNldCBhbmQgY291bGRuJ3QgZmluZCBh
            IGRlZmF1bHQsCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGp1c3QgcmV0dXJu
            IHNpbGVudGx5CiAgICAgICAgICAgICAgICByZXR1cm4KCiAgICAgICAgaWYgbWF0Y2hpbmdfanNv
            bl9vdXQ6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSA9IHN0cihtYXRjaGluZ19qc29uX291dFsibmFtZSJdKQogICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19tZXRhZGF0YV9wb29sX25hbWUgPSBzdHIoCiAgICAgICAgICAgICAgICBt
            YXRjaGluZ19qc29uX291dFsibWV0YWRhdGFfcG9vbCJdCiAgICAgICAgICAgICkKCiAgICAgICAg
            aWYgaXNpbnN0YW5jZShtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdLCBsaXN0KToKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBoYXMgYWxyZWFkeSBwcm92aWRlZCBkYXRhLXBvb2wtbmFt
            ZSwKICAgICAgICAgICAgIyB0aHJvdWdoIC0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lCiAgICAgICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAg
            ICAgICAgIyBpZiB0aGUgcHJvdmlkZWQgbmFtZSBpcyBub3QgbWF0Y2hpbmcgd2l0aCB0aGUgb25l
            IGluIHRoZSBsaXN0CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgICAgICAgICBu
            b3QgaW4gbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXQogICAgICAgICAgICAgICAgKToK
            ICAgICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lOiAne3NlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfScsICIKICAgICAgICAgICAgICAgICAgICAg
            ICAgImRvZXNuJ3QgbWF0Y2ggZnJvbSB0aGUgZGF0YS1wb29scyBsaXN0OiAiCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGYie1tzdHIoeCkgZm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFf
            cG9vbHMnXV19IgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiBkYXRhX3Bv
            b2wgbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgICAgICMgdGhlbiB0cnkgdG8gZmluZCBh
            IGRlZmF1bHQgZGF0YSBwb29sIG5hbWUKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAg
            ICMgaWYgbm8gZGF0YV9wb29scyBleGlzdCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgICAg
            ICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPT0gMDoKICAgICAgICAg
            ICAgICAgICAgICByZXR1cm4KICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25f
            b3V0WyJkYXRhX3Bvb2xzIl1bMF0KICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiB0
            aGVyZSBhcmUgbW9yZSB0aGFuIG9uZSAnZGF0YV9wb29scycgZXhpc3QsCiAgICAgICAgICAgICMg
            dGhlbiB3YXJuIHRoZSB1c2VyIHRoYXQgd2UgYXJlIHVzaW5nIHRoZSBzZWxlY3RlZCBuYW1lCiAg
            ICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdKSA+IDE6CiAg
            ICAgICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgICAgICAiV0FSTklORzogTXVsdGlw
            bGUgZGF0YSBwb29scyBkZXRlY3RlZDogIgogICAgICAgICAgICAgICAgICAgIGYie1tzdHIoeCkg
            Zm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFfcG9vbHMnXV19XG4iCiAgICAgICAgICAg
            ICAgICAgICAgZiJVc2luZyB0aGUgZGF0YS1wb29sOiAne3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lfSdcbiIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2No
            ZWNrZXJLZXkoc2VsZiwgdXNlcik6CiAgICAgICAgY2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fw
            c19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21kX2pzb24gPSB7CiAgICAgICAgICAgICJwcmVm
            aXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAgICAgICAgICAgImVudGl0eSI6IGVudGl0eSwK
            ICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBjYXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMo
            KSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAg
            ICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oCiAgICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAg
            ICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAg
            ICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAgICAgICApCiAgICAg
            ICAgIyBjaGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYu
            Y2hlY2tfdXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAg
            ICAgICAgIHJldHVybiB1c2VyX2tleQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21z
            ZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVy
            ZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHtzZWxmLnJ1bl9h
            c191c2VyfScgY29tbWFuZCBmYWlsZWRcbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJf
            bXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAg
            ICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0pCgogICAgZGVmIGdl
            dF9jZXBoX2Rhc2hib2FyZF9saW5rKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgi
            OiAibWdyIHNlcnZpY2VzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiAr
            IGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1
            bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25f
            b3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIGlmICJkYXNoYm9hcmQi
            IG5vdCBpbiBqc29uX291dDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICByZXR1cm4g
            anNvbl9vdXRbImRhc2hib2FyZCJdCgogICAgZGVmIGNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIo
            c2VsZik6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAg
            ICAgICAgICJ1c2VyIiwKICAgICAgICAgICAgImNyZWF0ZSIsCiAgICAgICAgICAgICItLXVpZCIs
            CiAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAg
            ICAgICAgICItLWRpc3BsYXktbmFtZSIsCiAgICAgICAgICAgICJSb29rIFJHVyBBZG1pbiBPcHMg
            dXNlciIsCiAgICAgICAgICAgICItLWNhcHMiLAogICAgICAgICAgICAiYnVja2V0cz0qO3VzZXJz
            PSo7dXNhZ2U9cmVhZDttZXRhZGF0YT1yZWFkO3pvbmU9cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgIiAiLmpvaW4oY21kKSkKICAgICAgICB0cnk6CiAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBhbHJlYWR5IGV4aXN0cywgd2UganVzdCBxdWVyeSBpdAog
            ICAgICAgICAgICBpZiBleGVjRXJyLnJldHVybmNvZGUgPT0gZXJybm8uRUVYSVNUOgogICAgICAg
            ICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAg
            ICAgICAgICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICAgICAgICAgImluZm8iLAogICAg
            ICAgICAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5FWFRFUk5B
            TF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUs
            CiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICAgICAg
            ICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICAgICAgICAgIF0KICAgICAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVy
            cj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxs
            ZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgICAgICBlcnJfbXNnID0g
            KAogICAgICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2Nt
            ZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShl
            cnJfbXNnKQogICAgICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0x
            IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4
            ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0
            dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0
            dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5m
            bz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19j
            YXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAg
            ICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAg
            ICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Jl
            YWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUi
            LAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQog
            ICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQo
            Y21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNh
            bGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZl
            cnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRh
            dGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2Fw
            czogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4
            ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2Fw
            X3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJf
            bXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7
            Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJD
            b2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAg
            ICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNv
            bm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAg
            IGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0
            cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9y
            dGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0
            cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3twb29sX25hbWV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYgaW5pdF9yYmRfcG9vbChzZWxmLCBy
            YmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlS
            YWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9w
            ZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAg
            ICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkKCiAgICBkZWYgdmFsaWRhdGVfcmFk
            b3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmJkX21l
            dGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQogICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYgcmFkb3NfbmFt
            ZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRvc19uYW1lc3Bh
            Y2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNl
            LCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRhaW5zIHVwcGVy
            IGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAgICAg
            IGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICBp
            ZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3BhY2UpIGlzIEZh
            bHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFj
            ZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9vbCAne3JiZF9w
            b29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRlX3N1YnZvbHVt
            ZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUpOgog
            ICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAg
            ICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIsCiAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19v
            dXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9j
            ZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAg
            ICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAgICJzdWJ2b2x1
            bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgICAgIGNlcGhm
            c19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3Mu
            Y2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmInN1YnZvbHVt
            ZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3JlYXRlZCIKICAg
            ICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBzZWxmLCBzdWJ2
            b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBwaW5fc2V0dGlu
            ZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAgICAgICAgICAg
            ImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgInBpbiIsCiAg
            ICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9n
            cm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0aW5nLAogICAg
            ICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1
            dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3Mu
            Q2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZvbHVtZV9ncm91
            cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRf
            cmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNzX2tleSA9IHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAgICAgIHNlY3Jl
            dF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0KICAg
            ICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAg
            IGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIvYWRtaW4vaW5m
            bz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmVxdWVzdF91
            cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToKICAgICAgICAg
            ICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJsLAogICAgICAg
            ICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3X2VuZHBvaW50
            KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkKICAgICAgICBl
            eGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90IGFibGUgdG8g
            Y2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAgIGlmIHIxIGlz
            IE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAgICAgICAgKQog
            ICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICAgICAi
            LTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJzdG9yYWdlX2Jh
            Y2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X2VuZHBv
            aW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXIn
            IGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91
            dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVt
            bXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJndyBlbmRwb2lu
            dCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9IHNlbGYuX2lu
            dmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cg
            ZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBv
            aW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd190
            bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2VsZi5lbmRwb2lu
            dF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAgIGlmIGVyciAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBpZiB0aGUgcmd3
            IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMgb25seSBjaGVj
            ayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2FwX3N1cHBvcnRl
            ZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAgICByZ3dfZnNp
            ZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAgICAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAgICAgICAgICBp
            ZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAg
            ICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7c2VsZi5fYXJn
            X3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlkYXRpbmcgYnkg
            Y2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFuZCB2YWxpZGF0
            aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUgY2VwaCBjbHVz
            dGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9vbCBleGlzdAog
            ICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFzc2VkIGVsc2Ug
            aXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0aGVzZSBkZWZh
            dWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggIT0g
            ImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBbCiAgICAgICAg
            ICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3Lm1ldGEiLAog
            ICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAgICAgICAgZiJ7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAgICAgICAgICBd
            CiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bvb2xzX3RvX3Zh
            bGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0cyhf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2wsICd7X3Jnd19w
            b29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4gIiIKCiAgICBk
            ZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1l
            LCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0ZV9jb25maWcg
            IT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWlu
            IiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAgICAgICAgICAg
            ImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVfY29uZmlnLAog
            ICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAgICAgICAgXQog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgIGVycl9t
            c2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtj
            bWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNv
            ZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAg
            ICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAgZGVmIGNvbnZl
            cnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAgICByZXR1cm4g
            dmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5
            X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAgICgKICAgICAg
            ICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIKICAgICAgICAg
            ICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAgIG9yIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIiCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICAgICAg
            ICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxhZ3MgLS10b3Bv
            bG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10b3BvbG9neS1m
            YWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfdG9w
            b2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6CiAgICAgICAg
            aWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJv
            dmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIKICAgICAgICAg
            ICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9JywiCiAgICAg
            ICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3BvbG9neV9wb29s
            cyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAgICAgICAgICkK
            ICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYs
            IHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bv
            b2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgogICAgZGVmIGlu
            aXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAg
            Zm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLmluaXRfcmJk
            X3Bvb2wocG9vbCkKCiAgICAjIHRoaXMgd2lsbCByZXR1cm4gdGhlIGZpbmFsIGFyZ3MgdGhhdCBz
            Y3JpcHQgdXNlcyB0byBwcm9jZXNzCiAgICAjIHRoZSBwcmlvcml0eSB0byBzZXQgYSBwYXJ0aWN1
            bGFyIHZhbHVlIGlzLAogICAgIyBjb21tYW5kLWxpbmUtYXJncyA+IGNvbmZpZy5pbmkgZmlsZSB2
            YWx1ZXMgPiBkZWZhdWx0IHZhbHVlcwogICAgZGVmIGdldEZpbmFsVXNlZEFyZ3Moc2VsZik6CiAg
            ICAgICAgYXJndW1lbnQgPSBmIltDb25maWd1cmF0aW9uc11cbiIKICAgICAgICBmb3IgYXJnIGlu
            IHZhcnMoc2VsZi5fYXJnX3BhcnNlcik6CiAgICAgICAgICAgIGlmIHN0cihnZXRhdHRyKHNlbGYu
            X2FyZ19wYXJzZXIsIGFyZykpOgogICAgICAgICAgICAgICAgIyBweXRob24gdHJlYXRzIGZsYWct
            bmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28gY29udmVydGluZyBiYWNrIHRvIGZsYWct
            bmFtZSwKICAgICAgICAgICAgICAgICMgc28gd2UgY2FuIGdldCB0aG9zZSB2YWx1ZXMgZnJvbSBj
            b25maWcgZmlsZQogICAgICAgICAgICAgICAgYXJnVmFsdWUgPSBhcmcucmVwbGFjZSgiXyIsICIt
            IikKICAgICAgICAgICAgICAgICMgZG8gbm90IGFkZCB0aGUgY29uZmlnLWZpbGUgZmxhZyBhbmQg
            YWxzbyBub3QgYWRkIHRoZSBib29sZWFuIGZsYWdzIHdoaWNoIGFyZSBzZXQgdG8gRmFsc2UKICAg
            ICAgICAgICAgICAgICMgYmVjYXVzZSBjb25maWcuaW5pIGZpbGUgdHJlYXRzIGJvb2xlYW4gZmxh
            Z3MgYXMgVHJ1ZSBhbHdheXMKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBhcmdWYWx1ZSAhPSAiY29uZmlnLWZpbGUiCiAgICAgICAgICAgICAgICAgICAgYW5kIGdldGF0
            dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSAhPSBGYWxzZQogICAgICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgICAgICBhcmd1bWVudCArPSBmInthcmdWYWx1ZX0gPSB7c3RyKGdldGF0dHIo
            c2VsZi5fYXJnX3BhcnNlciwgYXJnKSl9XG4iCiAgICAgICAgcmV0dXJuIGFyZ3VtZW50CgogICAg
            ZGVmIF9nZW5fb3V0cHV0X21hcChzZWxmKToKICAgICAgICBpZiBzZWxmLm91dF9tYXA6CiAgICAg
            ICAgICAgIHJldHVybgogICAgICAgICMgc3VwcG9ydCBsZWdhY3kgZmxhZyB3aXRoIHVwZ3JhZGVz
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWU6CiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuY2x1c3Rl
            cl9uYW1lCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gKAogICAg
            ICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUubG93ZXIoKQogICAgICAg
            ICkgICMgYWx3YXlzIGNvbnZlcnQgY2x1c3RlciBuYW1lIHRvIGxvd2VyY2FzZSBjaGFyYWN0ZXJz
            CiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZSkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2woc2VsZi5fYXJnX3BhcnNlci5y
            YmRfZGF0YV9wb29sX25hbWUpCiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uo
            KQogICAgICAgIHNlbGYuX2V4Y2x1ZGVkX2tleXMuYWRkKCJLOFNfQ0xVU1RFUl9OQU1FIikKICAg
            ICAgICBzZWxmLmdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoKQogICAgICAgICMgZG91Ymxl
            IHN0cmluZyBuZWVkZWQgZm9yIHVwc3RyZWFtIGV4cG9ydHMgb2YgZmxhZ3MKICAgICAgICBzZWxm
            Lm91dF9tYXBbIkFSR1MiXSA9IGYnIntzZWxmLmdldEZpbmFsVXNlZEFyZ3MoKX0iJwogICAgICAg
            IHNlbGYub3V0X21hcFsiTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFyc2VyLm5hbWVzcGFjZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiSzhTX0NMVVNURVJfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0ZT
            SUQiXSA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSTkFNRSJdID0gc2VsZi5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0gPSBzZWxmLmdldF9jZXBoX2V4dGVybmFsX21vbl9k
            YXRhKCkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSA9
            IHNlbGYuY3JlYXRlX2NoZWNrZXJLZXkoCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FS
            RF9MSU5LIl0gPSBzZWxmLmdldF9jZXBoX2Rhc2hib2FyZF9saW5rKCkKICAgICAgICAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICBz
            ZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxm
            LmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1ub2RlIikKICAgICAg
            ICAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVu
            dC5jc2ktcmJkLXByb3Zpc2lvbmVyIikKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09M
            X05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDRVBIRlNfTUVUQURBVEFfUE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAg
            IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRhX3Bvb2xfbmFtZQogICAgICAgICkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVU
            SF9QRVJNSVNTSU9OIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3Rl
            ZF9hdXRoX3Blcm1pc3Npb24KICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQURPU19O
            QU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXSA9IHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1l
            X2dyb3VwCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0gPSAi
            IgogICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXSA9
            ICIiCiAgICAgICAgIyBjcmVhdGUgQ2VwaEZTIG5vZGUgYW5kIHByb3Zpc2lvbmVyIGtleXJpbmcg
            b25seSB3aGVuIE1EUyBleGlzdHMKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19O
            QU1FIl0gYW5kIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAo
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwK
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1F
            Il0sCiAgICAgICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGll
            bnQuY3NpLWNlcGhmcy1ub2RlIikKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLWNl
            cGhmcy1wcm92aXNpb25lciIpCiAgICAgICAgICAgICMgY3JlYXRlIHRoZSBkZWZhdWx0ICJjc2ki
            IHN1YnZvbHVtZWdyb3VwCiAgICAgICAgICAgIHNlbGYuZ2V0X29yX2NyZWF0ZV9zdWJ2b2x1bWVf
            Z3JvdXAoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmls
            ZXN5c3RlbV9uYW1lCiAgICAgICAgICAgICkKICAgICAgICAgICAgIyBwaW4gdGhlIGRlZmF1bHQg
            ImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5waW5fc3Vidm9sdW1lKAogICAg
            ICAgICAgICAgICAgImNzaSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSwgImRpc3RyaWJ1dGVkIiwgIjEiCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXToKICAgICAgICAgICAgICAgIHNlbGYuZ2V0X29yX2Ny
            ZWF0ZV9zdWJ2b2x1bWVfZ3JvdXAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAg
            c2VsZi5waW5fc3Vidm9sdW1lKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            c3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAiZGlzdHJpYnV0ZWQiLAog
            ICAgICAgICAgICAgICAgICAgICIxIiwKICAgICAgICAgICAgICAgICkKICAgICAgICBzZWxmLm91
            dF9tYXBbIlJHV19UTFNfQ0VSVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJ
            TkdfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBP
            SU5UX1BPUlQiXSA9ICIiCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIuc2tpcF9tb25p
            dG9yaW5nX2VuZHBvaW50OgogICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIk1PTklUT1JJTkdfRU5EUE9JTlQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsi
            TU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICkgPSBzZWxmLmdldF9hY3Rp
            dmVfYW5kX3N0YW5kYnlfbWdycygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1F
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoKQogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzCiAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJd
            ID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWlu
            X2xhYmVsCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lf
            ZmFpbHVyZV9kb21haW5fdmFsdWVzCiAgICAgICAgKQogICAgICAgIGlmICgKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICBhbmQgc2Vs
            Zi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAgICAg
            ICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMg
            IT0gIiIKICAgICAgICApOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3ZhbHVl
            cygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXko
            c2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKSwKICAgICAgICAgICAgICAgIHNlbGYuY29u
            dmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXQogICAgICAgICAgICAgICAg
            KSwKICAgICAgICAgICAgKQogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3JiZF9w
            b29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJy
            YXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIHNlbGYuaW5pdF90b3BvbG9neV9yYmRfcG9vbHMoCiAgICAgICAgICAgICAgICBzZWxmLmNv
            bnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSkKICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHNlbGYucmFpc2Vf
            ZXhjZXB0aW9uX2lmX2FueV90b3BvbG9neV9mbGFnX2lzX21pc3NpbmcoKQoKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJHV19QT09MX1BSRUZJWCJdID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXgKICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gIiIKICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludDoKICAgICAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICAgICAgc2VsZi5jcmVhdGVfcmd3X2FkbWluX29w
            c191c2VyKCkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAh
            PSAiIgogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZh
            bGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUsICJyZWFsbSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwgInpv
            bmVncm91cCIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgZXJyID0g
            c2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsICJ6b25lIgogICAgICAgICAgICAgICAgICAgICkK
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgICkgb3Ig
            KAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgIT0g
            IiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3Vw
            X25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgICgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19B
            RE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIGluZm9f
            Y2FwX3N1cHBvcnRlZCwKICAgICAgICAgICAgICAgICAgICAgICAgZXJyLAogICAgICAgICAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICAg
            ICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50KGluZm9fY2FwX3N1cHBvcnRl
            ZCkKICAgICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9w
            YXRoOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9p
            bnRfdGxzX2NlcnQoKQogICAgICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciwgc2V0IHRoZSBSR1dfRU5EUE9JTlQKICAgICAgICAg
            ICAgICAgICAgICBpZiBlcnIgIT0gIi0xIjoKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJSR1dfRU5EUE9JTlQiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIGVyciA9ICJQbGVhc2UgcHJv
            dmlkZSBhbGwgdGhlIFJHVyBtdWx0aXNpdGUgcGFyYW1ldGVycyBvciBub25lIG9mIHRoZW0iCiAg
            ICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnIpCgogICAgZGVmIGdlbl9zaGVs
            bF9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAgICAgIHNoT3V0
            SU8gPSBTdHJpbmdJTygpCiAgICAgICAgZm9yIGssIHYgaW4gc2VsZi5vdXRfbWFwLml0ZW1zKCk6
            CiAgICAgICAgICAgIGlmIHYgYW5kIGsgbm90IGluIHNlbGYuX2V4Y2x1ZGVkX2tleXM6CiAgICAg
            ICAgICAgICAgICBzaE91dElPLndyaXRlKGYiZXhwb3J0IHtrfT17dn17TElORVNFUH0iKQogICAg
            ICAgIHNoT3V0ID0gc2hPdXRJTy5nZXR2YWx1ZSgpCiAgICAgICAgc2hPdXRJTy5jbG9zZSgpCiAg
            ICAgICAgcmV0dXJuIHNoT3V0CgogICAgZGVmIGdlbl9qc29uX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICBqc29uX291dCA9IFsKICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgIm5hbWUiOiAiZXh0ZXJuYWwtY2x1c3Rlci11c2VyLWNvbW1hbmQi
            LAogICAgICAgICAgICAgICAgImtpbmQiOiAiQ29uZmlnTWFwIiwKICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICJhcmdzIjogc2VsZi5vdXRfbWFwWyJBUkdTIl0s
            CiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAg
            ICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uLWVuZHBvaW50cyIsCiAgICAgICAgICAgICAg
            ICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAg
            ICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05f
            REFUQSJdLAogICAgICAgICAgICAgICAgICAgICJtYXhNb25JZCI6ICIwIiwKICAgICAgICAgICAg
            ICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9
            LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwK
            ICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4tc2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAg
            ICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJ
            RCJdLAogICAgICAgICAgICAgICAgICAgICJtb24tc2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9yLWNyZWRzIiwKICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxf
            VVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAg
            XQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBh
            ZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENsdXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAg
            ICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBz
            ZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9pbnQiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdFbmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAgICAgICAiTW9uaXRvcmluZ1Bv
            cnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgIyBp
            ZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1y
            YmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lf
            UkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJLZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAg
            ICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JC
            RF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJk
            LXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNS
            RVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJJRCI6
            IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQog
            ICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQn
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNy
            ZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSdd
            fSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAg
            ICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lP
            TkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAg
            ICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfQ0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMs
            IHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhmcy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAg
            ICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewog
            ICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBI
            RlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2Vj
            cmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAg
            ICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAg
            ICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRf
            TElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNl
            Y3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElO
            SyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxp
            bmsiLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0Un
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJBRE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAg
            ICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAg
            ICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0K
            ICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAg
            ICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJyYWRv
            c05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAg
            ICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09M
            X05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92
            aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9y
            YWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5v
            dXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            IGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xf
            TkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6
            IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXIt
            c2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9T
            RUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9t
            YXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICAgICAgfQogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAog
            ICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2Vw
            aC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1l
            IjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUn
            XX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgICAgICkKCiAgICAg
            ICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwn
            LCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhl
            biBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VDbGFzcwogICAgICAgIGlmICgKICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91
            dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTIl0KICAgICAgICApOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9wb2xvZ3kiLAogICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYu
            b3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RP
            TUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAgICAgICAgXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZhbHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAg
            ICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAg
            ICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhw
            YW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycgU3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2Vs
            Zi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwK
            ICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5v
            dXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6
            IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAg
            ICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtz
            ZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAg
            ICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5k
            LXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBz
            ZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ci
            LAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJlbmRwb2ludCI6IHNl
            bGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            UHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZ
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJzZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2VwaC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0
            LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJj
            ZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImNl
            cnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdLAogICAgICAgICAgICAgICAgICAgIH0s
            CiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVt
            cHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRlZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25z
            KHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMt
            bm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAg
            ICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwKICAgICAgICAgICAgImNsaWVudC5jc2ktcmJk
            LXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAg
            ICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNlciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNl
            ciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVzZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2Vy
            KQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgogICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNl
            cl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBnZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBn
            cmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAgICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRo
            Y2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBhZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBw
            b29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAgICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBu
            YW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIgdXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcg
            Y29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdldCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMg
            cmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAg
            ICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0oLio/KVwucmd3XC5tZXRhIgogICAgICAgICAg
            ICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVybiwgY2FwcykKICAgICAgICAgICAgICAgIGlm
            IG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wgbmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAg
            ICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2Vy
            KToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhlIGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAog
            ICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNl
            cn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNn
            ID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBwcmludChmInVzZXIge3Vz
            ZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIpCiAgICAgICAgICAgIHJldHVybgogICAgICAg
            IGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsiY2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3
            X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2VyLCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAg
            ICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNh
            cF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIsICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQog
            ICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlzOgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVl
            cyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBuZXdfY2FwX3Zh
            bHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBjdXJfY2FwX3Blcm1f
            bGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVl
            cy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBdCiAgICAgICAgICAgIG5ld19j
            YXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAgICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19j
            YXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJpcCgpCiAgICAgICAgICAgIF0KICAgICAgICAg
            ICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUg
            b3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2Nh
            cF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxpbWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1
            c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCByZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3Qg
            YW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIKICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0g
            W10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5hcHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9w
            ZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9saXN0XQogICAgICAgICAgICBleGlzdGluZ19j
            YXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19jYXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhp
            c3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAgICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXAp
            CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChleGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAg
            ICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1dGggY2FwcyIsCiAgICAg
            ICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAgICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAg
            ICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQogICAgICAgIHJldF92YWwsIGpzb25fb3V0
            LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBp
            ZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNhcHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5c
            biBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVk
            IHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoKICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAg
            IGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFk
            ZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBl
            bGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJh
            dGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19w
            YXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNl
            bGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0
            OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBzZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0
            ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0gb3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2Rl
            PSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAgICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRf
            b3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3NlKCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19tYWluX18iOgogICAgcmpPYmogPSBSYWRvc0pT
            T04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1haW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAgIHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDog
            e2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAgZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAg
            ICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9z
            RXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hpbGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0
            YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAgICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:35e88716ace3576683a9c4aa686df5f7a8b5bb6ee7a144fd2d87e3c9f5bf3961
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        version: 4.17.13-rhodf
      entries:
      - name: rook-ceph-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: rook-ceph-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: rook-ceph-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: rook-ceph-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: rook-ceph-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: rook-ceph-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: rook-ceph-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: rook-ceph-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: rook-ceph-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: rook-ceph-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: rook-ceph-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: rook-ceph-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: rook-ceph-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: rook-ceph-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: rook-ceph-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9VHJ1ZSwgICMgZGVmYXVsdCBpcyBUcnVlIHRvIGVuYWJsZSB2MiBtb24g
            cG9ydCBpbiBkb3duc3RyZWFtCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9ydCgz
            MzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQLmFk
            ZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAgICAg
            ICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24iLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJhc2gp
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3ZpZGVk
            IGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmlsZXN5
            c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAg
            ICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZzIG1l
            dGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50
            KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRl
            cyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9vbC1u
            YW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2Us
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRhIHBv
            b2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBpbiB0
            aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAgKQog
            ICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1lbmRw
            b2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+OjxQ
            T1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAg
            IHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50
            IFRMUyBjZXJ0aWZpY2F0ZSAob3IgaW50ZXJtZWRpYXRlIHNpZ25pbmcgY2VydGlmaWNhdGUpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAg
            IGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBUTFMgY2VydGlmaWNhdGlv
            biB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0ZSBpcyBwcm92aWRlZCAo
            Tk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRl
            ZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJD
            ZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMgKGNvbW1hIHNlcGFyYXRl
            ZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBvciBgPEZRRE4+YCkgZW50
            cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAgICkKICAgICAgICBvdXRw
            dXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25pdG9yaW5nLWVuZHBvaW50
            LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgcG9y
            dCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFs
            c2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRv
            IG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0aGUgQ2VwaCBjbHVzdGVy
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0
            aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29sIiwKICAgICAgICApCiAg
            ICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tZHJ5LXJ1biIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1
            ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBleGVjdXRlZCBjb21tYW5k
            cyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAg
            IGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxw
            PSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFtZXNwYWNlcywgdXNlZCBm
            b3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSAoc2hv
            dWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwKICAgICAgICAgICAgZGVm
            YXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9InBy
            b3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAogICAgICAgICkKICAgICAg
            ICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0tbmFt
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3LXJlYWxtIiwKICAg
            ICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0t
            cmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3
            LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhl
            IG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1wb29scyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25zdHJhaW5lZCByYmQgcG9v
            bHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAg
            ICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJrOHMg
            Y2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9uZSwgcmFjaywgb3IgaG9z
            dCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBjZXBoIGRvbWFpbiIsCiAg
            ICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICIt
            LXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBh
            cmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gdmFsdWVzIGNvcnJl
            c3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3BvbG9neS1wb29sc2AgbGlz
            dCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJnUC5hZGRfYXJndW1lbnRf
            Z3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAg
            ICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iVXBncmFkZXMgdGhlIGNl
            cGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIp
            IGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGggbmV3IHBlcm1pc3Npb25z
            IG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9sZGVyIHBlcm1pc3Npb24g
            d2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhv
            bjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3Jh
            ZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3NpIHVzZXJzKG5vbi1yZXN0
            cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNlcnMoRm9yIGV4YW1wbGU6
            IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMpLCB1
            c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24gZmxhZyBuZWVk
            IHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJtYW5kYXRvcnkgZmxhZ3M6
            ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1lIGFuZCAtLXJ1bi1hcy11
            c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsgImluIGNhc2Ugb2YgY2Vw
            aGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUgZmxh
            ZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5nIGl0IHdpbGwgYmUgbWFu
            ZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBo
            L2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3JhZGUgLS1yYmQtZGF0
            YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlICAt
            LXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3JhZ2UtcmVwbGljYXBvb2xg
            IgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3RyaWN0ZWQgdXNlciBjYW5u
            b3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVwZ3JhZGluZy4iCiAgICAg
            ICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2VkIHRvIGFwcGVuZCBuZXcg
            cGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2VkIGZvciBjaGFuZ2luZyB1
            c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBsZSB5b3Ugc2hvdWxkbid0
            IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAgICAgICAgKQoKICAgICAg
            ICAjIEFkZCBjb21tYW5kLWxpbmUgYXJndW1lbnRzCiAgICAgICAgY29uZmlnX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoImNvbmZpZyIpCiAgICAgICAgY29uZmlnX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY29uZmlnLWZpbGUiLAogICAgICAgICAgICB0eXBlPXN0
            ciwKICAgICAgICAgICAgaGVscD0iUGF0aCB0byB0aGUgY29uZmlndXJhdGlvbiBmaWxlLCBQcmlv
            cml0eTogY29tbWFuZC1saW5lLWFyZ3MgPiBjb25maWcuaW5pIGZpbGUgdmFsdWVzID4gZGVmYXVs
            dCB2YWx1ZXMiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICApCgogICAgICAgIGlmIGFyZ3NfdG9fcGFyc2U6CiAgICAgICAgICAgIGFz
            c2VydCAoCiAgICAgICAgICAgICAgICB0eXBlKGFyZ3NfdG9fcGFyc2UpID09IGxpc3QKICAgICAg
            ICAgICAgKSwgIkFyZ3VtZW50IHRvICdnZW5fYXJnX3BhcnNlcicgc2hvdWxkIGJlIGEgbGlzdCIK
            ICAgICAgICBlbHNlOgogICAgICAgICAgICBhcmdzX3RvX3BhcnNlID0gc3lzLmFyZ3ZbMTpdCiAg
            ICAgICAgcmV0dXJuIGFyZ1AucGFyc2VfYXJncyhhcmdzX3RvX3BhcnNlKQoKICAgIGRlZiB2YWxp
            ZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xf
            bmFtZQogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0
            YV9wb29sX25hbWUKCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJGbGFnICctLXJiZC1kYXRhLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAg
            ICAgICAgICAgICAgICkKCiAgICAgICAgICAgIGlmIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUg
            PT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAog
            ICAgICAgICAgICAgICAgICAgICJGbGFnICctLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBjbWRfanNv
            biA9IHsicHJlZml4IjogIm9zZCBkdW1wIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAg
            cmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNt
            ZF9qc29uKQogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAw
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICBmIntjbWRfanNvblsncHJlZml4J119IGNvbW1hbmQgZmFpbGVkLlxuIgog
            ICAgICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNl
            IHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBt
            ZXRhZGF0YV9wb29sX2V4aXN0LCBwb29sX2V4aXN0ID0gRmFsc2UsIEZhbHNlCgogICAgICAgICAg
            ICBmb3Iga2V5IGluIGpzb25fb3V0WyJwb29scyJdOgogICAgICAgICAgICAgICAgIyBpZiBlcmFz
            dXJlX2NvZGVfcHJvZmlsZSBpcyBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IHJl
            cGxpY2EgcG9vbAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGtleVsi
            ZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBrZXlb
            InBvb2xfbmFtZSJdID09IHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgICAg
            ICk6CiAgICAgICAgICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCA9IFRydWUKICAgICAg
            ICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgbm90IGVtcHR5IGFuZCBwb29s
            IG5hbWUgZXhpc3RzIHRoZW4gaXQgaXMgZWMgcG9vbAogICAgICAgICAgICAgICAgaWYga2V5WyJl
            cmFzdXJlX2NvZGVfcHJvZmlsZSJdIGFuZCBrZXlbInBvb2xfbmFtZSJdID09IHJiZF9wb29sX25h
            bWU6CiAgICAgICAgICAgICAgICAgICAgcG9vbF9leGlzdCA9IFRydWUKCiAgICAgICAgICAgIGlm
            IG5vdCBtZXRhZGF0YV9wb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9u
            RmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiUHJvdmlkZWQgcmJkX2VjX21l
            dGFkYXRhX3Bvb2wgbmFtZSwiCiAgICAgICAgICAgICAgICAgICAgZiIge3JiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgaWYgbm90IHBvb2xfZXhpc3Q6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgcmJkX2RhdGFfcG9v
            bCBuYW1lLCB7cmJkX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICByZXR1cm4gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQoKICAgIGRlZiBk
            cnlfcnVuKHNlbGYsIG1zZyk6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICBwcmludCgiRXhlY3V0ZTogIiArICInIiArIG1zZyArICInIikKCiAgICBkZWYg
            dmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgIHdpdGggb3BlbihzZWxmLl9h
            cmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoLCBlbmNvZGluZz0idXRmOCIpIGFzIGY6CiAgICAg
            ICAgICAgICAgICBjb250ZW50cyA9IGYucmVhZCgpCiAgICAgICAgICAgICAgICByZXR1cm4gY29u
            dGVudHMucnN0cmlwKCkKCiAgICBkZWYgX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoc2VsZik6
            CiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZSBhbmQgbm90IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIkVpdGhlciAnLS11cGdyYWRlJyBvciAn
            LS1yYmQtZGF0YS1wb29sLW5hbWUgPHBvb2xfbmFtZT4nIHNob3VsZCBiZSBzcGVjaWZpZWQiCiAg
            ICAgICAgICAgICkKCiAgICBkZWYgX2ludmFsaWRfZW5kcG9pbnQoc2VsZiwgZW5kcG9pbnRfc3Ry
            KToKICAgICAgICAjIGV4dHJhY3QgdGhlIHBvcnQgYnkgZ2V0dGluZyB0aGUgbGFzdCBzcGxpdCBv
            biBgOmAgZGVsaW1pdGVyCiAgICAgICAgdHJ5OgogICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAs
            IHBvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICBleGNlcHQgVmFsdWVF
            cnJvcjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk5vdCBh
            IHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iKQoKICAgICAgICB0cnk6CiAgICAgICAg
            ICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSA9PSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2lu
            dF9zdHJfaXAgPSBlbmRwb2ludF9zdHJfaXBbMSA6IGxlbihlbmRwb2ludF9zdHJfaXApIC0gMV0K
            ICAgICAgICAgICAgaXBfdHlwZSA9ICgKICAgICAgICAgICAgICAgICJJUHY0IiBpZiB0eXBlKGlw
            X2FkZHJlc3MoZW5kcG9pbnRfc3RyX2lwKSkgaXMgSVB2NEFkZHJlc3MgZWxzZSAiSVB2NiIKICAg
            ICAgICAgICAgKQogICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICBpcF90eXBl
            ID0gIkZRRE4iCiAgICAgICAgaWYgbm90IHBvcnQuaXNkaWdpdCgpOgogICAgICAgICAgICByYWlz
            ZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiUG9ydCBub3QgdmFsaWQ6IHtwb3J0fSIpCiAg
            ICAgICAgaW50UG9ydCA9IGludChwb3J0KQogICAgICAgIGlmIGludFBvcnQgPCAxIG9yIGludFBv
            cnQgPiAyKioxNiAtIDE6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZiJPdXQgb2YgcmFuZ2UgcG9ydCBudW1iZXI6IHtwb3J0fSIpCgogICAgICAgIHJldHVybiBp
            cF90eXBlCgogICAgZGVmIGVuZHBvaW50X2RpYWwoc2VsZiwgZW5kcG9pbnRfc3RyLCBpcF90eXBl
            LCB0aW1lb3V0PTMsIGNlcnQ9Tm9uZSk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXInIGluc3Rh
            bmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91dCB0byB0
            aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlSYWRv
            cyk6CiAgICAgICAgICAgIHJldHVybiAiIiwgIiIsICIiCiAgICAgICAgaWYgaXBfdHlwZSA9PSAi
            SVB2NiI6CiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCwg
            ZW5kcG9pbnRfc3RyX3BvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICAg
            ICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiTm90IGEgcHJvcGVyIGVuZHBvaW50
            OiB7ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgZW5kcG9p
            bnRfc3RyX2lwWzBdICE9ICJbIjoKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9ICJb
            IiArIGVuZHBvaW50X3N0cl9pcCArICJdIgogICAgICAgICAgICBlbmRwb2ludF9zdHIgPSAiOiIu
            am9pbihbZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydF0pCgogICAgICAgIHByb3Rv
            Y29scyA9IFsiaHR0cCIsICJodHRwcyJdCiAgICAgICAgcmVzcG9uc2VfZXJyb3IgPSBOb25lCiAg
            ICAgICAgZm9yIHByZWZpeCBpbiBwcm90b2NvbHM6CiAgICAgICAgICAgIHRyeToKICAgICAgICAg
            ICAgICAgIGVwID0gZiJ7cHJlZml4fTovL3tlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAg
            dmVyaWZ5ID0gTm9uZQogICAgICAgICAgICAgICAgIyBJZiB2ZXJpZnkgaXMgc2V0IHRvIGEgcGF0
            aCB0byBhIGRpcmVjdG9yeSwKICAgICAgICAgICAgICAgICMgdGhlIGRpcmVjdG9yeSBtdXN0IGhh
            dmUgYmVlbiBwcm9jZXNzZWQgdXNpbmcgdGhlIGNfcmVoYXNoIHV0aWxpdHkgc3VwcGxpZWQgd2l0
            aCBPcGVuU1NMLgogICAgICAgICAgICAgICAgaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3NraXBfdGxzOgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9IEZh
            bHNlCiAgICAgICAgICAgICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGlt
            ZW91dCwgdmVyaWZ5PUZhbHNlKQogICAgICAgICAgICAgICAgZWxpZiBwcmVmaXggPT0gImh0dHBz
            IiBhbmQgY2VydDoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBjZXJ0CiAgICAgICAgICAg
            ICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCwgdmVyaWZ5PWNl
            cnQpCiAgICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQpCiAgICAgICAgICAgICAgICBpZiByLnN0YXR1c19j
            b2RlID09IDIwMDoKICAgICAgICAgICAgICAgICAgICByZXR1cm4gcHJlZml4LCB2ZXJpZnksICIi
            CiAgICAgICAgICAgIGV4Y2VwdCBFeGNlcHRpb24gYXMgZXJyOgogICAgICAgICAgICAgICAgcmVz
            cG9uc2VfZXJyb3IgPSBlcnIKICAgICAgICAgICAgICAgIGNvbnRpbnVlCiAgICAgICAgc3lzLnN0
            ZGVyci53cml0ZSgKICAgICAgICAgICAgZiJ1bmFibGUgdG8gY29ubmVjdCB0byBlbmRwb2ludDog
            e2VuZHBvaW50X3N0cn0sIGZhaWxlZCBlcnJvcjoge3Jlc3BvbnNlX2Vycm9yfSIKICAgICAgICAp
            CiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICIiLAogICAgICAg
            ICAgICAoIi0xIiksCiAgICAgICAgKQoKICAgIGRlZiBwYXJzZV9jb25maWdfZmlsZShzZWxmLCBj
            b25maWdfZmlsZSk6CiAgICAgICAgY29uZmlnID0gY29uZmlncGFyc2VyLkNvbmZpZ1BhcnNlcigp
            CiAgICAgICAgY29uZmlnLnJlYWQoY29uZmlnX2ZpbGUpCiAgICAgICAgZm9yIGFyZyBpbiBsaXN0
            KHZhcnMoc2VsZi5fYXJnX3BhcnNlcikpOgogICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxh
            Zy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRlcm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxh
            Zy1uYW1lLAogICAgICAgICAgICAjIHNvIHdlIGNhbiBnZXQgdGhvc2UgdmFsdWVzIGZyb20gY29u
            ZmlnIGZpbGUKICAgICAgICAgICAgYXJndW1lbnQgPSBhcmcucmVwbGFjZSgiXyIsICItIikKICAg
            ICAgICAgICAgYXJndW1lbnRWYWx1ZSA9IHN0cihnZXRhdHRyKHNlbGYuX2FyZ19wYXJzZXIsIGFy
            ZykpCiAgICAgICAgICAgIGNvbmZpZ192YWx1ZSA9IGNvbmZpZy5nZXQoIkNvbmZpZ3VyYXRpb25z
            IiwgYXJndW1lbnQsIGZhbGxiYWNrPU5vbmUpCiAgICAgICAgICAgICMgZ2l2ZSBwcmlvcml0eSB0
            byBjb21tYW5kIGxpbmUgYXJndW1lbnQsIGlmIGNvbW1hbmQgbGluZSBhcmd1bWVudCBpcyBub3Qg
            cHJlc2VudCB1c2UgY29uZmlnLmluaSB2YWx1ZSwKICAgICAgICAgICAgaWYgKHN0cihzeXMuYXJn
            dikuZmluZChhcmd1bWVudCkgPT0gLTEpIGFuZCAoCiAgICAgICAgICAgICAgICAoY29uZmlnX3Zh
            bHVlICE9IE5vbmUpIGFuZCAoY29uZmlnX3ZhbHVlICE9ICIiKQogICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5fX3NldGF0dHJfXyhhcmcsIGNvbmZpZ192YWx1
            ZSkKCiAgICAgICAgcmV0dXJuIGNvbmZpZwoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhcmdfbGlz
            dD1Ob25lKToKICAgICAgICBzZWxmLm91dF9tYXAgPSB7fQogICAgICAgIHNlbGYuX2V4Y2x1ZGVk
            X2tleXMgPSBzZXQoKQogICAgICAgIHNlbGYuX2FyZ19wYXJzZXIgPSBzZWxmLmdlbl9hcmdfcGFy
            c2VyKGFyZ3NfdG9fcGFyc2U9YXJnX2xpc3QpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            b25maWdfZmlsZToKICAgICAgICAgICAgc2VsZi5jb25maWcgPSBzZWxmLnBhcnNlX2NvbmZpZ19m
            aWxlKHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGUpCiAgICAgICAgc2VsZi5fY2hlY2tfY29u
            ZmxpY3Rpbmdfb3B0aW9ucygpCiAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuX2FyZ19w
            YXJzZXIucnVuX2FzX3VzZXIKICAgICAgICBzZWxmLm91dHB1dF9maWxlID0gc2VsZi5fYXJnX3Bh
            cnNlci5vdXRwdXQKICAgICAgICBzZWxmLmNlcGhfY29uZiA9IHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aF9jb25mCiAgICAgICAgc2VsZi5jZXBoX2tleXJpbmcgPSBzZWxmLl9hcmdfcGFyc2VyLmtleXJp
            bmcKICAgICAgICAjIGlmIHVzZXIgbm90IHByb3ZpZGVkLCBnaXZlIGEgZGVmYXVsdCB1c2VyCiAg
            ICAgICAgaWYgbm90IHNlbGYucnVuX2FzX3VzZXIgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGU6CiAgICAgICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLkVYVEVSTkFMX1VTRVJf
            TkFNRQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCBhbmQg
            bm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggPSBzZWxmLkRFRkFVTFRfUkdXX1BPT0xfUFJFRklYCiAgICAgICAg
            aWYgc2VsZi5jZXBoX2NvbmY6CiAgICAgICAgICAgIGt3YXJncyA9IHt9CiAgICAgICAgICAgIGlm
            IHNlbGYuY2VwaF9rZXlyaW5nOgogICAgICAgICAgICAgICAga3dhcmdzWyJjb25mIl0gPSB7Imtl
            eXJpbmciOiBzZWxmLmNlcGhfa2V5cmluZ30KICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFk
            b3MuUmFkb3MoY29uZmZpbGU9c2VsZi5jZXBoX2NvbmYsICoqa3dhcmdzKQogICAgICAgIGVsc2U6
            CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKCkKICAgICAgICAgICAgc2Vs
            Zi5jbHVzdGVyLmNvbmZfcmVhZF9maWxlKCkKICAgICAgICBzZWxmLmNsdXN0ZXIuY29ubmVjdCgp
            CgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIGlmIHNlbGYuY2x1c3Rlci5zdGF0ZSA9
            PSAiY29ubmVjdGVkIjoKICAgICAgICAgICAgc2VsZi5jbHVzdGVyLnNodXRkb3duKCkKCiAgICBk
            ZWYgZ2V0X2ZzaWQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoIGZzaWQiKQogICAgICAgIHJldHVy
            biBzdHIoc2VsZi5jbHVzdGVyLmdldF9mc2lkKCkpCgogICAgZGVmIF9jb21tb25fY21kX2pzb25f
            Z2VuKHNlbGYsIGNtZF9qc29uKToKICAgICAgICBjbWQgPSBqc29uLmR1bXBzKGNtZF9qc29uLCBz
            b3J0X2tleXM9VHJ1ZSkKICAgICAgICByZXRfdmFsLCBjbWRfb3V0LCBlcnJfbXNnID0gc2VsZi5j
            bHVzdGVyLm1vbl9jb21tYW5kKGNtZCwgYiIiKQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            dmVyYm9zZToKICAgICAgICAgICAgcHJpbnQoZiJDb21tYW5kIElucHV0OiB7Y21kfSIpCiAgICAg
            ICAgICAgIHByaW50KAogICAgICAgICAgICAgICAgZiJSZXR1cm4gVmFsOiB7cmV0X3ZhbH1cbkNv
            bW1hbmQgT3V0cHV0OiB7Y21kX291dH1cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3IgTWVzc2Fn
            ZToge2Vycl9tc2d9XG4tLS0tLS0tLS0tXG4iCiAgICAgICAgICAgICkKICAgICAgICBqc29uX291
            dCA9IHt9CiAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciAoaS5lOyByZXRfdmFsIGlzIFpF
            Uk8pIGFuZCAnY21kX291dCcgaXMgbm90IGVtcHR5CiAgICAgICAgIyB0aGVuIGNvbnZlcnQgJ2Nt
            ZF9vdXQnIHRvIGEganNvbiBvdXRwdXQKICAgICAgICBpZiByZXRfdmFsID09IDAgYW5kIGNtZF9v
            dXQ6CiAgICAgICAgICAgIGpzb25fb3V0ID0ganNvbi5sb2FkcyhjbWRfb3V0KQogICAgICAgIHJl
            dHVybiByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZwoKICAgIGRlZiBnZXRfY2VwaF9leHRlcm5h
            bF9tb25fZGF0YShzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogInF1b3J1bV9z
            dGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5
            X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25b
            InByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29t
            bW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3Vj
            Y2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQp
            ID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAiJ3F1b3J1bV9zdGF0dXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAg
            ICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9P
            VVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICBxX2xlYWRlcl9uYW1lID0ganNvbl9v
            dXRbInF1b3J1bV9sZWFkZXJfbmFtZSJdCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHt9CiAg
            ICAgICAgcV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCA9IFsKICAgICAgICAgICAgbCBmb3IgbCBpbiBq
            c29uX291dFsibW9ubWFwIl1bIm1vbnMiXSBpZiBsWyJuYW1lIl0gPT0gcV9sZWFkZXJfbmFtZQog
            ICAgICAgIF0KICAgICAgICBpZiBsZW4ocV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCkgPT0gMDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gbWF0Y2hpbmcgJ21v
            bicgZGV0YWlscyBmb3VuZCIpCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3RbMF0KICAgICAgICAjIGdldCB0aGUgYWRkcmVzcyB2ZWN0b3Igb2YgdGhlIHF1
            b3J1bS1sZWFkZXIKICAgICAgICBxX2xlYWRlcl9hZGRydmVjID0gcV9sZWFkZXJfZGV0YWlscy5n
            ZXQoInB1YmxpY19hZGRycyIsIHt9KS5nZXQoImFkZHJ2ZWMiLCBbXSkKICAgICAgICBpcF9hZGRy
            ID0gc3RyKHFfbGVhZGVyX2RldGFpbHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYg
            cV9sZWFkZXJfYWRkcnZlY1swXVsidHlwZSJdID09ICJ2MiI6CiAgICAgICAgICAgICAgICBpcF9h
            ZGRyID0gcV9sZWFkZXJfYWRkcnZlY1swXVsiYWRkciJdCiAgICAgICAgICAgIGVsaWYgbGVuKHFf
            bGVhZGVyX2FkZHJ2ZWMpID4gMSBhbmQgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09ICJ2
            MiI6CiAgICAgICAgICAgICAgICBpcF9hZGRyID0gcV9sZWFkZXJfYWRkcnZlY1sxXVsiYWRkciJd
            CiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAg
            ICAgICAgICAgICAgICAgICIndjInIGFkZHJlc3MgdHlwZSBub3QgcHJlc2VudCwgYW5kICd2Mi1w
            b3J0LWVuYWJsZScgZmxhZyBpcyBwcm92aWRlZCIKICAgICAgICAgICAgICAgICkKCiAgICAgICAg
            cmV0dXJuIGYie3N0cihxX2xlYWRlcl9uYW1lKX09e2lwX2FkZHJ9IgoKICAgIGRlZiBfY29udmVy
            dF9ob3N0bmFtZV90b19pcChzZWxmLCBob3N0X25hbWUsIHBvcnQsIGlwX3R5cGUpOgogICAgICAg
            ICMgaWYgJ2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgdHlwZSwKICAgICAgICAjIGNhbGwg
            dGhlIGR1bW15IGluc3RhbmNlJ3MgImNvbnZlcnQiIG1ldGhvZAogICAgICAgIGlmIG5vdCBob3N0
            X25hbWU6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIkVtcHR5
            IGhvc3RuYW1lIHByb3ZpZGVkIikKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3Rlciwg
            RHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybiBzZWxmLmNsdXN0ZXIuX2NvbnZlcnRfaG9z
            dG5hbWVfdG9faXAoaG9zdF9uYW1lKQoKICAgICAgICBpZiBpcF90eXBlID09ICJGUUROIjoKICAg
            ICAgICAgICAgIyBjaGVjayB3aGljaCBpcCBGUUROIHNob3VsZCBiZSBjb252ZXJ0ZWQgdG8sIElQ
            djQgb3IgSVB2NgogICAgICAgICAgICAjIGNoZWNrIHRoZSBob3N0IGlwLCB0aGUgZW5kcG9pbnQg
            aXAgdHlwZSB3b3VsZCBiZSBzaW1pbGFyIHRvIGhvc3QgaXAKICAgICAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJvcmNoIGhvc3QgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVt
            cHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICInb3JjaCBob3N0IGxzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAg
            ICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZ
            X09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaG9zdF9hZGRyID0g
            anNvbl9vdXRbMF1bImFkZHIiXQogICAgICAgICAgICAjIGFkZCA6ODAgc2FtcGxlIHBvcnQgaW4g
            aXBfdHlwZSwgYXMgX2ludmFsaWRfZW5kcG9pbnQgYWxzbyB2ZXJpZnkgcG9ydAogICAgICAgICAg
            ICBob3N0X2lwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KGhvc3RfYWRkciArICI6ODAi
            KQogICAgICAgICAgICBpbXBvcnQgc29ja2V0CgogICAgICAgICAgICBpcCA9IFtdCiAgICAgICAg
            ICAgICMgZXhhbXBsZSBvdXRwdXQgWyg8QWRkcmVzc0ZhbWlseS5BRl9JTkVUOiAyPiwgPFNvY2tl
            dEtpbmQuU09DS19TVFJFQU06IDE+LCA2LCAnJywgKCc5My4xODQuMjE2LjM0JywgODApKSwgLi4u
            XQogICAgICAgICAgICAjIHdlIG5lZWQgdG8gZ2V0IDkzLjE4NC4yMTYuMzQgc28gaXQgd291bGQg
            YmUgaXBbMF1bNF1bMF0KICAgICAgICAgICAgaWYgaG9zdF9pcF90eXBlID09ICJJUHY2IjoKICAg
            ICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAgICAgICAgICAgICAgICAg
            IGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVUNiwgcHJvdG89c29ja2V0LklQ
            UFJPVE9fVENQCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsaWYgaG9zdF9pcF90eXBl
            ID09ICJJUHY0IjoKICAgICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAg
            ICAgICAgICAgICAgICAgIGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVULCBw
            cm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZGVs
            IHNvY2tldAogICAgICAgICAgICByZXR1cm4gaXBbMF1bNF1bMF0KICAgICAgICByZXR1cm4gaG9z
            dF9uYW1lCgogICAgZGVmIGdldF9hY3RpdmVfYW5kX3N0YW5kYnlfbWdycyhzZWxmKToKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIiwgc2Vs
            Zi5kcnlfcnVuKCJjZXBoIHN0YXR1cyIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0
            ID0gc2VsZi5fYXJnX3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5k
            cG9pbnQKICAgICAgICBzdGFuZGJ5X21ncnMgPSBbXQogICAgICAgIGlmIG5vdCBtb25pdG9yaW5n
            X2VuZHBvaW50X2lwX2xpc3Q6CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAic3Rh
            dHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVy
            cl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAj
            IGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRf
            dmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ21nciBzZXJ2aWNl
            cycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9t
            c2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnQgPSAoCiAgICAgICAgICAg
            ICAgICBqc29uX291dC5nZXQoIm1ncm1hcCIsIHt9KS5nZXQoInNlcnZpY2VzIiwge30pLmdldCgi
            cHJvbWV0aGV1cyIsICIiKQogICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9y
            aW5nX2VuZHBvaW50OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiY2FuJ3QgZmluZCBtb25pdG9yaW5nX2VuZHBvaW50
            LCBwcm9tZXRoZXVzIG1vZHVsZSBtaWdodCBub3QgYmUgZW5hYmxlZCwgIgogICAgICAgICAgICAg
            ICAgICAgICJlbmFibGUgdGhlIG1vZHVsZSBieSBydW5uaW5nICdjZXBoIG1nciBtb2R1bGUgZW5h
            YmxlIHByb21ldGhldXMnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIG5vdyBjaGVj
            ayB0aGUgc3RhbmQtYnkgbWdyLXMKICAgICAgICAgICAgc3RhbmRieV9hcnIgPSBqc29uX291dC5n
            ZXQoIm1ncm1hcCIsIHt9KS5nZXQoInN0YW5kYnlzIiwgW10pCiAgICAgICAgICAgIGZvciBlYWNo
            X3N0YW5kYnkgaW4gc3RhbmRieV9hcnI6CiAgICAgICAgICAgICAgICBpZiAibmFtZSIgaW4gZWFj
            aF9zdGFuZGJ5LmtleXMoKToKICAgICAgICAgICAgICAgICAgICBzdGFuZGJ5X21ncnMuYXBwZW5k
            KGVhY2hfc3RhbmRieVsibmFtZSJdKQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBw
            YXJzZWRfZW5kcG9pbnQgPSB1cmxwYXJzZShtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgICAg
            ICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50OiB7bW9u
            aXRvcmluZ19lbmRwb2ludH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3Jp
            bmdfZW5kcG9pbnRfaXBfbGlzdCA9IHBhcnNlZF9lbmRwb2ludC5ob3N0bmFtZQogICAgICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICAgICAgbW9uaXRv
            cmluZ19lbmRwb2ludF9wb3J0ID0gc3RyKHBhcnNlZF9lbmRwb2ludC5wb3J0KQoKICAgICAgICAj
            IGlmIG1vbml0b3JpbmcgZW5kcG9pbnQgcG9ydCBpcyBub3Qgc2V0LCBwdXQgYSBkZWZhdWx0IG1v
            biBwb3J0CiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoKICAgICAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0ID0gc2VsZi5ERUZBVUxUX01PTklUT1JJTkdfRU5E
            UE9JTlRfUE9SVAoKICAgICAgICAjIHVzZXIgY291bGQgZ2l2ZSBjb21tYSBhbmQgc3BhY2Ugc2Vw
            YXJhdGVkIGlucHV0cyAobGlrZSAtLW1vbml0b3JpbmctZW5kcG9pbnQ9IjxpcDE+LCA8aXAyPiIp
            CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcF9saXN0LnJlcGxhY2UoIiwiLCAiICIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0X3NwbGl0ID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0LnNwbGl0KCkKICAgICAg
            ICAjIGlmIG1vbml0b3JpbmctZW5kcG9pbnQgY291bGQgbm90IGJlIGZvdW5kLCByYWlzZSBhbiBl
            cnJvcgogICAgICAgIGlmIGxlbihtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQpID09
            IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vICdtb25p
            dG9yaW5nLWVuZHBvaW50JyBmb3VuZCIpCiAgICAgICAgIyBmaXJzdCBpcCBpcyB0cmVhdGVkIGFz
            IHRoZSBtYWluIG1vbml0b3JpbmctZW5kcG9pbnQKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X2lwID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzBdCiAgICAgICAgIyByZXN0
            IG9mIHRoZSBpcC1zIGFyZSBhZGRlZCB0byB0aGUgJ3N0YW5kYnlfbWdycycgbGlzdAogICAgICAg
            IHN0YW5kYnlfbWdycy5leHRlbmQobW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzE6
            XSkKICAgICAgICBmYWlsZWRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwCgogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnQgPSAiOiIuam9pbigKICAgICAgICAgICAgW21vbml0b3JpbmdfZW5k
            cG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydF0KICAgICAgICApCiAgICAgICAgaXBf
            dHlwZSA9IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQobW9uaXRvcmluZ19lbmRwb2ludCkKICAgICAg
            ICB0cnk6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBzZWxmLl9jb252ZXJ0
            X2hvc3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCwg
            bW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgIyBjb2xsZWN0IGFsbCB0aGUgJ3N0YW5kLWJ5JyBtZ3IgaXBzCiAgICAgICAgICAgIG1ncl9p
            cHMgPSBbXQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5X21nciBpbiBzdGFuZGJ5X21ncnM6
            CiAgICAgICAgICAgICAgICBmYWlsZWRfaXAgPSBlYWNoX3N0YW5kYnlfbWdyCiAgICAgICAgICAg
            ICAgICBtZ3JfaXBzLmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9jb252ZXJ0X2hv
            c3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgICAgICAgICBlYWNoX3N0YW5kYnlfbWdyLCBt
            b25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgZXhjZXB0OgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJDb252ZXJzaW9uIG9mIGhvc3Q6
            IHtmYWlsZWRfaXB9IHRvIElQIGZhaWxlZC4gIgogICAgICAgICAgICAgICAgIlBsZWFzZSBlbnRl
            ciB0aGUgSVAgYWRkcmVzc2VzIG9mIGFsbCB0aGUgY2VwaC1tZ3JzIHdpdGggdGhlICctLW1vbml0
            b3JpbmctZW5kcG9pbnQnIGZsYWciCiAgICAgICAgICAgICkKCiAgICAgICAgXywgXywgZXJyID0g
            c2VsZi5lbmRwb2ludF9kaWFsKG1vbml0b3JpbmdfZW5kcG9pbnQsIGlwX3R5cGUpCiAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZXJyKQogICAgICAgICMgYWRkIHRoZSB2YWxpZGF0ZWQgYWN0aXZlIG1nciBJUCBpbnRvIHRo
            ZSBmaXJzdCBpbmRleAogICAgICAgIG1ncl9pcHMuaW5zZXJ0KDAsIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXApCiAgICAgICAgYWxsX21ncl9pcHNfc3RyID0gIiwiLmpvaW4obWdyX2lwcykKICAgICAg
            ICByZXR1cm4gYWxsX21ncl9pcHNfc3RyLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKCiAgICBk
            ZWYgY2hlY2tfdXNlcl9leGlzdChzZWxmLCB1c2VyKToKICAgICAgICBjbWRfanNvbiA9IHsicHJl
            Zml4IjogImF1dGggZ2V0IiwgImVudGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0K
            ICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoK
            ICAgICAgICAgICAgcmV0dXJuICIiCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5
            Il0pCgogICAgZGVmIGdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAg
            ICJvc2QiOiAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwKICAgICAgICAgICAgIm1k
            cyI6ICJhbGxvdyAqIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0
            cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rl
            cl9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiazhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNf
            ZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgog
            ICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAg
            ICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNf
            Y2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVzeXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1si
            b3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9e2NlcGhmc19maWxlc3lzdGVt
            fSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2VwaGZzX25vZGVf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3
            IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsCiAgICAgICAg
            ICAgICJtZHMiOiAiYWxsb3cgcnciLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4
            c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0g
            PT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0
            eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAg
            ICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyAqPXtjZXBoZnNfZmlsZXN5c3Rl
            bX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2VudGl0eSgKICAg
            ICAgICBzZWxmLAogICAgICAgIGVudGl0eSwKICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAg
            IGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICBy
            YWRvc19uYW1lc3BhY2UsCiAgICApOgogICAgICAgIGlmICgKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgb3IgcmJkX3Bvb2xfbmFtZS5jb3VudCgi
            XyIpICE9IDAKICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZSAhPSAiIgogICAgICAg
            ICAgICAjIGNoZWNraW5nIGFsaWFzX3JiZF9wb29sX25hbWUgaXMgbm90IGVtcHR5IGFzIHRoZXJl
            IG1heWJlIGEgc3BlY2lhbCBjaGFyYWN0ZXIgdXNlZCBvdGhlciB0aGFuIC4gb3IgXwogICAgICAg
            ICk6CiAgICAgICAgICAgIGlmIGFsaWFzX3JiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAg
            ICJwbGVhc2Ugc2V0IHRoZSAnLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgYXMgdGhl
            IHJiZCBkYXRhIHBvb2wgbmFtZSBjb250YWlucyAnLicgb3IgJ18nIgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLmNv
            dW50KCIuIikgIT0gMAogICAgICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3Vu
            dCgiXyIpICE9IDAKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIictLWFsaWFzLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyB2YWx1ZSBzaG91bGQgbm90IGNvbnRhaW4gJy4nIG9yICdfJyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0
            ZXJfbmFtZX0te2FsaWFzX3JiZF9wb29sX25hbWV9IgogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtyYmRfcG9vbF9uYW1lfSIK
            CiAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgIHJldHVybiBlbnRpdHkKCiAgICBkZWYgZ2V0
            X3JiZF9wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0g
            ImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAg
            ICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAg
            ICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSBy
            YmQiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNl
            ci5yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19j
            bHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAg
            ICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAg
            ICAgICAgICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBm
            bGFnIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFn
            IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIi
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAn
            LS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bnRpdHkgPSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAg
            ICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFt
            ZSwKICAgICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRv
            c19uYW1lc3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNl
            ICE9ICIiOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25h
            bWVzcGFjZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBjYXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgog
            ICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X3JiZF9ub2RlX2NhcHNfYW5k
            X2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtbm9kZSIKICAg
            ICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSByYmQiLAog
            ICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJt
            aXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19jbHVzdGVy
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgcmFk
            b3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICAg
            ICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5v
            dCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIgogICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAg
            ICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAg
            ICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMt
            Y2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkg
            PSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAgICAgICAg
            ICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAg
            ICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRvc19uYW1l
            c3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlICE9ICIi
            OgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgZiJw
            cm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25hbWVzcGFj
            ZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBj
            YXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAg
            IHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2Vu
            dGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAgY2Fw
            cyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9z
            dGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLAogICAgICAgICAgICAib3NkIjogZiJwcm9maWxlIHJiZC1yZWFkLW9u
            bHksIGFsbG93IHJ3eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9e3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2csIGFsbG93IHggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5idWNrZXRzLmluZGV4IiwKICAgICAgICAg
            ICAgIm1kcyI6ICJhbGxvdyAqIiwKICAgICAgICB9CgogICAgICAgIHJldHVybiBjYXBzLCBlbnRp
            dHkKCiAgICBkZWYgZ2V0X2NhcHNfYW5kX2VudGl0eShzZWxmLCB1c2VyX25hbWUpOgogICAgICAg
            IGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAg
            ICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fw
            c19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgaW4gdXNl
            cl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgIT0gdXNlcl9u
            YW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVy
            bWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19ub2RlX2Nh
            cHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiBp
            biB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X3JiZF9w
            cm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1u
            b2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1ub2RlIiAh
            PSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRf
            YXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX25v
            ZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiIGlu
            IHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5oZWFsdGhjaGVja2VyIiAhPSB1c2Vy
            X25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9w
            ZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfZGVmYXVsdFVzZXJf
            Y2Fwc19hbmRfZW50aXR5KCkKCiAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgZiJubyB1c2VyIGZvdW5kIHdpdGggdXNlcl9uYW1lOiB7dXNlcl9uYW1l
            fSwgIgogICAgICAgICAgICAiZ2V0X2NhcHNfYW5kX2VudGl0eSBjb21tYW5kIGZhaWxlZC5cbiIK
            ICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKHNlbGYsIHVzZXIp
            OgogICAgICAgICIiIgogICAgICAgIGNvbW1hbmQ6IGNlcGggYXV0aCBnZXQtb3ItY3JlYXRlIGNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIG1vbiAnYWxsb3cgcicgbWdyICdhbGxvdyBydycg
            b3NkICdhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSonCiAgICAgICAgIiIiCiAgICAgICAg
            Y2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fwc19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21k
            X2pzb24gPSB7CiAgICAgICAgICAgICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAg
            ICAgICAgICAgImVudGl0eSI6IGVudGl0eSwKICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBj
            YXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMoKSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAg
            ICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICBzZWxm
            LmRyeV9ydW4oCiAgICAgICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAgICAg
            ICsgY21kX2pzb25bInByZWZpeCJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgICAgICsgIiAi
            CiAgICAgICAgICAgICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAg
            ICAgICAgICAgKSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCiAgICAgICAgIyBj
            aGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYuY2hlY2tf
            dXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAgICAgICAg
            IHJldHVybiB1c2VyX2tleSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICAg
            ICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0eShjbGllbnQuY3NpLXJi
            ZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9t
            c2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhl
            cmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9y
            IGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7dXNlcn0nIGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0
            X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAg
            ICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSksIGYie2VudGl0eS5zcGxpdCgnLics
            IDEpWzFdfSIKICAgICAgICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0
            eShjbGllbnQuY3NpLXJiZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICBkZWYgZ2V0X2NlcGhm
            c19kYXRhX3Bvb2xfZGV0YWlscyhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4Ijog
            ImZzIGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29u
            WyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2Nv
            bW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1
            Y2Nlc3NmdWwgYXR0ZW1wdCwgcmVwb3J0IGFuIGVycm9yCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            OgogICAgICAgICAgICAjIGlmIGZzIGFuZCBkYXRhX3Bvb2wgYXJndW1lbnRzIGFyZSBub3Qgc2V0
            LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPT0gIiIKICAgICAgICAg
            ICAgKToKICAgICAgICAgICAgICAgIHJldHVybgogICAgICAgICAgICAjIGlmIHVzZXIgaGFzIHBy
            b3ZpZGVkIGFueSBvZiB0aGUKICAgICAgICAgICAgIyAnLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1l
            JyBvciAnLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUnIGFyZ3VtZW50cywKICAgICAgICAgICAgIyBy
            YWlzZSBhbiBleGNlcHRpb24gYXMgd2UgYXJlIHVuYWJsZSB0byB2ZXJpZnkgdGhlIGFyZ3MKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiJ2ZzIGxzJyBjZXBoIGNhbGwgZmFpbGVkIHdpdGggZXJyb3I6IHtlcnJfbXNnfSIKICAgICAg
            ICAgICAgKQoKICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IHt9CiAgICAgICAgIyBpZiAnLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lJyBhcmd1bWVudCBpcyBwcm92aWRlZCwKICAgICAgICAjIGNo
            ZWNrIHdoZXRoZXIgdGhlIHByb3ZpZGVkIGZpbGVzeXN0ZW0tbmFtZSBleGlzdHMgb3Igbm90CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lOgogICAgICAg
            ICAgICAjIGdldCB0aGUgbWF0Y2hpbmcgbGlzdAogICAgICAgICAgICBtYXRjaGluZ19qc29uX291
            dF9saXN0ID0gWwogICAgICAgICAgICAgICAgbWF0Y2hlZAogICAgICAgICAgICAgICAgZm9yIG1h
            dGNoZWQgaW4ganNvbl9vdXQKICAgICAgICAgICAgICAgIGlmIHN0cihtYXRjaGVkWyJuYW1lIl0p
            ID09IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgdW5hYmxlIHRvIGZpbmQgYSBtYXRjaGluZyBmcy1uYW1lLCByYWlzZSBh
            biBlcnJvcgogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRfbGlzdCkgPT0gMDoK
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgZiJGaWxlc3lzdGVtIHByb3ZpZGVkLCAne3NlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgZiJpcyBub3QgZm91
            bmQgaW4gdGhlIGZzLWxpc3Q6IHtbc3RyKHhbJ25hbWUnXSkgZm9yIHggaW4ganNvbl9vdXRdfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBtYXRjaGlu
            Z19qc29uX291dF9saXN0WzBdCiAgICAgICAgIyBpZiBjZXBoZnMgZmlsZXN5c3RlbSBuYW1lIGlz
            IG5vdCBwcm92aWRlZCwKICAgICAgICAjIHRyeSB0byBnZXQgYSBkZWZhdWx0IGZzIG5hbWUgYnkg
            ZG9pbmcgdGhlIGZvbGxvd2luZwogICAgICAgIGVsc2U6CiAgICAgICAgICAgICMgYS4gY2hlY2sg
            aWYgdGhlcmUgaXMgb25seSBvbmUgZmlsZXN5c3RlbSBpcyBwcmVzZW50CiAgICAgICAgICAgIGlm
            IGxlbihqc29uX291dCkgPT0gMToKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0g
            anNvbl9vdXRbMF0KICAgICAgICAgICAgIyBiLiBvciBlbHNlLCBjaGVjayBpZiBkYXRhX3Bvb2wg
            bmFtZSBpcyBwcm92aWRlZAogICAgICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgIyBhbmQgaWYgcHJlc2VudCwgY2hlY2sg
            d2hldGhlciB0aGVyZSBleGlzdHMgYSBmcyB3aGljaCBoYXMgdGhlIGRhdGFfcG9vbAogICAgICAg
            ICAgICAgICAgZm9yIGVhY2hKIGluIGpzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lIGluIGVhY2hKWyJkYXRhX3Bvb2xz
            Il06CiAgICAgICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0gZWFjaEoKICAg
            ICAgICAgICAgICAgICAgICAgICAgYnJlYWsKICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            bm8gbWF0Y2hpbmcgZnMgZXhpc3RzLCB0aGF0IG1lYW5zIHByb3ZpZGVkIGRhdGFfcG9vbCBuYW1l
            IGlzIGludmFsaWQKICAgICAgICAgICAgICAgIGlmIG5vdCBtYXRjaGluZ19qc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGFfcG9vbCBuYW1lLCB7c2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9LCIKICAgICAgICAgICAgICAgICAgICAgICAgIiBk
            b2VzIG5vdCBleGlzdHMiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGMuIGlm
            IG5vdGhpbmcgaXMgc2V0IGFuZCBjb3VsZG4ndCBmaW5kIGEgZGVmYXVsdCwKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgICMganVzdCByZXR1cm4gc2lsZW50bHkKICAgICAgICAgICAg
            ICAgIHJldHVybgoKICAgICAgICBpZiBtYXRjaGluZ19qc29uX291dDoKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID0gc3RyKG1hdGNoaW5nX2pzb25f
            b3V0WyJuYW1lIl0pCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0WyJtZXRh
            ZGF0YV9wb29sIl0KICAgICAgICAgICAgKQoKICAgICAgICBpZiBpc2luc3RhbmNlKG1hdGNoaW5n
            X2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0sIGxpc3QpOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGhhcyBhbHJlYWR5IHByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lLAogICAgICAgICAgICAjIHRocm91
            Z2ggLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUKICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGlmIHRoZSBwcm92aWRl
            ZCBuYW1lIGlzIG5vdCBtYXRjaGluZyB3aXRoIHRoZSBvbmUgaW4gdGhlIGxpc3QKICAgICAgICAg
            ICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19k
            YXRhX3Bvb2xfbmFtZQogICAgICAgICAgICAgICAgICAgIG5vdCBpbiBtYXRjaGluZ19qc29uX291
            dFsiZGF0YV9wb29scyJdCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYi
            UHJvdmlkZWQgZGF0YS1wb29sLW5hbWU6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWV9JywgIgogICAgICAgICAgICAgICAgICAgICAgICAiZG9lc24ndCBtYXRjaCBmcm9t
            IHRoZSBkYXRhLXBvb2xzIGxpc3Q6ICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4
            KSBmb3IgeCBpbiBtYXRjaGluZ19qc29uX291dFsnZGF0YV9wb29scyddXX0iCiAgICAgICAgICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGlmIGRhdGFfcG9vbCBuYW1lIGlzIG5vdCBwcm92aWRl
            ZCwKICAgICAgICAgICAgIyB0aGVuIHRyeSB0byBmaW5kIGEgZGVmYXVsdCBkYXRhIHBvb2wgbmFt
            ZQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBpZiBubyBkYXRhX3Bvb2xzIGV4
            aXN0LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dFsiZGF0YV9wb29scyJdKSA9PSAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPSBzdHIo
            CiAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXVswXQog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGlmIHRoZXJlIGFyZSBtb3JlIHRoYW4gb25l
            ICdkYXRhX3Bvb2xzJyBleGlzdCwKICAgICAgICAgICAgIyB0aGVuIHdhcm4gdGhlIHVzZXIgdGhh
            dCB3ZSBhcmUgdXNpbmcgdGhlIHNlbGVjdGVkIG5hbWUKICAgICAgICAgICAgaWYgbGVuKG1hdGNo
            aW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID4gMToKICAgICAgICAgICAgICAgIHByaW50KAog
            ICAgICAgICAgICAgICAgICAgICJXQVJOSU5HOiBNdWx0aXBsZSBkYXRhIHBvb2xzIGRldGVjdGVk
            OiAiCiAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4KSBmb3IgeCBpbiBtYXRjaGluZ19qc29u
            X291dFsnZGF0YV9wb29scyddXX1cbiIKICAgICAgICAgICAgICAgICAgICBmIlVzaW5nIHRoZSBk
            YXRhLXBvb2w6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9J1xuIgog
            ICAgICAgICAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2hlY2tlcktleShzZWxmLCB1c2VyKToK
            ICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAg
            ICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVh
            dGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtj
            YXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0
            XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigK
            ICAgICAgICAgICAgICAgICJjZXBoICIKICAgICAgICAgICAgICAgICsgY21kX2pzb25bInByZWZp
            eCJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50
            aXR5Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICArICIgIi5qb2luKGNt
            ZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            CgogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBh
            dHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmIidhdXRoIGdldC1vci1jcmVhdGUge3NlbGYucnVuX2FzX3VzZXJ9JyBjb21tYW5kIGZhaWxl
            ZFxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVs
            c2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            c2VsZik6CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJtZ3Igc2VydmljZXMiLCAiZm9y
            bWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQog
            ICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAog
            ICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAg
            IHJldHVybiBOb25lCiAgICAgICAgaWYgImRhc2hib2FyZCIgbm90IGluIGpzb25fb3V0OgogICAg
            ICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIHJldHVybiBqc29uX291dFsiZGFzaGJvYXJkIl0K
            CiAgICBkZWYgY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcihzZWxmKToKICAgICAgICBjbWQgPSBb
            CiAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAgICAgICAgICAgInVzZXIiLAogICAgICAg
            ICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgIi0tdWlkIiwKICAgICAgICAgICAgc2VsZi5FWFRF
            Uk5BTF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgIi0tZGlzcGxheS1uYW1l
            IiwKICAgICAgICAgICAgIlJvb2sgUkdXIEFkbWluIE9wcyB1c2VyIiwKICAgICAgICAgICAgIi0t
            Y2FwcyIsCiAgICAgICAgICAgICJidWNrZXRzPSo7dXNlcnM9Kjt1c2FnZT1yZWFkO21ldGFkYXRh
            PXJlYWQ7em9uZT1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmVn
            cm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAog
            ICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyAiICIuam9pbihjbWQp
            KQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNz
            LkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGFscmVhZHkgZXhpc3RzLCB3ZSBqdXN0IHF1ZXJ5IGl0CiAgICAgICAgICAgIGlmIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSBlcnJuby5FRVhJU1Q6CiAgICAgICAgICAgICAgICBjbWQgPSBbCiAgICAg
            ICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgICAgICJ1c2Vy
            IiwKICAgICAgICAgICAgICAgICAgICAiaW5mbyIsCiAgICAgICAgICAgICAgICAgICAgIi0tdWlk
            IiwKICAgICAgICAgICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAi
            LS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgICAg
            ICAgICAgXQogICAgICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVj
            RXJyOgogICAgICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAg
            ICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5v
            dXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJu
            Y29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCiAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRv
            IGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAg
            ICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVj
            RXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2Us
            ICItMSIKCiAgICAgICAgIyBzZXBhcmF0ZWx5IGFkZCBpbmZvPXJlYWQgY2FwcyBmb3Igcmd3LWVu
            ZHBvaW50IGlwIHZhbGlkYXRpb24KICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQgPSBUcnVlCiAg
            ICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICJj
            YXBzIiwKICAgICAgICAgICAgImFkZCIsCiAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAg
            IHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICItLWNh
            cHMiLAogICAgICAgICAgICAiaW5mbz1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAg
            Ii0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVn
            cm91cF9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAg
            ICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNz
            LlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4
            ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIGNlcGggdmVyc2lvbiBub3Qgc3VwcG9ydGVkIGZv
            ciBhZGRpbmcgYGluZm89cmVhZGAgY2FwKHJnd192YWxpZGF0aW9uKQogICAgICAgICAgICBpZiAo
            CiAgICAgICAgICAgICAgICAiY291bGQgbm90IGFkZCBjYXBzOiB1bmFibGUgdG8gYWRkIGNhcHM6
            IGluZm89cmVhZFxuIgogICAgICAgICAgICAgICAgaW4gZXhlY0Vyci5zdGRlcnIuZGVjb2RlKCJ1
            dGYtOCIpCiAgICAgICAgICAgICAgICBhbmQgZXhlY0Vyci5yZXR1cm5jb2RlID09IDI0NAogICAg
            ICAgICAgICApOgogICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gRmFsc2UKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICBqc29ub3V0cHV0ID0ganNvbi5sb2Fkcyhv
            dXRwdXQpCiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJhY2Nlc3Nfa2V5Il0sCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsic2VjcmV0
            X2tleSJdLAogICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAgICIiLAog
            ICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX3Bvb2woc2VsZiwgcG9vbF9uYW1lKToKICAg
            ICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKHBvb2xfbmFtZSk6CiAgICAgICAg
            ICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRo
            ZSBwcm92aWRlZCBwb29sLCAne3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAg
            ICAgKQoKICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYsIHJiZF9wb29sX25hbWUpOgogICAgICAg
            IGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0
            dXJuCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1l
            KQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0
            KGlvY3R4LCBUcnVlKQoKICAgIGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAg
            ICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgog
            ICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFf
            ZWNfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5y
            YWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAg
            ICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScs
            ICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMgdXBwZXIgY2FzZSIKICAgICAgICAgICAgKQog
            ICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIu
            b3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9l
            eGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkgaXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRl
            ZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAg
            IGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAg
            ICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVt
            ZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSk6CiAgICAgICAgY21kID0gWwogICAgICAg
            ICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIsCiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91
            cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1f
            bmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgIF0KICAgICAgICB0cnk6
            CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJw
            cm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9y
            OgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            ICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgXQogICAgICAgICAgICB0
            cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRl
            cnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQ
            cm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vidm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3Jv
            dXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVkIgogICAgICAgICAgICAgICAgKQoKICAgIGRl
            ZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9zZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9
            IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAic3Vi
            dm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGluIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVz
            eXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICBwaW5f
            dHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcsCiAgICAgICAgXQogICAgICAgIHRyeToKICAg
            ICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nl
            c3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQg
            cGlubmVkIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3Vy
            bCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5f
            T1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAgc2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsi
            UkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXQogICAgICAgIHJnd19lbmRwb2ludCA9IHNl
            bGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAgICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCAr
            ICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1pbi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7
            ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5j
            b2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAgICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAog
            ICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAgICAgICAgICAgICAgICBhdXRoPVMzQXV0aChh
            Y2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5kcG9pbnQpLAogICAgICAgICAgICAgICAgdmVy
            aWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRp
            b25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgICAg
            ICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJsZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3
            X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAgICAgICAgICByZXR1cm4gIiIsICItMSIKICAg
            ICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYgcjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8i
            KSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAne3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiAoCiAg
            ICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAgICAgICItMSIsCiAgICAgICAgICAgICkKCiAg
            ICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3JhZ2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9p
            ZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3Vw
            cG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBv
            bmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAg
            ICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAg
            cmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVuZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBp
            cyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5k
            cG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUK
            ICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19z
            a2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAg
            ICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBi
            YXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBp
            cF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYgZXJyICE9ICIiOgogICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0
            aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3Vw
            cG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3VwcG9ydGVkOgogICAgICAgICAgICBmc2lkID0g
            c2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJnd19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3df
            ZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAg
            ICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAg
            ICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgICAgIGYiVGhl
            IHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0n
            LCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGluZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBh
            cGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZhbGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3ty
            Z3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBoIGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgICMgY2hl
            Y2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRl
            IGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQgZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2
            YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNlIGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAg
            IHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSIsCiAgICAgICAgICAgICAgICAiLnJndy5y
            b290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19w
            b29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3df
            cG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNfdG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAg
            ICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6
            CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tfcmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHJl
            dHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlz
            aXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmlnX25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmln
            KToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2NvbmZpZyAhPSAiIjoKICAgICAgICAgICAgY21k
            ID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgcmd3
            X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICAiZ2V0IiwKICAgICAgICAgICAgICAg
            ICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICByZ3dfbXVs
            dGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAgICBdCiAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9j
            ZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJv
            ciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAg
            ICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIu
            b3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29k
            ZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0dXJuICIt
            MSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYgY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9f
            YXJyYXkoc2VsZiwgdmFsdWUpOgogICAgICAgIHJldHVybiB2YWx1ZS5zcGxpdCgiLCIpCgogICAg
            ZGVmIHJhaXNlX2V4Y2VwdGlvbl9pZl9hbnlfdG9wb2xvZ3lfZmxhZ19pc19taXNzaW5nKHNlbGYp
            OgogICAgICAgIGlmICgKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9m
            YWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAg
            KQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9t
            YWluX3ZhbHVlcyA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIG9yICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVy
            ZV9kb21haW5fdmFsdWVzICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAgICAgICAgICAg
            ICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICApOgogICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgInBy
            b3ZpZGUgYWxsIHRoZSB0b3BvbG9neSBmbGFncyAtLXRvcG9sb2d5LXBvb2xzLCAtLXRvcG9sb2d5
            LWZhaWx1cmUtZG9tYWluLWxhYmVsLCAtLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIK
            ICAgICAgICAgICAgKQoKICAgIGRlZiB2YWxpZGF0ZV90b3BvbG9neV92YWx1ZXMoc2VsZiwgdG9w
            b2xvZ3lfcG9vbHMsIHRvcG9sb2d5X2ZkKToKICAgICAgICBpZiBsZW4odG9wb2xvZ3lfcG9vbHMp
            ICE9IGxlbih0b3BvbG9neV9mZCk6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCB0b3BvbG9neSBwb29scywg
            J3t0b3BvbG9neV9wb29sc30nLCBhbmQgIgogICAgICAgICAgICAgICAgZiJ0b3BvbG9neSBmYWls
            dXJlIGRvbWFpbiwgJ3t0b3BvbG9neV9mZH0nLCIKICAgICAgICAgICAgICAgIGYiYXJlIG9mIGRp
            ZmZlcmVudCBsZW5ndGgsICd7bGVuKHRvcG9sb2d5X3Bvb2xzKX0nIGFuZCAne2xlbih0b3BvbG9n
            eV9mZCl9JyByZXNwY3RpdmVseSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybgoKICAgIGRl
            ZiB2YWxpZGF0ZV90b3BvbG9neV9yYmRfcG9vbHMoc2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToK
            ICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9yYmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYu
            dmFsaWRhdGVfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgaW5pdF90b3BvbG9neV9yYmRfcG9vbHMo
            c2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToKICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9y
            YmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYuaW5pdF9yYmRfcG9vbChwb29sKQoKICAgICMgdGhp
            cyB3aWxsIHJldHVybiB0aGUgZmluYWwgYXJncyB0aGF0IHNjcmlwdCB1c2VzIHRvIHByb2Nlc3MK
            ICAgICMgdGhlIHByaW9yaXR5IHRvIHNldCBhIHBhcnRpY3VsYXIgdmFsdWUgaXMsCiAgICAjIGNv
            bW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmluaSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVz
            CiAgICBkZWYgZ2V0RmluYWxVc2VkQXJncyhzZWxmKToKICAgICAgICBhcmd1bWVudCA9IGYiW0Nv
            bmZpZ3VyYXRpb25zXVxuIgogICAgICAgIGZvciBhcmcgaW4gdmFycyhzZWxmLl9hcmdfcGFyc2Vy
            KToKICAgICAgICAgICAgaWYgc3RyKGdldGF0dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSk6CiAg
            ICAgICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxhZy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRl
            cm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxhZy1uYW1lLAogICAgICAgICAgICAgICAg
            IyBzbyB3ZSBjYW4gZ2V0IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAg
            ICAgICBhcmdWYWx1ZSA9IGFyZy5yZXBsYWNlKCJfIiwgIi0iKQogICAgICAgICAgICAgICAgIyBk
            byBub3QgYWRkIHRoZSBjb25maWctZmlsZSBmbGFnIGFuZCBhbHNvIG5vdCBhZGQgdGhlIGJvb2xl
            YW4gZmxhZ3Mgd2hpY2ggYXJlIHNldCB0byBGYWxzZQogICAgICAgICAgICAgICAgIyBiZWNhdXNl
            IGNvbmZpZy5pbmkgZmlsZSB0cmVhdHMgYm9vbGVhbiBmbGFncyBhcyBUcnVlIGFsd2F5cwogICAg
            ICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGFyZ1ZhbHVlICE9ICJjb25maWct
            ZmlsZSIKICAgICAgICAgICAgICAgICAgICBhbmQgZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBh
            cmcpICE9IEZhbHNlCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIGFyZ3Vt
            ZW50ICs9IGYie2FyZ1ZhbHVlfSA9IHtzdHIoZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcp
            KX1cbiIKICAgICAgICByZXR1cm4gYXJndW1lbnQKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAg
            IyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBj
            bHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRl
            X3JiZF9wb29sKHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNl
            bGYuaW5pdF9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVk
            ZWRfa2V5cy5hZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19k
            YXRhX3Bvb2xfZGV0YWlscygpCiAgICAgICAgIyBkb3VibGUgc3RyaW5nIG5lZWRlZCBmb3IgdXBz
            dHJlYW0gZXhwb3J0cyBvZiBmbGFncwogICAgICAgIHNlbGYub3V0X21hcFsiQVJHUyJdID0gZici
            e3NlbGYuZ2V0RmluYWxVc2VkQXJncygpfSInCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BB
            Q0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJL
            OFNfQ0xVU1RFUl9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgp
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1
            bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RB
            VEEiXSA9IHNlbGYuZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0
            X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktl
            eSgKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZGFzaGJvYXJkX2xpbmsoKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoImNsaWVudC5jc2ktcmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2Vs
            Zi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIi
            KQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19N
            RVRBREFUQV9QT09MX05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfbWV0YWRhdGFfcG9vbF9uYW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQ
            SEZTX0ZTX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAg
            ICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJn
            X3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9H
            Uk9VUCJdID0gc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBD
            ZXBoRlMgbm9kZSBhbmQgcHJvdmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFw
            WyJDRVBIRlNfUE9PTF9OQU1FIl06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNl
            bGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQog
            ICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJP
            VklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBI
            RlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAg
            ICAgICAgICAgIyBjcmVhdGUgdGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAg
            ICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAg
            ICJjc2kiLCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAg
            ICAgKQogICAgICAgICAgICAjIHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAog
            ICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9V
            UCJdOgogICAgICAgICAgICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUs
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAg
            ICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAog
            ICAgICAgICAgICAgICAgICAgICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEi
            LAogICAgICAgICAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0g
            PSAiIgogICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAg
            ICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lO
            VCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BP
            UlQiXSwKICAgICAgICAgICAgKSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkK
            ICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNf
            UE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2Vj
            X3Bvb2xfbmFtZSgpCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSA9IHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9t
            YXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMK
            ICAgICAgICApCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X3Bvb2xzICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5
            X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2Vs
            Zi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZ
            X1BPT0xTIl0pLAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90
            b19hcnJheSgKICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX1ZBTFVFUyJdCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAg
            c2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9M
            T0dZX1BPT0xTIl0pCiAgICAgICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5
            X3JiZF9wb29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRf
            dG9fYXJyYXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAg
            ICAgICAgZWxzZToKICAgICAgICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9s
            b2d5X2ZsYWdfaXNfbWlzc2luZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJF
            RklYIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0
            X21hcFsiUkdXX0VORFBPSU5UIl0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X2VuZHBvaW50OgogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAg
            ICAgICAgICAgICBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5f
            YXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAp
            OgogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwg
            InJlYWxtIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBz
            ZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0
            aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVf
            bmFtZSwgInpvbmUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAg
            IGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAg
            ICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAg
            ICBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAg
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9L
            RVkiXSwKICAgICAgICAgICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAgICAgICAgICAgICBlcnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxp
            ZGF0ZV9yZ3dfZW5kcG9pbnQoaW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVy
            cm9yLCBzZXQgdGhlIFJHV19FTkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAi
            LTEiOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRp
            c2l0ZSBwYXJhbWV0ZXJzIG9yIG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMu
            c3RkZXJyLndyaXRlKGVycikKCiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAg
            ICBmb3IgaywgdiBpbiBzZWxmLm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQg
            ayBub3QgaW4gc2VsZi5fZXhjbHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3Jp
            dGUoZiJleHBvcnQge2t9PXt2fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdl
            dHZhbHVlKCkKICAgICAgICBzaE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAg
            ICBkZWYgZ2VuX2pzb25fb3V0KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkK
            ICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAi
            IgogICAgICAgIGpzb25fb3V0ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJleHRlcm5hbC1jbHVzdGVyLXVzZXItY29tbWFuZCIsCiAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgImFyZ3MiOiBzZWxmLm91dF9tYXBbIkFSR1MiXSwKICAgICAgICAgICAgICAgIH0sCiAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2st
            Y2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZpZ01hcCIs
            CiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHNl
            bGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAgICAgICAg
            ICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5nIjogInt9
            IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAg
            ICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZnNp
            ZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAgICAgICAg
            ICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgt
            b3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91
            dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAgICAidXNl
            cktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYgJ01PTklU
            T1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1lbmRwb2lu
            dCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05J
            VE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9OSVRPUklO
            R19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9uaXRvcmlu
            Zy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0ZXIiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiTW9u
            aXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAg
            ICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVfU0VDUkVU
            JyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicgU2VjcmV0
            CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNS
            RVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JC
            RF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAog
            ICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgK
            ICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAg
            ICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ1c2Vy
            S2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAj
            IGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            ICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAg
            ICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMg
            aWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBq
            c29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5h
            bWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119
            IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAg
            ICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0
            X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRoZW4gb25s
            eSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9M
            SU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAg
            ICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            IHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAgICAgICAg
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0
            X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQgbm90IHNl
            bGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6IHNlbGYu
            b3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            Ijogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmImNlcGgt
            cmJkLXJhZG9zLW5hbWVzcGFjZS17c2VsZi5vdXRfbWFwWydSQURPU19OQU1FU1BBQ0UnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAg
            ICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRf
            bWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsn
            Q1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIs
            CiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAg
            ICAgICkKICAgICAgICBlbHNlOgogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJCRF9NRVRB
            REFUQV9FQ19QT09MX05BTUUiXToKICAgICAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAg
            ICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgt
            cmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICAiZGF0YVBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9P
            TF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlv
            L3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0K
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJuYW1lIjogImNlcGgtcmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3Rv
            cmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25l
            ci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICMgaWYgJ1RPUE9MT0dZX1BPT0xTJywgJ1RPUE9MT0dZX0ZBSUxVUkVf
            RE9NQUlOX0xBQkVMJywgJ1RPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUycgIGV4aXN0cywK
            ICAgICAgICAjIHRoZW4gb25seSBhZGQgJ3RvcG9sb2d5JyBTdG9yYWdlQ2xhc3MKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXQogICAgICAgICAg
            ICBhbmQgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXRvcG9sb2d5IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAg
            ICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidG9wb2xvZ3lGYWlsdXJlRG9tYWlu
            TGFiZWwiOiBzZWxmLm91dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xP
            R1lfRkFJTFVSRV9ET01BSU5fTEFCRUwiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5WYWx1ZXMiOiBzZWxmLm91
            dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xPR1lfRkFJTFVSRV9ET01B
            SU5fVkFMVUVTIgogICAgICAgICAgICAgICAgICAgICAgICBdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAidG9wb2xvZ3lQb29scyI6IHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNy
            ZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9j
            b250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lf
            UkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBI
            RlNfRlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZz
            TmFtZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAg
            ICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFt
            ZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQ
            SEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JH
            V19FTkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNz
            CiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3Mi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            ZW5kcG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAgICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAi
            a2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAg
            ICAgICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VT
            RVJfQUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2Vs
            Zi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAg
            ICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdS
            R1dfVExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0
            IiBzZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5k
            IjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJl
            dHVybiBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vy
            c19wZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlz
            aW9uZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNl
            bGYucnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2Vs
            Zi5ydW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2Vs
            Zi51cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25h
            bWVfZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAi
            Y2xpZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBw
            cm92aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0
            aGUgcmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2Fz
            IHVzZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1
            bHQnIHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0
            YSIKICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAg
            ICAgICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3Jh
            ZGUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lv
            bnMoc2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4
            aXN0cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVu
            dGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29u
            X291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAg
            ICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJp
            bnQoZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICBy
            ZXR1cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAg
            IHNlbGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5n
            X2NhcHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNl
            cikKICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAg
            ICBjYXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAg
            Y3VyX2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAg
            ICAgbmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            Y3VyX2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4g
            Y3VyX2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAg
            ICAgICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBm
            b3IgeCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAg
            ICBdCiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8g
            bWFpbnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3Qu
            ZXh0ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNh
            dGVzIHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1z
            IGluIHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5l
            d19jYXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4
            IGluIGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAg
            ICAgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAg
            ICAgICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFw
            cGVuZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tl
            YWNoQ2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRo
            IGNhcHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBj
            YXBzLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRf
            dmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pz
            b24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29t
            bWFuZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHBy
            aW50KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihz
            ZWxmKToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9u
            cygpCiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAg
            ICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRl
            ZF9vdXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3Vw
            cG9ydGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQog
            ICAgICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmls
            ZSBhbmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRw
            dXRfZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3Jp
            dGUoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJq
            T2JqID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2Vw
            dCBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1
            dGlvbiBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJv
            ciBhcyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQg
            T1NFcnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBv
            dXRwdXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0
            ZG93bigpCg==
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:62cd98508ab41f9526f10c55f6c8054475a1b029230f255fe7461fb750fd1def
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        version: 4.18.11-rhodf
      entries:
      - name: rook-ceph-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: rook-ceph-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: rook-ceph-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: rook-ceph-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: rook-ceph-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: rook-ceph-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: rook-ceph-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: rook-ceph-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: rook-ceph-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: rook-ceph-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: rook-ceph-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: rook-ceph-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:bbfddd7cb4a6cb7a131a0cef174054890fd39956d63e737c92fe8233ad8f98e4
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5a35593357fbf280561a1d4a64ecc0d5c2044409acff3ce6cb490a4a61462a4e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:791ac512370af5cde1e089de3def760e7dbce69badc03137051520ce316d63e2
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5dc66b7c486d1604c4bbb04c85544147c43264840e597958e6c5ca83e7e34c0e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:ddfdf850c9238a6988889e1840efe3175f03dad2489a5651e21cac62ac44e23e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.connect.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:19b163f055da43946643057c07500784210d02db868d2aa10f25a6501d767d3d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.connect.redhat.com/cpopen/common-service-operator@sha256:13f5b6699ab505d02550b09e100f7e205c7fde3f2bc36317025bc795be216761
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - registry.connect.redhat.com/isovalent/clife-bundle@sha256:6d2ba514e40559fe3568dd0107f5e6648d43b543b3dbe4319e4114ca05fe20a6
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/vault-config-operator@sha256:696bfe3769605870286e228f463dc2582202d6aeb8f199d78d87f153c68d7401
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-emrcontainers-controller@sha256:cde2c51ca18d0969be0549e37ecd03cc430be3b4746d53b8744cb552a2c475df
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:29ec87ce725d2c277824312607915bb7de92e18015c388543173dbdadd5834a5
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:c906c822fbab57c1eb95880b0dfb9cd04100c86222a5d5cfe6342298e3bb1b9e
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:e26bc4b183f69a448552f9fe740292d4e803165d341d7d66047688d01fcd060c
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
        version: 4.17.13-rhodf
      entries:
      - name: ocs-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:a4b5197db1ccda9934f30bfd63e41b0bef430ea9bd224cd3dc022466742b82b5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:0fcc80541d8966f3ea3502da1c4fdda24de277d8d73d0752b46e7d69fbcf7038
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
        - registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: ocs-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:9fcdfa4615b91a5bb919aff71fc4df046365b2ee6d540b5f5f47f9ccaeb9877c
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-003
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:9f5b63888cbf4b1c90e3803d0017a01ad9c0c22693b978c9ee8299bfc5337832
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:212ab9249707ca8ff835186d88e7888af18118fcbe98d948a3156b9700307a79
        version: 7.6.12-opr-003
      entries:
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/ipm-bundle@sha256:68c38ac59cf0d6cc145eb694e5b5a724da733caddb82c5585ef132e880dde544
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/community-operator-pipeline-prod/cluster-manager@sha256:a4c8a54148538b371e5cea3b2f2d4109f7b6bc4a7f2e5b52a840c9340fbf5e47
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-recyclebin-controller@sha256:9d916ed0af37d60d90f9b4711b1544b9373bd247476278c9c4041c9dbce6acc9
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/openshift-community-operators/oadp-operator@sha256:f8c5c6e59f13db8847080ea6d964da6e50e56f7d373bcc4e4278b1a8a80395e4
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/openshift-community-operators/oadp-operator@sha256:b7b4d340157513da2ebef6bc7bc262967d0e825e2baa1f2bc2ab2e170457e015
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a6399dfd6fe0e8dc9f59fac4382320d23f9e1709e48ed6da7db69e76354b2646
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:0a86a1808d9906eb57886d3d3cd5c04cf9578a80860342132d36ecfc411bf807
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:b02bab3af52a2469ced96618f1d122464568ba21ebc0a3a945b78ec4e6bdc80d
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a614ec94c2468e34e129e9ecd4bc0d1895eb67b96e77eb5ad94cd8bd507860f5
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Net AI
      provider-url: https://netai.tech
    name: netai-demo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netai-demo-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deployments.netai.tech/v1alpha1",
                "kind": "AntennaDeployment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netai-demo-operator"
                  },
                  "name": "antennadeployment-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Networking
          com.redhat.openshift.versions: v4.18
          containerImage: ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
          createdAt: "2025-07-09T11:33:08Z"
          description: Net AI Operator that uses EnergAIze engine forecasts to scale
            Red Hat OpenShift resources.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Net AI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntennaDeployment is the Schema for the antennadeployments
              API
            displayName: Antenna Deployment
            kind: AntennaDeployment
            name: antennadeployments.deployments.netai.tech
            version: v1alpha1
          - description: Demo is the Schema for the Demo API
            displayName: Demo
            kind: Demo
            name: demos.deployments.netai.tech
            version: v1alpha1
        description: POC Operator that uses saved EnergAize forecasts to manage OpenShift
          Custom Resources
        displayName: netai-demo-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - energaize
        - scaling
        - scale
        - manage
        - forecast
        - cells
        - 5g
        - networking
        - net
        - ai
        - network
        - telecommunication
        - xapp
        links:
        - name: Netai Demo Operator
          url: https://netai-demo-operator.domain
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Net AI
          url: https://netai.tech
        relatedImages:
        - registry.access.redhat.com/ubi8/ubi@sha256:68fecea0d255ee253acbf0c860eaebb7017ef5ef007c25bee9eeffd29ce85b29
        - registry.connect.redhat.com/net-ai/demo-operator-bundle@sha256:b1a30de1e984da0487a7d042bd940876bedb5a2f43dd73c7895e684a3a35d81e
        - ghcr.io/net-ai-tech/netai-demo-operator-bundle@sha256:0783aba7b039a569b49de2a61ef6dc6ff1d1f301d93c26d623b7f652321871c5
        - ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
        version: 0.0.1
      entries:
      - name: netai-demo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: netai-demo-operator
    provider:
      name: Net AI
      url: https://netai.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:42102a2e462e27dd1de28ca3f7db02d227c44736f2d57c05037a83ecbff6dd9f
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:b405d6148aa680bf31bdc6185c1702916c3219983d8482f27d89ed2f93659aa7
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/ipfs-operator@sha256:03a61ef328feb626408d42904cdfb9d45f66e022d5a1e1088a15b74455487743
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:c60664e5da8eb23bbd918b1c13b83565b1f33a1aa8961b425232fa1ef4d1a6f1
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:82cafd48db8107dd141064ce7530468b11804c8c54235a317eb450a7606bac3e
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:5d9b56300a3b17178467954b533bee19622f8e96fe9d4ac9745ead2c59b9ec5f
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-efs-controller@sha256:338a28600b08d1ead95ed409e3326e03f3c5e02ed840d3be56a950db6b1de061
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-apigateway-controller@sha256:44d34788b43d103530115fc53b0252bfdbe6e07a692847a3d32aa21e45892ff9
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.connect.redhat.com/yugabytedb/yugabyte-platform-operator-bundle@sha256:4cee86e973206b5300e88af088aca700cde3b2e2e5e1c70b9a20289e9f592349
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3e0fde49a8e8c5f780a76087bbc42507a42fea0e1fea572ae05ba66d27338365
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3b36dcea36ebf4b2e95fac6691f6fc8f1242f7e5c38195a9af28bfad514c28cb
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:36dc3710ac11d86e1cf0a3c6994ae0e6feb6c8ff1e8110997f27f4e573c259bc
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:64cadb49077eff7863dd678c4aec70f4101f17ee8e0bf7ca64ce7010c9bf4b9b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:4981d96081cc3d10cd26f20aab792e9c195b05bdcd609367011b7e9c7e6759b1
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - registry.connect.redhat.com/loadcore/loadcore@sha256:7fca6883b092ac7955bb42f893f163403960c81b83e360cfd1eac802e57ea839
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        - registry.connect.redhat.com/tigera/operator@sha256:4d0d814fcfeb628aae8a05cb7a0a416ae85395197950085c907b6cdcf2fb4317
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        - registry.connect.redhat.com/tigera/operator@sha256:b1a8b56ff5e1bb5326e1bde4931303358410da9768f0ebf9c08e2c2d8661d742
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        - registry.connect.redhat.com/tigera/operator@sha256:cb7f3367ff064cd01dc636a43fd89d5f6c4e6a7004cb7c94b1758b2398e36d71
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        - registry.connect.redhat.com/tigera/operator@sha256:de9724b84ca493d3b14c8f801c1049ac4713166b00d9bf3ea0b1986aac5040fa
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        - registry.connect.redhat.com/tigera/operator@sha256:8e0c026f369f80e8b36d0c94f1fd60f7f53656e57dc5be0b8b12189546874909
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        - registry.connect.redhat.com/tigera/operator@sha256:1fb1007f7aad744f0e1b0f6a32022c2571ee8558e9db53c85976bbd2a891a6b0
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - registry.connect.redhat.com/tigera/operator@sha256:42a4d0d0e24752dbb51daebcfca4d4850d04e79578bfb8b41cdc2a2a289df4a2
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        - registry.connect.redhat.com/tigera/operator@sha256:7894b770bb02b2271e1414d67f300218cc646534c9f2a6917e51dc8bff8ba731
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/openshift-community-operators/awss3-operator-registry@sha256:9ab3a2cd00b8704a8bae8e9b75f2a95d0a3e60fa3cb52589db8a4e7641b73027
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/openshift-community-operators/syndesis@sha256:9e8c08793aebc8857dcd451fa2fde2b3f05b9ff4f3d3cc604f9f752193f58d42
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/verticadb-operator@sha256:b7691da906474a8ab829966bb3da4a846eaf12909c97bd0d9cb6a1e5e7ff6c03
        - docker.io/vertica/verticadb-operator:1.11.2
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        - quay.io/community-operator-pipeline-prod/verticadb-operator@sha256:be706a582a4b8cae487835e920288b41229834cd3477b7389adb363db4b4f2ca
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              },
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "ServiceFunctionChain",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "servicefunctionchain-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "servicefunctionchain",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "servicefunctionchain-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-10-15T19:06:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-dpu-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
          - description: ServiceFunctionChain is the Schema for the servicefunctionchains
              API
            displayName: Service Function Chain
            kind: ServiceFunctionChain
            name: servicefunctionchains.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-dpu-operator-bundle@sha256:4e98c0f629e938a0c9d011e6a675d45589b044a02b3f9551fc0e8f7cd172c8a0
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:acb5ab931f4d7a46bb195c59d156441af5e9092b2739bcb0379b89d47ccf4e81
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:5dedc4733081fc21bcb77dbf4f4c032bf418f00bc4371c29c7ce028f51833c0b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: dpu-operator.v4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: dpu-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: dpu-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: dpu-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: dpu-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: dpu-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: dpu-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: dpu-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: dpu-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: dpu-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: dpu-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: dpu-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: dpu-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: dpu-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: dpu-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:4ab17b92016288f1ebfb2106c826a17e75d041f0dba7eaaf7545193c3305a408
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - registry.connect.redhat.com/streamnative/function-mesh-bundle@sha256:e4377187c20a70d768a253bc0794f3152aa6fcee22ddbf2fc88f45756e9681fe
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.connect.redhat.com/cpopen/odlm@sha256:064b597625550ca75dbc9f3c8fd6b2de6e0b9cb7f42d604e871bfc668ad05ac1
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:eccacdf9ac2b0b1f6061aea457c705319f268bfa30631ed9919ae393f0ef7f71
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:cf330a9e405a9641c213a74baf28307a6e247c4f03cc0523d071dad8f5f9e2e9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:412403b096566965b2c92a0a55beed8ebd6d8113a7e1fdf5eb3334769929ca41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:c89d8be470c2fd052c5be7a7dbfb133c9e3528bfa5b3681ffb8b3223c4c6bac2
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:6c9e04df3f5ddc45788333a85f6aeff8953b008a4b165e9cc2d5a079c6c73ece
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:5611a2fd53859d55d5dfdfdd117e80a0f30be0a4489ad838d571e872d3b5158a
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        - quay.io/openshift-community-operators/horreum-operator@sha256:a9d82ce368b015e45b93f917d4069978a0ba323f3dd684d470cdfd262e0c9f05
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:6d9eabdb4f645aa1b8b6c08ca94f9ec89466cdc35b3cb02aaba22e160d8fd9ea
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:5484f8ecf079ee126d24c05e1bcb06467146b3b8e751e90e0fd68f3e933125f0
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:11ce1c35f9c3d9b5ef9539b4402eb8358df843b29ffd94af7a453e5d765e4767
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-wafv2-controller@sha256:cb9a426baec724225c0132973d5b90a83672f9b0b75451adc6d7fd858c79ab58
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-glue-controller@sha256:2e1e23de186f3c61be8c7a30f3b295fe3b8b101509cf30bd6b7bff055ab74c82
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.109.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
          createdAt: "2025-09-18T09:13:25Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/eclipse/che-dashboard@sha256:8e2285991a725e87d2a9131ea23e08ce97cd5bdf1118f53a55a140c2c97ca25d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:1df1b75f702952b650e21717b04de5f39c84ca8f55f4afa0f32644ef8638d03a
        - quay.io/che-incubator/che-code@sha256:3b5c7e1ae664c82ddb3eaedc8161eb72a3a1a22c6c5e3f342f4aad972a6282a5
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/community-operator-pipeline-prod/eclipse-che@sha256:545d98c512c99bc7bf4ccd02c1eb554bbc4bca7da8cc66e0e26934edbfd50c59
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/che-idea-dev-server@sha256:0344c9574a6544c4c0ef93a798a9705ca3a4624508a55c8f94c2bae7196f914a
        - quay.io/eclipse/che-plugin-registry@sha256:bd71aa6817eeda0449056e78a6e832b82dd37b65ec6cba2bac3ffb9defa4048e
        - quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/che-incubator/configbump@sha256:77bd0fc269ff3724583b43ec68e158fcd300e825e1b44e24e2b9e67ed28df6da
        - quay.io/devfile/universal-developer-image@sha256:f76ea662cfdb9e80b93bcb58a393aed8b52aa72e7da73762d30a3ad22c3f06d9
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/eclipse/che-server@sha256:e5a542da0484eb2223ac292e6005cf7e116bec15d377557f92949728e1f5c8bf
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        version: 7.109.0
      entries:
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/group-sync-operator@sha256:fe6fda2de67fc4996d3f388e11faa0130fee19453111c829ba0aaa5be5410d01
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:e6d5ef5583ae26a8312dc331bd6003fda6c7c248de15b6790a128d0cd03965dd
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:35e0ee6b3ec5cb654b709c5e4a76a1f08182d826b47f077031d9139d4c2cc308
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b876ec88e1e64c17572e956b1c7e4020f5a423c1275f7f57025823ff7f749c61
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b3a45189ae8fc5302fb8a9e15d5747bacce044537923621edf21e9d3b267a27e
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-bundle@sha256:6cd4cba02a9e5c22062108363f6a052734c398019a19f84afc18a7b05a71d578
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        - registry.connect.redhat.com/loadcore/keysight-wap-stats-dashboards-bundle@sha256:f69f61eb818a1f0f4035ae79487402237b739c55eb104fc4643a7d3ef8d7d0c4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ram-controller@sha256:a90b07270f0cfa8929e01569895ab9a209ee44dcc452a3e1f6e0e9a52c5f13b9
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecs-controller@sha256:280d5ec57ad4deb45c145e702242468a4cc04d23112c33f4e72722a269c900e8
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-object-store-operator@sha256:c5af6adc5e7f9bb2d5bf6025b1b8c3e33ce5b9349a0246df1a26b7ec91b2ea2f
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lightbits Labs
      provider-url: https://www.lightbitslabs.com
    name: lightbits-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightbits-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "CSIDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "csidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "csidriver",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "csi.lightbitslabs.com",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "controller": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "tag": "v1.21.0-ubi9"
                  },
                  "node": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "sidecars": {
                      "lb-nvme-discovery-client": {
                        "registry": "docker.lightbitslabs.com",
                        "repository": "lightbits-operator/lb-nvme-discovery-client",
                        "tag": "v1.21.0-ubi9"
                      }
                    },
                    "tag": "v1.21.0-ubi9"
                  }
                }
              },
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "cluster1",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cluster",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "cluster1",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "addresses": "192.168.40.46,192.168.40.47,192.168.40.48",
                  "apiService": "192.168.40.46:443,192.168.40.47:443,192.168.40.48:443",
                  "clusterAdminSecretName": "lb-secret",
                  "clusterName": "cluster1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          com.redhat.openshift.versions: v4.17
          containerImage: docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
          createdAt: "2025-08-20T11:35:15Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/supported-versions: v4.17
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '[]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/suggested-namespace: lightbits-operator
          repository: https://github.com/LightBitsLabs/lightbits-operator
          support: Lightbits Labs <support@lightbitslabs.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster is the Schema for the clusters API
            displayName: Lightbits Cluster
            kind: Cluster
            name: clusters.storage.lightbitslabs.com
            version: v1alpha1
          - description: CSIDriver is the Schema for the csidrivers API
            displayName: Lightbits CSI Driver
            kind: CSIDriver
            name: csidrivers.storage.lightbitslabs.com
            version: v1alpha1
        description: |-
          Operator for deploying and managing Lightbits CSI driver.

          By using the operator users can quickly and easily deploy and upgrade Lightbits CSI Driver.

          ## Requirements

          - Kernel support for NVMe-TCP
          - Kernel support for NVMe Asymmetric Namespace Access (ANA)

          ## CSI driver Supported Features

          - Filesystem Volumes (xfs/ext4)
          - Raw Block Volumes
          - Dynamic/Static Provisioning
          - Volume Snapshots
          - Volume Clones
          - Multi-Tenancy & Credentials
          - Volume Expansion (Online/Offline)
          - Storage Capacity Tracking
          - Host-Side Encryption (LUKS)
          - Publish ReadOnly volumes

          ## How to install

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/install_lightbits_operator.html#installation)

          ## Documentation

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/)
        displayName: Lightbits Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - storage
        - lightbits
        links:
        - name: Lightbits Operator
          url: https://github.com/lightbitslabs/lightbits-operator
        maintainers:
        - email: k8s@lightbitslabs.com
          name: Lightbits Labs
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Lightbits Labs
          url: https://www.lightbitslabs.com
        relatedImages:
        - docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - registry.connect.redhat.com/lightbitslabs/lightbits-operator@sha256:53db496f7a254632a25acf8974aca4041cbddb109499871622ffc2182e7aaa83
        version: 1.2.0
      entries:
      - name: lightbits-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: lightbits-operator
    provider:
      name: Lightbits Labs
      url: https://www.lightbitslabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        - quay.io/community-operator-pipeline-prod/trident-operator:25.6.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-09-30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
          createdAt: "2025-09-30T03:27:25Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.41429998

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - quay.io/community-operator-pipeline-prod/sailoperator:1.28.0-nightly-2025-09-30
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-testing/proxyv2@sha256:28f058629e7969e8441b677b308a09bf5d4968dbd5024d54df8472d1f43e1a8e
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-testing/ztunnel@sha256:f4895a66a99ca0c87e54260d6d5784068a48fe7fbfa25144debad420ea7223fa
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-testing/pilot@sha256:42d64978872f6be8540027002a361933e4fe69590458353e39d2643b2459bb3c
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-testing/install-cni@sha256:a8c8f6fb3048461e08b86056000a3034f7945df067d5ebf43cca9bcb0ffa3916
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.28.0-nightly-2025-09-30
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      - name: sailoperator.v1.28.0-nightly-2025-09-21
        version: 1.28.0-nightly-2025-09-21
      - name: sailoperator.v1.28.0-nightly-2025-09-19
        version: 1.28.0-nightly-2025-09-19
      - name: sailoperator.v1.28.0-nightly-2025-09-18
        version: 1.28.0-nightly-2025-09-18
      - name: sailoperator.v1.28.0-nightly-2025-09-17
        version: 1.28.0-nightly-2025-09-17
      - name: sailoperator.v1.28.0-nightly-2025-09-16
        version: 1.28.0-nightly-2025-09-16
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/install-cni:1.23.0
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/pilot:1.23.0
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:a5698068f57fb27c6e4b8c385878c4651967ae5f17412c6ea6cf8bd34bfe76a1
        - docker.io/istio/proxyv2:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/pilot:1.24.1
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:5416c5fc62c7d9498349ed771bf93bfc0f712ad8da0b8ee109dc49d1e684fa56
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.3
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/proxyv2:1.24.2
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:9cc4d5a9f46e4d05f8adb5232459002a073c101cc813330325afb6682e2138ae
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.4
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - quay.io/community-operator-pipeline-prod/sailoperator:1.26.3
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-sagemaker-controller@sha256:c962ee90891a05197edc08b4cf015fce1e2431cced90bd274e89bb87708420f3
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-bedrockagent-controller@sha256:6688b26f71500af0e2ffbd0ccc53926b0c41adaaf817a88f0fdfd6fb398075e7
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
          createdAt: "2025-08-26T06:53:59Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/hpestorage/xspc-csi-operator-bundle@sha256:658dd12fe1b92bf38dd24601043cb84e27a27b0d987a36a0ec7d99b0514574e2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:4c289e5023fd3e59b2cf40335eba9eb53caf562472f16dd97b370773a0ba608b
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: xspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.32.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90",
                  "kernelVfDrivers": [
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
          description: This operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Intel device
            plugins for Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              GpuDevicePlugin is the Schema for the gpudeviceplugins API. It represents
              the GPU device plugin responsible for advertising Intel GPU hardware resources to
              the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              QatDevicePlugin is the Schema for the qatdeviceplugins API. It represents the QAT device
              plugin responsible for advertising Intel QuickAssist Technology hardware resources
              to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              SgxDevicePlugin is the Schema for the sgxdeviceplugins API. It represents
              the SGX device plugin responsible for advertising SGX device nodes to
              the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fpga
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4
        - registry.connect.redhat.com/intel/intel-device-plugins-operator-bundle@sha256:4a422ca897caeabb9fc3be3dfef3746819868de1674ea61c657cad257fbddf12
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
        version: 0.32.1
      entries:
      - name: intel-device-plugins-operator.v0.32.1
        version: 0.32.1
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        - quay.io/community-operator-pipeline-prod/flink-kubernetes-operator@sha256:7c919847b88a5969d0c0830e4154512bbca28015cf6fb7fa70cd20824a726685
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:13013220d7376d23df4780fe65c3bea325bdeb7b7a89d5704aa52e948ed7a2d5
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-operator-bundle@sha256:3fb13020c6a34b329825d268dd70c3c9310075780768312dc52cac1e7134a030
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:aa2267c1c871699db37763e938125f290ace4aff2aca6f668b2ce01f7df92b8d
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        version: 4.18.0-202509100344
      entries:
      - name: aws-efs-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: aws-efs-csi-driver-operator.v4.18.0-202508260754
        version: 4.18.0-202508260754
      - name: aws-efs-csi-driver-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: aws-efs-csi-driver-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: aws-efs-csi-driver-operator.v4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: aws-efs-csi-driver-operator.v4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: aws-efs-csi-driver-operator.v4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: aws-efs-csi-driver-operator.v4.18.0-202505280604
        version: 4.18.0-202505280604
      - name: aws-efs-csi-driver-operator.v4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: aws-efs-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: aws-efs-csi-driver-operator.v4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: aws-efs-csi-driver-operator.v4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: aws-efs-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: aws-efs-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: aws-efs-csi-driver-operator.v4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: aws-efs-csi-driver-operator.v4.18.0-202503190402
        version: 4.18.0-202503190402
      - name: aws-efs-csi-driver-operator.v4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: aws-efs-csi-driver-operator.v4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: aws-efs-csi-driver-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: aws-efs-csi-driver-operator.v4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-operator-bundle@sha256:7f3d41672fa8c60725304f0d45e468a4f01defcab7352501737ffddfd32edded
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-operator-bundle@sha256:75cb78f7a40d3daedbc11cbbb5505e029328b4120f0de11d4f6423de943e92d8
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:279f5d817f9dc90b202924fc2b2ab59e5303e9dddc1d78d71017d65cdd66d565
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle@sha256:8cba51274389030d632c8b82255399217d485146f4797907134e3c4ee55cb4d2
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:3bc76999ece42fcb3acd913b32be3fa0e0da056832922819b81119dcfb5f1582
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:2aeba259864bbddaa8cdd46ab6d8c3495450a0081469a27b9acde53008d91bed
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:f58f60283f6cf9ea745f4a25378ef6690d98af8fda0eb63f663066631e71450b
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:48b37c9f9417ef442d896e79851b176505045b487bddf77bb4e6d91fd5ad3dd1
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        - quay.io/community-operator-pipeline-prod/node-discovery-operator@sha256:899baace103c58e7337ae5cd75e3764e6de26966aab7340037c612c26ca7cf87
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:53777788253eaa1199e1854d14782619aeec0df9844d86b83742fe0005671cde
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:e3a13ea8b84b9e1abd158d394e11113e5f6c04d879547fa1d3e8638f1ac979d8
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:31258400bb37f4b8837ef294820f8a728ae83f8031d4e01ff0392690a315a30f
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        - quay.io/openshift-community-operators/move2kube-operator@sha256:cdf288be5ea779e03900af9b57c4d922b8760b88da4fcbf88437f8031d679aef
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:236a53d6fad949285b6abafcdd1a4541a7748f952c0c96fa2e71500079579ba4
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:38d03dacca046dda3ba08c39943ec2e7db33f1c7bb897e84fe1308a0982c63f1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/community-windows-machine-config-operator@sha256:2b79f0e6b8bc069fac7d6cb5c86783fdc98ccae30d279ca0492f1eca96b6ea3f
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        - registry.redhat.io/rhcl-1/limitador-operator-bundle@sha256:4c56cd8462531cf8ba39316a0c6f7493240bfa4a6ace77352fe7ac1597e2fd90
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        - registry.connect.redhat.com/netapp/trident-operator@sha256:53a4889e1036af9f858c9fdb1bd58f59b758a301bc773379572cf279f930f5c6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - quay.io/openshift-community-operators/ocm-operator@sha256:ca6a3b24ca4ecfcab52ee0402e007d6f14b23fa7f3408d740923806989b170e4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-eventbridge-controller@sha256:3a284fb7884e50ed9976368f312c75fc1f1cb2ec361297b38ee72d0ac6b66fc7
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/global-load-balancer-operator@sha256:50771c320c8b7faea144f765535e89979f9f81e61d34c8dd2e86139fa68e222e
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:1f4fd5ed8e398b713e733fb552e7294f22b3927d4e017f2ac15ea20024d3de32
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:759dc7d6836f8c8830ec145e3088bddc1b63bd40ff3047079fa8f0e986e74c30
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/openshift-community-operators/konveyor-operator@sha256:04e63481c5104bd034283d5d4b0567167f62480d2c61154b1d16dbd41c344dfb
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/openshift-community-operators/konveyor-operator@sha256:d4c590033b1a2bcd67375465298e5cb6d136fe091abb12b8d009aa683c23ef62
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:848db711d9e8a329b689b40003293fce19c43faedd31afd80d9ae9d5c11cda49
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:e29f175a799597c84f0663a66767bba24feb78f802516056b35b5beb14d9eb41
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:2d58f12044b0951c87c851527240196db148acb9797e6ac49f9474a6e6039d71
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:7b8bdecbe3b3def1ec810ad53f604c7a71daebcfa7433cbb3a0db937580d087c
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:16c4f165143b8bab7f3eb08c815c2cae0dbf7fffd9ce7a625357be5806c24c5c
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:c28522543b039454d2d97b48198d099ac0b8211cbc58f13468d5f70b9f69b037
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:18ba5c587dc613f9c1a8bc9e66081e52f3a37eedeea5e75881e628e9c35ef460
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - registry.connect.redhat.com/nvidia/nim-operator-bundle@sha256:3dbd307d32a39e37a02ac8fa9c9fb15a0c2fd6dc5aea12d55abb061e86365124
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-directory-operator@sha256:9d93e3aeef3cfbba0b655de7783ba9fd1f4cb4ee5de4389005b7c31d971d4947
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opentelemetry-operator@sha256:5dc347a5085c10823a122fab0bd42afd5d9b9aa19ba17e8992e9492000886629
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/aerospike-kubernetes-operator@sha256:bfa83ff729ef8b3a4ebca27b02e223327d3e1351d8bc78a06c18873d288dfb02
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        - quay.io/community-operator-pipeline-prod/aerospike-kubernetes-operator@sha256:63cdb7ea6cd6bda4c6ab2c1f3467489c2cef6c8a95836763b9c88b20c43e8168
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - quay.io/openshift-community-operators/sap-data-intelligence-observer-operator@sha256:5217360266f62fc8ba087fbc730074c14aacf62dbd124fdc3aa46650d51a243b
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sap-data-intelligence-observer-operator@sha256:79aa33cb7e4b570fc9c962fc6bd59708c7e28e81e2b188608780fa085db91442
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/kueue-operator
    name: kueue-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kueue-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "policy": "Disabled"
                    },
                    "integrations": {
                      "frameworks": [
                        "AppWrapper"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "FairSharing"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameOptional"
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob",
                        "Pod",
                        "Deployment",
                        "StatefulSet"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "byWorkload": "Parallel",
                      "policy": "ByWorkload"
                    },
                    "integrations": {
                      "frameworks": [
                        "RayJob",
                        "RayCluster",
                        "AppWrapper",
                        "PyTorchJob"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "Classical"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameRequired"
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-05-01T20:58:29Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.operator.openshift.io
            version: v1alpha1
        description: Kueue Operator description. TODO.
        displayName: Kueue Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:de644d640f355815f3aa40ced5456bb5b0e958aeba8bfbdc4ab360fcdb90f26d
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:0c4d9cd97f7579adbf2afd238c953ca5f170408cf27e3b2bf4aa16c108a76881
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:dd765c446564d4397570f4bc056b55989f6c0bae0d70514874e67765c9c77889
        version: 0.1.0
      entries:
      - name: kueue-operator.v0.1.0
        version: 0.1.0
      name: stable-v0.1
    - currentCSV: kueue-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-06-24T02:48:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:30659b42f5053dd756190211454c01f049bbc7689dddda3ecd8c406b157873a7
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:9a0b85b0752135694193befe00d10f7abdc0c99c6b9f295d59b237fa0fe3e6d2
        - registry.redhat.io/kueue-tech-preview/kueue-must-gather-rhel9@sha256:e5306c95bf6c5f55be6e791fcb66fe9b34f4f982536f4f90f8ed74acf3f6c31d
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:6f0b5f62f6e549c6f6769dd9a358a14668d7244f9c20753d1690d3e09948588c
        version: 0.2.1
      entries:
      - name: kueue-operator.v0.2.1
        version: 0.2.1
      - name: kueue-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: kueue-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-07-15T01:52:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue/kueue-rhel9-operator@sha256:5db0af8088cce6e9e7a88faf80855b6eaa1dd8679c6a56dd5ee98c50a3a8e2e4
        - registry.redhat.io/kueue/kueue-rhel9@sha256:16007bd38c480b607b63c82b3ee8721441cc374ee31c04b9e329426314590195
        - registry.redhat.io/kueue/kueue-must-gather-rhel9@sha256:f5f624ef8dbb3676ed4e37335e226e76e5f886c1990c6e0e8c0e4762740ed98d
        - registry.redhat.io/kueue/kueue-operator-bundle@sha256:1bafac48c00fb47972f512ccd5d846efbbccccd2259395de17537b39834d89df
        version: 1.0.1
      entries:
      - name: kueue-operator.v1.0.1
        version: 1.0.1
      - name: kueue-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: kueue-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/kueue-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-bundle-certified@sha256:a9b8102ab0ec81e5bea84332ef20b023d06660b410e7e2deadb00904d2098613
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle@sha256:261b77cfb315c608505e5659b2d01dc4e1ea96cd37fae25f417cd7df148f4ce4
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        - quay.io/openshift-community-operators/snyk-operator@sha256:97908144406700ae9c038316661dc7904f3e0f4e01b1bbdf68259b59dfb86d6c
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cadee58c1c1040fad5a46fc9be5fae4a1944845c5e736c50fa192192d5c17a36
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cf782e9e2a62a2fc71c81032e6bf40f9ba15d37c33feff2df551edb111c4e17c
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:60a9e4b05e7eb17ebfe58dd55db96ab2c68580d9282117f74f21869cb63e4a73
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:1461576c2e06c2ee8d84fd40836184566a346ed596b99cb81fe6a08d0ad99096
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        - quay.io/openshift-community-operators/camel-k@sha256:2b8369f729c18384f45a458e88c9f551eb00f26792b40dbf286e6ea2837e9400
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:6154ff0f72c9c49f1b7ea0edbe93e0a3b04f6e63abbd851d0e6706796050d672
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:21ce00e81ce495c142cbe9611f494cfb26b52b3c09a507bee0572c1e1e3333d7
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        - quay.io/openshift-community-operators/camel-k@sha256:26df9d918e942f7fc637212df87a114b5edd52570da18cc212b3499354cdb993
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:11dac0881f5ed66b5d42b2d69b13aba31cd56e6b32157cb3ce977be197189b79
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        - quay.io/openshift-community-operators/camel-k@sha256:cba8cde86a870b31f7dae9a554f4c496ed81731f91a8e64cd62b0caa88efaa38
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:06d31bb777b37f8f7c8981efc24ca26a2f9c56a1c579243d0c7bdddb34c76a98
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-dependencies
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-dependencies.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          createdAt: "2025-09-11T15:25:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          deploys dependencies.
        displayName: Data Foundation Dependencies
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-dependencies-operator-bundle@sha256:6c2223357b672b1a59d22eeab742d3c241d87d5524935974876c9371ad650afa
        version: 4.18.11-rhodf
      entries:
      - name: odf-dependencies.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-dependencies.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-dependencies.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-dependencies.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-dependencies.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-dependencies.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-dependencies.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-dependencies.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-dependencies.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-dependencies.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-dependencies.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-dependencies.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-dependencies
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "debug": false,
                  "ebpfProgramManagerMode": false,
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
          createdAt: "2024-09-13T12:46:10Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:577ab5d4780ba01e130ca5713fb7b3764244cdcb937b6b00e6bb17638f15fea2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ingress-node-firewall-operator-bundle@sha256:6c6a76287353a02c05164f9a0a3e07d27b75f0e5d403eb1b47f1e9dc7c9883d8
        version: 4.18.0-202509100344
      entries:
      - name: ingress-node-firewall.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ingress-node-firewall.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ingress-node-firewall.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ingress-node-firewall.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ingress-node-firewall.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: ingress-node-firewall.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: ingress-node-firewall.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ingress-node-firewall.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ingress-node-firewall.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: ingress-node-firewall.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ingress-node-firewall.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ingress-node-firewall.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: ingress-node-firewall.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ingress-node-firewall.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ingress-node-firewall.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:e39a0b57e73d92c5db330ac9b8a184eeb400397f14c05c1ef75c952d91e4f5b2
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:5024be9a3190288983e3c42e8a2dbb49aa44cd596237ad9ed0e72de5090dc9b9
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:ece58f1e360eccf7d970dca2f9da0c4d99a692648f165f677f11ff080649cf62
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:6bd423056b555d39337c0e3a9c31e7732e80f34d0c01f0766a6aadb2aff586f8
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        - quay.io/community-operator-pipeline-prod/beegfs-csi-driver-operator@sha256:1a9f2aab026c2d1549be5b4624c56d65e813c619fdd5beefb062d5c9fb8981b8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        - quay.io/openshift-community-operators/nexus-operator-m88i@sha256:46cd06d477b4d09dbfca7ca4434ea58f881314138ed744c6390033bd6fb7dc0f
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-secretsmanager-controller@sha256:c92ea9b492b6585a816668bd2a77d65471e339e02e25e44c9558a34b352db593
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:151e7d9bc500b7252af46d0856e3e16559ff310a1d0efedcc5d914164bbdef55
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:309ba89dece3ee9901ed66c40bd28d71f20c7d91a7feff552e28ea0c772b9178
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:63bd0faf0b64e4672014205f9dcfeec6c436edd7dc2aaf83cd053d6e4569d90a
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:1b464c682cd29c8fb7bbc566a1575b9002ad61163c2b6e94138e0e4b0ce51140
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:8223bda16ac5ffac4b7407c849c4c30ee006ddafd34cad8a03d2af8314d56f58
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:d7333e377a5aace0d42e42ca936fa7a84d288b181eda4747f9367a9539f2bedc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-bundle@sha256:0064fcdc252a04010e605706be377417de4c9be4289358ec0f331d6a37befb0b
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        - registry.connect.redhat.com/loadcore/keysight-nats-operator-bundle@sha256:89bde716c2aa4dba0738d2481a2c4707c59a748f84349f9fb01cb430e69d7beb
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        - quay.io/community-operator-pipeline-prod/percona-server-mongodb-operator@sha256:8659c9f25245e541609079970c8193fc6a58a7d323c249a81d743de696dc15d4
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-rhmp-bundle@sha256:b66e3cfad24cf28e0a738179cab062667b0103c52cf2e84abb88d5e018e12b2c
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:27c5d91d62371f3cf0dea4d50044d8a6e0a6015a453c4216a030683ea2474c64
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:9fe87e1b0c49dd00a6ab0c35b68ebf641f347ac5b405d31f3c5ace02a7a30540
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:35d2c7c2849f89474fd5648a0d2781dde28131303cc8a597ce243a3e314fe0af
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:dd0ab94059f782aa31696b15e1b25787cd58bab4905fe64faec7e49bbae3d289
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator-bundle@sha256:0f20db33b1eb069a767089862412846aacb723dc22b44adeb9793495ad20ea50
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle@sha256:506afdc804dc30e0a720e8e6f151b75a32148844de42d90d8c0648e9108d9ea7
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        - registry.connect.redhat.com/elastic/eck@sha256:063a1f405b87751c5102693451ae5c2e9ad203c6b8a31b062a5b403bb8b2febc
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/fence-agents-remediation@sha256:6ab21048f86979bcc34a84965dc2a97467a79507ab03469fe8c6deed224881f7
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        - quay.io/openshift-community-operators/oci-ccm-operator@sha256:eaf63c6f9ee9e9c3562d6ced4cff091c1e401daa826ee3a92b321a5cc8f13da2
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:18:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:1a98d55c28a0679134e24b6aafb3c4dc9db0098094c8d2bcfaa41a63c33308e1
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: odr-cluster-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-cluster-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-cluster-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-cluster-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-cluster-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-cluster-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-cluster-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-cluster-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-cluster-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-cluster-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-cluster-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-cluster-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-cluster-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-cluster-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-cluster-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:16f820575acdca39c0b7e85edff7b910809f6d184a9a479de2bb2f687c94a6f2
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-cluster-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-cluster-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-cluster-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-cluster-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-cluster-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-cluster-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-cluster-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-cluster-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-cluster-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-cluster-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-cluster-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-cluster-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a93b382258a34e0b56ef6616a8490e5f44ef1c6aa5d29c0d08d2147ae5b440e6
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3170da029869d15a246ad31d0debd4b9c03ba13855fb877a754cfdffb2286b80
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a7ce92f7f4aadca2781c44839bc357615395587bc8b68a735642967cf8694362
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:093c6a66d4faa1d980e4319a048db8b2869eba8f13f40f02615d6696257e6719
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:5a19032dda9a0f584b41cbe9e729ed23e6003d90ebbb026eb9e7bb9769d4a4e4
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3b2cf972e1ced0a46d3559da4eb1a1b32c50457293ede4958a13538a84b45316
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/proactive-node-scaling-operator@sha256:0ae01e9adc67ff7341fa3faf51a67c02206c41342374cab0bb2e6079d84fef4a
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:857c1bdff46d1a1d5abb51f75a80b7944bb461687b179b6f1622103947b8c3b1
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:d101e2b9ab4d8ca7c87c86ea4e3630dcc82a9c5a78885179c8b16773c8aac587
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:ecdd5c5fc609128e0af751f87b797b71e4ab0200a9abaaca835ff9a3f293fb10
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:be099e23851cb4b6d15e9c44c4da8fb6cba74aab005597d305757d1dae8b0122
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:96119f61731cec7eb037c3a4a7c2c8bfdf9ef11fd5b08abc25e3d24ddf342b06
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:2ea267aac67763b4f160c4e6876c0226d2fa135ee00614814a59b4de1d91f4c4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:218905f7902e6ac80366856d6c5ae37cd6d167c81e9521df7c3c3d6619f6c425
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:09fe022370abfcefcf3abfe8b6debc2f201a0ac8b8dacaf1278b8a6bee2b5273
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:5b865231214649649a2a9fb06934f931503f68ed740e150be1c7c8fcbb6f1317
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:16245de1e88228c231ab2edb4e9380753eff982bb72febadd00cdaad286a8eb3
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:491e7aea221eb0e1d2992310e53d50db287cbf32f62d090bd842236b603d38bb
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:7de09e5394985cdec87f5afab65b3343b8a37c0eb7f778774b21933ae69d691c
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle-marketplace@sha256:28f388286bc4b4981182ff1ab9e927f59744dcb755392907fe65cd3b96a44519
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        - registry.connect.redhat.com/loadcore/keysight-rest-api-browser-v1-bundle@sha256:89ef7cf0dc8b27e58a63a82a701894a7cc4d0643bf0c8c8e20c9bb59d1a1b1bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-bundle@sha256:80ca5ad4634305808ac5a3e8ff656ce6ff69b3c16dccc5e03f42763fa89bc762
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/community-operator-pipeline-prod/openshift-node-upgrade-mutex-operator@sha256:5ac79c9193934440464a595434fe4241e038505d32650843afcad5f5bc2ce8a7
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        - quay.io/openshift-community-operators/mercury-operator@sha256:0c253fd002cb389db40b2438d4d1c9e953a84125dbf0f4b6f5e3a6efd87cf3a6
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:9db455e082454734d2a17811dcfec53d3d60c28df6e7fe8bc979a45b28a35061
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:8650430afd813fd749fcd80ad5f6fd0aeaab3df048e8fe34e6aa72a07ab801fc
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:ed063f507fa8e2f378d0adbde823c110484d0212d124cc72d67efa987e86444e
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.8.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:68b310ed3cfd65db893ba015ef1d5442365201c0ced006c1915e90edb99933ea
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:bcc5ed9560667c7347f3044c7fa6148e3867c17e93e83595a62794339a5a6b0d
        version: 0.8.1
      entries:
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/3scale-apicast-operator-bundle@sha256:29fe669a56a69adb537798c91b9359fa3b051da5ffa9f27789a2067a3a72330a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.7.9-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:ea2229948fea841343fcaa337bb7e2854acf5f563d3ffb0af8bda1c21ec2a91f
        - registry.redhat.io/3scale-mas/apicast-operator-bundle@sha256:d3a608b37aa089ea46e803696ba4957585b875f8a810f517e891f385ad105937
        - registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
        version: 0.7.9-mas
      entries:
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        - registry.connect.redhat.com/loadcore/keysight-wap-storage-bundle@sha256:0955dbda03bd7307364a1a458ab8d6ac5a8c9feb6150706dcede9676a6513b0d
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:591fc4b5867dff26df1952653423b33889c0f56ba3f756057694d4c0e396eeb0
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:3225c49cf30a146784533496ce478466034d4549ff9721752958bac342fdec77
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:b0fa6b0356d83626509289b33dfa669bd7d32542aba986a18d296b85ce63389a
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:fe759127885502dc66423530ff03e37973355c5566a1c27e988b59d71528dd2b
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:c0336e5079f255ab6cd38e4e2e61d0cb4c6a35a976528fb44409bfde57e70dde
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/openshift-qiskit-operator@sha256:633ce4190b02034d70b52d7e624fc9d417ef997607f9710a36e73bbc0963fe4d
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a77d39ded03271f2deb8b438abd4328a1846ee0d4886cce14d869ba61f84d600
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a6f3a3a8aa52de4441ef3249a76dcff1b6b6defba417a9fdbba37eb458adee04
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:6208f64c7fe75f39ab7511e2b480c493d2ecc8d1e76a1ac498a93a6a24e8beb5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:11d64a921a98b7ec1a2126a03e446a055a5ae1705dc2005bb8b73cbbcf7ecf99
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        - quay.io/openshift-community-operators/hedvig-operator@sha256:81d956c12057f18bf920044a4e39c27c99b2c23afdccc0ff9468f0484e1961d7
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-operator-bundle@sha256:f40e7955ac649c383134ab6e77bdd521b75b21ecd9b476436405d0674c7fc855
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-operator-bundle@sha256:d04033c2cd2d236f6319a3e35880b89c8f0c60d53cd32f6f097d07d6d27e5f2c
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-ati-updates-bundle@sha256:3d06a58b2a277a67339c22f3fe10c1daac1c8eabaac8540b01177f6f5688784c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatchlogs-controller@sha256:94145cae2b8cef1410acbb4f1c1eb571e71408341716af1742f27ddf4dc350b6
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/coherence-operator:3.5.4
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
        - quay.io/community-operator-pipeline-prod/ndb-operator@sha256:5f33c2a8181ea8e91844afd1f4042affbd9eb878266b6b2bcc6906f0829ed00b
        version: 1.6.0
      entries:
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - quay.io/openshift-community-operators/keda@sha256:50816f675c76d36dbc62846579150477018ef87df325861751f3fc02a70c8781
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/community-operator-pipeline-prod/alloydb-omni-operator@sha256:08fafb811f709780c0ccb3c22e3c6417cb14a70a9305e46a7b50457647a82102
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:ed530db9b4a29fb591ea208cd0c6b3b83b49537ff5a0a4f501b54ce667f9b303
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:31bdff225bd825d02ce0f468b1edc77541ce85b7b9111f14a583c418c984232f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:80198f65f7e3f9022a34e0342e9423807b855dcb7bda810172eebc9a27afb5c2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:f04d565eb5162fd5a9c78cd8443af6a3b184a660277537f1d6d575caa2e85441
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:773592dfb9964970f444871a8812e8b74db3f205faaa0bf148defa1a2b87b649
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:0b1771f697a700263c5b827dbde56a5c9bab610e9d7df55ef93885a4e54cc0e9
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8441322a0fce02df407573675b5b5d92d56de97c8aec72541b33d2dadc4050d5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:be93be0341649215fb001725aac0c3c5925343adbf1ad00c979b8c3d489512d6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b22190de3162d3e03481cae8904ce563912756eea7ac7e7feec0acbbdae5e6f6
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:c2f39be75173b4cc57498c5845b14c464bf9af92b49716df9661c8135ac2cf7a
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b2d58f4dc20eb7e17e09af0e8c5de55fa04f9f4f697e10c6be81a7a6a08381b2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:27efa0ab1d9c69047a924bc1dcd2b56108f5f390a6da56afb6ab7a37cb4578e2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/volume-expander-operator@sha256:b775df99f891e2cee9c2169e9c948a8d61d65b4cc90f36752b46a56e2d4be602
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/apimatic-kubernetes-operator@sha256:a5f68205291caa9bb7903180afc974fe86252ba87e671978c8b49e0206976d06
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: sonar-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        - quay.io/openshift-community-operators/sonar-operator@sha256:5e7a87f52955162f090c1d7fc69f6d0917444b553355a11c061ab60f289a17d7
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    defaultChannel: alpha
    packageName: sonar-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-documentdb-controller@sha256:8a991c76f281ec30258b6317bd623894b78a4f39d9790a1cbb32e6f1b7ff63f5
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        - quay.io/community-operator-pipeline-prod/dynatrace-operator@sha256:c475f9cc0f6caa2bb316ff42a35a40b1d4bf8ef7c8a3384d9c78eb205bda7076
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        - quay.io/community-operator-pipeline-prod/kaoto-operator@sha256:4b6cc43eba8efbe17ffec746431d6cf6b045886ff865673a08a0820abecebdcd
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:5ea2dd102074798d1ee005dc03fbbd24f4324df831f279355dc2dfc262c9d880
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:e4dec1f39f8e24043cf16703efcac2cc3b953d3f4c37b3956fce0c65ef829764
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:ed8230e90f0ab0586a743d5e760a7469345b1136971aad17ef675c1beccf1982
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:08995782514e496c24d888352e9d1e6b788dd87adb761e0c2e745be2975ffcaf
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:7b324e897fbcc881ececa529822d7b738f82cd4c3adf4cc54b66080b888f5d9c
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:e2a329fc74ed5f17e395367126c7e96197178e8f886d475b657a7910a4a14152
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:0218d68f97e89b48bbdd9c7cd750d4a666cf33af1cb8a76507c067a332a7b6bd
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:2bc56e149c14920662b549a5367f98daf9afad0df29b4746b195f6afc0c9c597
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
          createdAt: "2025-09-19T22:01:50Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
        - quay.io/community-operator-pipeline-prod/ack-eks-controller@sha256:120566a02e487aaa8823323423fccfaa377135037e2e1cb9f3ff14313cf3bfeb
        version: 1.9.2
      entries:
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        - quay.io/community-operator-pipeline-prod/pelorus-operator@sha256:f54c42a30675897a971cac278d5d8ea24edf3121570264c07cde04f11cc4e7a8
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:57af9e32c31dad769bda449ebe35e09ea135e8a7d06cddb42be4d1a09005f518
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ab3cb65b88c34fd58524a2a51cd510c611924e5860834965bad0b4418a8b6dd3
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
        - registry.redhat.io/openshift4/ose-local-storage-operator-bundle@sha256:d303fc63f5b849f073e34a7791ea92cdc3cff269cd1b593e27480ae96cac1a15
        version: 4.18.0-202509100344
      entries:
      - name: local-storage-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: local-storage-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: local-storage-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: local-storage-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: local-storage-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: local-storage-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: local-storage-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: local-storage-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: local-storage-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: local-storage-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: local-storage-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: local-storage-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: local-storage-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: local-storage-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: local-storage-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: local-storage-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
          createdAt: "2024-09-04T17:53:42Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:07cafb83ba90378c1fa60338cf951a9d7145ec245710637e90fdb508875dcc45
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:1d86075469312e84f3d16379efcd6130ec9e8b23a1343e3945f071d82beff6e3
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:dc10544e1bb24883f38373a0445944745282e5a69ed1744f428f98871a7a39c9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:31b3658dd79c18111c21e99aea52cc198f8e95b458a1a4f8d2bfc4d21e9bdacb
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:1b8008a16a0f53c09ecc5a219964194689a1aeb01acee8313aa49a9c26259429
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:a539064234088a4ee13b8a9666ca9507208fe38189ee9b8b673a38a58aa75599
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:b137e1345038c9e566767f24cce60c03ec278f5a2efc28fa08e0bd8bfa016486
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:c001dc96a1478a58f30461ab018e40bede13a78f3cdce869a70dcb8b9bba22b1
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:92a3162c705537147005a065cdddbb410e1863d80c08d8b0f27b7ff5588500cd
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:8fbfffb0bc3adfb4d72961dad381e1daba414bf4a7084f03b5208f6be4191ceb
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:035855c3d4b32a29232846d73dc398e81f941fd90a03e33241ca06f6e80b2204
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:f86e96edb0418811df328debdee93b4a51cf7f651134ee35e9c50b9dbfc29fd8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:084f1e54ef0d7f457d5a7cfc57b567ad8d85237c12f707c82f7a30a54e3652ba
        version: 1.7.17
      entries:
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:9f5479ac795eb50fa5ea597828b2d50ddaa4878faa1a07b3de0e611cb4af5fc6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/operator-bundle@sha256:bc8a47d5f8ffe3a2fe48819f08c82525677131cfb4f3c0facd82470c0f10acee
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        - registry.connect.redhat.com/loadcore/keysight-nimbusmosaic-dusim-bundle@sha256:97d5f906ef83272a8d7a2a249f38f8b9b186f8e9b64729a049a7eac0dbda8b1a
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        - registry.connect.redhat.com/loadcore/keysight-wap-appsec-bundle@sha256:e5049593a52fb26c124de7005b45919f26fa0430293f18283a20c8cde981b025
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cli-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager/cli-manager-rhel9-operator@sha256:be0fc230fef111cff36a8b846bbac9aaa98aae94ad9200416e841aa487121de4
        - registry.redhat.io/cli-manager/cli-manager-rhel9@sha256:a6b7857c1acd5173c770e6ba1d03947eb13804d2512f7f7cd6bb96e5e5cdaa17
        - registry.redhat.io/cli-manager/cli-manager-operator-bundle@sha256:74c88f1cfef4cc31b794fc5ed8dcc10d3b0e6869c228d808f1d98819e69554b6
        version: 0.1.1
      entries:
      - name: cli-manager-operator.v0.1.1
        version: 0.1.1
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:a0e4c432590e65818e7ebaab127d483d68fdc5ccae9a13820665a8aef122fde7
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:242d754d5fe1f2a07091ae93da284b8912e0f4b4c973200cf6d1241e6abb88a9
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:437a786a301ea720cdb5651dff878a9f51470bab46ba28daae4928c7074c9157
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
          createdAt: "2024-11-21T23:48:37Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVSNetwork is the Schema for the ovsnetworks API
            displayName: OVSNetwork
            kind: OVSNetwork
            name: ovsnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-operator-bundle@sha256:4c8271d0bf6d94061fe8df8110b041cb0e1a840fb6263315f552166fc5053e73
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:2e71d4125a35f6e7bd2a04cdd8cb6aae058d745eda372ae3d3fd37783778f486
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:57a57f4f3a0adb8b7234683f2968352988ec9a414824f4f444cca096825a5df5
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:45835aba95af72aa8e73b630d1d2b4c1079217531f025f4715a2eb8e8856205f
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:8b412541a797aaff1b49aa229e310957a8d7da03a3d38e6bbfd65f2ca36c2bdd
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:6eb244a301ae5395b69b1bfcd0baac37c094f4c38a1a330c0a70fb81e71f6157
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:3b87d49b200ac4fdeac97f0e97a46001e91c8f4680e2975e90bd94ec37864403
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:0f18db6a5bbc952bee8f3e4ac9e93e807d1e85cd132affe94b1f4005c1e128ce
        - registry.redhat.io/openshift4/ose-sriov-rdma-cni-rhel9@sha256:65864bc235689ff38b63d2a7b2c72874ee155241a493fa9429fbe87d5100cfcf
        version: 4.18.0-202509100344
      entries:
      - name: sriov-network-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: sriov-network-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: sriov-network-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: sriov-network-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: sriov-network-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: sriov-network-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: sriov-network-operator.v4.18.0-202505260733
        version: 4.18.0-202505260733
      - name: sriov-network-operator.v4.18.0-202505121333
        version: 4.18.0-202505121333
      - name: sriov-network-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: sriov-network-operator.v4.18.0-202504221235
        version: 4.18.0-202504221235
      - name: sriov-network-operator.v4.18.0-202504160002
        version: 4.18.0-202504160002
      - name: sriov-network-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: sriov-network-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: sriov-network-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: sriov-network-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: sriov-network-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: sriov-network-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle-marketplace@sha256:4145f7a04d7dbdda226b75af4e4cabca63a803f290b1581ea4ae0dc223eca7df
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        - registry.connect.redhat.com/questlabs/alloy-operator@sha256:fe4e83fc4a50e1aa04e6a9ba2a57acc737077b86681c1d7c597646fad337f221
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - registry.connect.redhat.com/kove/kove-operator@sha256:04bf970760c4dfe910465b0ef6a18e9c17fa70f6388131f12ae0aa0cd646f663
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vertical-pod-autoscaler.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "autoscaling.k8s.io/v1",
                "kind": "VerticalPodAutoscaler",
                "metadata": {
                  "name": "myapp-vpa"
                },
                "spec": {
                  "resourcePolicy": {
                    "containerPolicies": [
                      {
                        "containerName": "*",
                        "controlledResources": [
                          "cpu",
                          "memory"
                        ],
                        "maxAllowed": {
                          "cpu": 1,
                          "memory": "500Mi"
                        },
                        "minAllowed": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    ]
                  },
                  "targetRef": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "myapp-deployment"
                  }
                }
              },
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "minReplicas": 2,
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "safetyMarginFraction": 0.15
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
          createdAt: "2024-12-05T23:41:17Z"
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1beta2
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1beta2
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: Vertical Pod Autoscaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - verticalpodautoscaler
        - vpa
        - autoscaling
        - scaling
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-operator-bundle@sha256:f5df5d878ef0a0b494c9bafe99c35f3152d30f6c426be789fcd636dd13eff6e6
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:5da084c4dc81251a707db1bd9347c13b39a1e01660120aa2e6a47f9d1d981fe9
        version: 4.18.0-202509100344
      entries:
      - name: vertical-pod-autoscaler.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: vertical-pod-autoscaler.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: vertical-pod-autoscaler.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: vertical-pod-autoscaler.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: vertical-pod-autoscaler.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: vertical-pod-autoscaler.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: vertical-pod-autoscaler.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: vertical-pod-autoscaler.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: vertical-pod-autoscaler.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: vertical-pod-autoscaler.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: vertical-pod-autoscaler.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: vertical-pod-autoscaler.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: vertical-pod-autoscaler.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: vertical-pod-autoscaler.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: vertical-pod-autoscaler.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator-metadata@sha256:d6d0e71f5451e8255edfff87ef28ba7743fe7057efd53b25bbbd01bf9e1d0307
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/lws-operator
    name: leader-worker-set
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: leader-worker-set.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "LeaderWorkerSetOperator",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-04-14T08:12:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lws-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/lws-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LeaderWorkerSetOperator
            name: leaderworkersetoperators.operator.openshift.io
            version: v1
        description: "Leader Worker Set Operator is based on the [LWS](https://lws.sigs.k8s.io/docs/)
          open source project.\nLeader Worker Set: An API for deploying a group of
          pods as a unit of replication. \nIt aims to address common deployment patterns
          of AI/ML inference workloads, \nespecially multi-host inference workloads
          where the LLM will be sharded and run across multiple devices on multiple
          nodes.\n"
        displayName: Leader Worker Set Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - leader-worker-set
        links:
        - name: Leader Worker Set
          url: https://lws.sigs.k8s.io/docs/overview/
        - name: Operand Source Code
          url: https://github.com/openshift/kubernetes-sigs-lws
        - name: Operator Source Code
          url: https://github.com/openshift/lws-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/lws-operator
        relatedImages:
        - registry.redhat.io/leader-worker-set/lws-rhel9@sha256:affb303b1173c273231bb50fef07310b0e220d2f08bfc0aa5912d0825e3e0d4f
        - registry.redhat.io/leader-worker-set/lws-operator-bundle@sha256:f641cf3b2d6659e26eb0ae49c67a3416152a2fbc2f26f490889a2bd169b9ea06
        - registry.redhat.io/leader-worker-set/lws-rhel9-operator@sha256:c202bfa15626262ff22682b64ac57539d28dd35f5960c490f5afea75cef34309
        version: 1.0.0
      entries:
      - name: leader-worker-set.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: leader-worker-set
    provider:
      name: Red Hat
      url: https://github.com/openshift/lws-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:72a54c9457e311001b25c1b414b7001c6d7f9a34b89efb1b4198084586386f31
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:7a04dfd687dfe1646bd59a39c9e88781ffb815085bbd15f6305f6eac5f59bc49
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:d1a9246def81537e3417a6d48d1e818459583db658b54a87ca5af6911c8a3f53
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:e88676c0346c245b315f1a3bf7f85b10515bae3620c786fedda0533b39d2a400
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:66bd3eba0b68c317fe128c6713488cb6374f9647095768d3f7eed8120bf02e77
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        - quay.io/openshift-community-operators/authorino-operator@sha256:8fe34f94f87c567401cc2bbbbcc3a3e719364b80a2b17851496f86c85af5c09a
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino:v0.20.0
        - quay.io/community-operator-pipeline-prod/authorino-operator@sha256:92ad93f27c34f1406fc05a248d04709b0109adf37f2f9424f48fc6a73d9d359e
        - quay.io/kuadrant/authorino-operator:v0.16.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/logging-operator@sha256:d3ce3322f7f25c7b8bfabdd6e9fd32b833239fc39592240e0ccff3a13757fb17
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:f252d6f89a7cdcc520c3af461ca57d0dfabf736b9c38277d77477c4c880e182c
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:97d03647567ddb6f04edea39244bc0814e8f7e557251060d9b241a5fabe04dfc
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:f3497acc3216d820014cce7332e9d0683f4ca6438782ace02210192f787ca097
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:23f2f66e9f6d79d05e2be5a10c394834f093f4677ab06d8935ed33980408e8d3
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:c6d6c3e0408b851cdc9589cc0ee642d77397cdfaa40f50cf42557c5aaf9c0690
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:ff314ab983699455c32ab0ef27bb552d00825974a2d7088df2000f09afd8a2e2
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:5f10cfc1bc8571f44a99d88755adade0743106219673bbdd2e4d3a04ee714778
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:9f90cd7ab8c0ab5da6eb12f7f0dbd177df62f07c38fb759a97989899cfcbe75d
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:71ff93b137b6569f8fdefbd446cf12cd95477e4294bdea98fb5318e5d7d1cdd6
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle@sha256:2a28b64f7321df23928c8226b3eb3161cf520d38cc57d875b3cbfc828c164edc
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        - quay.io/community-operator-pipeline-prod/hazelcast-platform-operator@sha256:1209876fc9103eb2d69c34e97e4e2e0b384a9a778ae7e6a53222b8552fbf2f26
        version: 5.13.0
      entries:
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kms-controller@sha256:a2c60484a6556949318aa1d730c601b5bb05b1795b945a607209ef3434cce72f
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - k8s.gcr.io/redis:v1
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - quay.io/openshift-community-operators/ripsaw@sha256:ff8d796d5bd6c8cb0e155b3ced631122397758eb472470223e30352eeddba480
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:68cdc37990bc7da8a6bbe3a4908378c8026e5ff980bf8c44c0277f689006fc51
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:97762fcd8bfeec64fe4d3d43a6e60bc705ff2c8773354357e545adf31c8f85c5
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:4fa7a30eb3f679a495aeb96794056ffd003237686b2b60e693c8cabf6bc15bae
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6dce519730a6eb73bc18bef8372cdde65c29af3cb52eb78afc808d8582526fd3
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:9f2c01e88b29154ba7a0f70a1c69a49c86a567808cf057cb5d115159278cf2cf
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:edec37e045212f74d160724ffc7b2c8ccf4e60eb95629aae80360494d78d5433
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:465f0bef1f167b1d16ba9ae5f2d9a1133d11786b6809698a84c5c962387dbf6f
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8be098242f971aefa61a4a24d2c6630ecaa7856ece842166c38887a66bce1bda
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6afbc7ffc98918e3a557b0b3d9b2c3aa80caa47b9a73205fac9e6572c66f4df7
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:eefdd05485167d490f0011d9e5d2bb6112db1daea25940c79e1f55b3af254548
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:05e66583757da895fd23c3af9830ae1ac6a84d107107e5b6bbaece57e4d79c9a
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:02f53d3395b2b98f92a96b466c83ca15c79a75d82451e2ed8ba921375e727c7f
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8abe0994052e2a328e54c9d30c920dc331e9b816a4606234a3023b9ef65ec4ed
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:ef3f85e1227e886a131d3ddec3fb431aef6d3daed46115b288246b9ace0e69e2
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8b9b06ead5e014b920c82034ded615413dca53cf36f804005408726c9bbda66f
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:51db956f1472ae0498c57b9488448c34cf7e24afab6b9997eb3689b001d73277
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:392f9299c75a3ca4f593eaebdfe515145696f12db9f990bb004cfa58dd6bf0ef
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:df4e4dce45f87e684e7e1044a7edb2456160141afb6ba57b9af28088ffbac407
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:fecad1a330421341e6946252f95735ed177a690eab6250be097413ea29ce9473
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:9dcf24fecf3ee0d999601c276ac15e8634c275590f2c42b8a6bcb83984bf1445
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a9950264bfefc536b4e0a0adc2348f92aa41da55dd50a2324ddad0253291b67c
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:4d59e60cc75de82bd3f3a063237cef8a5598d315f6d53e803d37589a94929b0e
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:40d17d4d0632c272779da152f710fb7c2de50acc3791eb0b9fe445a13c0cb102
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:8e321dc1bf74ba331b62d21bff4d8993956ab935cf5be32f01e7f7aa6897bf22
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:16fbdaaf303ab40202d49438737f3307926594a5259081eeeb39595b7ea41f8e
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-zero-trust-workload-identity-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zero-trust-workload-identity-manager.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "ZeroTrustWorkloadIdentityManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/part-of": "zero-trust-workload-identity-manager"
                  },
                  "name": "zerotrustworkloadidentitymanager-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
          createdAt: 2025-06-16T11:15:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: zero-trust-workload-identity-manager
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/zero-trust-workload-identity-manager
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterFederatedTrustDomain
            name: clusterfederatedtrustdomains.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterSPIFFEID
            name: clusterspiffeids.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterStaticEntry
            name: clusterstaticentries.spire.spiffe.io
            version: v1alpha1
          - kind: SpiffeCSIDriver
            name: spiffecsidrivers.operator.openshift.io
            version: v1alpha1
          - kind: SpireAgent
            name: spireagents.operator.openshift.io
            version: v1alpha1
          - kind: SpireOIDCDiscoveryProvider
            name: spireoidcdiscoveryproviders.operator.openshift.io
            version: v1alpha1
          - kind: SpireServer
            name: spireservers.operator.openshift.io
            version: v1alpha1
          - kind: ZeroTrustWorkloadIdentityManager
            name: zerotrustworkloadidentitymanagers.operator.openshift.io
            version: v1alpha1
        description: The Zero Trust Workload Identity Manager is an Openshift Day-2
          Operator designed to automate the setup and management of SPIFFE/SPIRE components
          (like spire-server, spire-agent, spiffe-csi-driver, and oidc-discovery-provider)
          within OpenShift clusters. It enables zero-trust security by dynamically
          issuing and rotating workload identities, enforcing strict identity-based
          authentication across workloads. This manager abstracts complex SPIRE configurations,
          streamlining workload onboarding with secure identities and improving the
          cluster's overall security posture.
        displayName: Zero Trust Workload Identity Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zero-trust-workload-identity-manager
        - ztwim
        - spire
        - spiffe-spire
        - security
        links:
        - name: Documentation
          url: https://github.com/openshift/zero-trust-workload-identity-manager/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-agent-rhel9@sha256:0281d6ab233a685ba0f739629584aacd9254e0b13a697fe8708efa41289e746e
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-operator-bundle@sha256:1974903a3e2b692eedab2602395fdc4a6e2ffbc79ed4e633705c05ce7771e1b7
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-helper-rhel9@sha256:a39c80c6e1495ea166a858c8ed03304c6455420c077d937c99bf8732a6829790
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-controller-manager-rhel9@sha256:67f0e6670195c84515e7d78139a2951e463db11445740fe1d9a449a7f338556c
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-oidc-discovery-provider-rhel9@sha256:5ba7e94b7ea06852d32002b055df556ff2a2fdc9a2e8f793b837c4856d773782
        - registry.access.redhat.com/ubi9@sha256:861e833044a903f689ecfa404424494a7e387ab39cf7949c54843285d13a9774
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:7d170ba9b3de2e88dc0d0a6b0c13630ca3b23eed96ea5e3c725516046d7c7fc7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-csi-driver-rhel9@sha256:a16ac0a4363f7cffada39a2f17aa42f5de3faf78b3688ae5ebf2ecf3a4759cb7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-server-rhel9@sha256:70f97c613feeaadfed99d37aabd62e1cb2247825292ad9be806bc2fd065c9166
        version: 0.1.0
      entries:
      - name: zero-trust-workload-identity-manager.v0.1.0
        version: 0.1.0
      name: tech-preview-v0.1
    defaultChannel: tech-preview-v0.1
    packageName: openshift-zero-trust-workload-identity-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-operator-bundle-rhel8@sha256:0fbbb0b32adadc5746251103d25bba6bd365b2ff60e289e4694061fc92c3ce6e
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        - registry.connect.redhat.com/paloalto/pcc-operator@sha256:c7200a8c652ee544b0e0bb4d9427eab3a3b2dee6ca61ee88abb8052316382673
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/sap-hana-express-operator@sha256:0fa4f29bd7669690be87a08405bed95d3ead2293a125b28149de2b10b02d2cd6
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/community-operator-pipeline-prod/falcon-operator@sha256:3916adf5083c83838564b1ed039decbadb2efa4ec288699ef0137a66983a3cc4
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.connect.redhat.com/hitachi/hnas-csi-operator-bundle@sha256:79e6ba3db78aa55a9f3839a172843d284d91e9cdaae9cb2c8e7e705f3d03115d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:60b293c23587072b3376f83afee95b0207228c5c7b80483b025adda572c3551d
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        - quay.io/community-operator-pipeline-prod/ack-rds-controller@sha256:80b0d6e33fc94471d05e05f6d370be3a84fcb5c2044486e007d57360eb69d9d7
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-messaging-topology-operator@sha256:4c62fb6d686a20a28437905d4e1a9b93ac05853b4dbf29fcce34421a43268724
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:b10c121395f5bc46d61fdeddd35a0c8d9ae4e67a9a77d8ada90e0ea53666323e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle-marketplace@sha256:9d84d46dd1eb74afa18466c73766402b46c2885ef483600aff83065f8065d8d1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        - registry.redhat.io/jboss-eap/eap-operator-bundle@sha256:0db70367917668964e2f4f5efdcef864ccc3f065848023e52caf4abdec728d9d
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:096946dd896b6d9d937eb6529c3a23a84ccb129649c5a0389d0ab83989c28b53
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/jhipster-online-operator@sha256:5f2436297c026832e5c15241d60090da581daa5502e352ea33d77f22ef3e7a52
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        - quay.io/community-operator-pipeline-prod/argocd-operator@sha256:483854621b95dcefa03d18a970600b271726422ae3362479f73dcfca9085c92d
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/mtr/mtr-operator-bundle@sha256:190a6fa2a6a9b6b182cdf9cab18814d05c203b54ca18dc533e553fa5dcca779e
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:2808a0397495982b4ea0001ede078803a043d5c9b0285662b08044fe4c11f243
        version: 1.0.2
      deprecation:
        message: |
          The preview-1.0 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:a91cee14f47824ce49759628d06bf4e48276e67dae00b50123d3233d78531720
        version: 1.1.0
      deprecation:
        message: |
          The preview-1.1 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:4477810692429ac498b6ec1b295e6ead0f8fea25e91f6ddf17255d7e349b82e1
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:ff2bb666c2696fed365df55de78141a02e372044647b8031e6d06e7583478af4
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        version: 1.2.2
      deprecation:
        message: |
          The stable-1.2 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:8da62ba1c19c905bc1b87a6233ead475b047a766dc2acb7569149ac5cfe7f0f1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        version: 1.3.3
      deprecation:
        message: |
          The stable-1.3 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-load-core-bundle@sha256:f62a7f8eb74d0d5309fede6c9c0201d1f0c3736f93f1aeec7446089d64fce28f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/openshift-community-operators/sigstore-helm-operator@sha256:ff4e548a8a003bd23e5076362bafb2e22f77b8445c5e3585d99cd5ce26fcaf56
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:5a5e2f775760aa68c85d0f69b700d23ef2a6be7bb415bd96401d03eb6077f90c
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:79a9191f0cd5d11af004ead87ba9eb5359c67fd658c346db4c3cea3b16b23987
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:54ea11add0a1d988e0c965968f9a65e011978190f76fa2f826f1378474bac1d4
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
          createdAt: "2024-10-11T14:05:47Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8337924489a1747d4bbca6dd6ac42210731cf721b549b5830846b271621e2ac1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:6d0aeffbccaf9df83b42c530bf82707aa7c1cc16f72a5ef30b66fbd562c01ab3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8c05a303be5c536f4051ef291475b578d4d38ca0946fb4b4dd8bca376ef432aa
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:e91c2e5473f3e24c185dd25e9843f333e6d7419d9c5f4d4396669717b5969683
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:c543dbbb66fcac02345e1d3c12333780721bf049dca0f0a27d4c55b15a5d3e14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:a65780c5680a6fbbd0e3d0eed423b14a9c9b3d7d17f801c1eb6eb6b777384ca5
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fd6fb83bbeb9010f7e91f0afc91beac0bf38cb79c2de8f533a2bd07b1b238262
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:936fd68d7659683de7789fcf34d2e5873bd4bd1534c01c2d80501647f7c3c442
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1dd8f0f08ff1323d910fea8adfa13cef64c2fd2f01968e4d9d55f71cf42835d3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:365115407964d182de7e7564438c6fd089e2c0bcc0f94ec48621ca0cc2bf0fe4
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4bea07e101740afc36a1617363a79b1c474efcacd584f6e475a53b3f2395a079
        version: 2.7.2
      entries:
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:3d0409c1a0b17c51d3dbd5aa0371edd597e3c5ba194aad987fe867dd804d1923
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    - currentCSV: mtv-operator.v2.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
          createdAt: "2025-09-09T18:06:51Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:41cc528bf93000095b4eb56aaff9a6dfb124d1711a2f31f0745163ff934ac96e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:0c8ebfdd52053dbcc221a3df7d603163a821dd9e0ac6f54977e081adc6cb4504
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a299acb18821691cddba52904140cef8bf12632fbf2e752b33118a574a5ff394
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:ad3b548405e8c4c075ef7d4c0eb0d0430684066ed7eab4563a1f8378eaf3d939
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4b37b1aecbbf087257be5ba2bb559f2de005aacb4cd371192cdde1a9891739cf
        - registry.redhat.io/migration-toolkit-virtualization/mtv-vsphere-xcopy-volume-populator-rhel9@sha256:c0a4a232ff8adfb7d57e0761251ad0d79a0f02d27c4bd813f62f6b943deb6585
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f90f4eaaaa02b289704b19dd171282f3d11f370ed6cea992e55964729ad5116e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:258efbf073f78f8b5b48fdda22ef055e2315046cfecc9644da51959289c836d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d57b39674260f3bb1ae4c164022c2c7be3cc9645f1c46a83e43f40965633451a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:aeaa9ffcb0ffbb0f69b2712a3031f7e2dfc0824273b5c9c0f6349bc701a0e89e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:e438c5fccc957ad26ecfc3633561342d3b6f2a6d5a4c7839109564008de62222
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:83a523392a9c715972bba7044c7c3d3944594cc89778449e1f59fbd6a2659baf
        version: 2.9.4
      entries:
      - name: mtv-operator.v2.9.4
        version: 2.9.4
      - name: mtv-operator.v2.9.3
        version: 2.9.3
      - name: mtv-operator.v2.9.2
        version: 2.9.2
      - name: mtv-operator.v2.9.1
        version: 2.9.1
      - name: mtv-operator.v2.9.0
        version: 2.9.0
      name: release-v2.9
    defaultChannel: release-v2.9
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:0ee0bd2449f183277af845e312257b4a1ced4de865831b15e1758a397d8d29eb
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:2cf058c819b40512094f84f4d374457455370ec4994d34cee6e39b9d87841f05
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        - quay.io/openshift-community-operators/camel-karavan-operator@sha256:6dc78ecdf6e1936872626c02d41986d6553c1500002d85692c97e86fa77c129d
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:f182078b0077b6df6d47ddcab0204e963c1475486bd61b8eccb73b46eb8b2995
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c8b3505ed491c1bf6a913f161715c8ac483d8d9adf1fd29916ea4edad987cd67
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:af5a13351f960c0938c68604b772f24743c0e997701838a7784c53406471e72a
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:0d9f5293d69eebd91fffd83ae371141e640cab8f9e2d559a83714ce0f94fee08
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:d6675617c19b4ac7c1913078eaf9ffbe3d03761b570dffa002d9de11f337f19f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:2a35a0ef852cb4d76e963df50449712a06ff13391c30675aaf931e21cb561b8a
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:00f7ce0148988e93679d4b53b45ae8d2ad54f308b908940e7b28504318f718c6
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:c352f5633beef2eaaa324dda66b9b484e5ea79efbbc3b781079f2e9d10c880a9
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-bundle@sha256:47a7bafff2715fbef6130ca0b84560b6c2358275821b72ae54ad46020e02bee5
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hive-operator:1.2.4953-c248a51
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:be43412a2e4bad2cd9d15e4fbabe0292585c5bbbd65da12d519f30f17d2fe594
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:822fd8e65c241cebedd835bfebf67645199561cfdd6148d5afffb8b35e499ae3
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        - quay.io/openshift-community-operators/hive-operator@sha256:9a3cb682afc66ad662045b3d8ff214203cc00be60f66f74564820f6dcf99cd69
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        - registry.connect.redhat.com/fortanix/ccm-node-agent-dcap-operator-bundle@sha256:545724c273dfd6a1fa923d468b2d054f39ec3186d01586a8b3fe494c0f22f8dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/ndmspc-operator@sha256:b1f47ce0cd6930d902dc3538a1a87140dbe9257dd50773a518fc2350185f90cc
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        - quay.io/community-operator-pipeline-prod/klusterlet@sha256:9efef9a034c9f12505a0d81bc9697c4a3ab7e3d05d1e7aaa4180490d70ddf6a2
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:30264b9f49912cc0e6bed59001869e3168caace20b292c9dad4cc2d4947d289d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:d7cc6d2575eb252fb3e2468106e801f192da7844b0b153dc3c6a12cf7cc0a8fe
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:60e99bb91309e448c652898dac070a6d920fda164f27879d94632c88a9e61ffa
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:11edce56874ff1ee666752ec5fa521f3f6ae78c52cf1f289ceb97a86e6c083aa
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:5e4c319ce871a0dcbab7fff23325e6233c9dbcb2592fdeb1a9ba7c57ac2a63b2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:d86bb21fdeec35b1fc9789023a558750da76e036e0790dcfce97dcd137dcb6fb
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:7d94521a5cbdd34f8307653a3e1b596ec86415b2165fabdbabc8ed8bed35473d
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:3d5ad015a953708cc53ac2ea92b1b1d99c870a801b26e790d41458adc0a958e5
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:fea5b977249013181616f60b95c5e6d224cd0f248717171d946ff0e4d6eed6c0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-memorydb-controller@sha256:7f2e65d817479fcc5658cac745cef325b2bfd6256eeed8b819f1713afd59fe68
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patch-operator@sha256:437f5b7c661d3e9bc2c8ac13b0c8f3565f685b9e5ef52791312bc64d55e2e476
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:21f4648e0e71b33ef24b4921ab34b9e2b5e90eac25ade735d51331ac4f7c4cf3
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:2ed6bfdf37e36c63cde3633443553dee410201200ad351e6d91a5651f8e20ecc
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - quay.io/helmoperators/cockroachdb:v5.0.4
        - quay.io/openshift-community-operators/cockroachdb@sha256:f42337e7b85a46d83c94694638e2312e10ca16a03542399a65ba783c94a32b63
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        - quay.io/openshift-community-operators/cockroachdb@sha256:d3016b1507515fc7712f9c47fd9082baf9ccb070aaab58ed0ef6e5abdedde8ba
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/netobserv-operator@sha256:73af4040707ce79baa3115deed1e9148436b248ef3570f1175a6d31fdf9596c4
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:394d122efcf0da16bb68bfef1292395cda1bdb21909ae55f8fcc54a16814f1cf
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:62f93c5989d1d8217a3bac076addac05c0906a1a1f1eeaacb87bcf256812fdf9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:ea8c060444aa8e89dfbd90d61d6ec3900ab4d6634f6aa8067face1e6974b9065
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        - registry.connect.redhat.com/loadcore/keysight-wap-notifications-bundle@sha256:9c2ebb98476bf3374b71a838c394017e7d31aaa045bd52b3a66924b6f6b7b067
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.connect.redhat.com/cpopen/ibm-licensing-operator@sha256:a9f68d204b6acb50ac5023e5f2314ad0d06e06483f87290381d1e92476a79579
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator-certified.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-operator-bundle@sha256:4512f786fed0d9b2285b5a2827ec3b597452f90c217e27398f2f0f552377d844
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:48b37460dd7245cf661009267c3353962f0f195efd6a3007c9cc70b51b91a9e3
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:9949a581df3960a770117b6cf6b9fee1c63ea12ca09ac2d677f1cbe0a83fbd93
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - quay.io/akaris/sosreport-operator:0.0.2
        - quay.io/openshift-community-operators/sosreport-operator@sha256:f0d28a3e2ac2e1cbd0c307797b728ad89ac33375135d72054a7b765e274e4bea
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/node-maintenance-operator@sha256:80dfaab6f7a5c34cebb45c874ce134bf60e3697e14a95b69f1972dd98ecc172a
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=4.11.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:75eaf7b8b71e1a15e002f6cc5430e5e50039699d7b52c3150f5f71b1595f00c5
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-operator-bundle@sha256:41ece1ea773fd610cf8a57483a4d2846e0bb3ce81db701021a3e0dd91e13b7c8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:f7159e880ec55da49ac774db0c454800055e7cd3f37ba94fbcb629d08aac1146
        version: 4.18.0-202509100344
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-operator-bundle@sha256:4f8908ba97517826fda14cab3624ed42a6348fb96f0be4abf9cd2a7b68f7d3e9
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-wap-migration-bundle@sha256:3deff8b48ab1a6f0e5fd0f8315bc1e191da55525cb3cf0f315f2f7afdb47a6c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - registry.connect.redhat.com/noiro/aci-operator-bundle@sha256:80e25268ab5aa6ea418deae5ccc197fbffe2b9cdf4f27894f5beeb9315f6b123
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:20a4f2c56cb186fff4e8d31021d84b690512cc65c3c9629ffe623a0213ca9b0a
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        - quay.io/community-operator-pipeline-prod/mariadb-operator@sha256:3cbbff7ae75fb4f4629c7b8b58ca85ce67b23d5da558241301266d010e8b1188
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acm-controller@sha256:4d63542b39f41c458dd1eec28e44edce3a919f4dd9137ff64a9452cac1a1f954
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - docker.io/absaoss/k8gb:v0.11.4
        - quay.io/openshift-community-operators/k8gb@sha256:9743f957b00e5ad344fed0c062d616d8222990139ac4afbf14871fd9813f0ef0
        - absaoss/k8s_crd:v0.0.11
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/ucs-ci-solutions-operator@sha256:e284253263183eb2cd313f6fb009bb830b9a79a4d4e2481468821c30a9cdce03
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-operator-bundle@sha256:67723df19ab7d7871a6b639da3975c230d95dd250fa2995a9f44bdbb4b141454
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:439aeb806c3264948a15e325f7b1e71037099a085c8d971646188ace77fcf5d3
        version: 4.18.0-202509100344
      entries:
      - name: secrets-store-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: secrets-store-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: secrets-store-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: secrets-store-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: secrets-store-csi-driver-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: secrets-store-csi-driver-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: secrets-store-csi-driver-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: secrets-store-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: secrets-store-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: secrets-store-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: secrets-store-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: secrets-store-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: secrets-store-csi-driver-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: secrets-store-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: secrets-store-csi-driver-operator.v4.18.0-202502280802
        version: 4.18.0-202502280802
      - name: secrets-store-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: secrets-store-csi-driver-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-09T14:17:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:c5bd4f10120088f285a3b65e85bba825f9f55964758901f3b5827ac31516675b
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:26935de5f21a7865aba484599136df6a251c009e8fa9781e1040da3b19e7dd33
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:e05074d46e062268461d2689f7048c1a5dc4ed89a039d93e382f5ebc0726a269
        version: 4.17.13-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-multicluster-orchestrator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-multicluster-orchestrator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-multicluster-orchestrator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-multicluster-orchestrator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-multicluster-orchestrator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-multicluster-orchestrator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-multicluster-orchestrator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-multicluster-orchestrator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-multicluster-orchestrator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-multicluster-orchestrator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-multicluster-orchestrator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-multicluster-orchestrator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-multicluster-orchestrator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-multicluster-orchestrator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-11T15:26:27Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:4b78c8df4f4aa4be805a442b059b5e2822464bb029826f2f0090cf8b6b747768
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:47c754d0e2782dd8d885758e68803173c32f1fe6319ac09c786dd24e5124df25
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:957805157e7690deedaae63856c9ce71e5c299610bca401524397a40e2820537
        version: 4.18.11-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-multicluster-orchestrator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-multicluster-orchestrator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-multicluster-orchestrator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-multicluster-orchestrator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-multicluster-orchestrator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-multicluster-orchestrator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-multicluster-orchestrator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-multicluster-orchestrator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-multicluster-orchestrator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-multicluster-orchestrator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-multicluster-orchestrator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:3b683338f5849b887e1bc81ea078a9b799d272f824abd4247ff87e70ca7acbf2
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle@sha256:333a6f476d03adcbfaa4ab38802859719a7dd8857783f8487c99b008d06815c5
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1a156d67fd5beb9ffddecab724490acb7143f8141ff3b392c6f519e76c3ad374
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:645fcb18684884aef8ca4d5299f0d053d496c3a5d3f37cadd66c51d38514a680
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:5a73ce3a8cc8422aeda37fd9ef97f5aecdbf8c69fb6ff488df26342e989b5c57
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1dfe9180a9fca4f773c1282dfd0006daa0bb48dfbac87e352661e051107f9033
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - registry.connect.redhat.com/a5f1f0cc7ac3db90370970f9b/tawon@sha256:3691c0948c8eb266ada348de504622227af6516f3b3d331eb09fd258dd0609a5
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    defaultChannel: alpha
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
          createdAt: "2025-09-19T13:48:41Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dapr-kubernetes-operator@sha256:8f2087df1987437f1c5e62aa9453fea220a41012077fec67b6ba8368ec267421
        - docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
        version: 0.0.11
      entries:
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-apigatewayv2-controller@sha256:1cf12140fcd67b0f3e8c5791a91c5b705bb3af8b2e03ea4ba2d4ce6f9f7a4809
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - registry.connect.redhat.com/mongodb/enterprise-operator-bundle@sha256:23c3c8387eb9377b52327d7193f59eeea5831f4050248b3cbb54cbcd9437fc71
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-websocket-service-bundle@sha256:da100dd9c5b542ba46de7166e783e28a0f7e675f9ac896f211405e5a9e2082ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2cc940711539edd93a23355dcb06365821d140346e1df248386a77667c0f47ea
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2418015d00846dd0d7a8aca11927f1e89b4d8d525e6ae936360e3e3b3bd9e22f
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "rawDeploymentServiceConfig": "Headed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.34.0
          createdAt: "2025-09-03T09:59:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Argo
          Workflows\n* Model Mesh - ModelMesh Serving is the Controller for managing
          ModelMesh, a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks\n* Feast - Feature
          Store is the fastest path to manage existing infrastructure to productionize
          analytic data for model training and online inference.\n* Llama Stack –
          unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opendatahub-operator:2.34.0
        - quay.io/opendatahub/opendatahub-operator:v2.34.0
        version: 2.34.0
      entries:
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        - quay.io/community-operator-pipeline-prod/opendatahub-operator@sha256:d3e3d5ae02b834522481fa9046bf2ad3ea06d706c5cb06a32d32e61e229e4a6c
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:2992841660e8b1664fa2ce06da894aa828d4ea1be3eb5b27476efdd73731486a
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:bf85a9da8a100a90deb3745497bc1505852910ef23dd12c624fa17de51bd8377
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:826354c9c06a06f998e522a88d6a81239e003e1d6a2633e9b7b27b8a61d5fe38
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-operator-bundle@sha256:a59435a22c2f470b149ff721240f14144d2e56200b141e59a4f1011267a6c925
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:1346dcee82f2120bd2af22be9f041e057bdc32f9ad1c3ef717d599ba6519faeb
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:44a5c4cc3086e4eac42dc78a72b7492d13df0098017288278d9006e767a480a1
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:b9dbbce235b7a4f45f42184cf0b908e91e17094e971bc29a8710f35127a6b06f
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:026418bae8b2d425d45b2b3481f65c241d89e25f455e3fbfcf2f79536d978bca
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        - registry.connect.redhat.com/nec/webotx-operator-bundle@sha256:7e655933a7fab3ac3aa88eb5e4ca9ac3a71b98875ef1ef5bc49e87284f160ff8
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-licensing-bundle@sha256:2d3abd0263183bc92718316581675b92073f37c3786784fae24b3587a9441def
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:329d6eaeff787bcde5db0ba4f9403eee533a601b45dc2d5d349b318705452591
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:2cac5fdb55664c08928a7d924eacb18cc4c03c609261e3b6d812590e4d166375
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:b35da8aa4e6746952e5cfdbbac72950a48f7ad32a3d762a470ce62118c78643a
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:e5a0cb8af7879bf13d940df278ccdb8a5c400e5d7108f23b4abef94c964e40d6
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:91aae0500cd3cfc9a4a536241feb4fe50288e2a68f69ab5c602c02169e5f5b2e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:40c7d896355dfe67ff3616f1e1df86548bfcd0524f7e56d4d72ab4f5f1739f6a
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:3f69a124f9c31ebc0c78f249872a69b1289622bf4f5aad4c739f6e8caf9cb85b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:33a2e21a069ff24af4bf0ffc51e83ddf398df5d9735d0123120768e9dc2c27f9
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/community-operator-pipeline-prod/gitwebhook-operator@sha256:2d33e4f4f4b607ee2006481efc7ad9341049085ef988777d23e586f7406a5109
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - quay.io/bpfman/bpfman-operator:v0.5.6
        - quay.io/community-operator-pipeline-prod/bpfman-operator@sha256:6b3603c7c1917c67b5aa57202d1d2fdc0d9ac6b3a9373afa7172f322b3b3fa9b
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:74775c7851ac98ba2903435c71b9d02cc504ca8ee98b65ad7282d671392756a6
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.connect.redhat.com/cpopen/ibm-elasticsearch-operator@sha256:0bedcf95b31e28d41b124780f3b97c68a87db5626ecf1d5e81bb06b8b3fcd91a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/federatorai-operator-bundle@sha256:d0ac978d749654ddf4cc20823805ba7d33ee62e9a7ec98fdd53d9926b697c347
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/community-operator-pipeline-prod/cert-utils-operator@sha256:92a0c0b893b56330593e0a255b8df61031e92a61d172aecc58be5805d847798a
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle-marketplace@sha256:19c9d44cb9a0c7f52501fcb589750f3e85e270d066423bcadac593dd79e48ff4
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        - quay.io/openshift-community-operators/github-arc-operator@sha256:d8d97d19cea466c4c240097b79bfb7de90312d937ad4db63f9e90901686e54da
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/self-node-remediation@sha256:549585ca18d331cfbd5911fa3e79912428a0be3308588be2438ece3247746bd8
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |2

          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler,
          **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: kaliu@redhat.com
          name: KaiYiLiu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-operator-bundle@sha256:5ebbc06088e99b937327cd457324461cafdda9649b2dd2db1bcb892be8556eed
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:141b069ff9925390607315b0cf675003abbdac9555b29700e3b999fadd6c89f2
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:971174692a1aae68ac07ffa132254eb3c018a52507eb32621b6dc286fd25ca0d
        version: 0.5.0
      entries:
      - name: power-monitoring-operator.v0.5.0
        version: 0.5.0
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.2-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
          createdAt: "2025-09-11T18:04:40Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.2'
          olm.substitutesFor: cryostat-operator.v4.0.2
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:b6a6e401042daf1780ae6be19aedc551ed176d1a13bf1f0f48359552c638a156
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:8fd30d86f5468227899560071a28e92197c5790ea820e918c47559f0695cfb8a
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:d2788c8732fec62c7a27e7bd1497fea1ab3ff605e7e913bbe17d8ef97567329b
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:550de94fb20462c5cc5c52166820319544e774b8d45fc7beed7a095f5c1300aa
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:bbee66a44c2e3a5525c2fa81db72e8528bccf40583d0bd665841644473308310
        - registry.redhat.io/ubi9/nginx-124@sha256:126903359331eb9f2c87950f952442db456750190956b9301d00acfcadbc19ed
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:67e313e50c6ef100d6938dc1a734728661f328ab1f9757667c7a06de11689967
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:d9720fefa3e34d05a5536fe1a66e23b30fe0845ae2b96b9c9be4ee705be9633b
        - registry.redhat.io/cryostat/cryostat-operator-bundle@sha256:455ecde240a152987fcfe0013abee992014510de9e89c014559a5d0665a56512
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:9ba156d802c9c3eafc948afef88e0eaca6b5e73bf32ba80ea7b0651ba7ce923d
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:2a5e1d2d39a7d9d2bc74316110e1a65e9e18f3a9fb7f0a498c7457dd9e6684ee
        version: 4.0.2+4
      entries:
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cognitoidentityprovider-controller@sha256:093670fcaee6ef20e471c0df45c9ea7f40e5f0c43208c9447623324fdb39dbe0
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
          createdAt: "2024-11-27T00:00:00Z"
          description: The deployment validation operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        - registry.redhat.io/dvo/deployment-validation-operator-bundle@sha256:c04bae05ee0c24fe5bf94257b0dd2cb020f5c9b5f173ceffffdd041b6dff4ee9
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.11
        version: 0.7.11
      - name: deployment-validation-operator.0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.0.7.5
        version: 0.7.5
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CephCSI Community
      provider-url: ""
    name: cephcsi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cephcsi-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:12:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:b023d95637934b16ec8342994edd4e3198579517b68817ce6b4cc14958746b9b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:10cfe59ee2d625ed67f3804ba7e471f10fa1e12bc4f96de5e2f453e5cdefcf89
        version: 4.17.13-rhodf
      entries:
      - name: cephcsi-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: cephcsi-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: cephcsi-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: cephcsi-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: cephcsi-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      name: stable-4.17
    - currentCSV: cephcsi-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:20:52Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:903e47fa88119b5507a0d773d58318236bbce53201b5e6a18a16311f704b1120
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:562703a8b197f03b29d0564df329ecfda192a94b7818c6fd5f919329a5cc9bd7
        version: 4.18.11-rhodf
      entries:
      - name: cephcsi-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: cephcsi-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: cephcsi-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: cephcsi-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: cephcsi-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: cephcsi-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: cephcsi-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: cephcsi-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: cephcsi-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: cephcsi-operator
    provider:
      name: CephCSI Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/openshift-community-operators/pulp-operator@sha256:1926e96b185443ff0c7a2861cefc2e3aded0c0493e0d7f7d78e45447e06750d4
        - quay.io/pulp/pulp-minimal:stable
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:1e75649538f13cf3bb2e6131160f038a70689d6f92a110bdcec348e865040c4a
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    - currentCSV: pulp-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 1
                  },
                  "custom_pulp_settings": "settings",
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "ingress_type": "nodeport",
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "example-pulp",
                  "postgres_configuration_secret": "example-pulp-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "example-pulp"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.1.0
          createdAt: "2025-09-18T18:00:01Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          metadata.annotations.support: Community
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://pulpproject.org/](https://pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://pulpproject.org/pulp-oci-images/docs/admin/reference/available-images/single-process-images/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://pulpproject.org/pulp_operator/](https://pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:92a3e95cb3b1e50100933ea08cb44074fc505f6c0e7fa24d028ec46fa9e95590
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.1.0
        - quay.io/pulp/pulp-web:stable
        version: 1.1.0
      entries:
      - name: pulp-operator.v1.1.0
        version: 1.1.0
      - name: pulp-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:fbf847c41e1736fb6efab7a6acdc03acc562fa7e667306773880b277883bc4e5
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-operator-bundle@sha256:7c808f6e4ea469d88703300bdcc6bdce160634833cbd10bd110dc69d05be588c
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
        version: 4.18.0-202509100344
      entries:
      - name: clusterresourceoverride-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: clusterresourceoverride-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: clusterresourceoverride-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: clusterresourceoverride-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: clusterresourceoverride-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: clusterresourceoverride-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: clusterresourceoverride-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: clusterresourceoverride-operator.v4.18.0-202505070803
        version: 4.18.0-202505070803
      - name: clusterresourceoverride-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: clusterresourceoverride-operator.v4.18.0-202504251302
        version: 4.18.0-202504251302
      - name: clusterresourceoverride-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: clusterresourceoverride-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: clusterresourceoverride-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: clusterresourceoverride-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: clusterresourceoverride-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: clusterresourceoverride-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: clusterresourceoverride-operator.v4.18.0-202501231202
        version: 4.18.0-202501231202
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator-metadata@sha256:a2f0513cd8b96d71855464384cad7f1cd8c39f8d445be1ff1c1b91f22008d16d
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-operator-bundle@sha256:c2229c2cfad9bbbca9179cb6617bce0033f5be40731641a69b41f9a90e034c8b
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        - registry.connect.redhat.com/hazelcast/hazelcast-platform-operator-bundle@sha256:0aabd3e7f5f993c1288450678d55e8f7b36c8b30894e5d844521a474aa6b2714
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudtrail-controller@sha256:20fe3c3ba7688906de40d7fd36c00552be280595708935ff800ad5228829f781
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/smart-gateway-operator-bundle@sha256:f12b2f1dd34c73b9c8a5ad461e771c0ce56c9ea4eb88f03225be1462a42c5a78
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:0cda4be8f74da7524a72c961c8762927c66e145d82e59af92073a54179bc9936
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:4bab095c5dd6f8c95a17a8ff2bf70ff3f53faea48b56eea8baa9cf9a42a5227a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/keycloak-operator@sha256:a4da3bcbb4e99d60b8217d9e62742eeb83eef163dfefb2f7efda131f312e46e3
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        - quay.io/openshift-community-operators/keycloak-operator@sha256:65fe00f310289b70538a6dde897ef663731d0da44a7bb1d0f55a3001a50e53fa
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/keycloak-operator@sha256:70bbd44f6ecf86aa8e9710175a96dd3c1d813aec40667a09800241ce30afd93f
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
          createdAt: "2025-06-19T11:07:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-zookeeper@sha256:c8487e18318a92a64e84a5ab29f844a0973f3b8b2aa291723014731471459176
        - docker.io/confluentinc/cp-server@sha256:072719eef75ee61294027ba77f7b9cd821437158761495a1644b5687af9305a8
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:7a271565db9c500f8c5755de397e8093251673d2fc6b8a10974f927f382b4704
        - confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2520a514b848b53338c411f0e0dfa68491596761f8bbbb090d652f71eda010af
        - docker.io/confluentinc/cp-kafka-rest@sha256:9c19e7900669e555810d26676393737a7d573b39c31e284c0ad78807bfe05ce3
        - docker.io/confluentinc/cp-server-connect@sha256:05e77fe8e19747521e1f1d5630bbf384b0fe0ec9efe8dc798de6e4a3eb0bfa2d
        - docker.io/confluentinc/cp-schema-registry@sha256:5b6bf5b971dd8fe2c821e227391858f58f6ae2128f1ae1faa988c3ea59b86801
        - docker.io/confluentinc/confluent-init-container@sha256:550a6b92fc5aca81ad8e0d3d47c0063af9da877e45c1ecda1a9cc44a70e5d87a
        - docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df1f6a99ecfe8d8acc0af222848aad13349f06071fc2bf1360d6d20f99769394
        version: 2.10.2
      entries:
      - name: confluent-for-kubernetes.v2.10.2
        version: 2.10.2
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.2",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
          createdAt: "2025-06-19T11:14:20Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:b7cde5736b5275abafafbb81f4c917890abfe8f400f436c3a1f712cf200832f8
        - docker.io/confluentinc/cp-schema-registry@sha256:4f6cabeea11ae532cb47905eadaa3192d7113ea6d0a0da310cdfd5eb7142fc45
        - docker.io/confluentinc/confluent-init-container@sha256:37161a90b89dfad625bf61ba7d299707d7b2a96271e9667044c3eaba3ffc5f66
        - docker.io/confluentinc/cp-server@sha256:dfed43a1618de1fea6d202921ca4cc2168f7455c3d37fe046f3e64febf042a87
        - docker.io/confluentinc/cp-zookeeper@sha256:f1e6d3d698a6e2757e3c225b5c6f8f007251838226c46f652a4cbb4314b50277
        - confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:2baa0cb5df8017d9d14c2f0c7e02ce371ba92462d8d1e93a3ca1bc548c20a74e
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:507e1538a64c4232ffda86504594def45cb20bbcb6b13fc6196c1f41e71f93bf
        - docker.io/confluentinc/cp-ksqldb-server@sha256:744f5d844ac223c857116f3994c5be2105898ea3896d53026eb62e76c7240c4f
        - docker.io/confluentinc/cp-server-connect@sha256:73822ffff10d386b98c9ca2bb295b4d079deb210116c8e4283cdc26ce304526f
        version: 2.11.2
      entries:
      - name: confluent-for-kubernetes.v2.11.2
        version: 2.11.2
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.0",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
          createdAt: "2025-06-13T13:02:11Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:ba49d46654770994c302d586a609afdfc58a7cc05f204aa090b643f0680a5014
        - docker.io/confluentinc/cp-server-connect@sha256:e4cbc1836402497ea42b2c961c5d6b4420c25b3c6a5f6f71be3e99fcc0f09702
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:a283f2a72b6346febaf711e71dee1e0e3666d7265320f3f69307fc2c0b95d5ef
        - docker.io/confluentinc/cp-schema-registry@sha256:3dcb9652a4c83d4566a7334041c1ac3ec359665a9450a70c5d7b41adaea55c90
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:35f2799f71c308d87e6a7dada00263abb9e08e895b10f10c815e7a9edf19dd49
        - docker.io/confluentinc/confluent-init-container@sha256:82277228368a783ba84367c920a9fa15fad6844723df1e16b07c9dcc829a54ad
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-server@sha256:ffd5961412da42c81c9b20f35e334aa8bdf351dcb452610e84156ed511116ef5
        version: 3.0.0
      entries:
      - name: confluent-for-kubernetes.v3.0.0
        version: 3.0.0
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator-bundle@sha256:21b69b5eb896a37691a8e086f3a51c0f68086a0ff9370fd66667acc6c935a850
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/mta/mta-operator-bundle@sha256:65c0932bfda32b1e62be9edd0f96bb4d6e41e302a2e8080fc6cc794ee5c52276
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-operator-bundle@sha256:f3d585e24d65905b2e9f594dcdb67721aa95057ab7587ae531c3487aa2f94e53
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/mta/mta-operator-bundle@sha256:a13643117c2867351718a872f7f1b2350c67855ca73a727a1dc140754ffe6589
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/mta/mta-operator-bundle@sha256:9eeb43af2bcab84f5261d1575f7c897903a7696dba011d256abffa1fe850eba2
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-operator-bundle@sha256:6aef520983fe01351886e108d32b998b20e74116fc0280c754a914bfad03d0bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ed10e35f45a140e21d826bee71b3ab96e67f1b4dc8c2f7f1676380fa51b67fe
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        - registry.redhat.io/mta/mta-operator-bundle@sha256:68953a2a721309fd103e09f3171d0c9d50020413631b575f1a44602a5d2ac855
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ec977b0c9c26551db110dcf2d5069647b1a5495f869ff48a0319cb496b84cdc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - registry.connect.redhat.com/eginnovations/universal-agent-operator@sha256:65f851074e0e1db57a62b559c0f10f9d30543011727aec3b832f9a35468139e9
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator-bundle@sha256:8d130a5f2407931dbc73e45a53610b95ceb04cb507b8f6519f630cada1c0da45
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:9009198860c7a1e0de4f429261d442b75b5b2893a943bb77914b14d643db8829
        - registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/openshift-community-operators/sosivio@sha256:110d7f4fb993d65ff497eca57c42206faab07dc5f5b1f43b1b682387e7cd3a10
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        - quay.io/openshift-community-operators/l5-operator@sha256:e4cb13801a80753c99a47f009ca0a29416b3e04afe5314d07d97c2464a6f0d75
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:712d31214423b326ab8f107435c85431ce8d99c2e52fd8d615005a4274e0ff98
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/operator@sha256:2cb5ec186af3264cba42333245616dd01f0f9868a849129b1a31afbc51ee1410
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - registry.connect.redhat.com/harness/gitops-operator@sha256:f97cdc2245b7133a04235c55540d3a87165e080b1d5b9572a663731f6ccbe597
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        - quay.io/openshift-community-operators/tf-controller@sha256:cdaeb470d9b331a3fa08167e5c904befb00be983e1354651670e21e13560bf9b
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.18.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
          - description: L2ServiceStatus is the Schema for the l2servicestatus API
            kind: ServiceL2Status
            name: servicel2statuses.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-metallb-operator-bundle@sha256:c80c3467cf52fc4b2a55e444fba9fdc8cf3b1efedad5d1276a889d07aed44e68
        - registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:6f23dd225f2bccfe39d05428a51de990a4a9338c49da079384eb30fb470c04f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509100344
      entries:
      - name: metallb-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: metallb-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: metallb-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: metallb-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: metallb-operator.v4.18.0-202506251003
        version: 4.18.0-202506251003
      - name: metallb-operator.v4.18.0-202506070804
        version: 4.18.0-202506070804
      - name: metallb-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: metallb-operator.v4.18.0-202505081435
        version: 4.18.0-202505081435
      - name: metallb-operator.v4.18.0-202504291304
        version: 4.18.0-202504291304
      - name: metallb-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: metallb-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: metallb-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: metallb-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: metallb-operator.v4.18.0-202503181802
        version: 4.18.0-202503181802
      - name: metallb-operator.v4.18.0-202503110933
        version: 4.18.0-202503110933
      - name: metallb-operator.v4.18.0-202502261233
        version: 4.18.0-202502261233
      - name: metallb-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: metallb-operator.v4.18.0-202502041632
        version: 4.18.0-202502041632
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:latest"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:latest
          createdAt: "2025-01-03T19:50:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The eBPF Manager Operator is a Kubernetes Operator for deploying
          [bpfman](https://bpfman.netlify.app/), a system daemon\nfor managing eBPF
          programs. It deploys bpfman itself along with CRDs to make deploying\neBPF
          programs in Kubernetes much easier.\n\n## Quick Start\n\nTo get bpfman up
          and running quickly simply click 'install' to deploy the bpfman-operator
          in the bpfman namespace via operator-hub.\n## Configuration\n\nThe `bpfman-config`
          configmap is automatically created in the `bpfman` namespace and used to
          configure the bpfman deployment.\n\nTo edit the config simply run\n\n```bash\nkubectl
          edit cm bpfman-config\n```\n\nThe following fields are adjustable\n\n- `bpfman.agent.image`:
          The image used for the bpfman-agent`\n- `bpfman.image`: The image used for
          bpfman`\n- `bpfman.log.level`: the log level for bpfman, currently supports
          `debug`, `info`, `warn`, `error`, and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`:
          the log level for the bpfman-agent currently supports `info`, `debug`, and
          `trace` \n\nThe bpfman operator deploys eBPF programs via CRDs. The following
          CRDs are currently available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n
          - KprobeProgram\n- TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n-
          XdpProgram\n - BpfNsApplication\n- TcProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n\n ## More information\n\nPlease checkout the [bpfman community
          website](https://bpfman.io/) for more information."
        displayName: eBPF Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - registry.redhat.io/bpfman/bpfman-operator-bundle@sha256:b6d43f1fec6ba7336c23d1fbceab15ebb2ed9e9e2eebb08c78d7541ee47b4204
        - registry.redhat.io/bpfman/bpfman-rhel9-operator@sha256:c3d878d77091bed03d99f72943d9b78e210e2f6194ea4b5cd2aa23f8d48f3f61
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      name: latest
    defaultChannel: latest
    packageName: bpfman-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-applicationautoscaling-controller@sha256:4f1b92eac30d72a66d8fe8c2c44339c7a5fd6af65d48c0d8697459c7af290741
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Agents",
                "metadata": {
                  "name": "loadcore-agents-sample",
                  "namespace": "keysight-loadcore-agents"
                },
                "spec": {
                  "agents": [
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent1",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "replicas": 2,
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    },
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent2",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    }
                  ],
                  "image": {
                    "registry": "prod.harbor.keysight.digital",
                    "repository": "loadcore/loadcore-agent",
                    "tag": "UBI8-0.1-7d990d7254"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "regcred"
                    }
                  ],
                  "networks": [
                    {
                      "config": "{ \"cniVersion\": \"0.3.1\", \"type\": \"ipvlan\", \"master\": \"ens192\", \"mode\": \"l2\", \"ipam\": { \"type\": \"whereabouts\", \"range\": \"192.168.1.0/24\", \"routes\": [ { \"dst\": \"0.0.0.0/0\" } ] } }",
                      "name": "loadcore-agent-network"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:53:55Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:f1d1e9f2aca11505407785bc33905744b667bc8aea49df73fcc50db379c0b375
        - registry.connect.redhat.com/loadcore/keysight-loadcore-agents-bundle@sha256:f16fcea424cb7b1eb647f3e9d3c08bd93c8a1b8bb7eab71a4e756da72a7a56a0
        version: 0.1.21
      entries:
      - name: keysight-loadcore-agents.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:a58c9c8454ac337c5a359dfe5b7c6b06983b7050ad62a0d11054f06fc17e92c1
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.89.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.89.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:f6c37a0808c137f226b6afb6cac1e2ce05184d6a01b9234969ce63056571e991
        version: 42.89.0
      entries:
      - name: t8c-operator.v42.89.0
        version: 42.89.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - quay.io/community-operator-pipeline-prod/gitlab-runner-operator@sha256:3e0314a2695c071fe0a34f3dc7684b43b5a31b6705f42982b4dbaacabe360d2b
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-operator-bundle@sha256:2bab9fa63b128a3ba0a6031f70f0342f95d17dd53a3ec2930b19940b5b2e6f47
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - registry.connect.redhat.com/eweb/cloudnfv@sha256:0bb427954fb8df5de197353908a5189922f514e121ae8cdb8242d125b1b3807d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - quay.io/openshift-community-operators/odoo-operator@sha256:616e1067ebf8deb66636d57cae2ac1f9981b83419d8e7f023bc7df1c1b745536
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/docling-operator@sha256:9a3717ee195572b29d1ee31b96bb95561f5521b032dbaa09a000413b227b5f12
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
          createdAt: "2025-09-19T21:59:41Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatch-controller@sha256:254df5f0e6821a65277bc3d029e7af87b10800b192f9724997996224525a9943
        version: 1.2.1
      entries:
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        - registry.connect.redhat.com/scaleops-sh/scaleops-operator-bundle@sha256:efd483d9c826e25b0d1cd79deff4263772260c72da79bad149ef7985cc00a44b
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:54bb383ac6c5f57ae77dcd96cdc33cea5c4c1aff9f9d356c1721591aa820e82d
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:948ad5eb47b7809ba5dd78cb6bb30b1bc64163617f229dc33cec251c1d5bb20e
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        - quay.io/community-operator-pipeline-prod/lbconfig-operator@sha256:b0bb724a5d9905633962a3e8267e75d371f754bce866a29cba3ac0d0c671ebff
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        - quay.io/community-operator-pipeline-prod/apicurio-registry-3@sha256:e207a4f7f53486b78f95ad958f7e0053a8a8561c3dcb1463a7176f27eb9ea082
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:610a5e66cf1b6c7f5451c9a0f5e0a6cc9435f2548335fbd50f44b8656adadd99
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:bc0d8b62d5814407e08817ece44626400cc3d9d29b14aa35dfe711b907fddf21
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.connect.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:15e8cbce994b783fa6dac737ac28f1b713d1c75891dbce4987d2c09ff295376b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:12d6ae3f139daf33dbf104499c195d9915470b957f401b5f622cfde92350a6c1
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:965ec7875eccc6a5fc56c3b8f0da668d826d34580117ba1438a607a9d0d13b73
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:f8a8cfcebba4b17fa6cfa99b3ef430af272c5237476dba4665803a90adfd5d2a
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        version: 3.74.9
      deprecation:
        message: |
          The `latest` channel is no longer supported.  Please switch to the `stable` channel.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:7fd7595e6a61352088f9a3a345be03a6c0b9caa0bbc5ddd8c61ba1d38b2c3b8e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        version: 3.62.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:5187b04f8010a274d6bc017fc82f1b00d886124e520c1b7de8b4da8cd2f37bc0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        version: 3.64.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:d3bf33899ca8b199ca8782978618ca343164e43fd0ae2ac6a0cfbf697efa6428
        version: 3.65.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6536ee795e8aaeab9321ee4820e55933195983a768ba24fab40508d03e50a6e4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        version: 3.66.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:ad88773ab141f45fe24c61e83ec736aab27f652fb82ca2a85fee6f885568c5b9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        version: 3.67.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6a10a4859527946f61f497ff84b80f973cb7be5982d428423836b5aca81a1f5d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:297e7f863106c6b41f882bbd5b691a12c797c24d832473a04e604d4639e4a68e
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        version: 3.69.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:9b35ae6df2c19935783735d213483f3fcecead2f49f873a4f7844ef9fb5c7f1a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        version: 3.70.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e001b9f844a6c298dd74f1ab683f3dbbf0ad66044cec419c373ef4f97d47be73
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        version: 3.71.3
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:878ed77d07d3f40f049a594af44376797305786a2a264dae3d0c50401643bfa5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        version: 3.72.4
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:459a8d91beff91638872d377a8b82c279ad2905ce3f4ad4bfc00f6822418ac57
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        version: 3.73.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        version: 3.74.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:a0d9cd31d953aaf7e84f5a39f405c5aa2ea62e68d6c728a936ca1ffc4dc1b93e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        version: 4.0.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:b26133843687cc50e2c52d6067d8b3369b5cb28dfbfd15dd4b6062096673195a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        version: 4.1.6
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:08aebe6e8275d3314b6756d522ec568ccdfb9c52df98c9822f187f6c5a3af19c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        version: 4.2.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:30f9384244819da013358325cec364bb210b3b33b4cf9680698531fe50562aac
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        version: 4.3.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:123c0108bd7a190dd2accfd21786600f787147f040e57fb5f840861e2caa441d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        version: 4.4.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:41351e88688d23a443ad4f141caf0433b2c2789e6a1a85b4f5b36abeb5656bb1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        version: 4.5.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6d6b2696a12d49cf3403e338a0e028187473b6f43ef2b0bd72c7701200ac8e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:c7f894ab773d1d389656d13ed88fbf575faddd95c5b8396c6593ab3f993f331e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-bundle-rhm@sha256:5d14e0c44ffb543b7cf59817fe587c8e819542d8edfa261a44e58aea0ecdb79b
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        - registry.connect.redhat.com/randoli/app-insights-operator@sha256:21ed3d9b8b8466665c1cb784d0bc244279d0730f6319625caf46af111c25d333
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
          createdAt: "2025-05-09T05:49:47Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-operator-bundle@sha256:f49b7a2c60e79edaba1a5faabc3983f87f9123b6d84121a438d179ccbb8102e3
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
        version: 1.1.1
      entries:
      - name: multiarch-tuning-operator.v1.1.1
        version: 1.1.1
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.9-r3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
          createdAt: "2025-08-05"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:327a9899adfe89af9aebdcfac651a9eddec9ecf8a345caee1cef1b4a1fb279d5
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:df0493f7edbdedf754c70e4efc256fa4d246a24d6c8882c562e56ef3844278c1
        - registry.redhat.io/apicurio/apicurio-registry-3-operator-bundle@sha256:728354aee3729a1fba233c2aa92cacb21d27efd4869706ddd214a2d3688e33b2
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
        version: 3.0.9-r3
      entries:
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:d9930900b0d2f92ae813246fc322290234ad8bbd165b10e36671b53852228718
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:87448055a162e215a4d4982a8494fec6e7cc59fe6704494ddfdcfea076a2100c
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:a7c6f1c9260c79ebbdfb25ea996516ba1e2bcbedf25401e9e6848b955a2d3d78
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-operator-bundle@sha256:9ab3d5eb70c654270fef9e8e9d4e5b68e1ac3c86bd0e0dcd36cc6f8741be1603
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:36561e5888cd34f8fb83ceff2910141fceaa6ea290bd246125930fe183e2ebc3
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:9e8bcbff46b97f3bc1071146b8db233abd1630016e6fc0efb679df8243c6fd03
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
          createdAt: "2024-07-10 00:59:51"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:518691d157b3e3b5031a2b1e4cc3f415839536079bb8570ae8690ac92d302ba0
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:a64f67f29a4eb8ce426ed842f1a5a5fcd83fd9d91acebb064cf65c89c9d8bfb6
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/keepalived-operator@sha256:5a46fcb72f02a1f0d9cbc72f44948ef4b09149a4b3fe2e45fd31df90d20b3c56
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-keyspaces-controller@sha256:6e56400162867a4f55a70c340aeeb074726fd3dca13b6da60c395d6f35a48b48
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/automotive-infra@sha256:6d4180ff6899d4de8ba2cd4a06ece234bbd8f0ec1bd0b4b0ae4b7a0c047121da
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        - quay.io/openshift-community-operators/aws-efs-operator@sha256:6372847df81faddc23c938928157c3ac280e537a4eec9f7edd599480260bc566
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/cluster-impairment-operator@sha256:6fd114eb8f525f670e505373c2999a4e66b53beda988e99b267204d2fe0e1b4f
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:8197e4a96ac8949d43cda62ac2257a9b1e71f0b54f42283ee4925f88c369f873
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:728074151b245f5e0d65bd8956149f05ba22a1b6438120ebca2d85b0d7d0c837
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:4109698147632b858fe2787964772a1339cdebee72b75b42d9b1c5dbbbecb45c
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:088d0eebf60c8f6961ca72c268f61d12648d575aa41552ae6398fefc6d9700cb
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d6d444e642bfd567eb40c7e084f5df743c23b47935e52e35ef541c67fcc00433
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:b3469cd953ef2015fad2f7f96d960cee5ecaa6c19df631a9c882154ab5b3de98
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:f49d3eddbfcec1018fd3716cd0ec2addef9961d2d308099636541868566979b0
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        - quay.io/openshift-community-operators/t8c@sha256:403b29a6262ba0e06de117f788aea7224268b798868c3e5b738b05e96df20456
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4bfaf97ab09a1aff1316536b99957b1f82763b009dde6d08fc9eea1fdd290fbf
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        - quay.io/openshift-community-operators/t8c@sha256:61f1ea9b3851d7272703f5bc21203ff03bb35fd52ed9780d0cf996379c78a00b
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:48a5a9fbfa92b4b87b3653613467a3fcceacadbbabd274e367b40f6871039cc8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d65834d8f9946c2764d45666c90c50f428b74afe7b066d666c18e9545994c9a8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4964196aba839fde569ce58ab7e4eef438c9638ae4057194f2d7a58eaa5a98d9
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        - quay.io/openshift-community-operators/t8c@sha256:46c815fa635d58820757f41b1d8f04333638c39471f4d6c901a6500379680b57
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d8a575a84dd86d241acaaec53af1e164bc1ff1b95b5eb710851ae698bbed6bf5
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:dfdc8fbdfc0ee76b2d5c62f6fd5deac2c50756661871d8b4357e25043e5ce66a
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:081d7f4f20332599fdbbd07f22cc18eca6c5625b619efeae6e861239cde4b12a
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:cf7ccc1a60d6787ebcd1d76415cca4cc55c2c3ee18d52c0511e66ca6bf9310f8
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: loki-operator.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
          createdAt: "2025-06-11T13:21:40Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b311677919183c6c91c6da40b0ddb2529f92ed52b2c46a07e399a4bad5dea280
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:fce287241be64b7b2c20a971c4f2d468c079ae81556aaab756572ebad27a17f5
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:ff98532bd8f21ecf375062d465f2aee4cdcd70b647d258c1e09f7e2286f998c9
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:ea2d995744cd4a4bd6e7467c2d6927bfc89a47d551ef0d6b048c0c57652f81be
        version: 6.3.1
      entries:
      - name: loki-operator.v6.3.1
        version: 6.3.1
      - name: loki-operator.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b793aef88881aa75d47a22377d8932efc0b497bb01b2f3f376d64891d1dace54
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-operator-bundle@sha256:722cf3c05eb5663a737cb1385b971b6d1f0c8b2595d5ac425f7b100c25e2fd26
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-operator-bundle@sha256:ed97b11fa0f0f7aadf193810a5a599bf26015a0af8ee6fd90061110dcb0d50f4
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f19ae051041c46fe7587f903a9c6221bcbe8be2c4cc0235739f7592eb93ae164
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:28e37ec35de1307405264cd4f79aaa6ce4d4ecbd6cbc27f4fef5db415c678d64
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        - registry.redhat.io/quay/quay-operator-bundle@sha256:4a004584fa6c7187481e8c8480f17e0e4edd18c32f52a2e271c68d02f1a9655f
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f16c0a135dd7380e2a489f963b38bd76a63a973f482d06ece8d3656aff223703
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b499eebed32c0261db414e307f96f8aef9e743035544c80c048526e7ba4fbdc2
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-operator-bundle@sha256:55a5e6297109aea029bd40fac09a9fbc10c28a8c5558f9afc9b9b86c5dc7a4d4
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-operator-bundle@sha256:e3f63ea56de7ca5388a20c17ac63e16f07935ab09bcdcd38a86639c9d59c8a7d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-operator-bundle@sha256:53fcddadaa31a5f7d74c38f0bbd0d80b4c09796a580ef8494f60ff82fb6fe6c1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-bundle@sha256:9d0ed6d279800a189c1466e0b32f80af2f8bad96a65b7dc0322c64337dddf39c
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:9d21b738b0d357ef064690eae66bd4b3810664bcde57237822f1e90e9fd6ad03
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - quay.io/openshift-community-operators/reportportal-operator@sha256:67a7ffaf6a303b92541170f266d7b2871f5c2fe73f6b6c9a2b7b33dbc4555fa5
        - quay.io/waynesun09/rp5-operator:v0.0.8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/openshift-community-operators/etcd@sha256:e42581eacbc8349f40c6a4ed6d7d82d6168518e15b1a24cba36817205bd8688f
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:1d308347e5e290ac4bb7dd7aab0a1c0659d661153541d81dd30733b52d1fbf83
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:f1757deff8c4048ae42cb033f3ddd7686dec2968e5a4cd61ff8f74fda69a8a09
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - docker.io/grafana/tempo-query:2.8.2
        - docker.io/grafana/tempo:2.8.2
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - quay.io/community-operator-pipeline-prod/tempo-operator@sha256:14282a0ca95401ab3e4c11dc2b611a6f57c00c461254f52ddf03ef74c725767b
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/jboss-webserver-5/jws5-operator-bundle@sha256:ce4060138117dbc8fc21abb60f0b8ef1bfa5ebc0dd8eb803c0e79d7447341964
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9y
            dCgzMzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQ
            LmFkZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAg
            ICAgICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24i
            LAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJh
            c2gpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3Zp
            ZGVkIGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAog
            ICAgICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmls
            ZXN5c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IG1ldGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBk
            ZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2Vw
            aGZzIGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1l
            bnQoCiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92
            aWRlcyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9v
            bC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRh
            IHBvb2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBp
            biB0aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAg
            KQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1l
            bmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZh
            bHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+
            OjxQT1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCki
            LAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAg
            ICAiLS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAg
            ICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBv
            aW50IFRMUyBjZXJ0aWZpY2F0ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRf
            YXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1za2lwLXRscyIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJJ
            Z25vcmUgVExTIGNlcnRpZmljYXRpb24gdmFsaWRhdGlvbiB3aGVuIGEgc2VsZi1zaWduZWQgY2Vy
            dGlmaWNhdGUgaXMgcHJvdmlkZWQgKE5PVCBSRUNPTU1FTkRFRCIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW1vbml0b3JpbmctZW5k
            cG9pbnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgZW5k
            cG9pbnRzIChjb21tYSBzZXBhcmF0ZWQgbGlzdCBvZiAoZm9ybWF0IGA8SVB2ND5gIG9yIGA8W0lQ
            djZdPmAgb3IgYDxGUUROPmApIGVudHJpZXMgb2YgYWN0aXZlIGFuZCBzdGFuZGJ5IG1ncnMpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tbW9uaXRvcmluZy1lbmRwb2ludC1wb3J0IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAg
            ICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkNlcGggTWFuYWdlciBw
            cm9tZXRoZXVzIGV4cG9ydGVyIHBvcnQiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1za2lwLW1vbml0b3JpbmctZW5kcG9pbnQiLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBhY3Rpb249InN0b3JlX3RydWUi
            LAogICAgICAgICAgICBoZWxwPSJEbyBub3QgY2hlY2sgZm9yIGEgbW9uaXRvcmluZyBlbmRwb2lu
            dCBmb3IgdGhlIENlcGggY2x1c3RlciIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgZXJhc3VyZSBjb2RlZCBSQkQgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWRyeS1ydW4iLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAg
            ICAgICBhY3Rpb249InN0b3JlX3RydWUiLAogICAgICAgICAgICBoZWxwPSJEcnkgcnVuIHByaW50
            cyB0aGUgZXhlY3V0ZWQgY29tbWFuZHMgd2l0aG91dCBydW5uaW5nIHRoZW0iLAogICAgICAgICkK
            ICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yYWRvcy1u
            YW1lc3BhY2UiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1G
            YWxzZSwKICAgICAgICAgICAgaGVscD0iRGl2aWRlcyBhIHBvb2wgaW50byBzZXBhcmF0ZSBsb2dp
            Y2FsIG5hbWVzcGFjZXMsIHVzZWQgZm9yIGNyZWF0aW5nIFJCRCBQVkMgaW4gYSBDZXBoQmxvY2tQ
            b29sUmFkb3NOYW1lc3BhY2UgKHNob3VsZCBiZSBsb3dlciBjYXNlKSIsCiAgICAgICAgKQogICAg
            ICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXN1YnZvbHVtZS1n
            cm91cCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgc3Vidm9sdW1lIGdy
            b3VwIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tcmd3LXJlYWxtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAg
            ICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhlIG5hbWUg
            b2YgdGhlIHJndy1yZWFsbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXJndy16b25lLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlk
            ZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cC1uYW1lIiwKICAg
            ICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAg
            ICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSByZ3ctem9uZWdyb3VwIiwKICAgICAg
            ICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9w
            b2xvZ3ktcG9vbHMiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJl
            ZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdG9wb2xv
            Z3ktY29uc3RyYWluZWQgcmJkIHBvb2xzIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3Vw
            LmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9wb2xvZ3ktZmFpbHVyZS1kb21haW4tbGFi
            ZWwiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwK
            ICAgICAgICAgICAgaGVscD0iazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gbGFiZWwgKGV4YW1w
            bGU6IHpvbmUsIHJhY2ssIG9yIGhvc3QpIGZvciB0aGUgdG9wb2xvZ3ktcG9vbHMgdGhhdCBtYXRj
            aCB0aGUgY2VwaCBkb21haW4iLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdGhlIGs4cyBjbHVzdGVyIGZhaWx1
            cmUgZG9tYWluIHZhbHVlcyBjb3JyZXNwb25kaW5nIHRvIGVhY2ggb2YgdGhlIHBvb2xzIGluIHRo
            ZSBgdG9wb2xvZ3ktcG9vbHNgIGxpc3QiLAogICAgICAgICkKCiAgICAgICAgdXBncmFkZV9ncm91
            cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikKICAgICAgICB1cGdyYWRlX2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFkZSIsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAg
            IGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9yIGV4YW1wbGU6IGNsaWVudC5j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0aGNoZWNrZXIgY2VwaCB1c2Vy
            cyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBuZXcgY2x1c3RlciB2ZXJzaW9u
            IGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBwbGllZC4iCiAgICAgICAgICAg
            ICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rl
            ci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVwZ3JhZGUgYWxsIHRoZSBkZWZh
            dWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAgICAgICsgIkZvciByZXN0cmlj
            dGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVu
            c2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2luZyAtLXJlc3RyaWN0ZWQtYXV0
            aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9yeSBmbGFncyIKICAgICAgICAg
            ICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29sLW5hbWUsIC0tazhzLWNsdXN0
            ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUgdXBncmFkaW5nIgogICAgICAg
            ICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3UgaGF2ZSBwYXNzZWQgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcgdXNlciB0aGVuIHdoaWxlIHVw
            Z3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAgICAgICAgICsgIlNhbXBsZSBy
            dW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMu
            cHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxpY2Fwb29sIC0tazhzLWNsdXN0
            ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGllbnQuY3NpLXJiZC1ub2RlLXJv
            b2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAiUFM6IEFuIGV4aXN0aW5nIG5v
            bi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0byBhIHJlc3RyaWN0ZWQgdXNl
            ciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRlIGZsYWcgc2hvdWxkIG9ubHkg
            YmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVzZXJzLCBpdCBzaG91bGRuJ3Qg
            YmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxpZWQgcGVybWlzc2lvbiwgZm9y
            IGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2ggcG9vbCB1c2VyIGhhcyBhY2Nl
            c3MiLAogICAgICAgICkKCiAgICAgICAgIyBBZGQgY29tbWFuZC1saW5lIGFyZ3VtZW50cwogICAg
            ICAgIGNvbmZpZ19ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb25maWciKQogICAg
            ICAgIGNvbmZpZ19ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWNvbmZpZy1maWxl
            IiwKICAgICAgICAgICAgdHlwZT1zdHIsCiAgICAgICAgICAgIGhlbHA9IlBhdGggdG8gdGhlIGNv
            bmZpZ3VyYXRpb24gZmlsZSwgUHJpb3JpdHk6IGNvbW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmlu
            aSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgKQoKICAgICAgICBpZiBhcmdzX3Rv
            X3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAogICAgICAgICAgICAgICAgdHlwZShhcmdzX3Rv
            X3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICksICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJz
            ZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgYXJnc190b19w
            YXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJldHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190
            b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6
            CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCgogICAgICAgICAgICBpZiByYmRfcG9v
            bF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBpZiByYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0
            YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAg
            KQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQi
            OiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            IG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCdd
            fSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21z
            ZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZh
            bHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtleSBpbiBqc29uX291dFsicG9vbHMiXToKICAg
            ICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wg
            bmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBvb2wKICAgICAgICAgICAgICAgIGlmICgKICAg
            ICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAg
            ICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bv
            b2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxl
            IGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAg
            ICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29s
            X25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgICAgIHBvb2xfZXhpc3Qg
            PSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAg
            ICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9wb29sIG5hbWUsIgogICAgICAgICAgICAgICAg
            ICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBwb29sX2V4aXN0OgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            IlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwge3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBl
            eGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihzZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIg
            KyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoOgogICAgICAg
            ICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rp
            bmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAgICAgY29udGVudHMgPSBmLnJlYWQoKQogICAg
            ICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJzdHJpcCgpCgogICAgZGVmIF9jaGVja19jb25m
            bGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJF
            aXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRhdGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBz
            aG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAgICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBv
            aW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAgICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdl
            dHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRlbGltaXRlcgogICAgICAgIHRyeToKICAgICAg
            ICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEp
            CiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikK
            CiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgog
            ICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4o
            ZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAgICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAg
            ICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNzKGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRB
            ZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAgICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJv
            cjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUUROIgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGln
            aXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQg
            bm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGludFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBp
            ZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioqMTYgLSAxOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9y
            dH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoKICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYs
            IGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMg
            aWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0
            IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBvaW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShz
            ZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gIiIsICIiLCAiIgog
            ICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJz
            cGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            Ik5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSAhPSAiWyI6CiAgICAgICAgICAgICAg
            ICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRwb2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAg
            ICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2VuZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3Ry
            X3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBbImh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJl
            c3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZvciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlcCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRf
            c3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9IE5vbmUKICAgICAgICAgICAgICAgICMgSWYg
            dmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBkaXJlY3RvcnksCiAgICAgICAgICAgICAgICAj
            IHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4gcHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFz
            aCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNTTC4KICAgICAgICAgICAgICAgIGlmIHByZWZp
            eCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RsczoKICAgICAgICAg
            ICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1GYWxzZSkKICAgICAgICAgICAgICAg
            IGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNlcnQ6CiAgICAgICAgICAgICAgICAgICAgdmVy
            aWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1l
            b3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAgICAgICAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10aW1lb3V0KQogICAgICAg
            ICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAyMDA6CiAgICAgICAgICAgICAgICAgICAgcmV0
            dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAgICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVy
            cjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vycm9yID0gZXJyCiAgICAgICAgICAgICAgICBj
            b250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRv
            IGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNw
            b25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICIiLAog
            ICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCItMSIpLAogICAgICAgICkKCiAgICBkZWYgcGFy
            c2VfY29uZmlnX2ZpbGUoc2VsZiwgY29uZmlnX2ZpbGUpOgogICAgICAgIGNvbmZpZyA9IGNvbmZp
            Z3BhcnNlci5Db25maWdQYXJzZXIoKQogICAgICAgIGNvbmZpZy5yZWFkKGNvbmZpZ19maWxlKQog
            ICAgICAgIGZvciBhcmcgaW4gbGlzdCh2YXJzKHNlbGYuX2FyZ19wYXJzZXIpKToKICAgICAgICAg
            ICAgIyBweXRob24gdHJlYXRzIGZsYWctbmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28g
            Y29udmVydGluZyBiYWNrIHRvIGZsYWctbmFtZSwKICAgICAgICAgICAgIyBzbyB3ZSBjYW4gZ2V0
            IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAgIGFyZ3VtZW50ID0gYXJn
            LnJlcGxhY2UoIl8iLCAiLSIpCiAgICAgICAgICAgIGFyZ3VtZW50VmFsdWUgPSBzdHIoZ2V0YXR0
            cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcpKQogICAgICAgICAgICBjb25maWdfdmFsdWUgPSBjb25m
            aWcuZ2V0KCJDb25maWd1cmF0aW9ucyIsIGFyZ3VtZW50LCBmYWxsYmFjaz1Ob25lKQogICAgICAg
            ICAgICAjIGdpdmUgcHJpb3JpdHkgdG8gY29tbWFuZCBsaW5lIGFyZ3VtZW50LCBpZiBjb21tYW5k
            IGxpbmUgYXJndW1lbnQgaXMgbm90IHByZXNlbnQgdXNlIGNvbmZpZy5pbmkgdmFsdWUsCiAgICAg
            ICAgICAgIGlmIChzdHIoc3lzLmFyZ3YpLmZpbmQoYXJndW1lbnQpID09IC0xKSBhbmQgKAogICAg
            ICAgICAgICAgICAgKGNvbmZpZ192YWx1ZSAhPSBOb25lKSBhbmQgKGNvbmZpZ192YWx1ZSAhPSAi
            IikKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuX19zZXRh
            dHRyX18oYXJnLCBjb25maWdfdmFsdWUpCgogICAgICAgIHJldHVybiBjb25maWcKCiAgICBkZWYg
            X19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAgICAgICAgc2VsZi5vdXRfbWFwID0ge30K
            ICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0KCkKICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3RvX3BhcnNlPWFyZ19saXN0KQogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGU6CiAgICAgICAgICAgIHNlbGYuY29uZmln
            ID0gc2VsZi5wYXJzZV9jb25maWdfZmlsZShzZWxmLl9hcmdfcGFyc2VyLmNvbmZpZ19maWxlKQog
            ICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoKQogICAgICAgIHNlbGYucnVu
            X2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRw
            dXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAgICAgc2VsZi5jZXBoX2NvbmYg
            PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNlbGYuY2VwaF9rZXlyaW5nID0g
            c2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1c2VyIG5vdCBwcm92aWRlZCwg
            Z2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxmLnJ1bl9hc191c2VyIGFuZCBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLnJ1bl9hc191c2Vy
            ID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGU6CiAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gc2VsZi5ERUZBVUxUX1JH
            V19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25mOgogICAgICAgICAgICBrd2Fy
            Z3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmluZzoKICAgICAgICAgICAgICAg
            IGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBoX2tleXJpbmd9CiAgICAgICAg
            ICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxlPXNlbGYuY2VwaF9jb25mLCAq
            Kmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIgPSByYWRvcy5S
            YWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3JlYWRfZmlsZSgpCiAgICAgICAg
            c2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToKICAgICAgICBp
            ZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAgICAgICAgICAgIHNlbGYuY2x1
            c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2Vw
            aCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rlci5nZXRfZnNpZCgpKQoKICAg
            IGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNvbik6CiAgICAgICAgY21kID0g
            anNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAgICAgICAgcmV0X3ZhbCwgY21k
            X291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFuZChjbWQsIGIiIikKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAgICAgIHByaW50KGYiQ29tbWFu
            ZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgIGYiUmV0
            dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2NtZF9vdXR9XG4iCiAgICAgICAg
            ICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0tLS0tLS0tLVxuIgogICAgICAg
            ICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJy
            b3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQnIGlzIG5vdCBlbXB0eQogICAg
            ICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24gb3V0cHV0CiAgICAgICAgaWYg
            cmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBqc29uX291dCA9IGpzb24ubG9h
            ZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cKCiAg
            ICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6CiAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9y
            dW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            IyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9ydW1fc3RhdHVzJyBjb21tYW5k
            IGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwg
            IT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICApCiAgICAgICAg
            cV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVyX25hbWUiXQogICAgICAgIHFf
            bGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21hdGNoaW5nX2xpc3QgPSBbCiAg
            ICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJdWyJtb25zIl0gaWYgbFsibmFt
            ZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQiKQogICAgICAgIHFfbGVhZGVy
            X2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAgICAgICAgIyBnZXQgdGhlIGFk
            ZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAgICAgcV9sZWFkZXJfYWRkcnZl
            YyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMiLCB7fSkuZ2V0KCJhZGRydmVj
            IiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9kZXRhaWxzWyJwdWJsaWNfYWRk
            ciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudjJfcG9ydF9l
            bmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIi
            OgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bImFkZHIiXQog
            ICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+IDEgYW5kIHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVh
            ZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAg
            c3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUg
            bm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAg
            ICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9h
            ZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBw
            b3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15
            IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRo
            b2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNp
            bnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYg
            aXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91
            bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUg
            aG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlw
            CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1h
            dCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlz
            IGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBm
            YWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3Zh
            bCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBh
            ZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28g
            dmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRw
            b2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAg
            ICAgICAgaXAgPSBbXQogICAgICAgICAgICAjIGV4YW1wbGUgb3V0cHV0IFsoPEFkZHJlc3NGYW1p
            bHkuQUZfSU5FVDogMj4sIDxTb2NrZXRLaW5kLlNPQ0tfU1RSRUFNOiAxPiwgNiwgJycsICgnOTMu
            MTg0LjIxNi4zNCcsIDgwKSksIC4uLl0KICAgICAgICAgICAgIyB3ZSBuZWVkIHRvIGdldCA5My4x
            ODQuMjE2LjM0IHNvIGl0IHdvdWxkIGJlIGlwWzBdWzRdWzBdCiAgICAgICAgICAgIGlmIGhvc3Rf
            aXBfdHlwZSA9PSAiSVB2NiI6CiAgICAgICAgICAgICAgICBpcCA9IHNvY2tldC5nZXRhZGRyaW5m
            bygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZhbWlseT1zb2NrZXQuQUZf
            SU5FVDYsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbGlmIGhvc3RfaXBfdHlwZSA9PSAiSVB2NCI6CiAgICAgICAgICAgICAgICBpcCA9IHNv
            Y2tldC5nZXRhZGRyaW5mbygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZh
            bWlseT1zb2NrZXQuQUZfSU5FVCwgcHJvdG89c29ja2V0LklQUFJPVE9fVENQCiAgICAgICAgICAg
            ICAgICApCiAgICAgICAgICAgIGRlbCBzb2NrZXQKICAgICAgICAgICAgcmV0dXJuIGlwWzBdWzRd
            WzBdCiAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQoKICAgIGRlZiBnZXRfYWN0aXZlX2FuZF9zdGFu
            ZGJ5X21ncnMoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAg
            ICAgICAgICByZXR1cm4gIiIsIHNlbGYuZHJ5X3J1bigiY2VwaCBzdGF0dXMiKQogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gc2VsZi5fYXJn
            X3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50CiAgICAgICAgc3RhbmRieV9tZ3JzID0gW10KICAg
            ICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0OgogICAgICAgICAgICBjbWRf
            anNvbiA9IHsicHJlZml4IjogInN0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1w
            dCwKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAg
            ICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICAgICAgIidtZ3Igc2VydmljZXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAg
            ICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlf
            T1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtb25pdG9yaW5nX2Vu
            ZHBvaW50ID0gKAogICAgICAgICAgICAgICAganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0
            KCJzZXJ2aWNlcyIsIHt9KS5nZXQoInByb21ldGhldXMiLCAiIikKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgImNhbid0IGZp
            bmQgbW9uaXRvcmluZ19lbmRwb2ludCwgcHJvbWV0aGV1cyBtb2R1bGUgbWlnaHQgbm90IGJlIGVu
            YWJsZWQsICIKICAgICAgICAgICAgICAgICAgICAiZW5hYmxlIHRoZSBtb2R1bGUgYnkgcnVubmlu
            ZyAnY2VwaCBtZ3IgbW9kdWxlIGVuYWJsZSBwcm9tZXRoZXVzJyIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgIyBub3cgY2hlY2sgdGhlIHN0YW5kLWJ5IG1nci1zCiAgICAgICAgICAgIHN0
            YW5kYnlfYXJyID0ganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0KCJzdGFuZGJ5cyIsIFtd
            KQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5IGluIHN0YW5kYnlfYXJyOgogICAgICAgICAg
            ICAgICAgaWYgIm5hbWUiIGluIGVhY2hfc3RhbmRieS5rZXlzKCk6CiAgICAgICAgICAgICAgICAg
            ICAgc3RhbmRieV9tZ3JzLmFwcGVuZChlYWNoX3N0YW5kYnlbIm5hbWUiXSkKICAgICAgICAgICAg
            dHJ5OgogICAgICAgICAgICAgICAgcGFyc2VkX2VuZHBvaW50ID0gdXJscGFyc2UobW9uaXRvcmlu
            Z19lbmRwb2ludCkKICAgICAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYi
            aW52YWxpZCBlbmRwb2ludDoge21vbml0b3JpbmdfZW5kcG9pbnR9IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBwYXJzZWRfZW5kcG9p
            bnQuaG9zdG5hbWUKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoK
            ICAgICAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHN0cihwYXJzZWRfZW5k
            cG9pbnQucG9ydCkKCiAgICAgICAgIyBpZiBtb25pdG9yaW5nIGVuZHBvaW50IHBvcnQgaXMgbm90
            IHNldCwgcHV0IGEgZGVmYXVsdCBtb24gcG9ydAogICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQ6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYu
            REVGQVVMVF9NT05JVE9SSU5HX0VORFBPSU5UX1BPUlQKCiAgICAgICAgIyB1c2VyIGNvdWxkIGdp
            dmUgY29tbWEgYW5kIHNwYWNlIHNlcGFyYXRlZCBpbnB1dHMgKGxpa2UgLS1tb25pdG9yaW5nLWVu
            ZHBvaW50PSI8aXAxPiwgPGlwMj4iKQogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlz
            dCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdC5yZXBsYWNlKCIsIiwgIiAiKQogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCA9IG1vbml0b3JpbmdfZW5kcG9pbnRf
            aXBfbGlzdC5zcGxpdCgpCiAgICAgICAgIyBpZiBtb25pdG9yaW5nLWVuZHBvaW50IGNvdWxkIG5v
            dCBiZSBmb3VuZCwgcmFpc2UgYW4gZXJyb3IKICAgICAgICBpZiBsZW4obW9uaXRvcmluZ19lbmRw
            b2ludF9pcF9saXN0X3NwbGl0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKCJObyAnbW9uaXRvcmluZy1lbmRwb2ludCcgZm91bmQiKQogICAgICAgICMg
            Zmlyc3QgaXAgaXMgdHJlYXRlZCBhcyB0aGUgbWFpbiBtb25pdG9yaW5nLWVuZHBvaW50CiAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9z
            cGxpdFswXQogICAgICAgICMgcmVzdCBvZiB0aGUgaXAtcyBhcmUgYWRkZWQgdG8gdGhlICdzdGFu
            ZGJ5X21ncnMnIGxpc3QKICAgICAgICBzdGFuZGJ5X21ncnMuZXh0ZW5kKG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXBfbGlzdF9zcGxpdFsxOl0pCiAgICAgICAgZmFpbGVkX2lwID0gbW9uaXRvcmluZ19l
            bmRwb2ludF9pcAoKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50ID0gIjoiLmpvaW4oCiAgICAg
            ICAgICAgIFttb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnRd
            CiAgICAgICAgKQogICAgICAgIGlwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KG1vbml0
            b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBv
            aW50X2lwID0gc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQog
            ICAgICAgICAgICApCiAgICAgICAgICAgICMgY29sbGVjdCBhbGwgdGhlICdzdGFuZC1ieScgbWdy
            IGlwcwogICAgICAgICAgICBtZ3JfaXBzID0gW10KICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRi
            eV9tZ3IgaW4gc3RhbmRieV9tZ3JzOgogICAgICAgICAgICAgICAgZmFpbGVkX2lwID0gZWFjaF9z
            dGFuZGJ5X21ncgogICAgICAgICAgICAgICAgbWdyX2lwcy5hcHBlbmQoCiAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgICAgICAg
            ICAgZWFjaF9zdGFuZGJ5X21nciwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAg
            ICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiQ29udmVyc2lvbiBvZiBob3N0OiB7ZmFpbGVkX2lwfSB0byBJUCBmYWlsZWQuICIKICAgICAg
            ICAgICAgICAgICJQbGVhc2UgZW50ZXIgdGhlIElQIGFkZHJlc3NlcyBvZiBhbGwgdGhlIGNlcGgt
            bWdycyB3aXRoIHRoZSAnLS1tb25pdG9yaW5nLWVuZHBvaW50JyBmbGFnIgogICAgICAgICAgICAp
            CgogICAgICAgIF8sIF8sIGVyciA9IHNlbGYuZW5kcG9pbnRfZGlhbChtb25pdG9yaW5nX2VuZHBv
            aW50LCBpcF90eXBlKQogICAgICAgIGlmIGVyciA9PSAiLTEiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGVycikKICAgICAgICAjIGFkZCB0aGUgdmFsaWRhdGVk
            IGFjdGl2ZSBtZ3IgSVAgaW50byB0aGUgZmlyc3QgaW5kZXgKICAgICAgICBtZ3JfaXBzLmluc2Vy
            dCgwLCBtb25pdG9yaW5nX2VuZHBvaW50X2lwKQogICAgICAgIGFsbF9tZ3JfaXBzX3N0ciA9ICIs
            Ii5qb2luKG1ncl9pcHMpCiAgICAgICAgcmV0dXJuIGFsbF9tZ3JfaXBzX3N0ciwgbW9uaXRvcmlu
            Z19lbmRwb2ludF9wb3J0CgogICAgZGVmIGNoZWNrX3VzZXJfZXhpc3Qoc2VsZiwgdXNlcik6CiAg
            ICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJhdXRoIGdldCIsICJlbnRpdHkiOiBmInt1c2Vy
            fSIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgIGlmIHJldF92YWwgIT0gMCBv
            ciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIHJldHVy
            biBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaGZzX3Byb3Zpc2lvbmVy
            X2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAi
            YWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRh
            dGE9KiIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFk
            YXRhPXtjZXBoZnNfZmlsZXN5c3RlbX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAg
            ICBkZWYgZ2V0X2NlcGhmc19ub2RlX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRp
            dHkgPSAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIKICAgICAgICBjYXBzID0gewogICAgICAgICAg
            ICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAg
            ICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAiYWxsb3cgcncgdGFn
            IGNlcGhmcyAqPSoiLAogICAgICAgICAgICAibWRzIjogImFsbG93IHJ3IiwKICAgICAgICB9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAi
            azhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVy
            LW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lz
            dGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAg
            IGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7
            ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVz
            eXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgKj17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5Cgog
            ICAgZGVmIGdldF9lbnRpdHkoCiAgICAgICAgc2VsZiwKICAgICAgICBlbnRpdHksCiAgICAgICAg
            cmJkX3Bvb2xfbmFtZSwKICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgIGs4c19j
            bHVzdGVyX25hbWUsCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgKToKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgIG9y
            IHJiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgIG9yIGFsaWFzX3JiZF9w
            b29sX25hbWUgIT0gIiIKICAgICAgICAgICAgIyBjaGVja2luZyBhbGlhc19yYmRfcG9vbF9uYW1l
            IGlzIG5vdCBlbXB0eSBhcyB0aGVyZSBtYXliZSBhIHNwZWNpYWwgY2hhcmFjdGVyIHVzZWQgb3Ro
            ZXIgdGhhbiAuIG9yIF8KICAgICAgICApOgogICAgICAgICAgICBpZiBhbGlhc19yYmRfcG9vbF9u
            YW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICAgICAicGxlYXNlIHNldCB0aGUgJy0tYWxpYXMtcmJkLWRhdGEt
            cG9vbC1uYW1lJyBmbGFnIGFzIHRoZSByYmQgZGF0YSBwb29sIG5hbWUgY29udGFpbnMgJy4nIG9y
            ICdfJyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAg
            YWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgICAgIG9yIGFs
            aWFzX3JiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgICk6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICInLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgdmFsdWUgc2hvdWxkIG5vdCBj
            b250YWluICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9
            IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXthbGlhc19yYmRfcG9vbF9uYW1lfSIKICAg
            ICAgICBlbHNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfS17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZToKICAgICAg
            ICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICByZXR1
            cm4gZW50aXR5CgogICAgZGVmIGdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNf
            Y2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJz
            ZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJi
            ZC1kYXRhLXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYg
            azhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAg
            ICAgICAgZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAg
            ICAgIGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1l
            LAogICAgICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmIHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFt
            ZX0gbmFtZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBv
            b2w9e3JiZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVm
            IGdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktcmJkLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            b3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9
            IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xfbmFtZQog
            ICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rl
            cl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFk
            b3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgazhzX2Ns
            dXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBmb3VuZCwg
            cGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAgICAgICAg
            ZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAg
            ICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAgICAgIGlm
            IHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gKAog
            ICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0gbmFt
            ZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBvb2w9e3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9k
            ZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gc2VsZi5y
            dW5fYXNfdXNlcgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLAogICAg
            ICAgICAgICAibWdyIjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwKICAgICAgICAgICAgIm9zZCI6
            IGYicHJvZmlsZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD17c2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5yb290LCBhbGxv
            dyBydyBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2ws
            IGFsbG93IHJ4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubG9n
            LCBhbGxvdyB4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuYnVj
            a2V0cy5pbmRleCIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQoKICAg
            ICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jYXBzX2FuZF9lbnRpdHkoc2Vs
            ZiwgdXNlcl9uYW1lKToKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIi
            IGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lv
            bmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3Ry
            aWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRf
            Y2VwaGZzX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBz
            ZWxmLmdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            Y3NpLXJiZC1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50
            LmNzaS1yYmQtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KCkKICAgICAg
            ICBpZiAiY2xpZW50LmNzaS1yYmQtbm9kZSIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAi
            Y2xpZW50LmNzaS1yYmQtbm9kZSIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAg
            cmV0dXJuIHNlbGYuZ2V0X3JiZF9ub2RlX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2VudGl0eSgpCgogICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgIGYibm8gdXNlciBmb3VuZCB3aXRo
            IHVzZXJfbmFtZToge3VzZXJfbmFtZX0sICIKICAgICAgICAgICAgImdldF9jYXBzX2FuZF9lbnRp
            dHkgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcihzZWxmLCB1c2VyKToKICAgICAgICAiIiIKICAgICAgICBjb21tYW5kOiBjZXBo
            IGF1dGggZ2V0LW9yLWNyZWF0ZSBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciBtb24gJ2Fs
            bG93IHInIG1nciAnYWxsb3cgcncnIG9zZCAnYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            JwogICAgICAgICIiIgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2Vu
            dGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1
            dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAg
            ICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBj
            YXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0K
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4g
            KAogICAgICAgICAgICAgICAgc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgICAgICJjZXBo
            ICIKICAgICAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAg
            ICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50aXR5Il0KICAg
            ICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21k
            X2pzb25bImNhcHMiXSkKICAgICAgICAgICAgICAgICksCiAgICAgICAgICAgICAgICAiIiwKICAg
            ICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAg
            dXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJf
            a2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXksIGYie2VudGl0eS5zcGxpdCgn
            LicsIDEpWzFdfSIKICAgICAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4nLDEpWzFdIHRvIHJlbmFt
            ZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1ub2RlCgogICAgICAgIHJl
            dF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRf
            anNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAg
            ICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGdl
            dC1vci1jcmVhdGUge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYi
            RXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJ
            U1R9IgogICAgICAgICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0p
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4n
            LDEpWzFdIHRvIHJlbmFtZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1u
            b2RlCgogICAgZGVmIGdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoc2VsZik6CiAgICAgICAg
            Y21kX2pzb24gPSB7InByZWZpeCI6ICJmcyBscyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            aWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlf
            cnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsIHJlcG9ydCBhbiBlcnJvcgog
            ICAgICAgIGlmIHJldF92YWwgIT0gMDoKICAgICAgICAgICAgIyBpZiBmcyBhbmQgZGF0YV9wb29s
            IGFyZ3VtZW50cyBhcmUgbm90IHNldCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSA9
            PSAiIgogICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9v
            bF9uYW1lID09ICIiCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICByZXR1cm4KICAgICAg
            ICAgICAgIyBpZiB1c2VyIGhhcyBwcm92aWRlZCBhbnkgb2YgdGhlCiAgICAgICAgICAgICMgJy0t
            Y2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgb3IgJy0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lJyBhcmd1
            bWVudHMsCiAgICAgICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIGFzIHdlIGFyZSB1bmFibGUg
            dG8gdmVyaWZ5IHRoZSBhcmdzCiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIidmcyBscycgY2VwaCBjYWxsIGZhaWxlZCB3aXRoIGVy
            cm9yOiB7ZXJyX21zZ30iCiAgICAgICAgICAgICkKCiAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQg
            PSB7fQogICAgICAgICMgaWYgJy0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgYXJndW1lbnQgaXMg
            cHJvdmlkZWQsCiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBwcm92aWRlZCBmaWxlc3lzdGVt
            LW5hbWUgZXhpc3RzIG9yIG5vdAogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZToKICAgICAgICAgICAgIyBnZXQgdGhlIG1hdGNoaW5nIGxpc3QKICAgICAg
            ICAgICAgbWF0Y2hpbmdfanNvbl9vdXRfbGlzdCA9IFsKICAgICAgICAgICAgICAgIG1hdGNoZWQK
            ICAgICAgICAgICAgICAgIGZvciBtYXRjaGVkIGluIGpzb25fb3V0CiAgICAgICAgICAgICAgICBp
            ZiBzdHIobWF0Y2hlZFsibmFtZSJdKSA9PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWUKICAgICAgICAgICAgXQogICAgICAgICAgICAjIHVuYWJsZSB0byBmaW5kIGEgbWF0
            Y2hpbmcgZnMtbmFtZSwgcmFpc2UgYW4gZXJyb3IKICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5n
            X2pzb25fb3V0X2xpc3QpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiRmlsZXN5c3RlbSBwcm92aWRlZCwg
            J3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWV9JywgIgogICAgICAgICAg
            ICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBmcy1saXN0OiB7W3N0cih4WyduYW1lJ10p
            IGZvciB4IGluIGpzb25fb3V0XX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1hdGNo
            aW5nX2pzb25fb3V0ID0gbWF0Y2hpbmdfanNvbl9vdXRfbGlzdFswXQogICAgICAgICMgaWYgY2Vw
            aGZzIGZpbGVzeXN0ZW0gbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgIyB0cnkgdG8gZ2V0
            IGEgZGVmYXVsdCBmcyBuYW1lIGJ5IGRvaW5nIHRoZSBmb2xsb3dpbmcKICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAjIGEuIGNoZWNrIGlmIHRoZXJlIGlzIG9ubHkgb25lIGZpbGVzeXN0ZW0gaXMg
            cHJlc2VudAogICAgICAgICAgICBpZiBsZW4oanNvbl9vdXQpID09IDE6CiAgICAgICAgICAgICAg
            ICBtYXRjaGluZ19qc29uX291dCA9IGpzb25fb3V0WzBdCiAgICAgICAgICAgICMgYi4gb3IgZWxz
            ZSwgY2hlY2sgaWYgZGF0YV9wb29sIG5hbWUgaXMgcHJvdmlkZWQKICAgICAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICMg
            YW5kIGlmIHByZXNlbnQsIGNoZWNrIHdoZXRoZXIgdGhlcmUgZXhpc3RzIGEgZnMgd2hpY2ggaGFz
            IHRoZSBkYXRhX3Bvb2wKICAgICAgICAgICAgICAgIGZvciBlYWNoSiBpbiBqc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSBpbiBlYWNoSlsiZGF0YV9wb29scyJdOgogICAgICAgICAgICAgICAgICAgICAgICBtYXRjaGlu
            Z19qc29uX291dCA9IGVhY2hKCiAgICAgICAgICAgICAgICAgICAgICAgIGJyZWFrCiAgICAgICAg
            ICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIG1hdGNoaW5nIGZzIGV4aXN0cywgdGhhdCBtZWFucyBw
            cm92aWRlZCBkYXRhX3Bvb2wgbmFtZSBpcyBpbnZhbGlkCiAgICAgICAgICAgICAgICBpZiBub3Qg
            bWF0Y2hpbmdfanNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhX3Bv
            b2wgbmFtZSwge3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfSwiCiAgICAg
            ICAgICAgICAgICAgICAgICAgICIgZG9lcyBub3QgZXhpc3RzIgogICAgICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgIyBjLiBpZiBub3RoaW5nIGlzIHNldCBhbmQgY291bGRuJ3QgZmluZCBh
            IGRlZmF1bHQsCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGp1c3QgcmV0dXJu
            IHNpbGVudGx5CiAgICAgICAgICAgICAgICByZXR1cm4KCiAgICAgICAgaWYgbWF0Y2hpbmdfanNv
            bl9vdXQ6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSA9IHN0cihtYXRjaGluZ19qc29uX291dFsibmFtZSJdKQogICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19tZXRhZGF0YV9wb29sX25hbWUgPSBzdHIoCiAgICAgICAgICAgICAgICBt
            YXRjaGluZ19qc29uX291dFsibWV0YWRhdGFfcG9vbCJdCiAgICAgICAgICAgICkKCiAgICAgICAg
            aWYgaXNpbnN0YW5jZShtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdLCBsaXN0KToKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBoYXMgYWxyZWFkeSBwcm92aWRlZCBkYXRhLXBvb2wtbmFt
            ZSwKICAgICAgICAgICAgIyB0aHJvdWdoIC0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lCiAgICAgICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAg
            ICAgICAgIyBpZiB0aGUgcHJvdmlkZWQgbmFtZSBpcyBub3QgbWF0Y2hpbmcgd2l0aCB0aGUgb25l
            IGluIHRoZSBsaXN0CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgICAgICAgICBu
            b3QgaW4gbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXQogICAgICAgICAgICAgICAgKToK
            ICAgICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lOiAne3NlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfScsICIKICAgICAgICAgICAgICAgICAgICAg
            ICAgImRvZXNuJ3QgbWF0Y2ggZnJvbSB0aGUgZGF0YS1wb29scyBsaXN0OiAiCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGYie1tzdHIoeCkgZm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFf
            cG9vbHMnXV19IgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiBkYXRhX3Bv
            b2wgbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgICAgICMgdGhlbiB0cnkgdG8gZmluZCBh
            IGRlZmF1bHQgZGF0YSBwb29sIG5hbWUKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAg
            ICMgaWYgbm8gZGF0YV9wb29scyBleGlzdCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgICAg
            ICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPT0gMDoKICAgICAgICAg
            ICAgICAgICAgICByZXR1cm4KICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25f
            b3V0WyJkYXRhX3Bvb2xzIl1bMF0KICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiB0
            aGVyZSBhcmUgbW9yZSB0aGFuIG9uZSAnZGF0YV9wb29scycgZXhpc3QsCiAgICAgICAgICAgICMg
            dGhlbiB3YXJuIHRoZSB1c2VyIHRoYXQgd2UgYXJlIHVzaW5nIHRoZSBzZWxlY3RlZCBuYW1lCiAg
            ICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdKSA+IDE6CiAg
            ICAgICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgICAgICAiV0FSTklORzogTXVsdGlw
            bGUgZGF0YSBwb29scyBkZXRlY3RlZDogIgogICAgICAgICAgICAgICAgICAgIGYie1tzdHIoeCkg
            Zm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFfcG9vbHMnXV19XG4iCiAgICAgICAgICAg
            ICAgICAgICAgZiJVc2luZyB0aGUgZGF0YS1wb29sOiAne3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lfSdcbiIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2No
            ZWNrZXJLZXkoc2VsZiwgdXNlcik6CiAgICAgICAgY2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fw
            c19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21kX2pzb24gPSB7CiAgICAgICAgICAgICJwcmVm
            aXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAgICAgICAgICAgImVudGl0eSI6IGVudGl0eSwK
            ICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBjYXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMo
            KSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAg
            ICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oCiAgICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAg
            ICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAg
            ICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAgICAgICApCiAgICAg
            ICAgIyBjaGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYu
            Y2hlY2tfdXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAg
            ICAgICAgIHJldHVybiB1c2VyX2tleQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21z
            ZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVy
            ZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHtzZWxmLnJ1bl9h
            c191c2VyfScgY29tbWFuZCBmYWlsZWRcbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJf
            bXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAg
            ICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0pCgogICAgZGVmIGdl
            dF9jZXBoX2Rhc2hib2FyZF9saW5rKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgi
            OiAibWdyIHNlcnZpY2VzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiAr
            IGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1
            bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25f
            b3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIGlmICJkYXNoYm9hcmQi
            IG5vdCBpbiBqc29uX291dDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICByZXR1cm4g
            anNvbl9vdXRbImRhc2hib2FyZCJdCgogICAgZGVmIGNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIo
            c2VsZik6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAg
            ICAgICAgICJ1c2VyIiwKICAgICAgICAgICAgImNyZWF0ZSIsCiAgICAgICAgICAgICItLXVpZCIs
            CiAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAg
            ICAgICAgICItLWRpc3BsYXktbmFtZSIsCiAgICAgICAgICAgICJSb29rIFJHVyBBZG1pbiBPcHMg
            dXNlciIsCiAgICAgICAgICAgICItLWNhcHMiLAogICAgICAgICAgICAiYnVja2V0cz0qO3VzZXJz
            PSo7dXNhZ2U9cmVhZDttZXRhZGF0YT1yZWFkO3pvbmU9cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgIiAiLmpvaW4oY21kKSkKICAgICAgICB0cnk6CiAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBhbHJlYWR5IGV4aXN0cywgd2UganVzdCBxdWVyeSBpdAog
            ICAgICAgICAgICBpZiBleGVjRXJyLnJldHVybmNvZGUgPT0gZXJybm8uRUVYSVNUOgogICAgICAg
            ICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAg
            ICAgICAgICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICAgICAgICAgImluZm8iLAogICAg
            ICAgICAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5FWFRFUk5B
            TF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUs
            CiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICAgICAg
            ICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICAgICAgICAgIF0KICAgICAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVy
            cj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxs
            ZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgICAgICBlcnJfbXNnID0g
            KAogICAgICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2Nt
            ZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShl
            cnJfbXNnKQogICAgICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0x
            IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4
            ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0
            dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0
            dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5m
            bz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19j
            YXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAg
            ICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAg
            ICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Jl
            YWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUi
            LAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQog
            ICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQo
            Y21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNh
            bGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZl
            cnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRh
            dGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2Fw
            czogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4
            ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2Fw
            X3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJf
            bXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7
            Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJD
            b2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAg
            ICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNv
            bm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAg
            IGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0
            cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9y
            dGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0
            cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3twb29sX25hbWV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYgaW5pdF9yYmRfcG9vbChzZWxmLCBy
            YmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlS
            YWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9w
            ZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAg
            ICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkKCiAgICBkZWYgdmFsaWRhdGVfcmFk
            b3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmJkX21l
            dGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQogICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYgcmFkb3NfbmFt
            ZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRvc19uYW1lc3Bh
            Y2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNl
            LCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRhaW5zIHVwcGVy
            IGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAgICAg
            IGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICBp
            ZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3BhY2UpIGlzIEZh
            bHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFj
            ZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9vbCAne3JiZF9w
            b29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRlX3N1YnZvbHVt
            ZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUpOgog
            ICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAg
            ICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIsCiAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19v
            dXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9j
            ZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAg
            ICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAgICJzdWJ2b2x1
            bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgICAgIGNlcGhm
            c19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3Mu
            Y2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmInN1YnZvbHVt
            ZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3JlYXRlZCIKICAg
            ICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBzZWxmLCBzdWJ2
            b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBwaW5fc2V0dGlu
            ZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAgICAgICAgICAg
            ImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgInBpbiIsCiAg
            ICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9n
            cm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0aW5nLAogICAg
            ICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1
            dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3Mu
            Q2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZvbHVtZV9ncm91
            cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRf
            cmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNzX2tleSA9IHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAgICAgIHNlY3Jl
            dF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0KICAg
            ICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAg
            IGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIvYWRtaW4vaW5m
            bz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmVxdWVzdF91
            cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToKICAgICAgICAg
            ICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJsLAogICAgICAg
            ICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3X2VuZHBvaW50
            KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkKICAgICAgICBl
            eGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90IGFibGUgdG8g
            Y2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAgIGlmIHIxIGlz
            IE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAgICAgICAgKQog
            ICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICAgICAi
            LTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJzdG9yYWdlX2Jh
            Y2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X2VuZHBv
            aW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXIn
            IGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91
            dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVt
            bXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJndyBlbmRwb2lu
            dCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9IHNlbGYuX2lu
            dmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cg
            ZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBv
            aW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd190
            bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2VsZi5lbmRwb2lu
            dF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAgIGlmIGVyciAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBpZiB0aGUgcmd3
            IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMgb25seSBjaGVj
            ayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2FwX3N1cHBvcnRl
            ZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAgICByZ3dfZnNp
            ZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAgICAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAgICAgICAgICBp
            ZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAg
            ICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7c2VsZi5fYXJn
            X3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlkYXRpbmcgYnkg
            Y2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFuZCB2YWxpZGF0
            aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUgY2VwaCBjbHVz
            dGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9vbCBleGlzdAog
            ICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFzc2VkIGVsc2Ug
            aXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0aGVzZSBkZWZh
            dWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggIT0g
            ImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBbCiAgICAgICAg
            ICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3Lm1ldGEiLAog
            ICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAgICAgICAgZiJ7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAgICAgICAgICBd
            CiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bvb2xzX3RvX3Zh
            bGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0cyhf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2wsICd7X3Jnd19w
            b29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4gIiIKCiAgICBk
            ZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1l
            LCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0ZV9jb25maWcg
            IT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWlu
            IiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAgICAgICAgICAg
            ImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVfY29uZmlnLAog
            ICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAgICAgICAgXQog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgIGVycl9t
            c2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtj
            bWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNv
            ZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAg
            ICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAgZGVmIGNvbnZl
            cnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAgICByZXR1cm4g
            dmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5
            X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAgICgKICAgICAg
            ICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIKICAgICAgICAg
            ICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAgIG9yIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIiCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICAgICAg
            ICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxhZ3MgLS10b3Bv
            bG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10b3BvbG9neS1m
            YWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfdG9w
            b2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6CiAgICAgICAg
            aWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJv
            dmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIKICAgICAgICAg
            ICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9JywiCiAgICAg
            ICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3BvbG9neV9wb29s
            cyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAgICAgICAgICkK
            ICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYs
            IHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bv
            b2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgogICAgZGVmIGlu
            aXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAg
            Zm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLmluaXRfcmJk
            X3Bvb2wocG9vbCkKCiAgICAjIHRoaXMgd2lsbCByZXR1cm4gdGhlIGZpbmFsIGFyZ3MgdGhhdCBz
            Y3JpcHQgdXNlcyB0byBwcm9jZXNzCiAgICAjIHRoZSBwcmlvcml0eSB0byBzZXQgYSBwYXJ0aWN1
            bGFyIHZhbHVlIGlzLAogICAgIyBjb21tYW5kLWxpbmUtYXJncyA+IGNvbmZpZy5pbmkgZmlsZSB2
            YWx1ZXMgPiBkZWZhdWx0IHZhbHVlcwogICAgZGVmIGdldEZpbmFsVXNlZEFyZ3Moc2VsZik6CiAg
            ICAgICAgYXJndW1lbnQgPSBmIltDb25maWd1cmF0aW9uc11cbiIKICAgICAgICBmb3IgYXJnIGlu
            IHZhcnMoc2VsZi5fYXJnX3BhcnNlcik6CiAgICAgICAgICAgIGlmIHN0cihnZXRhdHRyKHNlbGYu
            X2FyZ19wYXJzZXIsIGFyZykpOgogICAgICAgICAgICAgICAgIyBweXRob24gdHJlYXRzIGZsYWct
            bmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28gY29udmVydGluZyBiYWNrIHRvIGZsYWct
            bmFtZSwKICAgICAgICAgICAgICAgICMgc28gd2UgY2FuIGdldCB0aG9zZSB2YWx1ZXMgZnJvbSBj
            b25maWcgZmlsZQogICAgICAgICAgICAgICAgYXJnVmFsdWUgPSBhcmcucmVwbGFjZSgiXyIsICIt
            IikKICAgICAgICAgICAgICAgICMgZG8gbm90IGFkZCB0aGUgY29uZmlnLWZpbGUgZmxhZyBhbmQg
            YWxzbyBub3QgYWRkIHRoZSBib29sZWFuIGZsYWdzIHdoaWNoIGFyZSBzZXQgdG8gRmFsc2UKICAg
            ICAgICAgICAgICAgICMgYmVjYXVzZSBjb25maWcuaW5pIGZpbGUgdHJlYXRzIGJvb2xlYW4gZmxh
            Z3MgYXMgVHJ1ZSBhbHdheXMKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBhcmdWYWx1ZSAhPSAiY29uZmlnLWZpbGUiCiAgICAgICAgICAgICAgICAgICAgYW5kIGdldGF0
            dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSAhPSBGYWxzZQogICAgICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgICAgICBhcmd1bWVudCArPSBmInthcmdWYWx1ZX0gPSB7c3RyKGdldGF0dHIo
            c2VsZi5fYXJnX3BhcnNlciwgYXJnKSl9XG4iCiAgICAgICAgcmV0dXJuIGFyZ3VtZW50CgogICAg
            ZGVmIF9nZW5fb3V0cHV0X21hcChzZWxmKToKICAgICAgICBpZiBzZWxmLm91dF9tYXA6CiAgICAg
            ICAgICAgIHJldHVybgogICAgICAgICMgc3VwcG9ydCBsZWdhY3kgZmxhZyB3aXRoIHVwZ3JhZGVz
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWU6CiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuY2x1c3Rl
            cl9uYW1lCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gKAogICAg
            ICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUubG93ZXIoKQogICAgICAg
            ICkgICMgYWx3YXlzIGNvbnZlcnQgY2x1c3RlciBuYW1lIHRvIGxvd2VyY2FzZSBjaGFyYWN0ZXJz
            CiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZSkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2woc2VsZi5fYXJnX3BhcnNlci5y
            YmRfZGF0YV9wb29sX25hbWUpCiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uo
            KQogICAgICAgIHNlbGYuX2V4Y2x1ZGVkX2tleXMuYWRkKCJLOFNfQ0xVU1RFUl9OQU1FIikKICAg
            ICAgICBzZWxmLmdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoKQogICAgICAgICMgZG91Ymxl
            IHN0cmluZyBuZWVkZWQgZm9yIHVwc3RyZWFtIGV4cG9ydHMgb2YgZmxhZ3MKICAgICAgICBzZWxm
            Lm91dF9tYXBbIkFSR1MiXSA9IGYnIntzZWxmLmdldEZpbmFsVXNlZEFyZ3MoKX0iJwogICAgICAg
            IHNlbGYub3V0X21hcFsiTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFyc2VyLm5hbWVzcGFjZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiSzhTX0NMVVNURVJfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0ZT
            SUQiXSA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSTkFNRSJdID0gc2VsZi5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0gPSBzZWxmLmdldF9jZXBoX2V4dGVybmFsX21vbl9k
            YXRhKCkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSA9
            IHNlbGYuY3JlYXRlX2NoZWNrZXJLZXkoCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FS
            RF9MSU5LIl0gPSBzZWxmLmdldF9jZXBoX2Rhc2hib2FyZF9saW5rKCkKICAgICAgICAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICBz
            ZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxm
            LmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1ub2RlIikKICAgICAg
            ICAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVu
            dC5jc2ktcmJkLXByb3Zpc2lvbmVyIikKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09M
            X05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDRVBIRlNfTUVUQURBVEFfUE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAg
            IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRhX3Bvb2xfbmFtZQogICAgICAgICkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVU
            SF9QRVJNSVNTSU9OIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3Rl
            ZF9hdXRoX3Blcm1pc3Npb24KICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQURPU19O
            QU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXSA9IHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1l
            X2dyb3VwCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0gPSAi
            IgogICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXSA9
            ICIiCiAgICAgICAgIyBjcmVhdGUgQ2VwaEZTIG5vZGUgYW5kIHByb3Zpc2lvbmVyIGtleXJpbmcg
            b25seSB3aGVuIE1EUyBleGlzdHMKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19O
            QU1FIl0gYW5kIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAo
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwK
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1F
            Il0sCiAgICAgICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGll
            bnQuY3NpLWNlcGhmcy1ub2RlIikKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLWNl
            cGhmcy1wcm92aXNpb25lciIpCiAgICAgICAgICAgICMgY3JlYXRlIHRoZSBkZWZhdWx0ICJjc2ki
            IHN1YnZvbHVtZWdyb3VwCiAgICAgICAgICAgIHNlbGYuZ2V0X29yX2NyZWF0ZV9zdWJ2b2x1bWVf
            Z3JvdXAoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmls
            ZXN5c3RlbV9uYW1lCiAgICAgICAgICAgICkKICAgICAgICAgICAgIyBwaW4gdGhlIGRlZmF1bHQg
            ImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5waW5fc3Vidm9sdW1lKAogICAg
            ICAgICAgICAgICAgImNzaSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSwgImRpc3RyaWJ1dGVkIiwgIjEiCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXToKICAgICAgICAgICAgICAgIHNlbGYuZ2V0X29yX2Ny
            ZWF0ZV9zdWJ2b2x1bWVfZ3JvdXAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAg
            c2VsZi5waW5fc3Vidm9sdW1lKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            c3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAiZGlzdHJpYnV0ZWQiLAog
            ICAgICAgICAgICAgICAgICAgICIxIiwKICAgICAgICAgICAgICAgICkKICAgICAgICBzZWxmLm91
            dF9tYXBbIlJHV19UTFNfQ0VSVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJ
            TkdfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBP
            SU5UX1BPUlQiXSA9ICIiCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIuc2tpcF9tb25p
            dG9yaW5nX2VuZHBvaW50OgogICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIk1PTklUT1JJTkdfRU5EUE9JTlQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsi
            TU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICkgPSBzZWxmLmdldF9hY3Rp
            dmVfYW5kX3N0YW5kYnlfbWdycygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1F
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoKQogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzCiAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJd
            ID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWlu
            X2xhYmVsCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lf
            ZmFpbHVyZV9kb21haW5fdmFsdWVzCiAgICAgICAgKQogICAgICAgIGlmICgKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICBhbmQgc2Vs
            Zi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAgICAg
            ICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMg
            IT0gIiIKICAgICAgICApOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3ZhbHVl
            cygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXko
            c2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKSwKICAgICAgICAgICAgICAgIHNlbGYuY29u
            dmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXQogICAgICAgICAgICAgICAg
            KSwKICAgICAgICAgICAgKQogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3JiZF9w
            b29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJy
            YXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIHNlbGYuaW5pdF90b3BvbG9neV9yYmRfcG9vbHMoCiAgICAgICAgICAgICAgICBzZWxmLmNv
            bnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSkKICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHNlbGYucmFpc2Vf
            ZXhjZXB0aW9uX2lmX2FueV90b3BvbG9neV9mbGFnX2lzX21pc3NpbmcoKQoKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJHV19QT09MX1BSRUZJWCJdID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXgKICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gIiIKICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludDoKICAgICAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICAgICAgc2VsZi5jcmVhdGVfcmd3X2FkbWluX29w
            c191c2VyKCkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAh
            PSAiIgogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZh
            bGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUsICJyZWFsbSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwgInpv
            bmVncm91cCIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgZXJyID0g
            c2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsICJ6b25lIgogICAgICAgICAgICAgICAgICAgICkK
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgICkgb3Ig
            KAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgIT0g
            IiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3Vw
            X25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgICgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19B
            RE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIGluZm9f
            Y2FwX3N1cHBvcnRlZCwKICAgICAgICAgICAgICAgICAgICAgICAgZXJyLAogICAgICAgICAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICAg
            ICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50KGluZm9fY2FwX3N1cHBvcnRl
            ZCkKICAgICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9w
            YXRoOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9p
            bnRfdGxzX2NlcnQoKQogICAgICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciwgc2V0IHRoZSBSR1dfRU5EUE9JTlQKICAgICAgICAg
            ICAgICAgICAgICBpZiBlcnIgIT0gIi0xIjoKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJSR1dfRU5EUE9JTlQiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIGVyciA9ICJQbGVhc2UgcHJv
            dmlkZSBhbGwgdGhlIFJHVyBtdWx0aXNpdGUgcGFyYW1ldGVycyBvciBub25lIG9mIHRoZW0iCiAg
            ICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnIpCgogICAgZGVmIGdlbl9zaGVs
            bF9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAgICAgIHNoT3V0
            SU8gPSBTdHJpbmdJTygpCiAgICAgICAgZm9yIGssIHYgaW4gc2VsZi5vdXRfbWFwLml0ZW1zKCk6
            CiAgICAgICAgICAgIGlmIHYgYW5kIGsgbm90IGluIHNlbGYuX2V4Y2x1ZGVkX2tleXM6CiAgICAg
            ICAgICAgICAgICBzaE91dElPLndyaXRlKGYiZXhwb3J0IHtrfT17dn17TElORVNFUH0iKQogICAg
            ICAgIHNoT3V0ID0gc2hPdXRJTy5nZXR2YWx1ZSgpCiAgICAgICAgc2hPdXRJTy5jbG9zZSgpCiAg
            ICAgICAgcmV0dXJuIHNoT3V0CgogICAgZGVmIGdlbl9qc29uX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICBqc29uX291dCA9IFsKICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgIm5hbWUiOiAiZXh0ZXJuYWwtY2x1c3Rlci11c2VyLWNvbW1hbmQi
            LAogICAgICAgICAgICAgICAgImtpbmQiOiAiQ29uZmlnTWFwIiwKICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICJhcmdzIjogc2VsZi5vdXRfbWFwWyJBUkdTIl0s
            CiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAg
            ICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uLWVuZHBvaW50cyIsCiAgICAgICAgICAgICAg
            ICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAg
            ICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05f
            REFUQSJdLAogICAgICAgICAgICAgICAgICAgICJtYXhNb25JZCI6ICIwIiwKICAgICAgICAgICAg
            ICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9
            LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwK
            ICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4tc2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAg
            ICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJ
            RCJdLAogICAgICAgICAgICAgICAgICAgICJtb24tc2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9yLWNyZWRzIiwKICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxf
            VVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAg
            XQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBh
            ZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENsdXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAg
            ICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBz
            ZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9pbnQiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdFbmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAgICAgICAiTW9uaXRvcmluZ1Bv
            cnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgIyBp
            ZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1y
            YmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lf
            UkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJLZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAg
            ICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JC
            RF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJk
            LXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNS
            RVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJJRCI6
            IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQog
            ICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQn
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNy
            ZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSdd
            fSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAg
            ICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lP
            TkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAg
            ICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfQ0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMs
            IHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhmcy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAg
            ICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewog
            ICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBI
            RlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2Vj
            cmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAg
            ICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAg
            ICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRf
            TElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNl
            Y3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElO
            SyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxp
            bmsiLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0Un
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJBRE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAg
            ICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAg
            ICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0K
            ICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAg
            ICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJyYWRv
            c05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAg
            ICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09M
            X05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92
            aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9y
            YWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5v
            dXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            IGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xf
            TkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6
            IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXIt
            c2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9T
            RUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9t
            YXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICAgICAgfQogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAog
            ICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2Vw
            aC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1l
            IjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUn
            XX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgICAgICkKCiAgICAg
            ICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwn
            LCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhl
            biBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VDbGFzcwogICAgICAgIGlmICgKICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91
            dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTIl0KICAgICAgICApOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9wb2xvZ3kiLAogICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYu
            b3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RP
            TUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAgICAgICAgXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZhbHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAg
            ICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAg
            ICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhw
            YW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycgU3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2Vs
            Zi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwK
            ICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5v
            dXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6
            IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAg
            ICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtz
            ZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAg
            ICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5k
            LXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBz
            ZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ci
            LAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJlbmRwb2ludCI6IHNl
            bGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            UHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZ
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJzZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2VwaC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0
            LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJj
            ZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImNl
            cnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdLAogICAgICAgICAgICAgICAgICAgIH0s
            CiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVt
            cHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRlZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25z
            KHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMt
            bm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAg
            ICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwKICAgICAgICAgICAgImNsaWVudC5jc2ktcmJk
            LXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAg
            ICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNlciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNl
            ciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVzZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2Vy
            KQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgogICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNl
            cl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBnZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBn
            cmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAgICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRo
            Y2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBhZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBw
            b29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAgICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBu
            YW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIgdXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcg
            Y29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdldCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMg
            cmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAg
            ICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0oLio/KVwucmd3XC5tZXRhIgogICAgICAgICAg
            ICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVybiwgY2FwcykKICAgICAgICAgICAgICAgIGlm
            IG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wgbmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAg
            ICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2Vy
            KToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhlIGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAog
            ICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNl
            cn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNn
            ID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBwcmludChmInVzZXIge3Vz
            ZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIpCiAgICAgICAgICAgIHJldHVybgogICAgICAg
            IGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsiY2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3
            X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2VyLCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAg
            ICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNh
            cF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIsICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQog
            ICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlzOgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVl
            cyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBuZXdfY2FwX3Zh
            bHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBjdXJfY2FwX3Blcm1f
            bGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVl
            cy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBdCiAgICAgICAgICAgIG5ld19j
            YXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAgICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19j
            YXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJpcCgpCiAgICAgICAgICAgIF0KICAgICAgICAg
            ICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUg
            b3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2Nh
            cF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxpbWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1
            c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCByZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3Qg
            YW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIKICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0g
            W10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5hcHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9w
            ZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9saXN0XQogICAgICAgICAgICBleGlzdGluZ19j
            YXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19jYXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhp
            c3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAgICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXAp
            CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChleGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAg
            ICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1dGggY2FwcyIsCiAgICAg
            ICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAgICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAg
            ICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQogICAgICAgIHJldF92YWwsIGpzb25fb3V0
            LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBp
            ZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNhcHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5c
            biBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVk
            IHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoKICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAg
            IGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFk
            ZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBl
            bGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJh
            dGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19w
            YXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNl
            bGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0
            OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBzZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0
            ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0gb3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2Rl
            PSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAgICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRf
            b3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3NlKCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19tYWluX18iOgogICAgcmpPYmogPSBSYWRvc0pT
            T04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1haW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAgIHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDog
            e2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAgZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAg
            ICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9z
            RXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hpbGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0
            YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAgICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:35e88716ace3576683a9c4aa686df5f7a8b5bb6ee7a144fd2d87e3c9f5bf3961
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        version: 4.17.13-rhodf
      entries:
      - name: rook-ceph-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: rook-ceph-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: rook-ceph-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: rook-ceph-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: rook-ceph-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: rook-ceph-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: rook-ceph-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: rook-ceph-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: rook-ceph-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: rook-ceph-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: rook-ceph-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: rook-ceph-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: rook-ceph-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: rook-ceph-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: rook-ceph-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9VHJ1ZSwgICMgZGVmYXVsdCBpcyBUcnVlIHRvIGVuYWJsZSB2MiBtb24g
            cG9ydCBpbiBkb3duc3RyZWFtCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9ydCgz
            MzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQLmFk
            ZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAgICAg
            ICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24iLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJhc2gp
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3ZpZGVk
            IGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmlsZXN5
            c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAg
            ICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZzIG1l
            dGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50
            KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRl
            cyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9vbC1u
            YW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2Us
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRhIHBv
            b2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBpbiB0
            aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAgKQog
            ICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1lbmRw
            b2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+OjxQ
            T1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAg
            IHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50
            IFRMUyBjZXJ0aWZpY2F0ZSAob3IgaW50ZXJtZWRpYXRlIHNpZ25pbmcgY2VydGlmaWNhdGUpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAg
            IGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBUTFMgY2VydGlmaWNhdGlv
            biB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0ZSBpcyBwcm92aWRlZCAo
            Tk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRl
            ZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJD
            ZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMgKGNvbW1hIHNlcGFyYXRl
            ZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBvciBgPEZRRE4+YCkgZW50
            cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAgICkKICAgICAgICBvdXRw
            dXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25pdG9yaW5nLWVuZHBvaW50
            LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgcG9y
            dCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFs
            c2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRv
            IG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0aGUgQ2VwaCBjbHVzdGVy
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0
            aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29sIiwKICAgICAgICApCiAg
            ICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tZHJ5LXJ1biIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1
            ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBleGVjdXRlZCBjb21tYW5k
            cyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAg
            IGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxw
            PSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFtZXNwYWNlcywgdXNlZCBm
            b3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSAoc2hv
            dWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwKICAgICAgICAgICAgZGVm
            YXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9InBy
            b3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAogICAgICAgICkKICAgICAg
            ICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0tbmFt
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3LXJlYWxtIiwKICAg
            ICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0t
            cmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3
            LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhl
            IG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1wb29scyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25zdHJhaW5lZCByYmQgcG9v
            bHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAg
            ICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJrOHMg
            Y2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9uZSwgcmFjaywgb3IgaG9z
            dCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBjZXBoIGRvbWFpbiIsCiAg
            ICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICIt
            LXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBh
            cmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gdmFsdWVzIGNvcnJl
            c3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3BvbG9neS1wb29sc2AgbGlz
            dCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJnUC5hZGRfYXJndW1lbnRf
            Z3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAg
            ICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iVXBncmFkZXMgdGhlIGNl
            cGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIp
            IGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGggbmV3IHBlcm1pc3Npb25z
            IG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9sZGVyIHBlcm1pc3Npb24g
            d2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhv
            bjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3Jh
            ZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3NpIHVzZXJzKG5vbi1yZXN0
            cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNlcnMoRm9yIGV4YW1wbGU6
            IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMpLCB1
            c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24gZmxhZyBuZWVk
            IHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJtYW5kYXRvcnkgZmxhZ3M6
            ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1lIGFuZCAtLXJ1bi1hcy11
            c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsgImluIGNhc2Ugb2YgY2Vw
            aGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUgZmxh
            ZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5nIGl0IHdpbGwgYmUgbWFu
            ZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBo
            L2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3JhZGUgLS1yYmQtZGF0
            YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlICAt
            LXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3JhZ2UtcmVwbGljYXBvb2xg
            IgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3RyaWN0ZWQgdXNlciBjYW5u
            b3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVwZ3JhZGluZy4iCiAgICAg
            ICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2VkIHRvIGFwcGVuZCBuZXcg
            cGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2VkIGZvciBjaGFuZ2luZyB1
            c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBsZSB5b3Ugc2hvdWxkbid0
            IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAgICAgICAgKQoKICAgICAg
            ICAjIEFkZCBjb21tYW5kLWxpbmUgYXJndW1lbnRzCiAgICAgICAgY29uZmlnX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoImNvbmZpZyIpCiAgICAgICAgY29uZmlnX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY29uZmlnLWZpbGUiLAogICAgICAgICAgICB0eXBlPXN0
            ciwKICAgICAgICAgICAgaGVscD0iUGF0aCB0byB0aGUgY29uZmlndXJhdGlvbiBmaWxlLCBQcmlv
            cml0eTogY29tbWFuZC1saW5lLWFyZ3MgPiBjb25maWcuaW5pIGZpbGUgdmFsdWVzID4gZGVmYXVs
            dCB2YWx1ZXMiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICApCgogICAgICAgIGlmIGFyZ3NfdG9fcGFyc2U6CiAgICAgICAgICAgIGFz
            c2VydCAoCiAgICAgICAgICAgICAgICB0eXBlKGFyZ3NfdG9fcGFyc2UpID09IGxpc3QKICAgICAg
            ICAgICAgKSwgIkFyZ3VtZW50IHRvICdnZW5fYXJnX3BhcnNlcicgc2hvdWxkIGJlIGEgbGlzdCIK
            ICAgICAgICBlbHNlOgogICAgICAgICAgICBhcmdzX3RvX3BhcnNlID0gc3lzLmFyZ3ZbMTpdCiAg
            ICAgICAgcmV0dXJuIGFyZ1AucGFyc2VfYXJncyhhcmdzX3RvX3BhcnNlKQoKICAgIGRlZiB2YWxp
            ZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xf
            bmFtZQogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0
            YV9wb29sX25hbWUKCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJGbGFnICctLXJiZC1kYXRhLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAg
            ICAgICAgICAgICAgICkKCiAgICAgICAgICAgIGlmIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUg
            PT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAog
            ICAgICAgICAgICAgICAgICAgICJGbGFnICctLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBjbWRfanNv
            biA9IHsicHJlZml4IjogIm9zZCBkdW1wIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAg
            cmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNt
            ZF9qc29uKQogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAw
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICBmIntjbWRfanNvblsncHJlZml4J119IGNvbW1hbmQgZmFpbGVkLlxuIgog
            ICAgICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNl
            IHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBt
            ZXRhZGF0YV9wb29sX2V4aXN0LCBwb29sX2V4aXN0ID0gRmFsc2UsIEZhbHNlCgogICAgICAgICAg
            ICBmb3Iga2V5IGluIGpzb25fb3V0WyJwb29scyJdOgogICAgICAgICAgICAgICAgIyBpZiBlcmFz
            dXJlX2NvZGVfcHJvZmlsZSBpcyBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IHJl
            cGxpY2EgcG9vbAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGtleVsi
            ZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBrZXlb
            InBvb2xfbmFtZSJdID09IHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgICAg
            ICk6CiAgICAgICAgICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCA9IFRydWUKICAgICAg
            ICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgbm90IGVtcHR5IGFuZCBwb29s
            IG5hbWUgZXhpc3RzIHRoZW4gaXQgaXMgZWMgcG9vbAogICAgICAgICAgICAgICAgaWYga2V5WyJl
            cmFzdXJlX2NvZGVfcHJvZmlsZSJdIGFuZCBrZXlbInBvb2xfbmFtZSJdID09IHJiZF9wb29sX25h
            bWU6CiAgICAgICAgICAgICAgICAgICAgcG9vbF9leGlzdCA9IFRydWUKCiAgICAgICAgICAgIGlm
            IG5vdCBtZXRhZGF0YV9wb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9u
            RmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiUHJvdmlkZWQgcmJkX2VjX21l
            dGFkYXRhX3Bvb2wgbmFtZSwiCiAgICAgICAgICAgICAgICAgICAgZiIge3JiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgaWYgbm90IHBvb2xfZXhpc3Q6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgcmJkX2RhdGFfcG9v
            bCBuYW1lLCB7cmJkX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICByZXR1cm4gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQoKICAgIGRlZiBk
            cnlfcnVuKHNlbGYsIG1zZyk6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICBwcmludCgiRXhlY3V0ZTogIiArICInIiArIG1zZyArICInIikKCiAgICBkZWYg
            dmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgIHdpdGggb3BlbihzZWxmLl9h
            cmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoLCBlbmNvZGluZz0idXRmOCIpIGFzIGY6CiAgICAg
            ICAgICAgICAgICBjb250ZW50cyA9IGYucmVhZCgpCiAgICAgICAgICAgICAgICByZXR1cm4gY29u
            dGVudHMucnN0cmlwKCkKCiAgICBkZWYgX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoc2VsZik6
            CiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZSBhbmQgbm90IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIkVpdGhlciAnLS11cGdyYWRlJyBvciAn
            LS1yYmQtZGF0YS1wb29sLW5hbWUgPHBvb2xfbmFtZT4nIHNob3VsZCBiZSBzcGVjaWZpZWQiCiAg
            ICAgICAgICAgICkKCiAgICBkZWYgX2ludmFsaWRfZW5kcG9pbnQoc2VsZiwgZW5kcG9pbnRfc3Ry
            KToKICAgICAgICAjIGV4dHJhY3QgdGhlIHBvcnQgYnkgZ2V0dGluZyB0aGUgbGFzdCBzcGxpdCBv
            biBgOmAgZGVsaW1pdGVyCiAgICAgICAgdHJ5OgogICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAs
            IHBvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICBleGNlcHQgVmFsdWVF
            cnJvcjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk5vdCBh
            IHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iKQoKICAgICAgICB0cnk6CiAgICAgICAg
            ICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSA9PSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2lu
            dF9zdHJfaXAgPSBlbmRwb2ludF9zdHJfaXBbMSA6IGxlbihlbmRwb2ludF9zdHJfaXApIC0gMV0K
            ICAgICAgICAgICAgaXBfdHlwZSA9ICgKICAgICAgICAgICAgICAgICJJUHY0IiBpZiB0eXBlKGlw
            X2FkZHJlc3MoZW5kcG9pbnRfc3RyX2lwKSkgaXMgSVB2NEFkZHJlc3MgZWxzZSAiSVB2NiIKICAg
            ICAgICAgICAgKQogICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICBpcF90eXBl
            ID0gIkZRRE4iCiAgICAgICAgaWYgbm90IHBvcnQuaXNkaWdpdCgpOgogICAgICAgICAgICByYWlz
            ZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiUG9ydCBub3QgdmFsaWQ6IHtwb3J0fSIpCiAg
            ICAgICAgaW50UG9ydCA9IGludChwb3J0KQogICAgICAgIGlmIGludFBvcnQgPCAxIG9yIGludFBv
            cnQgPiAyKioxNiAtIDE6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZiJPdXQgb2YgcmFuZ2UgcG9ydCBudW1iZXI6IHtwb3J0fSIpCgogICAgICAgIHJldHVybiBp
            cF90eXBlCgogICAgZGVmIGVuZHBvaW50X2RpYWwoc2VsZiwgZW5kcG9pbnRfc3RyLCBpcF90eXBl
            LCB0aW1lb3V0PTMsIGNlcnQ9Tm9uZSk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXInIGluc3Rh
            bmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91dCB0byB0
            aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlSYWRv
            cyk6CiAgICAgICAgICAgIHJldHVybiAiIiwgIiIsICIiCiAgICAgICAgaWYgaXBfdHlwZSA9PSAi
            SVB2NiI6CiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCwg
            ZW5kcG9pbnRfc3RyX3BvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICAg
            ICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiTm90IGEgcHJvcGVyIGVuZHBvaW50
            OiB7ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgZW5kcG9p
            bnRfc3RyX2lwWzBdICE9ICJbIjoKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9ICJb
            IiArIGVuZHBvaW50X3N0cl9pcCArICJdIgogICAgICAgICAgICBlbmRwb2ludF9zdHIgPSAiOiIu
            am9pbihbZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydF0pCgogICAgICAgIHByb3Rv
            Y29scyA9IFsiaHR0cCIsICJodHRwcyJdCiAgICAgICAgcmVzcG9uc2VfZXJyb3IgPSBOb25lCiAg
            ICAgICAgZm9yIHByZWZpeCBpbiBwcm90b2NvbHM6CiAgICAgICAgICAgIHRyeToKICAgICAgICAg
            ICAgICAgIGVwID0gZiJ7cHJlZml4fTovL3tlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAg
            dmVyaWZ5ID0gTm9uZQogICAgICAgICAgICAgICAgIyBJZiB2ZXJpZnkgaXMgc2V0IHRvIGEgcGF0
            aCB0byBhIGRpcmVjdG9yeSwKICAgICAgICAgICAgICAgICMgdGhlIGRpcmVjdG9yeSBtdXN0IGhh
            dmUgYmVlbiBwcm9jZXNzZWQgdXNpbmcgdGhlIGNfcmVoYXNoIHV0aWxpdHkgc3VwcGxpZWQgd2l0
            aCBPcGVuU1NMLgogICAgICAgICAgICAgICAgaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3NraXBfdGxzOgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9IEZh
            bHNlCiAgICAgICAgICAgICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGlt
            ZW91dCwgdmVyaWZ5PUZhbHNlKQogICAgICAgICAgICAgICAgZWxpZiBwcmVmaXggPT0gImh0dHBz
            IiBhbmQgY2VydDoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBjZXJ0CiAgICAgICAgICAg
            ICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCwgdmVyaWZ5PWNl
            cnQpCiAgICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQpCiAgICAgICAgICAgICAgICBpZiByLnN0YXR1c19j
            b2RlID09IDIwMDoKICAgICAgICAgICAgICAgICAgICByZXR1cm4gcHJlZml4LCB2ZXJpZnksICIi
            CiAgICAgICAgICAgIGV4Y2VwdCBFeGNlcHRpb24gYXMgZXJyOgogICAgICAgICAgICAgICAgcmVz
            cG9uc2VfZXJyb3IgPSBlcnIKICAgICAgICAgICAgICAgIGNvbnRpbnVlCiAgICAgICAgc3lzLnN0
            ZGVyci53cml0ZSgKICAgICAgICAgICAgZiJ1bmFibGUgdG8gY29ubmVjdCB0byBlbmRwb2ludDog
            e2VuZHBvaW50X3N0cn0sIGZhaWxlZCBlcnJvcjoge3Jlc3BvbnNlX2Vycm9yfSIKICAgICAgICAp
            CiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICIiLAogICAgICAg
            ICAgICAoIi0xIiksCiAgICAgICAgKQoKICAgIGRlZiBwYXJzZV9jb25maWdfZmlsZShzZWxmLCBj
            b25maWdfZmlsZSk6CiAgICAgICAgY29uZmlnID0gY29uZmlncGFyc2VyLkNvbmZpZ1BhcnNlcigp
            CiAgICAgICAgY29uZmlnLnJlYWQoY29uZmlnX2ZpbGUpCiAgICAgICAgZm9yIGFyZyBpbiBsaXN0
            KHZhcnMoc2VsZi5fYXJnX3BhcnNlcikpOgogICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxh
            Zy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRlcm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxh
            Zy1uYW1lLAogICAgICAgICAgICAjIHNvIHdlIGNhbiBnZXQgdGhvc2UgdmFsdWVzIGZyb20gY29u
            ZmlnIGZpbGUKICAgICAgICAgICAgYXJndW1lbnQgPSBhcmcucmVwbGFjZSgiXyIsICItIikKICAg
            ICAgICAgICAgYXJndW1lbnRWYWx1ZSA9IHN0cihnZXRhdHRyKHNlbGYuX2FyZ19wYXJzZXIsIGFy
            ZykpCiAgICAgICAgICAgIGNvbmZpZ192YWx1ZSA9IGNvbmZpZy5nZXQoIkNvbmZpZ3VyYXRpb25z
            IiwgYXJndW1lbnQsIGZhbGxiYWNrPU5vbmUpCiAgICAgICAgICAgICMgZ2l2ZSBwcmlvcml0eSB0
            byBjb21tYW5kIGxpbmUgYXJndW1lbnQsIGlmIGNvbW1hbmQgbGluZSBhcmd1bWVudCBpcyBub3Qg
            cHJlc2VudCB1c2UgY29uZmlnLmluaSB2YWx1ZSwKICAgICAgICAgICAgaWYgKHN0cihzeXMuYXJn
            dikuZmluZChhcmd1bWVudCkgPT0gLTEpIGFuZCAoCiAgICAgICAgICAgICAgICAoY29uZmlnX3Zh
            bHVlICE9IE5vbmUpIGFuZCAoY29uZmlnX3ZhbHVlICE9ICIiKQogICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5fX3NldGF0dHJfXyhhcmcsIGNvbmZpZ192YWx1
            ZSkKCiAgICAgICAgcmV0dXJuIGNvbmZpZwoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhcmdfbGlz
            dD1Ob25lKToKICAgICAgICBzZWxmLm91dF9tYXAgPSB7fQogICAgICAgIHNlbGYuX2V4Y2x1ZGVk
            X2tleXMgPSBzZXQoKQogICAgICAgIHNlbGYuX2FyZ19wYXJzZXIgPSBzZWxmLmdlbl9hcmdfcGFy
            c2VyKGFyZ3NfdG9fcGFyc2U9YXJnX2xpc3QpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            b25maWdfZmlsZToKICAgICAgICAgICAgc2VsZi5jb25maWcgPSBzZWxmLnBhcnNlX2NvbmZpZ19m
            aWxlKHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGUpCiAgICAgICAgc2VsZi5fY2hlY2tfY29u
            ZmxpY3Rpbmdfb3B0aW9ucygpCiAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuX2FyZ19w
            YXJzZXIucnVuX2FzX3VzZXIKICAgICAgICBzZWxmLm91dHB1dF9maWxlID0gc2VsZi5fYXJnX3Bh
            cnNlci5vdXRwdXQKICAgICAgICBzZWxmLmNlcGhfY29uZiA9IHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aF9jb25mCiAgICAgICAgc2VsZi5jZXBoX2tleXJpbmcgPSBzZWxmLl9hcmdfcGFyc2VyLmtleXJp
            bmcKICAgICAgICAjIGlmIHVzZXIgbm90IHByb3ZpZGVkLCBnaXZlIGEgZGVmYXVsdCB1c2VyCiAg
            ICAgICAgaWYgbm90IHNlbGYucnVuX2FzX3VzZXIgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGU6CiAgICAgICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLkVYVEVSTkFMX1VTRVJf
            TkFNRQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCBhbmQg
            bm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggPSBzZWxmLkRFRkFVTFRfUkdXX1BPT0xfUFJFRklYCiAgICAgICAg
            aWYgc2VsZi5jZXBoX2NvbmY6CiAgICAgICAgICAgIGt3YXJncyA9IHt9CiAgICAgICAgICAgIGlm
            IHNlbGYuY2VwaF9rZXlyaW5nOgogICAgICAgICAgICAgICAga3dhcmdzWyJjb25mIl0gPSB7Imtl
            eXJpbmciOiBzZWxmLmNlcGhfa2V5cmluZ30KICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFk
            b3MuUmFkb3MoY29uZmZpbGU9c2VsZi5jZXBoX2NvbmYsICoqa3dhcmdzKQogICAgICAgIGVsc2U6
            CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKCkKICAgICAgICAgICAgc2Vs
            Zi5jbHVzdGVyLmNvbmZfcmVhZF9maWxlKCkKICAgICAgICBzZWxmLmNsdXN0ZXIuY29ubmVjdCgp
            CgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIGlmIHNlbGYuY2x1c3Rlci5zdGF0ZSA9
            PSAiY29ubmVjdGVkIjoKICAgICAgICAgICAgc2VsZi5jbHVzdGVyLnNodXRkb3duKCkKCiAgICBk
            ZWYgZ2V0X2ZzaWQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoIGZzaWQiKQogICAgICAgIHJldHVy
            biBzdHIoc2VsZi5jbHVzdGVyLmdldF9mc2lkKCkpCgogICAgZGVmIF9jb21tb25fY21kX2pzb25f
            Z2VuKHNlbGYsIGNtZF9qc29uKToKICAgICAgICBjbWQgPSBqc29uLmR1bXBzKGNtZF9qc29uLCBz
            b3J0X2tleXM9VHJ1ZSkKICAgICAgICByZXRfdmFsLCBjbWRfb3V0LCBlcnJfbXNnID0gc2VsZi5j
            bHVzdGVyLm1vbl9jb21tYW5kKGNtZCwgYiIiKQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            dmVyYm9zZToKICAgICAgICAgICAgcHJpbnQoZiJDb21tYW5kIElucHV0OiB7Y21kfSIpCiAgICAg
            ICAgICAgIHByaW50KAogICAgICAgICAgICAgICAgZiJSZXR1cm4gVmFsOiB7cmV0X3ZhbH1cbkNv
            bW1hbmQgT3V0cHV0OiB7Y21kX291dH1cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3IgTWVzc2Fn
            ZToge2Vycl9tc2d9XG4tLS0tLS0tLS0tXG4iCiAgICAgICAgICAgICkKICAgICAgICBqc29uX291
            dCA9IHt9CiAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciAoaS5lOyByZXRfdmFsIGlzIFpF
            Uk8pIGFuZCAnY21kX291dCcgaXMgbm90IGVtcHR5CiAgICAgICAgIyB0aGVuIGNvbnZlcnQgJ2Nt
            ZF9vdXQnIHRvIGEganNvbiBvdXRwdXQKICAgICAgICBpZiByZXRfdmFsID09IDAgYW5kIGNtZF9v
            dXQ6CiAgICAgICAgICAgIGpzb25fb3V0ID0ganNvbi5sb2FkcyhjbWRfb3V0KQogICAgICAgIHJl
            dHVybiByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZwoKICAgIGRlZiBnZXRfY2VwaF9leHRlcm5h
            bF9tb25fZGF0YShzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogInF1b3J1bV9z
            dGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5
            X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25b
            InByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29t
            bW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3Vj
            Y2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQp
            ID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAiJ3F1b3J1bV9zdGF0dXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAg
            ICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9P
            VVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICBxX2xlYWRlcl9uYW1lID0ganNvbl9v
            dXRbInF1b3J1bV9sZWFkZXJfbmFtZSJdCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHt9CiAg
            ICAgICAgcV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCA9IFsKICAgICAgICAgICAgbCBmb3IgbCBpbiBq
            c29uX291dFsibW9ubWFwIl1bIm1vbnMiXSBpZiBsWyJuYW1lIl0gPT0gcV9sZWFkZXJfbmFtZQog
            ICAgICAgIF0KICAgICAgICBpZiBsZW4ocV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCkgPT0gMDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gbWF0Y2hpbmcgJ21v
            bicgZGV0YWlscyBmb3VuZCIpCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3RbMF0KICAgICAgICAjIGdldCB0aGUgYWRkcmVzcyB2ZWN0b3Igb2YgdGhlIHF1
            b3J1bS1sZWFkZXIKICAgICAgICBxX2xlYWRlcl9hZGRydmVjID0gcV9sZWFkZXJfZGV0YWlscy5n
            ZXQoInB1YmxpY19hZGRycyIsIHt9KS5nZXQoImFkZHJ2ZWMiLCBbXSkKICAgICAgICBpcF9hZGRy
            ID0gc3RyKHFfbGVhZGVyX2RldGFpbHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYg
            cV9sZWFkZXJfYWRkcnZlY1swXVsidHlwZSJdID09ICJ2MiI6CiAgICAgICAgICAgICAgICBpcF9h
            ZGRyID0gcV9sZWFkZXJfYWRkcnZlY1swXVsiYWRkciJdCiAgICAgICAgICAgIGVsaWYgbGVuKHFf
            bGVhZGVyX2FkZHJ2ZWMpID4gMSBhbmQgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09ICJ2
            MiI6CiAgICAgICAgICAgICAgICBpcF9hZGRyID0gcV9sZWFkZXJfYWRkcnZlY1sxXVsiYWRkciJd
            CiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAg
            ICAgICAgICAgICAgICAgICIndjInIGFkZHJlc3MgdHlwZSBub3QgcHJlc2VudCwgYW5kICd2Mi1w
            b3J0LWVuYWJsZScgZmxhZyBpcyBwcm92aWRlZCIKICAgICAgICAgICAgICAgICkKCiAgICAgICAg
            cmV0dXJuIGYie3N0cihxX2xlYWRlcl9uYW1lKX09e2lwX2FkZHJ9IgoKICAgIGRlZiBfY29udmVy
            dF9ob3N0bmFtZV90b19pcChzZWxmLCBob3N0X25hbWUsIHBvcnQsIGlwX3R5cGUpOgogICAgICAg
            ICMgaWYgJ2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgdHlwZSwKICAgICAgICAjIGNhbGwg
            dGhlIGR1bW15IGluc3RhbmNlJ3MgImNvbnZlcnQiIG1ldGhvZAogICAgICAgIGlmIG5vdCBob3N0
            X25hbWU6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIkVtcHR5
            IGhvc3RuYW1lIHByb3ZpZGVkIikKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3Rlciwg
            RHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybiBzZWxmLmNsdXN0ZXIuX2NvbnZlcnRfaG9z
            dG5hbWVfdG9faXAoaG9zdF9uYW1lKQoKICAgICAgICBpZiBpcF90eXBlID09ICJGUUROIjoKICAg
            ICAgICAgICAgIyBjaGVjayB3aGljaCBpcCBGUUROIHNob3VsZCBiZSBjb252ZXJ0ZWQgdG8sIElQ
            djQgb3IgSVB2NgogICAgICAgICAgICAjIGNoZWNrIHRoZSBob3N0IGlwLCB0aGUgZW5kcG9pbnQg
            aXAgdHlwZSB3b3VsZCBiZSBzaW1pbGFyIHRvIGhvc3QgaXAKICAgICAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJvcmNoIGhvc3QgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVt
            cHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICInb3JjaCBob3N0IGxzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAg
            ICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZ
            X09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaG9zdF9hZGRyID0g
            anNvbl9vdXRbMF1bImFkZHIiXQogICAgICAgICAgICAjIGFkZCA6ODAgc2FtcGxlIHBvcnQgaW4g
            aXBfdHlwZSwgYXMgX2ludmFsaWRfZW5kcG9pbnQgYWxzbyB2ZXJpZnkgcG9ydAogICAgICAgICAg
            ICBob3N0X2lwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KGhvc3RfYWRkciArICI6ODAi
            KQogICAgICAgICAgICBpbXBvcnQgc29ja2V0CgogICAgICAgICAgICBpcCA9IFtdCiAgICAgICAg
            ICAgICMgZXhhbXBsZSBvdXRwdXQgWyg8QWRkcmVzc0ZhbWlseS5BRl9JTkVUOiAyPiwgPFNvY2tl
            dEtpbmQuU09DS19TVFJFQU06IDE+LCA2LCAnJywgKCc5My4xODQuMjE2LjM0JywgODApKSwgLi4u
            XQogICAgICAgICAgICAjIHdlIG5lZWQgdG8gZ2V0IDkzLjE4NC4yMTYuMzQgc28gaXQgd291bGQg
            YmUgaXBbMF1bNF1bMF0KICAgICAgICAgICAgaWYgaG9zdF9pcF90eXBlID09ICJJUHY2IjoKICAg
            ICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAgICAgICAgICAgICAgICAg
            IGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVUNiwgcHJvdG89c29ja2V0LklQ
            UFJPVE9fVENQCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsaWYgaG9zdF9pcF90eXBl
            ID09ICJJUHY0IjoKICAgICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAg
            ICAgICAgICAgICAgICAgIGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVULCBw
            cm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZGVs
            IHNvY2tldAogICAgICAgICAgICByZXR1cm4gaXBbMF1bNF1bMF0KICAgICAgICByZXR1cm4gaG9z
            dF9uYW1lCgogICAgZGVmIGdldF9hY3RpdmVfYW5kX3N0YW5kYnlfbWdycyhzZWxmKToKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIiwgc2Vs
            Zi5kcnlfcnVuKCJjZXBoIHN0YXR1cyIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0
            ID0gc2VsZi5fYXJnX3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5k
            cG9pbnQKICAgICAgICBzdGFuZGJ5X21ncnMgPSBbXQogICAgICAgIGlmIG5vdCBtb25pdG9yaW5n
            X2VuZHBvaW50X2lwX2xpc3Q6CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAic3Rh
            dHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVy
            cl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAj
            IGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRf
            dmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ21nciBzZXJ2aWNl
            cycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9t
            c2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnQgPSAoCiAgICAgICAgICAg
            ICAgICBqc29uX291dC5nZXQoIm1ncm1hcCIsIHt9KS5nZXQoInNlcnZpY2VzIiwge30pLmdldCgi
            cHJvbWV0aGV1cyIsICIiKQogICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9y
            aW5nX2VuZHBvaW50OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiY2FuJ3QgZmluZCBtb25pdG9yaW5nX2VuZHBvaW50
            LCBwcm9tZXRoZXVzIG1vZHVsZSBtaWdodCBub3QgYmUgZW5hYmxlZCwgIgogICAgICAgICAgICAg
            ICAgICAgICJlbmFibGUgdGhlIG1vZHVsZSBieSBydW5uaW5nICdjZXBoIG1nciBtb2R1bGUgZW5h
            YmxlIHByb21ldGhldXMnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIG5vdyBjaGVj
            ayB0aGUgc3RhbmQtYnkgbWdyLXMKICAgICAgICAgICAgc3RhbmRieV9hcnIgPSBqc29uX291dC5n
            ZXQoIm1ncm1hcCIsIHt9KS5nZXQoInN0YW5kYnlzIiwgW10pCiAgICAgICAgICAgIGZvciBlYWNo
            X3N0YW5kYnkgaW4gc3RhbmRieV9hcnI6CiAgICAgICAgICAgICAgICBpZiAibmFtZSIgaW4gZWFj
            aF9zdGFuZGJ5LmtleXMoKToKICAgICAgICAgICAgICAgICAgICBzdGFuZGJ5X21ncnMuYXBwZW5k
            KGVhY2hfc3RhbmRieVsibmFtZSJdKQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBw
            YXJzZWRfZW5kcG9pbnQgPSB1cmxwYXJzZShtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgICAg
            ICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50OiB7bW9u
            aXRvcmluZ19lbmRwb2ludH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3Jp
            bmdfZW5kcG9pbnRfaXBfbGlzdCA9IHBhcnNlZF9lbmRwb2ludC5ob3N0bmFtZQogICAgICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICAgICAgbW9uaXRv
            cmluZ19lbmRwb2ludF9wb3J0ID0gc3RyKHBhcnNlZF9lbmRwb2ludC5wb3J0KQoKICAgICAgICAj
            IGlmIG1vbml0b3JpbmcgZW5kcG9pbnQgcG9ydCBpcyBub3Qgc2V0LCBwdXQgYSBkZWZhdWx0IG1v
            biBwb3J0CiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoKICAgICAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0ID0gc2VsZi5ERUZBVUxUX01PTklUT1JJTkdfRU5E
            UE9JTlRfUE9SVAoKICAgICAgICAjIHVzZXIgY291bGQgZ2l2ZSBjb21tYSBhbmQgc3BhY2Ugc2Vw
            YXJhdGVkIGlucHV0cyAobGlrZSAtLW1vbml0b3JpbmctZW5kcG9pbnQ9IjxpcDE+LCA8aXAyPiIp
            CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcF9saXN0LnJlcGxhY2UoIiwiLCAiICIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0X3NwbGl0ID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0LnNwbGl0KCkKICAgICAg
            ICAjIGlmIG1vbml0b3JpbmctZW5kcG9pbnQgY291bGQgbm90IGJlIGZvdW5kLCByYWlzZSBhbiBl
            cnJvcgogICAgICAgIGlmIGxlbihtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQpID09
            IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vICdtb25p
            dG9yaW5nLWVuZHBvaW50JyBmb3VuZCIpCiAgICAgICAgIyBmaXJzdCBpcCBpcyB0cmVhdGVkIGFz
            IHRoZSBtYWluIG1vbml0b3JpbmctZW5kcG9pbnQKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X2lwID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzBdCiAgICAgICAgIyByZXN0
            IG9mIHRoZSBpcC1zIGFyZSBhZGRlZCB0byB0aGUgJ3N0YW5kYnlfbWdycycgbGlzdAogICAgICAg
            IHN0YW5kYnlfbWdycy5leHRlbmQobW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzE6
            XSkKICAgICAgICBmYWlsZWRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwCgogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnQgPSAiOiIuam9pbigKICAgICAgICAgICAgW21vbml0b3JpbmdfZW5k
            cG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydF0KICAgICAgICApCiAgICAgICAgaXBf
            dHlwZSA9IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQobW9uaXRvcmluZ19lbmRwb2ludCkKICAgICAg
            ICB0cnk6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBzZWxmLl9jb252ZXJ0
            X2hvc3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCwg
            bW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgIyBjb2xsZWN0IGFsbCB0aGUgJ3N0YW5kLWJ5JyBtZ3IgaXBzCiAgICAgICAgICAgIG1ncl9p
            cHMgPSBbXQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5X21nciBpbiBzdGFuZGJ5X21ncnM6
            CiAgICAgICAgICAgICAgICBmYWlsZWRfaXAgPSBlYWNoX3N0YW5kYnlfbWdyCiAgICAgICAgICAg
            ICAgICBtZ3JfaXBzLmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9jb252ZXJ0X2hv
            c3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgICAgICAgICBlYWNoX3N0YW5kYnlfbWdyLCBt
            b25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgZXhjZXB0OgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJDb252ZXJzaW9uIG9mIGhvc3Q6
            IHtmYWlsZWRfaXB9IHRvIElQIGZhaWxlZC4gIgogICAgICAgICAgICAgICAgIlBsZWFzZSBlbnRl
            ciB0aGUgSVAgYWRkcmVzc2VzIG9mIGFsbCB0aGUgY2VwaC1tZ3JzIHdpdGggdGhlICctLW1vbml0
            b3JpbmctZW5kcG9pbnQnIGZsYWciCiAgICAgICAgICAgICkKCiAgICAgICAgXywgXywgZXJyID0g
            c2VsZi5lbmRwb2ludF9kaWFsKG1vbml0b3JpbmdfZW5kcG9pbnQsIGlwX3R5cGUpCiAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZXJyKQogICAgICAgICMgYWRkIHRoZSB2YWxpZGF0ZWQgYWN0aXZlIG1nciBJUCBpbnRvIHRo
            ZSBmaXJzdCBpbmRleAogICAgICAgIG1ncl9pcHMuaW5zZXJ0KDAsIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXApCiAgICAgICAgYWxsX21ncl9pcHNfc3RyID0gIiwiLmpvaW4obWdyX2lwcykKICAgICAg
            ICByZXR1cm4gYWxsX21ncl9pcHNfc3RyLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKCiAgICBk
            ZWYgY2hlY2tfdXNlcl9leGlzdChzZWxmLCB1c2VyKToKICAgICAgICBjbWRfanNvbiA9IHsicHJl
            Zml4IjogImF1dGggZ2V0IiwgImVudGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0K
            ICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoK
            ICAgICAgICAgICAgcmV0dXJuICIiCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5
            Il0pCgogICAgZGVmIGdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAg
            ICJvc2QiOiAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwKICAgICAgICAgICAgIm1k
            cyI6ICJhbGxvdyAqIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0
            cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rl
            cl9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiazhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNf
            ZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgog
            ICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAg
            ICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNf
            Y2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVzeXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1si
            b3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9e2NlcGhmc19maWxlc3lzdGVt
            fSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2VwaGZzX25vZGVf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3
            IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsCiAgICAgICAg
            ICAgICJtZHMiOiAiYWxsb3cgcnciLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4
            c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0g
            PT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0
            eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAg
            ICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyAqPXtjZXBoZnNfZmlsZXN5c3Rl
            bX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2VudGl0eSgKICAg
            ICAgICBzZWxmLAogICAgICAgIGVudGl0eSwKICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAg
            IGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICBy
            YWRvc19uYW1lc3BhY2UsCiAgICApOgogICAgICAgIGlmICgKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgb3IgcmJkX3Bvb2xfbmFtZS5jb3VudCgi
            XyIpICE9IDAKICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZSAhPSAiIgogICAgICAg
            ICAgICAjIGNoZWNraW5nIGFsaWFzX3JiZF9wb29sX25hbWUgaXMgbm90IGVtcHR5IGFzIHRoZXJl
            IG1heWJlIGEgc3BlY2lhbCBjaGFyYWN0ZXIgdXNlZCBvdGhlciB0aGFuIC4gb3IgXwogICAgICAg
            ICk6CiAgICAgICAgICAgIGlmIGFsaWFzX3JiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAg
            ICJwbGVhc2Ugc2V0IHRoZSAnLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgYXMgdGhl
            IHJiZCBkYXRhIHBvb2wgbmFtZSBjb250YWlucyAnLicgb3IgJ18nIgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLmNv
            dW50KCIuIikgIT0gMAogICAgICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3Vu
            dCgiXyIpICE9IDAKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIictLWFsaWFzLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyB2YWx1ZSBzaG91bGQgbm90IGNvbnRhaW4gJy4nIG9yICdfJyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0
            ZXJfbmFtZX0te2FsaWFzX3JiZF9wb29sX25hbWV9IgogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtyYmRfcG9vbF9uYW1lfSIK
            CiAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgIHJldHVybiBlbnRpdHkKCiAgICBkZWYgZ2V0
            X3JiZF9wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0g
            ImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAg
            ICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAg
            ICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSBy
            YmQiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNl
            ci5yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19j
            bHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAg
            ICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAg
            ICAgICAgICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBm
            bGFnIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFn
            IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIi
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAn
            LS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bnRpdHkgPSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAg
            ICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFt
            ZSwKICAgICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRv
            c19uYW1lc3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNl
            ICE9ICIiOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25h
            bWVzcGFjZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBjYXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgog
            ICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X3JiZF9ub2RlX2NhcHNfYW5k
            X2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtbm9kZSIKICAg
            ICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSByYmQiLAog
            ICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJt
            aXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19jbHVzdGVy
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgcmFk
            b3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICAg
            ICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5v
            dCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIgogICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAg
            ICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAg
            ICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMt
            Y2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkg
            PSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAgICAgICAg
            ICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAg
            ICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRvc19uYW1l
            c3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlICE9ICIi
            OgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgZiJw
            cm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25hbWVzcGFj
            ZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBj
            YXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAg
            IHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2Vu
            dGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAgY2Fw
            cyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9z
            dGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLAogICAgICAgICAgICAib3NkIjogZiJwcm9maWxlIHJiZC1yZWFkLW9u
            bHksIGFsbG93IHJ3eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9e3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2csIGFsbG93IHggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5idWNrZXRzLmluZGV4IiwKICAgICAgICAg
            ICAgIm1kcyI6ICJhbGxvdyAqIiwKICAgICAgICB9CgogICAgICAgIHJldHVybiBjYXBzLCBlbnRp
            dHkKCiAgICBkZWYgZ2V0X2NhcHNfYW5kX2VudGl0eShzZWxmLCB1c2VyX25hbWUpOgogICAgICAg
            IGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAg
            ICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fw
            c19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgaW4gdXNl
            cl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgIT0gdXNlcl9u
            YW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVy
            bWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19ub2RlX2Nh
            cHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiBp
            biB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X3JiZF9w
            cm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1u
            b2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1ub2RlIiAh
            PSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRf
            YXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX25v
            ZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiIGlu
            IHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5oZWFsdGhjaGVja2VyIiAhPSB1c2Vy
            X25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9w
            ZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfZGVmYXVsdFVzZXJf
            Y2Fwc19hbmRfZW50aXR5KCkKCiAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgZiJubyB1c2VyIGZvdW5kIHdpdGggdXNlcl9uYW1lOiB7dXNlcl9uYW1l
            fSwgIgogICAgICAgICAgICAiZ2V0X2NhcHNfYW5kX2VudGl0eSBjb21tYW5kIGZhaWxlZC5cbiIK
            ICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKHNlbGYsIHVzZXIp
            OgogICAgICAgICIiIgogICAgICAgIGNvbW1hbmQ6IGNlcGggYXV0aCBnZXQtb3ItY3JlYXRlIGNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIG1vbiAnYWxsb3cgcicgbWdyICdhbGxvdyBydycg
            b3NkICdhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSonCiAgICAgICAgIiIiCiAgICAgICAg
            Y2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fwc19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21k
            X2pzb24gPSB7CiAgICAgICAgICAgICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAg
            ICAgICAgICAgImVudGl0eSI6IGVudGl0eSwKICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBj
            YXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMoKSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAg
            ICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICBzZWxm
            LmRyeV9ydW4oCiAgICAgICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAgICAg
            ICsgY21kX2pzb25bInByZWZpeCJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgICAgICsgIiAi
            CiAgICAgICAgICAgICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAg
            ICAgICAgICAgKSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCiAgICAgICAgIyBj
            aGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYuY2hlY2tf
            dXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAgICAgICAg
            IHJldHVybiB1c2VyX2tleSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICAg
            ICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0eShjbGllbnQuY3NpLXJi
            ZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9t
            c2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhl
            cmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9y
            IGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7dXNlcn0nIGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0
            X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAg
            ICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSksIGYie2VudGl0eS5zcGxpdCgnLics
            IDEpWzFdfSIKICAgICAgICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0
            eShjbGllbnQuY3NpLXJiZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICBkZWYgZ2V0X2NlcGhm
            c19kYXRhX3Bvb2xfZGV0YWlscyhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4Ijog
            ImZzIGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29u
            WyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2Nv
            bW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1
            Y2Nlc3NmdWwgYXR0ZW1wdCwgcmVwb3J0IGFuIGVycm9yCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            OgogICAgICAgICAgICAjIGlmIGZzIGFuZCBkYXRhX3Bvb2wgYXJndW1lbnRzIGFyZSBub3Qgc2V0
            LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPT0gIiIKICAgICAgICAg
            ICAgKToKICAgICAgICAgICAgICAgIHJldHVybgogICAgICAgICAgICAjIGlmIHVzZXIgaGFzIHBy
            b3ZpZGVkIGFueSBvZiB0aGUKICAgICAgICAgICAgIyAnLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1l
            JyBvciAnLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUnIGFyZ3VtZW50cywKICAgICAgICAgICAgIyBy
            YWlzZSBhbiBleGNlcHRpb24gYXMgd2UgYXJlIHVuYWJsZSB0byB2ZXJpZnkgdGhlIGFyZ3MKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiJ2ZzIGxzJyBjZXBoIGNhbGwgZmFpbGVkIHdpdGggZXJyb3I6IHtlcnJfbXNnfSIKICAgICAg
            ICAgICAgKQoKICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IHt9CiAgICAgICAgIyBpZiAnLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lJyBhcmd1bWVudCBpcyBwcm92aWRlZCwKICAgICAgICAjIGNo
            ZWNrIHdoZXRoZXIgdGhlIHByb3ZpZGVkIGZpbGVzeXN0ZW0tbmFtZSBleGlzdHMgb3Igbm90CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lOgogICAgICAg
            ICAgICAjIGdldCB0aGUgbWF0Y2hpbmcgbGlzdAogICAgICAgICAgICBtYXRjaGluZ19qc29uX291
            dF9saXN0ID0gWwogICAgICAgICAgICAgICAgbWF0Y2hlZAogICAgICAgICAgICAgICAgZm9yIG1h
            dGNoZWQgaW4ganNvbl9vdXQKICAgICAgICAgICAgICAgIGlmIHN0cihtYXRjaGVkWyJuYW1lIl0p
            ID09IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgdW5hYmxlIHRvIGZpbmQgYSBtYXRjaGluZyBmcy1uYW1lLCByYWlzZSBh
            biBlcnJvcgogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRfbGlzdCkgPT0gMDoK
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgZiJGaWxlc3lzdGVtIHByb3ZpZGVkLCAne3NlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgZiJpcyBub3QgZm91
            bmQgaW4gdGhlIGZzLWxpc3Q6IHtbc3RyKHhbJ25hbWUnXSkgZm9yIHggaW4ganNvbl9vdXRdfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBtYXRjaGlu
            Z19qc29uX291dF9saXN0WzBdCiAgICAgICAgIyBpZiBjZXBoZnMgZmlsZXN5c3RlbSBuYW1lIGlz
            IG5vdCBwcm92aWRlZCwKICAgICAgICAjIHRyeSB0byBnZXQgYSBkZWZhdWx0IGZzIG5hbWUgYnkg
            ZG9pbmcgdGhlIGZvbGxvd2luZwogICAgICAgIGVsc2U6CiAgICAgICAgICAgICMgYS4gY2hlY2sg
            aWYgdGhlcmUgaXMgb25seSBvbmUgZmlsZXN5c3RlbSBpcyBwcmVzZW50CiAgICAgICAgICAgIGlm
            IGxlbihqc29uX291dCkgPT0gMToKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0g
            anNvbl9vdXRbMF0KICAgICAgICAgICAgIyBiLiBvciBlbHNlLCBjaGVjayBpZiBkYXRhX3Bvb2wg
            bmFtZSBpcyBwcm92aWRlZAogICAgICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgIyBhbmQgaWYgcHJlc2VudCwgY2hlY2sg
            d2hldGhlciB0aGVyZSBleGlzdHMgYSBmcyB3aGljaCBoYXMgdGhlIGRhdGFfcG9vbAogICAgICAg
            ICAgICAgICAgZm9yIGVhY2hKIGluIGpzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lIGluIGVhY2hKWyJkYXRhX3Bvb2xz
            Il06CiAgICAgICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0gZWFjaEoKICAg
            ICAgICAgICAgICAgICAgICAgICAgYnJlYWsKICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            bm8gbWF0Y2hpbmcgZnMgZXhpc3RzLCB0aGF0IG1lYW5zIHByb3ZpZGVkIGRhdGFfcG9vbCBuYW1l
            IGlzIGludmFsaWQKICAgICAgICAgICAgICAgIGlmIG5vdCBtYXRjaGluZ19qc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGFfcG9vbCBuYW1lLCB7c2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9LCIKICAgICAgICAgICAgICAgICAgICAgICAgIiBk
            b2VzIG5vdCBleGlzdHMiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGMuIGlm
            IG5vdGhpbmcgaXMgc2V0IGFuZCBjb3VsZG4ndCBmaW5kIGEgZGVmYXVsdCwKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgICMganVzdCByZXR1cm4gc2lsZW50bHkKICAgICAgICAgICAg
            ICAgIHJldHVybgoKICAgICAgICBpZiBtYXRjaGluZ19qc29uX291dDoKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID0gc3RyKG1hdGNoaW5nX2pzb25f
            b3V0WyJuYW1lIl0pCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0WyJtZXRh
            ZGF0YV9wb29sIl0KICAgICAgICAgICAgKQoKICAgICAgICBpZiBpc2luc3RhbmNlKG1hdGNoaW5n
            X2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0sIGxpc3QpOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGhhcyBhbHJlYWR5IHByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lLAogICAgICAgICAgICAjIHRocm91
            Z2ggLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUKICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGlmIHRoZSBwcm92aWRl
            ZCBuYW1lIGlzIG5vdCBtYXRjaGluZyB3aXRoIHRoZSBvbmUgaW4gdGhlIGxpc3QKICAgICAgICAg
            ICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19k
            YXRhX3Bvb2xfbmFtZQogICAgICAgICAgICAgICAgICAgIG5vdCBpbiBtYXRjaGluZ19qc29uX291
            dFsiZGF0YV9wb29scyJdCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYi
            UHJvdmlkZWQgZGF0YS1wb29sLW5hbWU6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWV9JywgIgogICAgICAgICAgICAgICAgICAgICAgICAiZG9lc24ndCBtYXRjaCBmcm9t
            IHRoZSBkYXRhLXBvb2xzIGxpc3Q6ICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4
            KSBmb3IgeCBpbiBtYXRjaGluZ19qc29uX291dFsnZGF0YV9wb29scyddXX0iCiAgICAgICAgICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGlmIGRhdGFfcG9vbCBuYW1lIGlzIG5vdCBwcm92aWRl
            ZCwKICAgICAgICAgICAgIyB0aGVuIHRyeSB0byBmaW5kIGEgZGVmYXVsdCBkYXRhIHBvb2wgbmFt
            ZQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBpZiBubyBkYXRhX3Bvb2xzIGV4
            aXN0LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dFsiZGF0YV9wb29scyJdKSA9PSAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPSBzdHIo
            CiAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXVswXQog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGlmIHRoZXJlIGFyZSBtb3JlIHRoYW4gb25l
            ICdkYXRhX3Bvb2xzJyBleGlzdCwKICAgICAgICAgICAgIyB0aGVuIHdhcm4gdGhlIHVzZXIgdGhh
            dCB3ZSBhcmUgdXNpbmcgdGhlIHNlbGVjdGVkIG5hbWUKICAgICAgICAgICAgaWYgbGVuKG1hdGNo
            aW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID4gMToKICAgICAgICAgICAgICAgIHByaW50KAog
            ICAgICAgICAgICAgICAgICAgICJXQVJOSU5HOiBNdWx0aXBsZSBkYXRhIHBvb2xzIGRldGVjdGVk
            OiAiCiAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4KSBmb3IgeCBpbiBtYXRjaGluZ19qc29u
            X291dFsnZGF0YV9wb29scyddXX1cbiIKICAgICAgICAgICAgICAgICAgICBmIlVzaW5nIHRoZSBk
            YXRhLXBvb2w6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9J1xuIgog
            ICAgICAgICAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2hlY2tlcktleShzZWxmLCB1c2VyKToK
            ICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAg
            ICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVh
            dGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtj
            YXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0
            XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigK
            ICAgICAgICAgICAgICAgICJjZXBoICIKICAgICAgICAgICAgICAgICsgY21kX2pzb25bInByZWZp
            eCJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50
            aXR5Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICArICIgIi5qb2luKGNt
            ZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            CgogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBh
            dHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmIidhdXRoIGdldC1vci1jcmVhdGUge3NlbGYucnVuX2FzX3VzZXJ9JyBjb21tYW5kIGZhaWxl
            ZFxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVs
            c2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            c2VsZik6CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJtZ3Igc2VydmljZXMiLCAiZm9y
            bWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQog
            ICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAog
            ICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAg
            IHJldHVybiBOb25lCiAgICAgICAgaWYgImRhc2hib2FyZCIgbm90IGluIGpzb25fb3V0OgogICAg
            ICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIHJldHVybiBqc29uX291dFsiZGFzaGJvYXJkIl0K
            CiAgICBkZWYgY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcihzZWxmKToKICAgICAgICBjbWQgPSBb
            CiAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAgICAgICAgICAgInVzZXIiLAogICAgICAg
            ICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgIi0tdWlkIiwKICAgICAgICAgICAgc2VsZi5FWFRF
            Uk5BTF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgIi0tZGlzcGxheS1uYW1l
            IiwKICAgICAgICAgICAgIlJvb2sgUkdXIEFkbWluIE9wcyB1c2VyIiwKICAgICAgICAgICAgIi0t
            Y2FwcyIsCiAgICAgICAgICAgICJidWNrZXRzPSo7dXNlcnM9Kjt1c2FnZT1yZWFkO21ldGFkYXRh
            PXJlYWQ7em9uZT1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmVn
            cm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAog
            ICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyAiICIuam9pbihjbWQp
            KQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNz
            LkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGFscmVhZHkgZXhpc3RzLCB3ZSBqdXN0IHF1ZXJ5IGl0CiAgICAgICAgICAgIGlmIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSBlcnJuby5FRVhJU1Q6CiAgICAgICAgICAgICAgICBjbWQgPSBbCiAgICAg
            ICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgICAgICJ1c2Vy
            IiwKICAgICAgICAgICAgICAgICAgICAiaW5mbyIsCiAgICAgICAgICAgICAgICAgICAgIi0tdWlk
            IiwKICAgICAgICAgICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAi
            LS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgICAg
            ICAgICAgXQogICAgICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVj
            RXJyOgogICAgICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAg
            ICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5v
            dXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJu
            Y29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCiAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRv
            IGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAg
            ICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVj
            RXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2Us
            ICItMSIKCiAgICAgICAgIyBzZXBhcmF0ZWx5IGFkZCBpbmZvPXJlYWQgY2FwcyBmb3Igcmd3LWVu
            ZHBvaW50IGlwIHZhbGlkYXRpb24KICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQgPSBUcnVlCiAg
            ICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICJj
            YXBzIiwKICAgICAgICAgICAgImFkZCIsCiAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAg
            IHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICItLWNh
            cHMiLAogICAgICAgICAgICAiaW5mbz1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAg
            Ii0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVn
            cm91cF9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAg
            ICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNz
            LlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4
            ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIGNlcGggdmVyc2lvbiBub3Qgc3VwcG9ydGVkIGZv
            ciBhZGRpbmcgYGluZm89cmVhZGAgY2FwKHJnd192YWxpZGF0aW9uKQogICAgICAgICAgICBpZiAo
            CiAgICAgICAgICAgICAgICAiY291bGQgbm90IGFkZCBjYXBzOiB1bmFibGUgdG8gYWRkIGNhcHM6
            IGluZm89cmVhZFxuIgogICAgICAgICAgICAgICAgaW4gZXhlY0Vyci5zdGRlcnIuZGVjb2RlKCJ1
            dGYtOCIpCiAgICAgICAgICAgICAgICBhbmQgZXhlY0Vyci5yZXR1cm5jb2RlID09IDI0NAogICAg
            ICAgICAgICApOgogICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gRmFsc2UKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICBqc29ub3V0cHV0ID0ganNvbi5sb2Fkcyhv
            dXRwdXQpCiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJhY2Nlc3Nfa2V5Il0sCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsic2VjcmV0
            X2tleSJdLAogICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAgICIiLAog
            ICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX3Bvb2woc2VsZiwgcG9vbF9uYW1lKToKICAg
            ICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKHBvb2xfbmFtZSk6CiAgICAgICAg
            ICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRo
            ZSBwcm92aWRlZCBwb29sLCAne3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAg
            ICAgKQoKICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYsIHJiZF9wb29sX25hbWUpOgogICAgICAg
            IGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0
            dXJuCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1l
            KQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0
            KGlvY3R4LCBUcnVlKQoKICAgIGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAg
            ICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgog
            ICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFf
            ZWNfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5y
            YWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAg
            ICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScs
            ICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMgdXBwZXIgY2FzZSIKICAgICAgICAgICAgKQog
            ICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIu
            b3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9l
            eGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkgaXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRl
            ZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAg
            IGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAg
            ICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVt
            ZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSk6CiAgICAgICAgY21kID0gWwogICAgICAg
            ICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIsCiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91
            cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1f
            bmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgIF0KICAgICAgICB0cnk6
            CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJw
            cm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9y
            OgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            ICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgXQogICAgICAgICAgICB0
            cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRl
            cnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQ
            cm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vidm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3Jv
            dXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVkIgogICAgICAgICAgICAgICAgKQoKICAgIGRl
            ZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9zZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9
            IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAic3Vi
            dm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGluIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVz
            eXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICBwaW5f
            dHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcsCiAgICAgICAgXQogICAgICAgIHRyeToKICAg
            ICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nl
            c3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQg
            cGlubmVkIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3Vy
            bCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5f
            T1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAgc2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsi
            UkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXQogICAgICAgIHJnd19lbmRwb2ludCA9IHNl
            bGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAgICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCAr
            ICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1pbi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7
            ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5j
            b2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAgICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAog
            ICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAgICAgICAgICAgICAgICBhdXRoPVMzQXV0aChh
            Y2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5kcG9pbnQpLAogICAgICAgICAgICAgICAgdmVy
            aWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRp
            b25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgICAg
            ICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJsZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3
            X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAgICAgICAgICByZXR1cm4gIiIsICItMSIKICAg
            ICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYgcjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8i
            KSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAne3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiAoCiAg
            ICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAgICAgICItMSIsCiAgICAgICAgICAgICkKCiAg
            ICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3JhZ2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9p
            ZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3Vw
            cG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBv
            bmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAg
            ICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAg
            cmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVuZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBp
            cyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5k
            cG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUK
            ICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19z
            a2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAg
            ICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBi
            YXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBp
            cF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYgZXJyICE9ICIiOgogICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0
            aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3Vw
            cG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3VwcG9ydGVkOgogICAgICAgICAgICBmc2lkID0g
            c2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJnd19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3df
            ZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAg
            ICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAg
            ICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgICAgIGYiVGhl
            IHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0n
            LCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGluZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBh
            cGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZhbGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3ty
            Z3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBoIGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgICMgY2hl
            Y2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRl
            IGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQgZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2
            YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNlIGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAg
            IHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSIsCiAgICAgICAgICAgICAgICAiLnJndy5y
            b290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19w
            b29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3df
            cG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNfdG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAg
            ICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6
            CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tfcmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHJl
            dHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlz
            aXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmlnX25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmln
            KToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2NvbmZpZyAhPSAiIjoKICAgICAgICAgICAgY21k
            ID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgcmd3
            X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICAiZ2V0IiwKICAgICAgICAgICAgICAg
            ICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICByZ3dfbXVs
            dGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAgICBdCiAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9j
            ZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJv
            ciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAg
            ICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIu
            b3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29k
            ZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0dXJuICIt
            MSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYgY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9f
            YXJyYXkoc2VsZiwgdmFsdWUpOgogICAgICAgIHJldHVybiB2YWx1ZS5zcGxpdCgiLCIpCgogICAg
            ZGVmIHJhaXNlX2V4Y2VwdGlvbl9pZl9hbnlfdG9wb2xvZ3lfZmxhZ19pc19taXNzaW5nKHNlbGYp
            OgogICAgICAgIGlmICgKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9m
            YWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAg
            KQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9t
            YWluX3ZhbHVlcyA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIG9yICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVy
            ZV9kb21haW5fdmFsdWVzICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAgICAgICAgICAg
            ICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICApOgogICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgInBy
            b3ZpZGUgYWxsIHRoZSB0b3BvbG9neSBmbGFncyAtLXRvcG9sb2d5LXBvb2xzLCAtLXRvcG9sb2d5
            LWZhaWx1cmUtZG9tYWluLWxhYmVsLCAtLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIK
            ICAgICAgICAgICAgKQoKICAgIGRlZiB2YWxpZGF0ZV90b3BvbG9neV92YWx1ZXMoc2VsZiwgdG9w
            b2xvZ3lfcG9vbHMsIHRvcG9sb2d5X2ZkKToKICAgICAgICBpZiBsZW4odG9wb2xvZ3lfcG9vbHMp
            ICE9IGxlbih0b3BvbG9neV9mZCk6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCB0b3BvbG9neSBwb29scywg
            J3t0b3BvbG9neV9wb29sc30nLCBhbmQgIgogICAgICAgICAgICAgICAgZiJ0b3BvbG9neSBmYWls
            dXJlIGRvbWFpbiwgJ3t0b3BvbG9neV9mZH0nLCIKICAgICAgICAgICAgICAgIGYiYXJlIG9mIGRp
            ZmZlcmVudCBsZW5ndGgsICd7bGVuKHRvcG9sb2d5X3Bvb2xzKX0nIGFuZCAne2xlbih0b3BvbG9n
            eV9mZCl9JyByZXNwY3RpdmVseSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybgoKICAgIGRl
            ZiB2YWxpZGF0ZV90b3BvbG9neV9yYmRfcG9vbHMoc2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToK
            ICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9yYmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYu
            dmFsaWRhdGVfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgaW5pdF90b3BvbG9neV9yYmRfcG9vbHMo
            c2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToKICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9y
            YmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYuaW5pdF9yYmRfcG9vbChwb29sKQoKICAgICMgdGhp
            cyB3aWxsIHJldHVybiB0aGUgZmluYWwgYXJncyB0aGF0IHNjcmlwdCB1c2VzIHRvIHByb2Nlc3MK
            ICAgICMgdGhlIHByaW9yaXR5IHRvIHNldCBhIHBhcnRpY3VsYXIgdmFsdWUgaXMsCiAgICAjIGNv
            bW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmluaSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVz
            CiAgICBkZWYgZ2V0RmluYWxVc2VkQXJncyhzZWxmKToKICAgICAgICBhcmd1bWVudCA9IGYiW0Nv
            bmZpZ3VyYXRpb25zXVxuIgogICAgICAgIGZvciBhcmcgaW4gdmFycyhzZWxmLl9hcmdfcGFyc2Vy
            KToKICAgICAgICAgICAgaWYgc3RyKGdldGF0dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSk6CiAg
            ICAgICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxhZy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRl
            cm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxhZy1uYW1lLAogICAgICAgICAgICAgICAg
            IyBzbyB3ZSBjYW4gZ2V0IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAg
            ICAgICBhcmdWYWx1ZSA9IGFyZy5yZXBsYWNlKCJfIiwgIi0iKQogICAgICAgICAgICAgICAgIyBk
            byBub3QgYWRkIHRoZSBjb25maWctZmlsZSBmbGFnIGFuZCBhbHNvIG5vdCBhZGQgdGhlIGJvb2xl
            YW4gZmxhZ3Mgd2hpY2ggYXJlIHNldCB0byBGYWxzZQogICAgICAgICAgICAgICAgIyBiZWNhdXNl
            IGNvbmZpZy5pbmkgZmlsZSB0cmVhdHMgYm9vbGVhbiBmbGFncyBhcyBUcnVlIGFsd2F5cwogICAg
            ICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGFyZ1ZhbHVlICE9ICJjb25maWct
            ZmlsZSIKICAgICAgICAgICAgICAgICAgICBhbmQgZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBh
            cmcpICE9IEZhbHNlCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIGFyZ3Vt
            ZW50ICs9IGYie2FyZ1ZhbHVlfSA9IHtzdHIoZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcp
            KX1cbiIKICAgICAgICByZXR1cm4gYXJndW1lbnQKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAg
            IyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBj
            bHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRl
            X3JiZF9wb29sKHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNl
            bGYuaW5pdF9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVk
            ZWRfa2V5cy5hZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19k
            YXRhX3Bvb2xfZGV0YWlscygpCiAgICAgICAgIyBkb3VibGUgc3RyaW5nIG5lZWRlZCBmb3IgdXBz
            dHJlYW0gZXhwb3J0cyBvZiBmbGFncwogICAgICAgIHNlbGYub3V0X21hcFsiQVJHUyJdID0gZici
            e3NlbGYuZ2V0RmluYWxVc2VkQXJncygpfSInCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BB
            Q0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJL
            OFNfQ0xVU1RFUl9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgp
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1
            bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RB
            VEEiXSA9IHNlbGYuZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0
            X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktl
            eSgKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZGFzaGJvYXJkX2xpbmsoKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoImNsaWVudC5jc2ktcmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2Vs
            Zi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIi
            KQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19N
            RVRBREFUQV9QT09MX05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfbWV0YWRhdGFfcG9vbF9uYW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQ
            SEZTX0ZTX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAg
            ICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJn
            X3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9H
            Uk9VUCJdID0gc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBD
            ZXBoRlMgbm9kZSBhbmQgcHJvdmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFw
            WyJDRVBIRlNfUE9PTF9OQU1FIl06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNl
            bGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQog
            ICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJP
            VklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBI
            RlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAg
            ICAgICAgICAgIyBjcmVhdGUgdGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAg
            ICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAg
            ICJjc2kiLCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAg
            ICAgKQogICAgICAgICAgICAjIHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAog
            ICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9V
            UCJdOgogICAgICAgICAgICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUs
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAg
            ICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAog
            ICAgICAgICAgICAgICAgICAgICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEi
            LAogICAgICAgICAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0g
            PSAiIgogICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAg
            ICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lO
            VCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BP
            UlQiXSwKICAgICAgICAgICAgKSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkK
            ICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNf
            UE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2Vj
            X3Bvb2xfbmFtZSgpCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSA9IHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9t
            YXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMK
            ICAgICAgICApCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X3Bvb2xzICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5
            X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2Vs
            Zi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZ
            X1BPT0xTIl0pLAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90
            b19hcnJheSgKICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX1ZBTFVFUyJdCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAg
            c2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9M
            T0dZX1BPT0xTIl0pCiAgICAgICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5
            X3JiZF9wb29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRf
            dG9fYXJyYXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAg
            ICAgICAgZWxzZToKICAgICAgICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9s
            b2d5X2ZsYWdfaXNfbWlzc2luZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJF
            RklYIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0
            X21hcFsiUkdXX0VORFBPSU5UIl0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X2VuZHBvaW50OgogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAg
            ICAgICAgICAgICBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5f
            YXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAp
            OgogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwg
            InJlYWxtIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBz
            ZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0
            aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVf
            bmFtZSwgInpvbmUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAg
            IGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAg
            ICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAg
            ICBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAg
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9L
            RVkiXSwKICAgICAgICAgICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAgICAgICAgICAgICBlcnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxp
            ZGF0ZV9yZ3dfZW5kcG9pbnQoaW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVy
            cm9yLCBzZXQgdGhlIFJHV19FTkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAi
            LTEiOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRp
            c2l0ZSBwYXJhbWV0ZXJzIG9yIG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMu
            c3RkZXJyLndyaXRlKGVycikKCiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAg
            ICBmb3IgaywgdiBpbiBzZWxmLm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQg
            ayBub3QgaW4gc2VsZi5fZXhjbHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3Jp
            dGUoZiJleHBvcnQge2t9PXt2fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdl
            dHZhbHVlKCkKICAgICAgICBzaE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAg
            ICBkZWYgZ2VuX2pzb25fb3V0KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkK
            ICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAi
            IgogICAgICAgIGpzb25fb3V0ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJleHRlcm5hbC1jbHVzdGVyLXVzZXItY29tbWFuZCIsCiAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgImFyZ3MiOiBzZWxmLm91dF9tYXBbIkFSR1MiXSwKICAgICAgICAgICAgICAgIH0sCiAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2st
            Y2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZpZ01hcCIs
            CiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHNl
            bGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAgICAgICAg
            ICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5nIjogInt9
            IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAg
            ICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZnNp
            ZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAgICAgICAg
            ICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgt
            b3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91
            dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAgICAidXNl
            cktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYgJ01PTklU
            T1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1lbmRwb2lu
            dCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05J
            VE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9OSVRPUklO
            R19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9uaXRvcmlu
            Zy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0ZXIiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiTW9u
            aXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAg
            ICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVfU0VDUkVU
            JyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicgU2VjcmV0
            CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNS
            RVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JC
            RF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAog
            ICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgK
            ICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAg
            ICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ1c2Vy
            S2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAj
            IGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            ICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAg
            ICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMg
            aWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBq
            c29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5h
            bWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119
            IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAg
            ICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0
            X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRoZW4gb25s
            eSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9M
            SU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAg
            ICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            IHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAgICAgICAg
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0
            X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQgbm90IHNl
            bGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6IHNlbGYu
            b3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            Ijogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmImNlcGgt
            cmJkLXJhZG9zLW5hbWVzcGFjZS17c2VsZi5vdXRfbWFwWydSQURPU19OQU1FU1BBQ0UnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAg
            ICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRf
            bWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsn
            Q1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIs
            CiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAg
            ICAgICkKICAgICAgICBlbHNlOgogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJCRF9NRVRB
            REFUQV9FQ19QT09MX05BTUUiXToKICAgICAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAg
            ICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgt
            cmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICAiZGF0YVBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9P
            TF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlv
            L3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0K
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJuYW1lIjogImNlcGgtcmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3Rv
            cmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25l
            ci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICMgaWYgJ1RPUE9MT0dZX1BPT0xTJywgJ1RPUE9MT0dZX0ZBSUxVUkVf
            RE9NQUlOX0xBQkVMJywgJ1RPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUycgIGV4aXN0cywK
            ICAgICAgICAjIHRoZW4gb25seSBhZGQgJ3RvcG9sb2d5JyBTdG9yYWdlQ2xhc3MKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXQogICAgICAgICAg
            ICBhbmQgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXRvcG9sb2d5IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAg
            ICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidG9wb2xvZ3lGYWlsdXJlRG9tYWlu
            TGFiZWwiOiBzZWxmLm91dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xP
            R1lfRkFJTFVSRV9ET01BSU5fTEFCRUwiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5WYWx1ZXMiOiBzZWxmLm91
            dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xPR1lfRkFJTFVSRV9ET01B
            SU5fVkFMVUVTIgogICAgICAgICAgICAgICAgICAgICAgICBdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAidG9wb2xvZ3lQb29scyI6IHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNy
            ZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9j
            b250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lf
            UkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBI
            RlNfRlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZz
            TmFtZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAg
            ICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFt
            ZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQ
            SEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JH
            V19FTkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNz
            CiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3Mi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            ZW5kcG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAgICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAi
            a2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAg
            ICAgICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VT
            RVJfQUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2Vs
            Zi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAg
            ICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdS
            R1dfVExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0
            IiBzZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5k
            IjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJl
            dHVybiBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vy
            c19wZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlz
            aW9uZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNl
            bGYucnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2Vs
            Zi5ydW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2Vs
            Zi51cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25h
            bWVfZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAi
            Y2xpZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBw
            cm92aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0
            aGUgcmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2Fz
            IHVzZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1
            bHQnIHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0
            YSIKICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAg
            ICAgICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3Jh
            ZGUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lv
            bnMoc2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4
            aXN0cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVu
            dGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29u
            X291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAg
            ICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJp
            bnQoZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICBy
            ZXR1cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAg
            IHNlbGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5n
            X2NhcHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNl
            cikKICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAg
            ICBjYXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAg
            Y3VyX2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAg
            ICAgbmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            Y3VyX2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4g
            Y3VyX2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAg
            ICAgICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBm
            b3IgeCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAg
            ICBdCiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8g
            bWFpbnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3Qu
            ZXh0ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNh
            dGVzIHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1z
            IGluIHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5l
            d19jYXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4
            IGluIGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAg
            ICAgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAg
            ICAgICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFw
            cGVuZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tl
            YWNoQ2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRo
            IGNhcHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBj
            YXBzLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRf
            dmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pz
            b24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29t
            bWFuZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHBy
            aW50KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihz
            ZWxmKToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9u
            cygpCiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAg
            ICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRl
            ZF9vdXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3Vw
            cG9ydGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQog
            ICAgICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmls
            ZSBhbmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRw
            dXRfZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3Jp
            dGUoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJq
            T2JqID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2Vw
            dCBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1
            dGlvbiBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJv
            ciBhcyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQg
            T1NFcnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBv
            dXRwdXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0
            ZG93bigpCg==
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:62cd98508ab41f9526f10c55f6c8054475a1b029230f255fe7461fb750fd1def
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        version: 4.18.11-rhodf
      entries:
      - name: rook-ceph-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: rook-ceph-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: rook-ceph-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: rook-ceph-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: rook-ceph-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: rook-ceph-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: rook-ceph-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: rook-ceph-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: rook-ceph-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: rook-ceph-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: rook-ceph-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: rook-ceph-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:bbfddd7cb4a6cb7a131a0cef174054890fd39956d63e737c92fe8233ad8f98e4
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5a35593357fbf280561a1d4a64ecc0d5c2044409acff3ce6cb490a4a61462a4e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:791ac512370af5cde1e089de3def760e7dbce69badc03137051520ce316d63e2
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5dc66b7c486d1604c4bbb04c85544147c43264840e597958e6c5ca83e7e34c0e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:ddfdf850c9238a6988889e1840efe3175f03dad2489a5651e21cac62ac44e23e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/application-services-metering-operator@sha256:c3600cc8cd789df1f3c8f48bb3b1bc4c9cbf0285b6f6ce06f1d563c52f50915e
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.connect.redhat.com/cpopen/common-service-operator@sha256:13f5b6699ab505d02550b09e100f7e205c7fde3f2bc36317025bc795be216761
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - registry.connect.redhat.com/isovalent/clife-bundle@sha256:6d2ba514e40559fe3568dd0107f5e6648d43b543b3dbe4319e4114ca05fe20a6
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/vault-config-operator@sha256:696bfe3769605870286e228f463dc2582202d6aeb8f199d78d87f153c68d7401
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-emrcontainers-controller@sha256:cde2c51ca18d0969be0549e37ecd03cc430be3b4746d53b8744cb552a2c475df
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:29ec87ce725d2c277824312607915bb7de92e18015c388543173dbdadd5834a5
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:c906c822fbab57c1eb95880b0dfb9cd04100c86222a5d5cfe6342298e3bb1b9e
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:e26bc4b183f69a448552f9fe740292d4e803165d341d7d66047688d01fcd060c
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
        version: 4.17.13-rhodf
      entries:
      - name: ocs-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:a4b5197db1ccda9934f30bfd63e41b0bef430ea9bd224cd3dc022466742b82b5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:0fcc80541d8966f3ea3502da1c4fdda24de277d8d73d0752b46e7d69fbcf7038
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
        - registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: ocs-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:9fcdfa4615b91a5bb919aff71fc4df046365b2ee6d540b5f5f47f9ccaeb9877c
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-003
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:9f5b63888cbf4b1c90e3803d0017a01ad9c0c22693b978c9ee8299bfc5337832
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:212ab9249707ca8ff835186d88e7888af18118fcbe98d948a3156b9700307a79
        version: 7.6.12-opr-003
      entries:
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/ipm-bundle@sha256:68c38ac59cf0d6cc145eb694e5b5a724da733caddb82c5585ef132e880dde544
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.connect.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:19b163f055da43946643057c07500784210d02db868d2aa10f25a6501d767d3d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/community-operator-pipeline-prod/cluster-manager@sha256:a4c8a54148538b371e5cea3b2f2d4109f7b6bc4a7f2e5b52a840c9340fbf5e47
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/openshift-community-operators/oadp-operator@sha256:f8c5c6e59f13db8847080ea6d964da6e50e56f7d373bcc4e4278b1a8a80395e4
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/openshift-community-operators/oadp-operator@sha256:b7b4d340157513da2ebef6bc7bc262967d0e825e2baa1f2bc2ab2e170457e015
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a6399dfd6fe0e8dc9f59fac4382320d23f9e1709e48ed6da7db69e76354b2646
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:0a86a1808d9906eb57886d3d3cd5c04cf9578a80860342132d36ecfc411bf807
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:b02bab3af52a2469ced96618f1d122464568ba21ebc0a3a945b78ec4e6bdc80d
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a614ec94c2468e34e129e9ecd4bc0d1895eb67b96e77eb5ad94cd8bd507860f5
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Net AI
      provider-url: https://netai.tech
    name: netai-demo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netai-demo-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deployments.netai.tech/v1alpha1",
                "kind": "AntennaDeployment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netai-demo-operator"
                  },
                  "name": "antennadeployment-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Networking
          com.redhat.openshift.versions: v4.18
          containerImage: ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
          createdAt: "2025-07-09T11:33:08Z"
          description: Net AI Operator that uses EnergAIze engine forecasts to scale
            Red Hat OpenShift resources.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Net AI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntennaDeployment is the Schema for the antennadeployments
              API
            displayName: Antenna Deployment
            kind: AntennaDeployment
            name: antennadeployments.deployments.netai.tech
            version: v1alpha1
          - description: Demo is the Schema for the Demo API
            displayName: Demo
            kind: Demo
            name: demos.deployments.netai.tech
            version: v1alpha1
        description: POC Operator that uses saved EnergAize forecasts to manage OpenShift
          Custom Resources
        displayName: netai-demo-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - energaize
        - scaling
        - scale
        - manage
        - forecast
        - cells
        - 5g
        - networking
        - net
        - ai
        - network
        - telecommunication
        - xapp
        links:
        - name: Netai Demo Operator
          url: https://netai-demo-operator.domain
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Net AI
          url: https://netai.tech
        relatedImages:
        - registry.access.redhat.com/ubi8/ubi@sha256:68fecea0d255ee253acbf0c860eaebb7017ef5ef007c25bee9eeffd29ce85b29
        - registry.connect.redhat.com/net-ai/demo-operator-bundle@sha256:b1a30de1e984da0487a7d042bd940876bedb5a2f43dd73c7895e684a3a35d81e
        - ghcr.io/net-ai-tech/netai-demo-operator-bundle@sha256:0783aba7b039a569b49de2a61ef6dc6ff1d1f301d93c26d623b7f652321871c5
        - ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
        version: 0.0.1
      entries:
      - name: netai-demo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: netai-demo-operator
    provider:
      name: Net AI
      url: https://netai.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:42102a2e462e27dd1de28ca3f7db02d227c44736f2d57c05037a83ecbff6dd9f
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:b405d6148aa680bf31bdc6185c1702916c3219983d8482f27d89ed2f93659aa7
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-recyclebin-controller@sha256:9d916ed0af37d60d90f9b4711b1544b9373bd247476278c9c4041c9dbce6acc9
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:c60664e5da8eb23bbd918b1c13b83565b1f33a1aa8961b425232fa1ef4d1a6f1
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:82cafd48db8107dd141064ce7530468b11804c8c54235a317eb450a7606bac3e
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:5d9b56300a3b17178467954b533bee19622f8e96fe9d4ac9745ead2c59b9ec5f
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-efs-controller@sha256:338a28600b08d1ead95ed409e3326e03f3c5e02ed840d3be56a950db6b1de061
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/ipfs-operator@sha256:03a61ef328feb626408d42904cdfb9d45f66e022d5a1e1088a15b74455487743
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.connect.redhat.com/yugabytedb/yugabyte-platform-operator-bundle@sha256:4cee86e973206b5300e88af088aca700cde3b2e2e5e1c70b9a20289e9f592349
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3e0fde49a8e8c5f780a76087bbc42507a42fea0e1fea572ae05ba66d27338365
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3b36dcea36ebf4b2e95fac6691f6fc8f1242f7e5c38195a9af28bfad514c28cb
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:36dc3710ac11d86e1cf0a3c6994ae0e6feb6c8ff1e8110997f27f4e573c259bc
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:64cadb49077eff7863dd678c4aec70f4101f17ee8e0bf7ca64ce7010c9bf4b9b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:4981d96081cc3d10cd26f20aab792e9c195b05bdcd609367011b7e9c7e6759b1
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-apigateway-controller@sha256:44d34788b43d103530115fc53b0252bfdbe6e07a692847a3d32aa21e45892ff9
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        - registry.connect.redhat.com/tigera/operator@sha256:4d0d814fcfeb628aae8a05cb7a0a416ae85395197950085c907b6cdcf2fb4317
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        - registry.connect.redhat.com/tigera/operator@sha256:b1a8b56ff5e1bb5326e1bde4931303358410da9768f0ebf9c08e2c2d8661d742
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        - registry.connect.redhat.com/tigera/operator@sha256:cb7f3367ff064cd01dc636a43fd89d5f6c4e6a7004cb7c94b1758b2398e36d71
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        - registry.connect.redhat.com/tigera/operator@sha256:de9724b84ca493d3b14c8f801c1049ac4713166b00d9bf3ea0b1986aac5040fa
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        - registry.connect.redhat.com/tigera/operator@sha256:8e0c026f369f80e8b36d0c94f1fd60f7f53656e57dc5be0b8b12189546874909
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        - registry.connect.redhat.com/tigera/operator@sha256:1fb1007f7aad744f0e1b0f6a32022c2571ee8558e9db53c85976bbd2a891a6b0
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - registry.connect.redhat.com/tigera/operator@sha256:42a4d0d0e24752dbb51daebcfca4d4850d04e79578bfb8b41cdc2a2a289df4a2
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        - registry.connect.redhat.com/tigera/operator@sha256:7894b770bb02b2271e1414d67f300218cc646534c9f2a6917e51dc8bff8ba731
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/openshift-community-operators/awss3-operator-registry@sha256:9ab3a2cd00b8704a8bae8e9b75f2a95d0a3e60fa3cb52589db8a4e7641b73027
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/openshift-community-operators/syndesis@sha256:9e8c08793aebc8857dcd451fa2fde2b3f05b9ff4f3d3cc604f9f752193f58d42
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/verticadb-operator@sha256:b7691da906474a8ab829966bb3da4a846eaf12909c97bd0d9cb6a1e5e7ff6c03
        - docker.io/vertica/verticadb-operator:1.11.2
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        - quay.io/community-operator-pipeline-prod/verticadb-operator@sha256:be706a582a4b8cae487835e920288b41229834cd3477b7389adb363db4b4f2ca
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              },
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "ServiceFunctionChain",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "servicefunctionchain-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "servicefunctionchain",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "servicefunctionchain-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-10-15T19:06:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-dpu-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
          - description: ServiceFunctionChain is the Schema for the servicefunctionchains
              API
            displayName: Service Function Chain
            kind: ServiceFunctionChain
            name: servicefunctionchains.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-dpu-operator-bundle@sha256:4e98c0f629e938a0c9d011e6a675d45589b044a02b3f9551fc0e8f7cd172c8a0
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:acb5ab931f4d7a46bb195c59d156441af5e9092b2739bcb0379b89d47ccf4e81
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:5dedc4733081fc21bcb77dbf4f4c032bf418f00bc4371c29c7ce028f51833c0b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: dpu-operator.v4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: dpu-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: dpu-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: dpu-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: dpu-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: dpu-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: dpu-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: dpu-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: dpu-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: dpu-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: dpu-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: dpu-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: dpu-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: dpu-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: dpu-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:4ab17b92016288f1ebfb2106c826a17e75d041f0dba7eaaf7545193c3305a408
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - registry.connect.redhat.com/streamnative/function-mesh-bundle@sha256:e4377187c20a70d768a253bc0794f3152aa6fcee22ddbf2fc88f45756e9681fe
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - registry.connect.redhat.com/loadcore/loadcore@sha256:7fca6883b092ac7955bb42f893f163403960c81b83e360cfd1eac802e57ea839
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.connect.redhat.com/cpopen/odlm@sha256:064b597625550ca75dbc9f3c8fd6b2de6e0b9cb7f42d604e871bfc668ad05ac1
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:eccacdf9ac2b0b1f6061aea457c705319f268bfa30631ed9919ae393f0ef7f71
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:cf330a9e405a9641c213a74baf28307a6e247c4f03cc0523d071dad8f5f9e2e9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:412403b096566965b2c92a0a55beed8ebd6d8113a7e1fdf5eb3334769929ca41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:c89d8be470c2fd052c5be7a7dbfb133c9e3528bfa5b3681ffb8b3223c4c6bac2
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:6c9e04df3f5ddc45788333a85f6aeff8953b008a4b165e9cc2d5a079c6c73ece
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:5611a2fd53859d55d5dfdfdd117e80a0f30be0a4489ad838d571e872d3b5158a
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        - quay.io/openshift-community-operators/horreum-operator@sha256:a9d82ce368b015e45b93f917d4069978a0ba323f3dd684d470cdfd262e0c9f05
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-wafv2-controller@sha256:cb9a426baec724225c0132973d5b90a83672f9b0b75451adc6d7fd858c79ab58
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-glue-controller@sha256:2e1e23de186f3c61be8c7a30f3b295fe3b8b101509cf30bd6b7bff055ab74c82
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.109.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
          createdAt: "2025-09-18T09:13:25Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/eclipse/che-dashboard@sha256:8e2285991a725e87d2a9131ea23e08ce97cd5bdf1118f53a55a140c2c97ca25d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:1df1b75f702952b650e21717b04de5f39c84ca8f55f4afa0f32644ef8638d03a
        - quay.io/che-incubator/che-code@sha256:3b5c7e1ae664c82ddb3eaedc8161eb72a3a1a22c6c5e3f342f4aad972a6282a5
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/community-operator-pipeline-prod/eclipse-che@sha256:545d98c512c99bc7bf4ccd02c1eb554bbc4bca7da8cc66e0e26934edbfd50c59
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/che-idea-dev-server@sha256:0344c9574a6544c4c0ef93a798a9705ca3a4624508a55c8f94c2bae7196f914a
        - quay.io/eclipse/che-plugin-registry@sha256:bd71aa6817eeda0449056e78a6e832b82dd37b65ec6cba2bac3ffb9defa4048e
        - quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/che-incubator/configbump@sha256:77bd0fc269ff3724583b43ec68e158fcd300e825e1b44e24e2b9e67ed28df6da
        - quay.io/devfile/universal-developer-image@sha256:f76ea662cfdb9e80b93bcb58a393aed8b52aa72e7da73762d30a3ad22c3f06d9
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/eclipse/che-server@sha256:e5a542da0484eb2223ac292e6005cf7e116bec15d377557f92949728e1f5c8bf
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        version: 7.109.0
      entries:
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/group-sync-operator@sha256:fe6fda2de67fc4996d3f388e11faa0130fee19453111c829ba0aaa5be5410d01
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:e6d5ef5583ae26a8312dc331bd6003fda6c7c248de15b6790a128d0cd03965dd
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:35e0ee6b3ec5cb654b709c5e4a76a1f08182d826b47f077031d9139d4c2cc308
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b876ec88e1e64c17572e956b1c7e4020f5a423c1275f7f57025823ff7f749c61
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b3a45189ae8fc5302fb8a9e15d5747bacce044537923621edf21e9d3b267a27e
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-bundle@sha256:6cd4cba02a9e5c22062108363f6a052734c398019a19f84afc18a7b05a71d578
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        - registry.connect.redhat.com/loadcore/keysight-wap-stats-dashboards-bundle@sha256:f69f61eb818a1f0f4035ae79487402237b739c55eb104fc4643a7d3ef8d7d0c4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:6d9eabdb4f645aa1b8b6c08ca94f9ec89466cdc35b3cb02aaba22e160d8fd9ea
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:5484f8ecf079ee126d24c05e1bcb06467146b3b8e751e90e0fd68f3e933125f0
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:11ce1c35f9c3d9b5ef9539b4402eb8358df843b29ffd94af7a453e5d765e4767
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecs-controller@sha256:280d5ec57ad4deb45c145e702242468a4cc04d23112c33f4e72722a269c900e8
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-object-store-operator@sha256:c5af6adc5e7f9bb2d5bf6025b1b8c3e33ce5b9349a0246df1a26b7ec91b2ea2f
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lightbits Labs
      provider-url: https://www.lightbitslabs.com
    name: lightbits-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightbits-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "CSIDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "csidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "csidriver",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "csi.lightbitslabs.com",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "controller": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "tag": "v1.21.0-ubi9"
                  },
                  "node": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "sidecars": {
                      "lb-nvme-discovery-client": {
                        "registry": "docker.lightbitslabs.com",
                        "repository": "lightbits-operator/lb-nvme-discovery-client",
                        "tag": "v1.21.0-ubi9"
                      }
                    },
                    "tag": "v1.21.0-ubi9"
                  }
                }
              },
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "cluster1",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cluster",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "cluster1",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "addresses": "192.168.40.46,192.168.40.47,192.168.40.48",
                  "apiService": "192.168.40.46:443,192.168.40.47:443,192.168.40.48:443",
                  "clusterAdminSecretName": "lb-secret",
                  "clusterName": "cluster1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          com.redhat.openshift.versions: v4.17
          containerImage: docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
          createdAt: "2025-08-20T11:35:15Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/supported-versions: v4.17
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '[]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/suggested-namespace: lightbits-operator
          repository: https://github.com/LightBitsLabs/lightbits-operator
          support: Lightbits Labs <support@lightbitslabs.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster is the Schema for the clusters API
            displayName: Lightbits Cluster
            kind: Cluster
            name: clusters.storage.lightbitslabs.com
            version: v1alpha1
          - description: CSIDriver is the Schema for the csidrivers API
            displayName: Lightbits CSI Driver
            kind: CSIDriver
            name: csidrivers.storage.lightbitslabs.com
            version: v1alpha1
        description: |-
          Operator for deploying and managing Lightbits CSI driver.

          By using the operator users can quickly and easily deploy and upgrade Lightbits CSI Driver.

          ## Requirements

          - Kernel support for NVMe-TCP
          - Kernel support for NVMe Asymmetric Namespace Access (ANA)

          ## CSI driver Supported Features

          - Filesystem Volumes (xfs/ext4)
          - Raw Block Volumes
          - Dynamic/Static Provisioning
          - Volume Snapshots
          - Volume Clones
          - Multi-Tenancy & Credentials
          - Volume Expansion (Online/Offline)
          - Storage Capacity Tracking
          - Host-Side Encryption (LUKS)
          - Publish ReadOnly volumes

          ## How to install

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/install_lightbits_operator.html#installation)

          ## Documentation

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/)
        displayName: Lightbits Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - storage
        - lightbits
        links:
        - name: Lightbits Operator
          url: https://github.com/lightbitslabs/lightbits-operator
        maintainers:
        - email: k8s@lightbitslabs.com
          name: Lightbits Labs
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Lightbits Labs
          url: https://www.lightbitslabs.com
        relatedImages:
        - docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - registry.connect.redhat.com/lightbitslabs/lightbits-operator@sha256:53db496f7a254632a25acf8974aca4041cbddb109499871622ffc2182e7aaa83
        version: 1.2.0
      entries:
      - name: lightbits-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: lightbits-operator
    provider:
      name: Lightbits Labs
      url: https://www.lightbitslabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ram-controller@sha256:a90b07270f0cfa8929e01569895ab9a209ee44dcc452a3e1f6e0e9a52c5f13b9
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-09-30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
          createdAt: "2025-09-30T03:27:25Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.41429998

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - quay.io/community-operator-pipeline-prod/sailoperator:1.28.0-nightly-2025-09-30
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-testing/proxyv2@sha256:28f058629e7969e8441b677b308a09bf5d4968dbd5024d54df8472d1f43e1a8e
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-testing/ztunnel@sha256:f4895a66a99ca0c87e54260d6d5784068a48fe7fbfa25144debad420ea7223fa
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-testing/pilot@sha256:42d64978872f6be8540027002a361933e4fe69590458353e39d2643b2459bb3c
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-testing/install-cni@sha256:a8c8f6fb3048461e08b86056000a3034f7945df067d5ebf43cca9bcb0ffa3916
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.28.0-nightly-2025-09-30
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      - name: sailoperator.v1.28.0-nightly-2025-09-21
        version: 1.28.0-nightly-2025-09-21
      - name: sailoperator.v1.28.0-nightly-2025-09-19
        version: 1.28.0-nightly-2025-09-19
      - name: sailoperator.v1.28.0-nightly-2025-09-18
        version: 1.28.0-nightly-2025-09-18
      - name: sailoperator.v1.28.0-nightly-2025-09-17
        version: 1.28.0-nightly-2025-09-17
      - name: sailoperator.v1.28.0-nightly-2025-09-16
        version: 1.28.0-nightly-2025-09-16
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/install-cni:1.23.0
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/pilot:1.23.0
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:a5698068f57fb27c6e4b8c385878c4651967ae5f17412c6ea6cf8bd34bfe76a1
        - docker.io/istio/proxyv2:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/pilot:1.24.1
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:5416c5fc62c7d9498349ed771bf93bfc0f712ad8da0b8ee109dc49d1e684fa56
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.3
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/proxyv2:1.24.2
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:9cc4d5a9f46e4d05f8adb5232459002a073c101cc813330325afb6682e2138ae
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.4
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - quay.io/community-operator-pipeline-prod/sailoperator:1.26.3
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-sagemaker-controller@sha256:c962ee90891a05197edc08b4cf015fce1e2431cced90bd274e89bb87708420f3
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-bedrockagent-controller@sha256:6688b26f71500af0e2ffbd0ccc53926b0c41adaaf817a88f0fdfd6fb398075e7
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        - quay.io/community-operator-pipeline-prod/trident-operator:25.6.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.32.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90",
                  "kernelVfDrivers": [
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
          description: This operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Intel device
            plugins for Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              GpuDevicePlugin is the Schema for the gpudeviceplugins API. It represents
              the GPU device plugin responsible for advertising Intel GPU hardware resources to
              the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              QatDevicePlugin is the Schema for the qatdeviceplugins API. It represents the QAT device
              plugin responsible for advertising Intel QuickAssist Technology hardware resources
              to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              SgxDevicePlugin is the Schema for the sgxdeviceplugins API. It represents
              the SGX device plugin responsible for advertising SGX device nodes to
              the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fpga
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4
        - registry.connect.redhat.com/intel/intel-device-plugins-operator-bundle@sha256:4a422ca897caeabb9fc3be3dfef3746819868de1674ea61c657cad257fbddf12
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
        version: 0.32.1
      entries:
      - name: intel-device-plugins-operator.v0.32.1
        version: 0.32.1
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        - quay.io/community-operator-pipeline-prod/flink-kubernetes-operator@sha256:7c919847b88a5969d0c0830e4154512bbca28015cf6fb7fa70cd20824a726685
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:13013220d7376d23df4780fe65c3bea325bdeb7b7a89d5704aa52e948ed7a2d5
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-operator-bundle@sha256:3fb13020c6a34b329825d268dd70c3c9310075780768312dc52cac1e7134a030
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:aa2267c1c871699db37763e938125f290ace4aff2aca6f668b2ce01f7df92b8d
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        version: 4.18.0-202509100344
      entries:
      - name: aws-efs-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: aws-efs-csi-driver-operator.v4.18.0-202508260754
        version: 4.18.0-202508260754
      - name: aws-efs-csi-driver-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: aws-efs-csi-driver-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: aws-efs-csi-driver-operator.v4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: aws-efs-csi-driver-operator.v4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: aws-efs-csi-driver-operator.v4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: aws-efs-csi-driver-operator.v4.18.0-202505280604
        version: 4.18.0-202505280604
      - name: aws-efs-csi-driver-operator.v4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: aws-efs-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: aws-efs-csi-driver-operator.v4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: aws-efs-csi-driver-operator.v4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: aws-efs-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: aws-efs-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: aws-efs-csi-driver-operator.v4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: aws-efs-csi-driver-operator.v4.18.0-202503190402
        version: 4.18.0-202503190402
      - name: aws-efs-csi-driver-operator.v4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: aws-efs-csi-driver-operator.v4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: aws-efs-csi-driver-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: aws-efs-csi-driver-operator.v4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-operator-bundle@sha256:7f3d41672fa8c60725304f0d45e468a4f01defcab7352501737ffddfd32edded
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-operator-bundle@sha256:75cb78f7a40d3daedbc11cbbb5505e029328b4120f0de11d4f6423de943e92d8
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:279f5d817f9dc90b202924fc2b2ab59e5303e9dddc1d78d71017d65cdd66d565
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
          createdAt: "2025-08-26T06:53:59Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/hpestorage/xspc-csi-operator-bundle@sha256:658dd12fe1b92bf38dd24601043cb84e27a27b0d987a36a0ec7d99b0514574e2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:4c289e5023fd3e59b2cf40335eba9eb53caf562472f16dd97b370773a0ba608b
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: xspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle@sha256:8cba51274389030d632c8b82255399217d485146f4797907134e3c4ee55cb4d2
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:3bc76999ece42fcb3acd913b32be3fa0e0da056832922819b81119dcfb5f1582
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:2aeba259864bbddaa8cdd46ab6d8c3495450a0081469a27b9acde53008d91bed
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:f58f60283f6cf9ea745f4a25378ef6690d98af8fda0eb63f663066631e71450b
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:48b37c9f9417ef442d896e79851b176505045b487bddf77bb4e6d91fd5ad3dd1
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        - quay.io/community-operator-pipeline-prod/node-discovery-operator@sha256:899baace103c58e7337ae5cd75e3764e6de26966aab7340037c612c26ca7cf87
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:53777788253eaa1199e1854d14782619aeec0df9844d86b83742fe0005671cde
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:e3a13ea8b84b9e1abd158d394e11113e5f6c04d879547fa1d3e8638f1ac979d8
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:31258400bb37f4b8837ef294820f8a728ae83f8031d4e01ff0392690a315a30f
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        - quay.io/openshift-community-operators/move2kube-operator@sha256:cdf288be5ea779e03900af9b57c4d922b8760b88da4fcbf88437f8031d679aef
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:236a53d6fad949285b6abafcdd1a4541a7748f952c0c96fa2e71500079579ba4
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:38d03dacca046dda3ba08c39943ec2e7db33f1c7bb897e84fe1308a0982c63f1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        - registry.redhat.io/rhcl-1/limitador-operator-bundle@sha256:4c56cd8462531cf8ba39316a0c6f7493240bfa4a6ace77352fe7ac1597e2fd90
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        - registry.connect.redhat.com/netapp/trident-operator@sha256:53a4889e1036af9f858c9fdb1bd58f59b758a301bc773379572cf279f930f5c6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - quay.io/openshift-community-operators/ocm-operator@sha256:ca6a3b24ca4ecfcab52ee0402e007d6f14b23fa7f3408d740923806989b170e4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/community-windows-machine-config-operator@sha256:2b79f0e6b8bc069fac7d6cb5c86783fdc98ccae30d279ca0492f1eca96b6ea3f
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-eventbridge-controller@sha256:3a284fb7884e50ed9976368f312c75fc1f1cb2ec361297b38ee72d0ac6b66fc7
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:1f4fd5ed8e398b713e733fb552e7294f22b3927d4e017f2ac15ea20024d3de32
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:759dc7d6836f8c8830ec145e3088bddc1b63bd40ff3047079fa8f0e986e74c30
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/openshift-community-operators/konveyor-operator@sha256:04e63481c5104bd034283d5d4b0567167f62480d2c61154b1d16dbd41c344dfb
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/openshift-community-operators/konveyor-operator@sha256:d4c590033b1a2bcd67375465298e5cb6d136fe091abb12b8d009aa683c23ef62
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:848db711d9e8a329b689b40003293fce19c43faedd31afd80d9ae9d5c11cda49
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:e29f175a799597c84f0663a66767bba24feb78f802516056b35b5beb14d9eb41
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:2d58f12044b0951c87c851527240196db148acb9797e6ac49f9474a6e6039d71
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:7b8bdecbe3b3def1ec810ad53f604c7a71daebcfa7433cbb3a0db937580d087c
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:16c4f165143b8bab7f3eb08c815c2cae0dbf7fffd9ce7a625357be5806c24c5c
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:c28522543b039454d2d97b48198d099ac0b8211cbc58f13468d5f70b9f69b037
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:18ba5c587dc613f9c1a8bc9e66081e52f3a37eedeea5e75881e628e9c35ef460
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - registry.connect.redhat.com/nvidia/nim-operator-bundle@sha256:3dbd307d32a39e37a02ac8fa9c9fb15a0c2fd6dc5aea12d55abb061e86365124
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-directory-operator@sha256:9d93e3aeef3cfbba0b655de7783ba9fd1f4cb4ee5de4389005b7c31d971d4947
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/global-load-balancer-operator@sha256:50771c320c8b7faea144f765535e89979f9f81e61d34c8dd2e86139fa68e222e
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/aerospike-kubernetes-operator@sha256:bfa83ff729ef8b3a4ebca27b02e223327d3e1351d8bc78a06c18873d288dfb02
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        - quay.io/community-operator-pipeline-prod/aerospike-kubernetes-operator@sha256:63cdb7ea6cd6bda4c6ab2c1f3467489c2cef6c8a95836763b9c88b20c43e8168
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - quay.io/openshift-community-operators/sap-data-intelligence-observer-operator@sha256:5217360266f62fc8ba087fbc730074c14aacf62dbd124fdc3aa46650d51a243b
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sap-data-intelligence-observer-operator@sha256:79aa33cb7e4b570fc9c962fc6bd59708c7e28e81e2b188608780fa085db91442
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/kueue-operator
    name: kueue-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kueue-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "policy": "Disabled"
                    },
                    "integrations": {
                      "frameworks": [
                        "AppWrapper"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "FairSharing"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameOptional"
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob",
                        "Pod",
                        "Deployment",
                        "StatefulSet"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "byWorkload": "Parallel",
                      "policy": "ByWorkload"
                    },
                    "integrations": {
                      "frameworks": [
                        "RayJob",
                        "RayCluster",
                        "AppWrapper",
                        "PyTorchJob"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "Classical"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameRequired"
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-05-01T20:58:29Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.operator.openshift.io
            version: v1alpha1
        description: Kueue Operator description. TODO.
        displayName: Kueue Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:de644d640f355815f3aa40ced5456bb5b0e958aeba8bfbdc4ab360fcdb90f26d
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:0c4d9cd97f7579adbf2afd238c953ca5f170408cf27e3b2bf4aa16c108a76881
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:dd765c446564d4397570f4bc056b55989f6c0bae0d70514874e67765c9c77889
        version: 0.1.0
      entries:
      - name: kueue-operator.v0.1.0
        version: 0.1.0
      name: stable-v0.1
    - currentCSV: kueue-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-06-24T02:48:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:30659b42f5053dd756190211454c01f049bbc7689dddda3ecd8c406b157873a7
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:9a0b85b0752135694193befe00d10f7abdc0c99c6b9f295d59b237fa0fe3e6d2
        - registry.redhat.io/kueue-tech-preview/kueue-must-gather-rhel9@sha256:e5306c95bf6c5f55be6e791fcb66fe9b34f4f982536f4f90f8ed74acf3f6c31d
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:6f0b5f62f6e549c6f6769dd9a358a14668d7244f9c20753d1690d3e09948588c
        version: 0.2.1
      entries:
      - name: kueue-operator.v0.2.1
        version: 0.2.1
      - name: kueue-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: kueue-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-07-15T01:52:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue/kueue-rhel9-operator@sha256:5db0af8088cce6e9e7a88faf80855b6eaa1dd8679c6a56dd5ee98c50a3a8e2e4
        - registry.redhat.io/kueue/kueue-rhel9@sha256:16007bd38c480b607b63c82b3ee8721441cc374ee31c04b9e329426314590195
        - registry.redhat.io/kueue/kueue-must-gather-rhel9@sha256:f5f624ef8dbb3676ed4e37335e226e76e5f886c1990c6e0e8c0e4762740ed98d
        - registry.redhat.io/kueue/kueue-operator-bundle@sha256:1bafac48c00fb47972f512ccd5d846efbbccccd2259395de17537b39834d89df
        version: 1.0.1
      entries:
      - name: kueue-operator.v1.0.1
        version: 1.0.1
      - name: kueue-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: kueue-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/kueue-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-bundle-certified@sha256:a9b8102ab0ec81e5bea84332ef20b023d06660b410e7e2deadb00904d2098613
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opentelemetry-operator@sha256:5dc347a5085c10823a122fab0bd42afd5d9b9aa19ba17e8992e9492000886629
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        - quay.io/openshift-community-operators/snyk-operator@sha256:97908144406700ae9c038316661dc7904f3e0f4e01b1bbdf68259b59dfb86d6c
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cadee58c1c1040fad5a46fc9be5fae4a1944845c5e736c50fa192192d5c17a36
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cf782e9e2a62a2fc71c81032e6bf40f9ba15d37c33feff2df551edb111c4e17c
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:60a9e4b05e7eb17ebfe58dd55db96ab2c68580d9282117f74f21869cb63e4a73
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:1461576c2e06c2ee8d84fd40836184566a346ed596b99cb81fe6a08d0ad99096
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        - quay.io/openshift-community-operators/camel-k@sha256:2b8369f729c18384f45a458e88c9f551eb00f26792b40dbf286e6ea2837e9400
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:6154ff0f72c9c49f1b7ea0edbe93e0a3b04f6e63abbd851d0e6706796050d672
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:21ce00e81ce495c142cbe9611f494cfb26b52b3c09a507bee0572c1e1e3333d7
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        - quay.io/openshift-community-operators/camel-k@sha256:26df9d918e942f7fc637212df87a114b5edd52570da18cc212b3499354cdb993
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:11dac0881f5ed66b5d42b2d69b13aba31cd56e6b32157cb3ce977be197189b79
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        - quay.io/openshift-community-operators/camel-k@sha256:cba8cde86a870b31f7dae9a554f4c496ed81731f91a8e64cd62b0caa88efaa38
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:06d31bb777b37f8f7c8981efc24ca26a2f9c56a1c579243d0c7bdddb34c76a98
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-dependencies
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-dependencies.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          createdAt: "2025-09-11T15:25:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          deploys dependencies.
        displayName: Data Foundation Dependencies
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-dependencies-operator-bundle@sha256:6c2223357b672b1a59d22eeab742d3c241d87d5524935974876c9371ad650afa
        version: 4.18.11-rhodf
      entries:
      - name: odf-dependencies.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-dependencies.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-dependencies.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-dependencies.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-dependencies.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-dependencies.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-dependencies.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-dependencies.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-dependencies.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-dependencies.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-dependencies.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-dependencies.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-dependencies
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "debug": false,
                  "ebpfProgramManagerMode": false,
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
          createdAt: "2024-09-13T12:46:10Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:577ab5d4780ba01e130ca5713fb7b3764244cdcb937b6b00e6bb17638f15fea2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ingress-node-firewall-operator-bundle@sha256:6c6a76287353a02c05164f9a0a3e07d27b75f0e5d403eb1b47f1e9dc7c9883d8
        version: 4.18.0-202509100344
      entries:
      - name: ingress-node-firewall.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ingress-node-firewall.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ingress-node-firewall.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ingress-node-firewall.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ingress-node-firewall.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: ingress-node-firewall.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: ingress-node-firewall.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ingress-node-firewall.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ingress-node-firewall.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: ingress-node-firewall.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ingress-node-firewall.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ingress-node-firewall.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: ingress-node-firewall.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ingress-node-firewall.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ingress-node-firewall.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:e39a0b57e73d92c5db330ac9b8a184eeb400397f14c05c1ef75c952d91e4f5b2
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:5024be9a3190288983e3c42e8a2dbb49aa44cd596237ad9ed0e72de5090dc9b9
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:ece58f1e360eccf7d970dca2f9da0c4d99a692648f165f677f11ff080649cf62
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:6bd423056b555d39337c0e3a9c31e7732e80f34d0c01f0766a6aadb2aff586f8
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle@sha256:261b77cfb315c608505e5659b2d01dc4e1ea96cd37fae25f417cd7df148f4ce4
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        - quay.io/community-operator-pipeline-prod/beegfs-csi-driver-operator@sha256:1a9f2aab026c2d1549be5b4624c56d65e813c619fdd5beefb062d5c9fb8981b8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        - quay.io/openshift-community-operators/nexus-operator-m88i@sha256:46cd06d477b4d09dbfca7ca4434ea58f881314138ed744c6390033bd6fb7dc0f
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-secretsmanager-controller@sha256:c92ea9b492b6585a816668bd2a77d65471e339e02e25e44c9558a34b352db593
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:151e7d9bc500b7252af46d0856e3e16559ff310a1d0efedcc5d914164bbdef55
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:309ba89dece3ee9901ed66c40bd28d71f20c7d91a7feff552e28ea0c772b9178
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:63bd0faf0b64e4672014205f9dcfeec6c436edd7dc2aaf83cd053d6e4569d90a
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:1b464c682cd29c8fb7bbc566a1575b9002ad61163c2b6e94138e0e4b0ce51140
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:8223bda16ac5ffac4b7407c849c4c30ee006ddafd34cad8a03d2af8314d56f58
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:d7333e377a5aace0d42e42ca936fa7a84d288b181eda4747f9367a9539f2bedc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-bundle@sha256:0064fcdc252a04010e605706be377417de4c9be4289358ec0f331d6a37befb0b
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        - registry.connect.redhat.com/loadcore/keysight-nats-operator-bundle@sha256:89bde716c2aa4dba0738d2481a2c4707c59a748f84349f9fb01cb430e69d7beb
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-rhmp-bundle@sha256:b66e3cfad24cf28e0a738179cab062667b0103c52cf2e84abb88d5e018e12b2c
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:27c5d91d62371f3cf0dea4d50044d8a6e0a6015a453c4216a030683ea2474c64
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        - quay.io/community-operator-pipeline-prod/percona-server-mongodb-operator@sha256:8659c9f25245e541609079970c8193fc6a58a7d323c249a81d743de696dc15d4
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:9fe87e1b0c49dd00a6ab0c35b68ebf641f347ac5b405d31f3c5ace02a7a30540
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:35d2c7c2849f89474fd5648a0d2781dde28131303cc8a597ce243a3e314fe0af
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:dd0ab94059f782aa31696b15e1b25787cd58bab4905fe64faec7e49bbae3d289
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator-bundle@sha256:0f20db33b1eb069a767089862412846aacb723dc22b44adeb9793495ad20ea50
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle@sha256:506afdc804dc30e0a720e8e6f151b75a32148844de42d90d8c0648e9108d9ea7
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        - registry.connect.redhat.com/elastic/eck@sha256:063a1f405b87751c5102693451ae5c2e9ad203c6b8a31b062a5b403bb8b2febc
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/fence-agents-remediation@sha256:6ab21048f86979bcc34a84965dc2a97467a79507ab03469fe8c6deed224881f7
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:18:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:1a98d55c28a0679134e24b6aafb3c4dc9db0098094c8d2bcfaa41a63c33308e1
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: odr-cluster-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-cluster-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-cluster-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-cluster-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-cluster-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-cluster-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-cluster-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-cluster-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-cluster-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-cluster-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-cluster-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-cluster-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-cluster-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-cluster-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-cluster-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:16f820575acdca39c0b7e85edff7b910809f6d184a9a479de2bb2f687c94a6f2
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-cluster-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-cluster-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-cluster-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-cluster-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-cluster-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-cluster-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-cluster-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-cluster-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-cluster-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-cluster-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-cluster-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-cluster-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a93b382258a34e0b56ef6616a8490e5f44ef1c6aa5d29c0d08d2147ae5b440e6
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3170da029869d15a246ad31d0debd4b9c03ba13855fb877a754cfdffb2286b80
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a7ce92f7f4aadca2781c44839bc357615395587bc8b68a735642967cf8694362
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:093c6a66d4faa1d980e4319a048db8b2869eba8f13f40f02615d6696257e6719
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:5a19032dda9a0f584b41cbe9e729ed23e6003d90ebbb026eb9e7bb9769d4a4e4
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3b2cf972e1ced0a46d3559da4eb1a1b32c50457293ede4958a13538a84b45316
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/proactive-node-scaling-operator@sha256:0ae01e9adc67ff7341fa3faf51a67c02206c41342374cab0bb2e6079d84fef4a
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        - quay.io/openshift-community-operators/oci-ccm-operator@sha256:eaf63c6f9ee9e9c3562d6ced4cff091c1e401daa826ee3a92b321a5cc8f13da2
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.connect.redhat.com/vastdataorg/csi-operator-bundle@sha256:925cc412650ec3a32ea4a55ffe65d16b76916d51f0b57ff62182b4ce12c5d2d8
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator-bundle@sha256:e0e7fc3a69575199a6fcf93985451dd66a5eb7d1ab8caecf059d7d52f3953dc4
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle-rhmp@sha256:d90f6bce87b783a80b1b81275cb61162c7c2a7898aa67ea85e3223deb1dfcb4b
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-operator-bundle@sha256:1a950c41ce2ae204d6bcba959b2d1a3df2d994bf1fefce4c7095c4ca20a5fd6d
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-iam-operator@sha256:d7e2507481b280118997ec9fe8062a9d56ed0867265732ecb5c89ef6fccd98fc
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator-metadata@sha256:7fb2750941193d64f8cca278ff7d81d3ca5e29a2ad0568dff7ddae4396548bb3
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:57e796822bcd705974c0efcd6c190ce39eac2e70b9d6a0c27f69d85dfe966db2
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:1e7781ff09c5cfb87ede98b96d31e58b760bedbf7a23f41778654faba3507591
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:f234e8af232f2d3a9472307ae82834f96756841b8d60b8c42ac4f9ca3fd17545
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-lambda-controller@sha256:9b2943aa770ebd5638927b19f05f629575e1f8fef129f8520e01c2238771d7dc
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-operator-bundle@sha256:bcc3521d20a1662d4032901f46e3bd4744c9910b060f52bdf10c6a8dfea1276b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/sap-commerce-operator@sha256:9b6a3c9c7a3af1a91578fa06c202d128fc7bfff3ae21300a4382cffb11def17a
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.connect.redhat.com/ogaye/ogaye@sha256:f35a21fb3936b259aabe27b453612616fca4e6327c9042361b4d00ef332905c0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - registry.connect.redhat.com/hpestorage/csi-driver-operator-bundle@sha256:39fedc5650ff2fef6ed1d47a5745851101b7dd8f6ce1f8b72ac3a78bf3c1fd6a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        - quay.io/community-operator-pipeline-prod/cass-operator-community@sha256:428effade29fcb3e8981e84f7df945f9bc49fe1a9b56217573db8b387706c26b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - quay.io/openshift-community-operators/seldon-operator@sha256:864e87ef21e81624010a1cbcee6b17c8c816dcb0a699c66badd32eec3baf39fd
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
          createdAt: "2025-09-23 17:13:29"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.16-3,hyperconverged-cluster-operator-test-rhel9:v4.18.16-1,kubevirt-common-instancetypes-test-rhel9:v4.18.16-1,kubevirt-tekton-tasks-test-rhel9:v4.18.16-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:86d21c945452fb380a895aa96630726fdba73b4e10549aa8ae353c7cac37fe44
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:aa24c4a937c9c1e52c30f229af5b26ce413b684375801d0f5f1ef2a50e8502e2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:6542fcf51604d3ba95ec0d42aa7bc19bdef940e1959808efc98f170a68a7d01c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:82dc19865ed64c129b5214b413f5f8da16d90da532c40d3e02b28f767d72f3c5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:a67782645abd4a869ca8db132953f5f4da72c853fe16f9521c77f241a0703bfa
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:134cb74301a9f9f704d35dea38ffa73aa3ee0ecf43b7fb3e73a89c9b432af9d4
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:6cbb8785d06e58af9b8fc1a66d0d5989a1b490b04d0191252cbb16ace14a3cdd
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:424ea484cd63823b4585e630ffc4053c32309e3f7dc24a212582afc60064e503
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:01082e3cdbac6fa50bec206401660dedd832a71f1f9e52c27281587f1f0122ba
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:a4b7120550d915cb74f03dc3530082f3a0d6850809062469f59934d8518479b9
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:0db7542ab7d67866046b1dde2f3ec8a04c47caa9c622ea9832dc59d127b4e919
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:fb9321f06c842dcb628f9082094fb697146e8740130c889bee81ba6941177303
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:cfba9b634c5a8871511ae6c1644e63560fa51d8b2729eb0db207b89ea9a1b3e9
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:0384141b5c99df83759f69c88895bc23f946da385d2c6db0fc75f80c0853a77c
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:2cffcd87d0a42748ffdbc63707751eb8212ce34ac33d9225690f04ce696178cb
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:42e5ad70d114edd15c6e4d13d48399f572ce71ead1f45a9d94de2262a77754db
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:3da5b2baa67e468fe72efcdb4caeeb4f3384365719d5edb5fff6656edcd74141
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:521b8f630c112e9d8eb4d5686ec7a6907fff8cf491628b99d6aa4a10253b3d78
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:350bc99751e2cac633b98c633b6c83e60bfd55dde4301282ff8eb61f85ea8d49
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:e21e94b2778c786d2b6aa0c107df267055b1f7d36c589a5900d313b3fe3404c6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8ab2c1ea24981ca684505c2f9e050cd938efbcbdedabd74897e0c6c0d5c5a433
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:68b318dd7fed09f60f6df215c9812b38791707b4e9af1de27acf91365766808f
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:c3a694c846fc85709e756a617d2581c06062d482b0bba88983192f1efbf8cb21
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:7e60c4e1fb9a0e10b039887747287cc06ca156f813d88c7421f3950eaf07bf31
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:710d24725c6aa3b8282d32a310d3e107b6a2e57708425d5e3acde9e82e33591f
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:deaf31fb3ac137556fedec55c44abb71d01c6927aaa21e5d23627ce139f12107
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:8c3f276434dfb7462c556f7be3f93a0c857a463e2fa2befcaef3a7c0859c7368
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:e51dfca6233c67b1b5372c682a2fc6e0fe54f2beef86e4c9451aabc80017a5fb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e3af0269b0a3225b2b5d43cce57a40aef6788830e8af8f4330bd36e4aaf8ec97
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:1c4bcdb3e2074b3f99c4dd0c57a264330c7989b8564efcc50bfe0e086d487f5c
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:27d1ea63ca9dc4fe073a28793f574da3a991a4c36e8da55fc2c7889cede187ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:9f8ab3e7e76ccad5a794de310ed0ba39e070e00647e42cf2b9a1bcdd8c377f3f
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:e0d239237d8052c36aff1509d65c00909b8563ce14341b7eb7fb79e9b84deace
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:af0b3d4305dcc5ccf00569deddf122aabf444bda0d3f0c2a5c4fad7e0c4117e2
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:071b2bf8f6470a6f87c3804eac3a10a947376469e42a55e4f19d3b2fc98c7e95
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:8339ad1cbefafbb91be533f1980ffb0e22114d52fd5c11c1b06cc55d9554c9f4
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:f514eb6d468fdbc6f5bc2b60cafdaff19ea55cb9a9c60f22b9db466f5b8f9000
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:c839c0eef642ac854e04129b007bdd529f9b380840e052fc306ecc0d62e4d055
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:d661d52528a89d87764e16b5d1dcb6ed328075477453e69c6cb9814e2a2d0ea9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:21c02925a871875a8eb70efda72f79d4c287da5cb9c0a4fc19de9ebfcb06e496
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:ae92990aa7ee3cf0bb98549ba78882c032afc8ed323fa83d1e8de24951b09d11
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:ecbdb919859e032ba4f2fe3af4dc79e87a5ad28545942d9d75c734aa814b7b23
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:e6507ebc8d71561772e573691b160f60ee6819b3f6404d5f0db123f0e4bc17e4
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:250cb53bbe119c5609a99ad32e97fc23572af9c08490e0f5212cf0977f555827
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:7eb05621f7cfe8162b08bff8a7411d6b05959affd7dc0b7729b4ccf1d90bb005
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:55996948c6d66abfafc630d7f6fe4b67201af0b61613a39e1747218393fb5e01
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:7f61b66615c9b22ac0e7e6ded50fb5a4f0a71ca228d34340977d82c8eb8322d4
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:09abeaf5bd5a7ac9f272759932e8442828efe6935dcd3b365eb42b7655c1de18
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:d32a745d70478945bb25f04ca85c17dbc1fefa0a791522e42591a6876f085029
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:db86fa7625d770a7eb8a4af977499f4569023992d33caac0caf8e3b71871f574
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:f5ebfeb2d0bcb0e86999d4efdd08d03c34c90daa2523b1382e02f9d4930ded67
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:e3f9e0c1b53a54354246d9c934ea48c038a547845d6a75c7bfac5e82f780ee8a
        version: 4.18.16
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.16
        version: 4.18.16
      - name: kubevirt-hyperconverged-operator.v4.18.15
        version: 4.18.15
      - name: kubevirt-hyperconverged-operator.v4.18.14
        version: 4.18.14
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.12
        version: 4.18.12
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.10
        version: 4.18.10
      - name: kubevirt-hyperconverged-operator.v4.18.9
        version: 4.18.9
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.7
        version: 4.18.7
      - name: kubevirt-hyperconverged-operator.v4.18.6
        version: 4.18.6
      - name: kubevirt-hyperconverged-operator.v4.18.5
        version: 4.18.5
      - name: kubevirt-hyperconverged-operator.v4.18.4
        version: 4.18.4
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:a32214d166ee692a4cae917fe97e0535a4a696155e1c878ae5335c77dadef0df
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.13
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
          createdAt: "2025-08-17 16:56:34"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.13-6,hyperconverged-cluster-operator-test-rhel9:v4.18.13-4,kubevirt-common-instancetypes-test-rhel9:v4.18.13-3,kubevirt-tekton-tasks-test-rhel9:v4.18.13-2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:a4d0418e49929e186e37f20f64c46f8eca20e314f6da4deb55f0b5407686c2e2
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:9fbf91cd986fe4064b99d6eb4bde7a79f58966763765a5001d9ae75ed22655f5
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:710dd1bcf4e171943d5260bd16121abd75058bbeb80bc8c237d24aadca99585c
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:396eec2f40ad14e9348b229498313e478903125dbb8c6f7c5f3dc3dad359d18f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:36a46082ca4ce5fd4784a9613d49c30585883bccf7433dc2ecd8de787eace937
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c73b90403d587fa6993e736fd84c08ed39381ac506502c3711da99dd4ea39b4b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:47d3263ddc6d80e844f8240502df81c0b71c2ed85a29a0d4afac7b5c98379091
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:6c39b35d0142e12757091831e2f9900c5f8ebcb1498e0eb75b62c86ee2e7238c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:8a69cfd1bcf5745779a019499632141ecdbc4b2d9a4858c66d28df11087359f3
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:b7771fb216f69af28f49c615fa050bde00d27067acdb505578d104ce1337febd
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:21ad0c18f57b964b86ced3498ac0c36a6daf8a60bdeece77a0b41b168dbd6586
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:877fe3539fe1a4d9f8cf25bfe2c1bb847ff9413693848ad539c26442472c1d3e
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:6ceaacb96b29fbd5a07dac9e8aeb038b76d53f6dc30188b88068755f35da68a9
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:4f806c32d077c6021033733674ed5701bd45086f9fdbfe90985813e2efc27ee9
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:1e73a912388b8b3de05e669128d30dd3081d5747932dc10287d613e19f59a49b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ca5969eb64a112ae5c12ae3b6752b6858dfb999d78d61daffd61db5e07a7cdb9
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:b0a4d68244205f839c0bb2c5cd4f4f305537ce57b42d24d0b46fa4fe2fe5cbba
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:960ab26612067d50aae973b63ffe482649c90ff8f42c2dde698ef548422460cd
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9600fd87e8331989db167870e83618bfb5f273e34e28c2006fa72b49aceb86e0
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:d90907bbc0d2810d592316f609bc8fd0f551de4c37ac02c1c03ccc014cdef056
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:c78c9ee196047ada7ecb614a42cbbfe3dec934534942ee5cacf34cf030e50378
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:f54b97303ba8579a2ef38d971035117efbab46d0eece22a759bc2fba98d11511
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:1d984a34589472f1c554bb4a53b6df470104affacd341eccd5fa1e334a7d8854
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:28d40c603ff685d922149ae30dc77ba1d9bc9af0ce2ab6d9038ff8272940ca08
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:c399069fc6a0d3b3e7093ab6cc2d540d05a04a68213a24a7a500a83d084246df
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:3c2706973b2bca2fbbf88d5af1de0b3c53ef1c7e5029b9a3ef42f961795ad65c
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:725286e0c9265a66e3f4b2a86fcfa7cbdda4e963863af9fb2af9e1226e10f34f
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:3e546a40d6a3e2cce2567750f93260b3db0978477e8ce34f63cd5d195c58c195
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:3765263eb3b193d860699d3e348346b762eabca51ecf31f8f97db69f3ca5e4bd
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:8985f049a49cae810ffb59dd4c94afca632808d3c8ba40809c5447a382b63bad
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:7095f50cb41030a9f49af2ad9a99d81407b5d8491c421c34934f788f02eb7528
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:f501a890dd87d37eb24cdd86d5889eadfad1e454fcc782e8ca2454150e8dc494
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:04b135c11a1e4518febbad5ee56e118ca711ad3454db19f1824cda0d4d497742
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ebc17f70a654181527a8a1d39e069076af82343cab55531fd7a86eaa85b12fd2
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:a9788ba6756715f286ede2fa04e460fd021b24f2a0df068c2aa626c89814a7b5
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:5794ed8b4af6cc37d278cd91f08f0418b89550269d69c9196772fcca1a4933e7
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:0c2c54d18dddf8ae7f2993dd42b3d07217f63266218c329119dda13e0c431fb6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:f15c753ca6687605d956d5200ec9104067c9bfe66e987dce7c32bc45ed79c57d
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:ee7ae60e4ba4b1d13b0aecb053e3ccd22b62d2277d28413f394d32e6a3ffe139
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:b5a6c60e663b81cef29e9622af8f0eb62838be99ff51f576ca1b275c6e5a204a
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:4c58b6e5356866deb394d55e2f96e4d83489d24c988e8016ccddd8738b895b7b
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:88d6db3caa017e4c2d8114af1abd28df4d0751b01f291d42b84586378e0c9484
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:0838f54f7ab4fbc04346ef67501fbca0b6c2097558bbd9c6fec2632ba0ce7bc8
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:9e46d9a0ca2e1678a02a52d14e2d8f0eb007b21ae105100106532cd12f317a41
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:a0821002f32c485c28d43dcc3afe5f447ecec4a567ef1b5f07b162f96245cbe6
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:fd17e015ad32d54f8464bc337c874b9131e0edca750adb4165161f0d382b8c3a
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:45c7a05260790a6ea78c809316dc58ed98b32cd94a884e82f24650051a27a163
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:2e6b6d7d04690a616eb210fe4b142372807bb9c59b6af6e7d6fa196f8dadff3e
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:8e69000f409c115b6af634c057106dbe894395708a694c3a4cb030b06e77407b
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:498282a9a9afceec8587eef445214f5db6ed6c09c2c3362049c4097a210e5aa5
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:11031e69287b4fafa77e2d54dc40d8c7de64e7ed8528abfca7a798b5aa8970b9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:faf31a311d34ea8bad50c83918d5ac1dc32277b9d8e356a866d166f97209e40d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:61f96cac775c4f636f0685900c419a6771fe3f3b44648d93a176b4ab5efa9092
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:70219973c00c12519a1e5abd9fa3061669de8e7273ea1dcf9f72d40ae683d244
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:f0eba5e54eced4a4478d395b098b3dca201b7b374f9e6b7f4a5c1669953c0991
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:be1b6c7e71d25c45b0f4bc2ac8507a1c215e30e39f51bc9f907436f6bcc96b4b
        version: 4.18.13
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      - name: kubevirt-hyperconverged-operator.v4.18.2
        version: 4.18.2
      - name: kubevirt-hyperconverged-operator.v4.18.1
        version: 4.18.1
      - name: kubevirt-hyperconverged-operator.v4.18.0
        version: 4.18.0
      - name: kubevirt-hyperconverged-operator.v4.17.4
        version: 4.17.4
      - name: kubevirt-hyperconverged-operator.v4.17.3
        version: 4.17.3
      - name: kubevirt-hyperconverged-operator.v4.17.2
        version: 4.17.2
      - name: kubevirt-hyperconverged-operator.v4.17.1
        version: 4.17.1
      - name: kubevirt-hyperconverged-operator.v4.17.0
        version: 4.17.0
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        - registry.connect.redhat.com/fortanix/ccm-node-agent-operator-bundle@sha256:12746020eefd383963e08a86e3542608aed1999b4e520435e366a307a6a9eebc
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - registry.connect.redhat.com/microsafe/operador-aims-mobile-bundle@sha256:1572f715c13d852c2a824b726912c4e5cc49aea43d0d91b6fd673f7198b0d038
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - quay.io/openshift-community-operators/ibm-security-verify-access-operator@sha256:ec3970c7a6a5ee3ac863dfe6f16234fe19cd7ea4a541fd0cbae9a8183d0763a2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.18.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-05-14T20:50:15Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.17.0 <10.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.18/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.18/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.18 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.18/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.18/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.18/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:f391499545a2981a3c3492766bbaae6a72a82db07d88251e85fda5d672a108e0
        - registry.redhat.io/openshift4-wincw/windows-machine-config-operator-bundle@sha256:4ebf4ace7dcaf0a935b83314b2a514baffca704b2f36cec5cd61dcf13e459946
        version: 10.18.1
      entries:
      - name: windows-machine-config-operator.v10.18.1
        version: 10.18.1
      - name: windows-machine-config-operator.v10.18.0
        version: 10.18.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:6e58f2d6a1619a06be05807ff5407924087b7b27611628f9baf999733b67645b
        - registry.redhat.io/web-terminal/web-terminal-operator-bundle@sha256:fb969c8b8ab53ab6639795086efad6870dd212f04c94934b75910598efcf3ca3
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:0b133afa920b5180a3c3abe3dd5c73a9cfc252a71346978f94adcb659d683404
        version: 1.13.0
      entries:
      - name: web-terminal.v1.13.0
        version: 1.13.0
      - name: web-terminal.v1.12.1
        version: 1.12.1
      - name: web-terminal.v1.12.1-0.1745393748.p
        version: 1.12.1+0.1745393748.p
      - name: web-terminal.v1.12.1-0.1743657317.p
        version: 1.12.1+0.1743657317.p
      - name: web-terminal.v1.12.1-0.1742484227.p
        version: 1.12.1+0.1742484227.p
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso-operators/openstack-operator-bundle@sha256:945271aec89541d6d8a749c41af2fff3fe24e41fd50e935e3944451c940bf940
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/customized-user-remediation@sha256:dc517e2349fe993e3417b858865fb73b98b7b163c0b4231d4d7be1f11992bb78
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pubsubplus-eventbroker-operator@sha256:2761ebb89dace09a1c1ccad92c28c877f3a6f75a674191ffc3788be7fa549926
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/ember-csi-community-operator@sha256:5dce9b3018cfb300cd651c7980353cedc6617383d5c525e6a4a39ffb8e105774
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator.2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v2",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "600m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "baseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c",
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "image": "registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733",
                    "resources": {
                      "limits": {
                        "cpu": "400m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "10Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "50Gi",
                    "analyticsvol": "50Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-02-15T09:21:28Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v2
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: ABot-Operator-v2.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-operator-bundle@sha256:35f90c08438e4e5aadd01de20bc82e59454c394c0e1d24ade6c56d6e02053bd4
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c
        - registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:e3b3c32d12223097a9d0eb466c3f49fbdab519a675123c7e69597b53ab4c79fe
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:50f1b4c65541b56b85f42bc9b3127bc7bc9faad1974067ce11481de7596e4c6e
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        version: 2.0.0
      entries:
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:4603bf04e53ce948bbf0ee5d239d5e265b40598044ef9e1f4539f4e466e59cb2
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:a178716e62ab8a752bcb9023ad9631af2dbc72357266cd13b43398e21f230828
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:c5288fea5b2ef8125f7b1ff7f7633eccd4e52d14ac9cfc5e473c2e317018ca3d
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:704925128f81d4ffadc4dcfaa69fc8011cf3d0f431d24222a8d7a03bad675331
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:8da482ee4c94b2546bf2b83345b1e7df6e70dcc63f68d8a2e839f10138951d2e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/openshift-community-operators/ibm-spectrum-scale-csi-operator@sha256:01d24dfa0d6b6146230459f8f7e74f2749d7fff3c4d75c40de386e014f9367d6
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        - quay.io/openshift-community-operators/openebs@sha256:ff179bdd6e63b8e2cb1ce3b96d5e03d20c1a2410296fed6fb331bf5bc2e6b14d
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:895fcf14b337398585f51e4b6e1f46a4a17c01a153ca889bd4374d691d89b9c8
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:dbf68e729484ba11aaafbcf1a590aca7f233de10e71b01d78c319dc5b5c6c278
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:e9824dd4ca39accfc674b88f7a2b76fba3006ff856e78402348c720f2aa8a6b7
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:767d55dc7343406637dd16f1b3a73bfaec353e0b058143fd68df2474271e3b76
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:43c90bbe3b20e95e1f352cdad097d63820a1282deedeb769352cb100da43ecc3
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:2f853ad4fece4510e847e96fe8085b53c56ee342e1f387677ff8cef345298e72
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:70c5237a251adb5f44a085d091ca6c24d5b313aa73f6545389a2b938740a9a20
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/socmmd@sha256:5adf9706daf49b35cfcf74b6fb07b056f87714b87047803b797a56f138c77486
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        - quay.io/openshift-community-operators/must-gather-operator@sha256:a71e5e25aaa92f03334776fee87316809bf2a46dfb34c497d63b228a194982f2
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: nbde-tang-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nbde-tang-server.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
          createdAt: "2025-04-24T18:06:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/nbde-tang-server
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tang Server allows definition of the NBDE Tang Server deployment
            displayName: Tang Server
            kind: TangServer
            name: nbde.openshift.io
            version: v1alpha1
        description: NBDE Tang Server operator provides Network-Bound Disk Encryption
          in OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NBDE
        - PBD
        - Installation
        - Network
        links:
        - name: NBDE Tang Server
          url: https://github.com/openshift/nbde-tang-server
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c15928ecc2f0ff370f567a2ca909b7a71e2becaff0af649275396e59232b7b8c
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-operator-bundle@sha256:79bf610b144cbdf4404a1153f1c395179ce1743274eec7102c862aae4f45890a
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
        version: 1.1.1
      entries:
      - name: nbde-tang-server.v1.1.1
        version: 1.1.1
      - name: nbde-tang-server.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: nbde-tang-server
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:6f0e07eb583201703f50ce9b3bf47e7aa63630074afcb588902caf05ff3ba2bf
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:35fae43ffc92770e13919f081195b35d8dd293f223fb9e577d75ae53f156305a
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:e135baa6b7eefcaf70780b6833acb689c5aef076b6280ff29c02949fbb33b06e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:6a447f455b8a3a3a8390787b36fe1d59a7d8bbf834f66a63fb0023f0557796e4
        version: 4.17.13-rhodf
      entries:
      - name: odf-prometheus-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-prometheus-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-prometheus-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-prometheus-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-prometheus-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-prometheus-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-prometheus-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-prometheus-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-prometheus-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-prometheus-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-prometheus-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-prometheus-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-prometheus-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-prometheus-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-prometheus-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:64c4cc3155e8ea6630bc7fecd392a41d71193cfb20f832b6334cd4795cf0735e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7916e81066baac027fb575a1edaab3c98a00505c73c4ac3c2530c111c11f0030
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:4352d0aecaac2efc3a15bf87925c08f7f85ec7baac9eebc29787f05af047514e
        version: 4.18.11-rhodf
      entries:
      - name: odf-prometheus-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-prometheus-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-prometheus-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-prometheus-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-prometheus-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-prometheus-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-prometheus-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-prometheus-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-prometheus-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-prometheus-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-prometheus-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-prometheus-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-operator-bundle@sha256:bd5bdc7085f3e9ce8494488f29b4d83a29796c47d536106cd6e14a7b67ab07ee
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/3scale-amp2/3scale-operator-bundle@sha256:934fd9123d1da778b9f8375854f2a61167f0efbc76371edfdc9962f7626e694a
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.11.8-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:9c2aba45971c924fd55b2aab50ee8e6861a9defd3808b38ff30316727af31f9a
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:4f807ee6e885e966853f70e1c40bbc179dc1f994b401ac868b7f2b921f9d17a0
        - registry.redhat.io/rhel8/mysql-80@sha256:7c05e61c0f4f5dad6aeda136e9a1af2f04f4de3c51004e997d4eacae3778da17
        - registry.redhat.io/3scale-mas/3scale-operator-bundle@sha256:a6c2eff926feb11da27eb9972cb86ee0ef9ca0b013afb6fe6af4f72439d557ef
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:419eb09d1b165363d5b90df01c1fa10f1f91f03649a8758620ffa49276685feb
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:a03a1ae7f01259192b8d7b4ecee6ddfcdc0cd9e2d24a8c2306accb93ee767135
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:7d07caf7b6c5a195ba4706e82f3d28c3793afab99b1a6d5a7020547c48534cdc
        - registry.redhat.io/3scale-mas/searchd-rhel7@sha256:1426e4bd4aa0eb28b479900ffc02621203886280d24997cd11818387ff30d9df
        - registry.redhat.io/3scale-mas/system-rhel7@sha256:b0fd85a4994502a3a2ae9ba6a7610c38af64b6cc897efac12feb4b115a484f77
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:22431acc2c991b6ee6b4de0219366dee6ea745122a49a0175d5fb58ed7d9205f
        version: 0.11.8-mas
      entries:
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator-bundle@sha256:4629bd129addb09478e48964a62acee88f5dd2746875801bd19d2d55cadd9da2
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        - registry.connect.redhat.com/loadcore/keysight-wap-test-core-bundle@sha256:16c938699ab0d71138693641ba7dcce2f4304da1d357a67a057c4effb4300401
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-cluster-operator@sha256:681e51da33e3cd9d6a4b617dca4338827237b70d1e8065322bc4135846039c56
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/cluster-aas-operator@sha256:aad6afc2f6f1f4e3b2c0d14b6c178d6f71ef318b0d3efc97faa5fdb736eaa721
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        - registry.redhat.io/openshift4/compliance-rhel8-operator-metadata@sha256:1c06868bce1f5d7145ef658bfcc284c44011aa24a17166aa124547b38f2259e4
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:8a2194d8d26dff03ec53ac0ab93c62c268e48675079a88db18bb14a977a391be
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:06e4abf4a94d44728b6bfa685b6ed027962b60fc2a0060a7256297857fd5f6fa
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:b1dc39979be1c347b123b2dd8070e4509179b7b499d4cf3fc7feb4552c6e2413
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:14fb2aa627ea7f38fba738cc4302cf1962660b10d3806c71161d8c7d64a8d060
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
          createdAt: "2025-08-26T06:53:32Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.1.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.1
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:498bcfc3c56bf642aaa70aec4db61b8610af0a22e9e378dfba02356770566198
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hitachi/hspc-operator-bundle@sha256:2d5e1ed3782a992034bec29d85d24befed31ad5eaf7a21ea44acdd775f90cb73
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: hspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        - registry.connect.redhat.com/opdev/operator-certification-operator-bundle@sha256:c390f32b40449fb09477e288fe5347319b691d3e89e020163e32007e6679c685
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/openshift-community-operators/tagger@sha256:a528169d305dbc2bc69794437665e4f940ea708dfd957c58d69bc79316b68612
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - registry.connect.redhat.com/amd/gpu-operator-v1@sha256:8a92a0d18a5ae6a837ca6590c2beb65d889cd744a098c22d3f01d2fffdc4dc57
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/community-operator-pipeline-prod/mongodb-atlas-kubernetes@sha256:f12108323fc990f869394e09b4ff6d9751d0a07f3c152fbd45b0f738746a468b
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
          createdAt: "2025-04-23T07:03:43Z"
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - kind: NodeFeatureGroup
            name: nodefeaturegroups.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.18/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-nfd-operator-bundle@sha256:ffb2d53dfdc63bfc9320bc4e927d037244111ab0cdbe5721ad69ce06792d9753
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:a1a60b53db81b32de0127dbb63f8c3a9ef7f70c95968aa53ffe8b49eced7dc43
        version: 4.18.0-202509100344
      entries:
      - name: nfd.4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: nfd.4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: nfd.4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: nfd.4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: nfd.4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: nfd.4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: nfd.4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: nfd.4.18.0-202505052133
        version: 4.18.0-202505052133
      - name: nfd.4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: nfd.4.18.0-202504232132
        version: 4.18.0-202504232132
      - name: nfd.4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: nfd.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: nfd.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: nfd.4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: nfd.4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: nfd.4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: nfd.4.18.0-202502101733
        version: 4.18.0-202502101733
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:f6de1136ced78ddf1ff48fdadc259c251919dd3a70819fe0e9247411999d51f3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
          createdAt: "2025-05-15 07:58:52"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.20.1'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:0885c82cd09105a078d4308b4ec1047b96f3cbf1fd555a1c90a8b053befee27e
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5783a00c386858d1eb45f224431d07d40bb9be746de965e7b129ee30969d4014
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5e822b12b5aadaa08a2f95b879be911455607a9c04922a641625515f955dc9ca
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:2217d9128fe50d61d1dff48b588113b5267002f72d025b1b6fbf87a1c2e670d1
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:90505e5ffaee7af330306c4e045d3755e0c74d30b9f45e1e7739c205d945872d
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:2e18b858f09e7595a89c4dc6ea5a85f8c5952f581b769093bfaab8156bab0813
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:6b4d34e7a40b21cf464e2f42cef21ef72a69ae8a3f4e7600a7bd85aa689117d9
        version: 0.20.1
      entries:
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:e533406d832461cd8889f5e95569ea9f8dbbd57b51dcd08b56050ba1e3775ad0
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:4530ed52522a74051aaedabd62b99d5c5274a0ab72eb28d4e75207d9e9d3391b
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:02e6baf8619cc2b782cbb1f3072e554a7638eaf3fd804abe1ba5f43f692e511b
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:df59ef548b82e9d72fecd06aa6b9c569f96485693c8a4ecb8ec278ae92df436f
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:e68814481ca33a422ed5c2b55b8a369e8b48092fbc71b138cb938a119d39e689
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:8ce258f36d3baab3175b29d5ed6df89588f128d249285bb1e7cb45539933c511
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/korrel8r@sha256:f00f8a7e2011bcc97d87fc7f5bb40a19d6b5737d1f5d7f3b2322ec39009c0618
        - quay.io/korrel8r/operator:0.1.8
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53resolver-controller@sha256:542229539932458d76180d23f1d5c5795e2de89169d8054f7b95be6f3700881b
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:3a96ab2db995d24a9ca064b5566821828d6e73ea20df77c9932d8972cb8e5780
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:c2b996e8dac2f4c81a3e48e4203499f48e4e15d0692e98d61dc8c97192e0a776
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:68eb666168d3a41cc98f91a0f2d74656af73cf76ec2cbe2cc308ae2ffab84875
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        - registry.redhat.io/rhosdt/opentelemetry-operator-bundle@sha256:1844430afea89707b70bba22cb9a2db98161cfc62b53b1086b63f109058028c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - registry.connect.redhat.com/citrix/admagent-operator-bundle@sha256:2a32d9a8cc3bd67ba7ebfe3f96b59e42009121d630918d417717d8ac9f2e40f4
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        - quay.io/openshift-community-operators/rhoas-operator@sha256:5299cd4c88012692b31d560ef981f80f994a0468552f43e03623adbf89b49ff6
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-networkfirewall-controller@sha256:96c35a47f58c48532fa3a5750e52a6d7bc2d530b54e807d57547f8780e9d7aea
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:20:34Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:85ba83a60fece2cc1c39ff245193712f918fbd55e1e283e940425fbb1b1f0e29
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: recipe.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: recipe.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: recipe.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: recipe.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: recipe.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: recipe.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: recipe.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: recipe.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: recipe.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: recipe.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: recipe.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: recipe.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: recipe.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: recipe.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: recipe.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:27:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:3c7033c68ff5e65533d2c3d0f2032cf34aecc153fd34f33ad3239331442d370e
        version: 4.18.11-rhodf
      entries:
      - name: recipe.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: recipe.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: recipe.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: recipe.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: recipe.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: recipe.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: recipe.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: recipe.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: recipe.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: recipe.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: recipe.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: recipe.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        - registry.connect.redhat.com/sonatype/nxiq-operator-bundle@sha256:48bccf9c8cc07127153af7974b949b8cd8429adf8714dcc866f026806e38fb08
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:9d24507aabd06856c7c35d73417b6ce065164f9c966053b494d68f14f0e260b7
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:00e6945419be8ac83aa7c888fe243c2fb67ee4ae5030cae117c111fdf80f5fe4
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-organizations-controller@sha256:76ea247239dde4e52e5016944a41908de8fdcc4173cad8dfff04413a0e3cdf48
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - quay.io/openshift-community-operators/universal-crossplane@sha256:b3104adb471083de96d49fa53f663dcf0e3e7ff5e3573bb1422aa6045277bea0
        - upbound/crossplane:v1.5.1-up.1
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
          createdAt: "2024-06-12"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:87c29bad5a920ea676a5aa8f6e582492547e391ea166694f84cbb75044718a19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:a29fc3bfed3add2768499e94435e8993bfd704a8a46268e6e63a8d20d2c4ca6c
        - registry.redhat.io/openshift4/ose-ptp-operator-bundle@sha256:7202697ea75c3503d4165b1d970802ac776c2977755a8b240094046f74c162c0
        version: 4.18.0-202509100344
      entries:
      - name: ptp-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ptp-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ptp-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ptp-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ptp-operator.v4.18.0-202506260833
        version: 4.18.0-202506260833
      - name: ptp-operator.v4.18.0-202506121732
        version: 4.18.0-202506121732
      - name: ptp-operator.v4.18.0-202505291136
        version: 4.18.0-202505291136
      - name: ptp-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ptp-operator.v4.18.0-202505071704
        version: 4.18.0-202505071704
      - name: ptp-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ptp-operator.v4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: ptp-operator.v4.18.0-202504170406
        version: 4.18.0-202504170406
      - name: ptp-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ptp-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ptp-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: ptp-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ptp-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ptp-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator-bundle@sha256:b473fba287414d3ccb09aaabc64f463af2c912c322ca2c41723020b216d98d14
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
          createdAt: 2025/06/23
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.0.0 <5.1.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kube-descheduler-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * LongLifecycle
          * CompactAndScale
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `mode` - Configures the descheduler to either evict pods or to simulate the eviction.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
          * `evictionLimits` - Restricts the number of evictions during each descheduling run.
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:4c653bee02c0135c8625fea6a01e18a046013ee330c5253ff66a7cd705ba08c1
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-operator-bundle@sha256:37afe091e3b9656c107196d7120ce171d8b0bf895264a9791c4559e97c81a00a
        version: 5.1.3
      entries:
      - name: clusterkubedescheduleroperator.v5.1.3
        version: 5.1.3
      - name: clusterkubedescheduleroperator.v5.1.2
        version: 5.1.2
      - name: clusterkubedescheduleroperator.v5.1.1
        version: 5.1.1
      - name: clusterkubedescheduleroperator.v5.1.0
        version: 5.1.0
      - name: clusterkubedescheduleroperator.v5.0.1
        version: 5.0.1
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:afbe40e3dd6894210741fa65a3f0bb6e4a3d00ca261eb6bc3adac6a22251a74e
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.connect.redhat.com/vfunction/vfunction-server-operator-bundle@sha256:f1784afe63b2e834a6b0bf6da254339ffe3597d02dc2d389f9162d806bedbe74
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator-bundle@sha256:b2bc3e161311f8f5d2f5fc967815259697ce30bf94e980214bc05aa0e1529d55
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/odf-node-recovery-operator@sha256:77b0ef6110a71da56216719ceed0f448257dec9633c98d7dfabc1038b23b6e29
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/hyperfoil-bundle@sha256:1d3625d12d95deb4f48660459fd0fc8b250b52b46d68136927d0f12465aaadf1
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        - quay.io/community-operator-pipeline-prod/wandb-operator@sha256:9bebaf0e72f0b3ce20dabf7ace62cb5ecd1015243438248e8101a238988cd60e
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:bf79fd57e1df791fcd45dd4745c923046dd7759ff5b3762cd7a329c5bf611316
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:b03f4e88ab1fed4744356a584d673fb2fdb15de7e1de7786bdd222b853690670
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:aa261c79b421bd49f7b46974a3bb5a790d5a1a73dd8dea4e1d8a4634fb68a51f
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:123488f0877c054a348f5b4135e68d77c17523bc4e5ea14dd73168372592f4ce
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:e66501276e3a10ae6d08dad3c4658fdb8b0b635e4b2fe72fa943b0e5300d7f2d
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle@sha256:47ad4efd37ad85bccf915a6d9448d404b7ce3164d7afb4c944648eeba9f0b04b
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-bundle@sha256:3e7ab287f64398b2aa434d9cfd5d99e2e9cf2d40c2e9fa1b1418537807b7aeea
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle@sha256:6c09391edf49d71b558f8ddbaf4e947dee17490202521539910d78b5ddcb3531
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        - quay.io/community-operator-pipeline-prod/ack-prometheusservice-controller@sha256:dd9c62f7493d43c6f28055e6f1aa7d586e25b2d6d359418782cf18de45da32e8
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/datatrucker-operator@sha256:4ab1d683d95d3eeba944b977594aff2e3b3e7203f56598b2fb2f7cff75354be3
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        - quay.io/community-operator-pipeline-prod/neuvector-community-operator@sha256:788f742c9f2def80d0b768eb96c5138f2cebdad767ea4ebb8cc26451771e23f5
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/community-operator-pipeline-prod/codeflare-operator@sha256:3dbd2b65d7841cc97c37db59ae8aff9991c4d0f0324c4391afd41c4de20c232a
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        - quay.io/community-operator-pipeline-prod/trustify-operator@sha256:49def1c9612621f90d2398cbe7385488c167cc8ccdcd951f098433eb42cb9ac5
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/susql-operator@sha256:4a47a8448610e54b8dc9700828c9250b84ee19126d6d4edbb62b35597e40bf37
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/grafana-operator@sha256:ae9aa1d05ab9dadcc9961ca4293d0878fd7656be79c869c7fdbee718f259ad86
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/orchestrator-operator-bundle@sha256:b82fd43d2dad2e525464177081869c923e239dcb09319458162f33a96370a6e6
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle-marketplace@sha256:b7f67aec4e911d89b3e6217e8718c349c20a859149b829a0390807c113e27196
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - registry.connect.redhat.com/kedify/kedify-agent-helm-operator@sha256:d176136a44e790298158ea3c4232e2aa7d0ad88ff18b6a5b39238162ed41226a
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        - quay.io/openshift-community-operators/tidb-operator@sha256:4f73b39de6129b906b5eb7bb0b7485f3da406cfad8616f81e375cc8ffcb156ed
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.829
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.829'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/aquasec-operator-bundle@sha256:59e62e5cbd479eed9ddd083db881469056c35253d76aef6bda9cbdb578e05b3f
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/console@sha256:05bd172efdb4729027ea6a491c874a1a3368156a506b7b5d228837b12282ee1d
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/gateway@sha256:cc8a3e950ae45fb40866b7369aa9393e92d1ab91dc0795ee682217fc421e9ec2
        - registry.connect.redhat.com/aquasec/enforcer@sha256:b48f026d7876656c0b365a90bcaaca7f4544f38b85400c110cf20e57c67ac8f6
        - registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:ce827db4a2b01b054e33f1091ee7046a14757b8781444bc9868783fe1efbf641
        - registry.connect.redhat.com/aquasec/scanner@sha256:d0ab5e0dad4e042324874de02ee36585a93b53ad1c3e905416af2fcb98f15e2f
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.829
      entries:
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:ad6b8e10abfff1cdac1ea35b44c456e3d59b4dd1372cd47c2e4221ed93e1f71f
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:167b9063a2dec7f9a5f1c5f3c02daf481c45c5e6a2f1b25353102e85921b6fd7
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:83c5e966addca5197f25c5f9c81a52648db44c2fd9e4969bf30b82c4880f0a40
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:bfbb9d0d5556eb067543f4e139a935000319b439f700240adaf393c68bd8244b
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:e0b1cff317bd4e35334429d1cc7c548dbda54065716aea96ef30610f0b91901b
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:eef6e7cf8f76ccd5e9ed16a5e4bdeb4398859121a6b0e86284b9766408cc2b30
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/openshift-community-operators/dbaas-operator@sha256:b2474b569406fb0f47a7042042429bbc5cd2edbbd119e701c6f440a139e1b3ac
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
          createdAt: "2025-09-09T14:18:23Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.18"}]'
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5189547a19ea81eb18142ad41e6412577eb9ac64d3987ba15b6665a1117a2e55
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:63d8b19d73e5cddcdce67c94e677666d3385ab8dfa1a534b9860fe1db33a2f91
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 4.17.13-rhodf
      entries:
      - name: odf-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
          createdAt: "2025-09-11T15:26:46Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:d45bf7fc74c58bc192a0c9d30157bd745f7d48b2a47820e3b768e8cf29b3c6a0
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:e76c59039b39b3caeabca73353a91501e8374e832c45eaf4005b1c08244e00b1
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odf-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:b4e4d262801b8ff616710075ff05689a8956e975e8ac563e50204a26d1725458
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:afc1e5bdade3ff0e3918fffb8fafb2700dfba7c473f99a2f347cfc7f889c8678
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: cluster-logging.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing, Observability
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
          createdAt: "2025-05-14T12:47:05Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.

              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:255a73d34b85f25b3655fb2b63954aa6b635b89080a8c249fc41568fd8365b10
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:2495de935f503808bf8700d9d66b8412d0f15fc754cf569d02abfbd74b5a6ee8
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:8c6decb25972dffb37917fb5f5a91ea05fb21b87823eb75807d8c601c12fc1bb
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
        version: 6.3.1
      entries:
      - name: cluster-logging.v6.3.1
        version: 6.3.1
      - name: cluster-logging.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - registry.connect.redhat.com/nginx/nginx-ingress-operator-bundle@sha256:4e032549f2c36c5a660994b5ec460ef214bb875a53f6f7e9b814d57049683a7a
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-bundle@sha256:31fd95c5232105d3354515102f099b9b662a8659f08da5d67d4e147d4f934fc7
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:17:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dd20cf6cd0b69404d2ddf5554e7f89870429c117740d2df51f3be5a8163b8db9
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:26c47d0498f8b8a61338cda3fffb10724c7e0f78d74d14537580763614130256
        version: 4.17.13-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-csi-addons-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-csi-addons-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-csi-addons-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-csi-addons-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-csi-addons-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-csi-addons-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-csi-addons-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-csi-addons-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-csi-addons-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-csi-addons-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-csi-addons-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-csi-addons-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-csi-addons-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-csi-addons-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:25:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationCronJob is the Schema for the encryptionkeyrotationcronjobs
              API
            displayName: EncryptionKeyRotation CronJob
            kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationJob is the Schema for the encryptionkeyrotationjobs
              API
            displayName: EncryptionKeyRotation Job
            kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFenceClass is the Schema for the networkfenceclasses
              API
            displayName: Network Fence Class
            kind: NetworkFenceClass
            name: networkfenceclasses.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationClass is the Schema for the volumegroupreplicationclasses
              API
            displayName: Volume Group Replication Class
            kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationContent is the Schema for the volumegroupreplicationcontents
              API
            displayName: Volume Group Replication Content
            kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dc3cbef1d890b1be5cd59d4e2fcb91f4f88e2aa53bf59c78a2cbe9af136ec3cd
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:c3e9d2deb1454fbd423ff9eeca34225df86f92b536c962300a75158505341678
        version: 4.18.11-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-csi-addons-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-csi-addons-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-csi-addons-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-csi-addons-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-csi-addons-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-csi-addons-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-csi-addons-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-csi-addons-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-csi-addons-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-csi-addons-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-csi-addons-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:a7b60f1c65a6fb4ee0109c7b2b0bd12fb9487e2f6edb059bd18cfd2337ea9116
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        - quay.io/community-operator-pipeline-prod/apicurio-api-controller@sha256:c783cd6db1ae5a6152848f5e4258e89b37077a177199be905d0cfe451a48e5ef
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        - quay.io/openshift-community-operators/fossul-operator@sha256:95c043ddc818f44a303e27508e0acdcbbcb3d9b1329712c85a8e3fd4cb647e9a
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        - quay.io/community-operator-pipeline-prod/victoriametrics-operator:0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        - registry.connect.redhat.com/flux/flux-operator@sha256:13c2f3b8092fc7b84548dd6383edbb8e41b7a8f3acefe63620f07f7a6454d8b9
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.connect.redhat.com/ibm/modelbuilder-bundle@sha256:4940d853ba477aeeec99550c35c132b69bab8bd7b6fbc1a2fe118c828c7ba088
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-elbv2-controller@sha256:f40f12efb734af5930c225925e99f31eb8db5fecece704a27ea50bd8661658c2
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - rasen708/cxl-operator:latest
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/ncn-community-operator@sha256:15edc1898fe9f8f56708da7d6de3548d06f1df3bdbb3f66a9cf4f7990fe9fad6
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - quay.io/observatorium/opa-openshift:latest
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - docker.io/grafana/loki-operator:0.8.0
        - docker.io/grafana/loki:3.4.2
        - quay.io/community-operator-pipeline-prod/loki-operator@sha256:b7e52fc9b9a988eed8b8338a2b2eba34b020cb24a189c2fa4db63110118dbbc1
        - quay.io/observatorium/api:latest
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
          createdAt: "2025-09-19T19:13:23Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sns-controller@sha256:f535d7861f6ef1488ccf95d0c5629a924033acb133d9452dc86e4c80352e69d2
        version: 1.2.1
      entries:
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
          createdAt: 2025/06/24
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-run-once-duration-override-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-operator-bundle@sha256:8ff7f1c707b40448a719137f37f6a3e09860812f7f07c3e61b7f73d1216399dc
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:42f73634df2c06216fac7b3742b2ac21ef5f12ee98d8694c06f7c9073773acd0
        version: 1.2.2
      entries:
      - name: runoncedurationoverrideoperator.v1.2.2
        version: 1.2.2
      - name: runoncedurationoverrideoperator.v1.2.1
        version: 1.2.1
      - name: runoncedurationoverrideoperator.v1.2.0
        version: 1.2.0
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-paygo-bundle-certified@sha256:e0a8c4d0de20c91b73957d6ea09fb94005e680a092c54389e6e411edbeb59296
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-access-operator@sha256:6e37f565564de371b5797895f3e96e0efd4f592d7d3e899ea8507438bdb60e78
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        - quay.io/community-operator-pipeline-prod/ack-ec2-controller@sha256:58d3249744b82a7f75ea2b1461adb03259350930da791da62841f05d91e94d01
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/egressip-ipam-operator@sha256:1e2443488e0aad361cccece51689e2b8da882862cb62c1208a14ab1b5ec2922d
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.connect.redhat.com/intel/sriov-fec-operator-bundle@sha256:c30849cbf451eb016786bc367e3716389b1d45676ad298c4de6a492c7bb00f4f
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:2338eaab863304250b3b1d267e9f2261c48eef1f97c280b7fa0af0ebc3ba208b
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:40802ba8bf531637f19adbbaeba3631e0b0557168ecf61f566fb5f2da25d1d4f
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        - registry.connect.redhat.com/loadcore/keysight-wap-db-bundle@sha256:3940b8cbf7579d9ce888f176f96c4d191a7cdfaec4be9251e8da4942e7897e97
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:2647a21dfb60364cfb7c4d1b81508adc13d859e97553bb2f710901e6ad59f4a3
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:dfaeb7ac2ccfed03042febea709f0ab66a4c57d62fa041b7ea887a0783d60a36
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:8d20352c0091985cf93e39cd18b2bb092bc15fc5e9900fe46ed5c6fc3c0af885
        version: 8.17.30
      entries:
      - name: kubeturbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - quay.io/openshift-community-operators/airflow-helm-operator@sha256:b3d05d2e2b4ff8ae64e4fc0338fcfa1ffdc345c6fdda8eaefed4e9108750fe58
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/kubernetes-imagepuller-operator@sha256:a72e3728dcf395667820c10c4c5ccb3f3d040ae5d04f50680e193398d3b72b88
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:5d4e4e62b03888bcbf6336934f2ddf5c5cb3d91bc8482a3aa1697e09625c6865
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-operator:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:df0939534c6aaa988714e64728ecdb387ff7dc4b32a5d97d56703ab143787171
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:ef65de2143df2c93aa73f099fa66d39e3239c30a9d547ed7bf258da0a19e8e35
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        - quay.io/openshift-community-operators/prometurbo@sha256:83444ab464bce4fb5021a2fca6d4d6ff72b40e90a6cfc9c5ac6c74357c775ad5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:032f053f21e273714e9f8a8f6854c83bc71cb84f5b32174704b490eca09eebc8
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:40b961a28ea9761c052a5cd17d2452fb72ea4786a92e13a0a1b589c7de31bf26
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:e4d902722d83f82372021d49c80f9880dfdb8bb312d4635e1d97980b60ca1fb6
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:961aa224525116fd575cff3a37a6465e9c652ee446927dd2bd03249b29469472
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:ad94f962d7aa064dc2b5b12ac2ea66ae5973909c82e6a96829165eaeff7ea952
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:504672b2db7fd25d1c85c9614f8242f98b4d2d04184a947dbaaada41ec205b45
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-codeartifact-controller@sha256:9bc95a0db7be356d2f657ae61c1c3098ff6e7e8bf1a675810dfb1b18805c6c8b
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:e80f87f289ea944a78b5c38df600c0a22fa470a2ce90f0b1e9453cd63586d75d
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:bddfd4ebd973089f7446cb16ed9668d418f91c3907a5ede22052be4ccbea6d0c
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:b31ebfab28ec75c0d7110d6f01e3b26fd973c36013308e12f69584ff38ab3779
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:7edf59729c2ae8b4506d5361eda01a98f33243987ac082abf758719c76a4804a
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:94bade4574c3679a425628f68067f9aead57bdd78e60f635041f421ed30341de
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:f5dda99364d07df3a3680a2d7031c27df071c8e80746b1ececb5487240e9bbb0
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:a4704a3977fccceb145e8972d2d7de074455bc228f2e925e5b427ffe3dcc623d
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:4f8f9dae947e9d34e86dc4237aec9167a9a25f83cecfae4982dc952494502486
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:6c0899610a78d2737c2096552e5146d9d5602b5568fab24b51b409dbffcf1961
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:84c6c826473b844293bd2048a498c5e0e8a558224a09f8e2e0cc85a2f10fcacf
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:c3913778f986ae4f3a7d94da2a9d0157233dbb676e3c1eecacbdbc02a864685a
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-s3control-controller@sha256:114ea0669459ae148ab32f5366f239da9cfde23cd6bd6a67c14e5844856837a4
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/openshift-community-operators/nsm-operator@sha256:330a930f885dacf65537beb600fe7f428e02687e83778e47dcda3dc2df15ae24
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:30375ef2cab46b967f6d2cd46862377c83e58bc6489eb0e75b495a173ebfd84f
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:466b9c30c02a3dfcd9cc94fe8eb432dca7076b8c79eb97ebe5802d716fd963f8
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:d5121af03b32a4b2286ff4f39499931a1ec04e05af45dccc6599dcc1b49e27ad
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:a0739ee7420b64e6195ec53057b032c356faf0ea4b8c1943725052f925eaeb19
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.connect.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:02d101a5929b60e612bf64a944e282ad9fd353c47d4ff28d2c9d5c0b84704ce5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        - registry.connect.redhat.com/coroot/coroot-operator@sha256:44aa4ea98d468f1d1ced671bd2e2bd5bf79ce467f8c572c2248abdce2b7c5acf
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acmpca-controller@sha256:c3929eb9a46af8feb2c830fd0b9225b6e3ad2559a9b5e04fda9239ef4cc07d9a
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        - quay.io/openshift-community-operators/machine-deletion-remediation@sha256:34d1e6cbfaf248a7b28c2a6e6a39f3fe46c183b9c38ac5fbb98eb3f421ba903c
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/machine-deletion-remediation@sha256:99b0b439d37ecdac8ec74433acf1a0ef787d0d6d1d7ef7cfe5f0a29f9dfe6071
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:90f512ce90433a33b50c58613d9b7136e00f116c48e179c4a1be89d6a35f9abc
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:53e73c53c31e0e06c990d3d966c2b4053d559a58a17a781155a3ceff3efac831
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:246712ce23cdd590ff83d84e9800b4ed761886ed5bab9f9cda4125a5d930e7e5
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:68e766235bdcd8f3335e2ab6e09942d98b183b3e498abb65acab6481664d46ed
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:91afe3cccd5eabbe4933459e63fd64a6bd87e399ea1aa81ae30dc3e8341c83e4
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:56db2fd54dbe982dde6a4558b664b19a676a8ac99bfbc8082d097cd6d71a03ea
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:d17af97462de060dc21f997dac973d9f895ce91f80821f1a810d93dbcbda31e3
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:783380361431859301477dfe63b6915240e934ffd07a2af2651094579be47a15
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:46f8c0ea6e383b742ab9cdaed1c8f624c2cf405bb1ce39c0ea6977940d52de07
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:e4ea95a86e77aa64da0b1800f00c2e0e598776cde7cfc7bb9941a60bac207ed9
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:aca3d91b7b9f8c1e8a1e0d045c42ba5f27444e0c1c0bc4a87c697b428c2fff78
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:5af7a32da67d39e7a011672de333792e63604cbfee997c19a495fc759d2fd739
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:74f0cf91e2feb60fe794f71b9f56fa8afc60e53379a0aef104701dd6692272e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:72724358b360ab74cc0857e2f18749d3d5b4c76bb6e930f31c5b04c7716e4c2d
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.connect.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:61f935c5985994994eb2f454fb12d74312f7d7d1627fc2b55f3d6ec9bdeb4bc7
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - registry.connect.redhat.com/mongodb/mongodb-kubernetes-bundle@sha256:354b16ea43cfcb9bff704b6265fb10d94037beb231138e1fb44963f3b348bb40
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sqs-controller@sha256:a22009a5d2037aacf52d223e45b74e6aa08020d4cfd1836ff137254b73e19bb9
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/namespace-configuration-operator@sha256:971991d2bf717001a98b175c50580cd7db9d230492e72670606c2ed50c53feb0
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ecr-controller@sha256:03ca7f898cfbf14f18a06cf3b510cddade3863275df934be891734b8c177c1e6
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        - quay.io/openshift-community-operators/eventing-kogito@sha256:abe8d5803d2d96ebbb9fe4cda0feaa0217e92def6a669404362bcf741d934690
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.21.0
          createdAt: "2025-09-15T11:39:08Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        - quay.io/community-operator-pipeline-prod/kepler-operator@sha256:37fe88e7234e5eb07c493a6412f27bb8df33896f7eab0a94b0e396fcdf870715
        - quay.io/sustainable_computing_io/kepler-operator:0.21.0
        - quay.io/sustainable_computing_io/kepler:v0.11.0
        version: 0.21.0
      entries:
      - name: kepler-operator.v0.21.0
        version: 0.21.0
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - quay.io/openshift-community-operators/eclipse-amlen-operator@sha256:bb89e2a7cb1c9172b080c72ad3227017d135fc30076d2a68b4e215fa8fa55846
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/eclipse-amlen-operator@sha256:5fbfb0bd4f764643cdc1021e99e489514a1f035d85131b13e56b11a3045883c2
        - controller:latest
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/microcks@sha256:1313bb08037e57cde02d9d48c4d8c9663eed08651a3c29f872b5ef85ac764112
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-community-operators/openshift-nfd-operator@sha256:880517267f12e0ca4dd9621aa196c901eb1f754e5ec990a1459d0869a8c17451
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:e5e0c3b3a25c13ea916d70fa6f6ddc20f2c730bd9909aa2654672280c27afda2
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:8c65e65bbaf0ebb10dba03655c5854c69c82ddc0e6b2a29293eee029222e5efa
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
          createdAt: "2025-03-03 15:18:52"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virtio-container-disk@sha256:3cff6989c7d7f72d2f0ac27e512d99873768766ba420d65a6b7794676d0c1e88
        - quay.io/kubevirt/virt-operator@sha256:d5fce85dd4232745408daebfbf70c2bc879a06f8340d1f7448b141d5ba3f1821
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-uploadserver@sha256:2b59d3c0a7c154fb6915cf3439b42eb4aaf48f94025ddb75fa9b73b16148fc13
        - quay.io/kubevirt/ssp-operator@sha256:61ac0c182528a9b54d84755fc0931ad8d408ae74e7e4b644316a5e9e630ab360
        - quay.io/kubevirt/libguestfs-tools@sha256:d655b9abbe8cce45bce3934e11d01715e170ddda4da78d7c51f356e7ee1e3263
        - quay.io/kubevirt/bridge-marker@sha256:e492ca4a6d1234781928aedefb096941d95babee4116baaba4d2a3834813826a
        - quay.io/kubevirt/virt-exportserver@sha256:ec1693cc6e8a14ab83f2feea87427643a968d65832c44b25a1bb911482403ff2
        - ghcr.io/k8snetworkplumbingwg/ovs-cni-plugin@sha256:516791acf430bc747f01366e2b748ac76c0b5e39ad8592be2b3dcc809429fa1e
        - ghcr.io/kubevirt/kubesecondarydns@sha256:8273cdbc438e06864eaa8e47947bea18fa5118a97cdaddc41b5dfa6e13474c79
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:66c55f5ffa4b44f2e58990baa7c672ec9776612a5f1c57fee5a5ab1c080f17bb
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:902af624e05eca764c17137c27813b1618f388e4b1a53b27b54665e8735251a2
        - quay.io/kubevirt/aaq-operator@sha256:5e53b755adaf6c66f59812b590a2eae2d84554fa658e93621ac977316843d918
        - quay.io/kubevirt/hostpath-csi-driver@sha256:f8d44aabe0024058b4380e1af42a57a90b7f8c66c8627091bf425fc384c0880f
        - quay.io/kubevirt/virt-exportproxy@sha256:55e38274e59bfe5a2cb5671534b8b6e75d90e43107abecc95fd463f7be8f0026
        - quay.io/kubevirt/cdi-apiserver@sha256:1849bc14687449d36561eb1c768fdbf0cd14ababf9c254b33a4ddb14abfba01f
        - registry.k8s.io/sig-storage/livenessprobe@sha256:33692aed26aaf105b4d6e66280cceca9e0463f500c81b5d8c955428a75438f32
        - quay.io/kubevirt/network-passt-binding@sha256:c794cc1b9173267f1e3d34e666d9e2f782b88d93b47fb4449675346b467f802c
        - quay.io/kubevirt/virt-controller@sha256:68a4967ef5925d4c24fb5eab0aa5acf0a82946beb7f1139bd45c85efe3653cec
        - quay.io/kubevirt/cdi-controller@sha256:94d16a6bd0d9776eddedf64165f39e63e4dfe4f8a26e858bd099db5d09136047
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:92f2ac10f28a7e80699bd984dbc4de897011b1f68c2be95f3fcb90df410b119d
        - quay.io/kubevirt/virt-artifacts-server@sha256:789f68f4e27cf947cdd4f1c0732042176e67b7faa5e37e1b503b8dace3a46125
        - quay.io/kubevirt/hostpath-provisioner@sha256:156ae46a03e2ff5550967773100122f9abc82e25bd7acaa55dbdec22238dd2ca
        - quay.io/kubevirt/sidecar-shim@sha256:cb4025f7275f8de2891c2195c5faf128ac1b68595cdb3eff776a2e5360b8f034
        - quay.io/kubevirt/cdi-importer@sha256:f9fff83c8307bca61a22e8296cd103b32e63d7c8b2a170c6781c6cd962211d26
        - quay.io/kubevirt/pr-helper@sha256:fd434b4ea1ca5682fc5b05f23de01130c0c10f77a0a2d9d089d8f3cb5b545aa6
        - quay.io/kubevirt/kubevirt-template-validator@sha256:b227679f8ccc7c53469574ee5f58c9cc39eb6c54ec01377e4b3df3205a2f4ab4
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:89ddc40700d52cda23104978498fb4d74a643e807450b630beaa71ffdfa1fc57
        - ghcr.io/kubevirt/ipam-controller@sha256:fd7c38e129ea4ed8a86d1801c9e1ca387802376fc4abc0b0d228d4278f448ecd
        - quay.io/kubevirt/aaq-controller@sha256:b9fd0818e4019cccfb38f0923db058166eafdaea8bb35a720aa22b5ce55daf26
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:16030e8320088cf74dc5fbc7dccfea40169f09722dfad15765fa28bceb8de439
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:ac955c4ad03c40d74341623b22f99e8f796a6ed2ef5cc539fba3c1504d159955
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:8061bd1276ff022fe52a0b07bc6fa8d27e5f6f20cf3bf764e76d347d2e3c929b
        - quay.io/kubevirt/cdi-uploadproxy@sha256:6a98ef32c3778e46812c698970e14736f178fd2a11b1f74ccc36c34ce2e34466
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/kubevirt/passt-binding-cni@sha256:f3fea2ec66a26850489197edf7c85d074eb78bbac52916b10016ac0de8ef7c14
        - quay.io/kubevirt/cdi-cloner@sha256:feefb3933780bf352ba03b62b900907d02c64e73f7e74b6d336f89daf2752f62
        - quay.io/kubevirt/cdi-operator@sha256:8297771144f2f358d893771a9dde208a6bedf6bfd67ab2927b16c27424ca3f26
        - quay.io/kubevirt/kubemacpool@sha256:6cc8ca860960c003b36754c6dde249f26f78d32eb22159a25b5647e197dbc9f9
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/aaq-server@sha256:18c23eb4e32ed083047c8808b35fa81a22294a7e8bbabc74e2612fd2eba97dd7
        - quay.io/kubevirt/virt-launcher@sha256:18508d8767f8530df4d6aca4ffe467c40f9891fbe052548b3b2316f8bfb1bc38
        - quay.io/kubevirt/virt-api@sha256:d92c78b3048756fd11d3c6f1a48f4ce8555dd39e866039313dc23bfa70668443
        - quay.io/kubevirt/virt-handler@sha256:4ddd39539282e79cf3602700b791eb91a6d79991bf2e4dc4cba24e3da9db3d1c
        version: 1.14.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.14.0
        version: 1.14.0
      - name: kubevirt-hyperconverged-operator.v1.13.0
        version: 1.13.0
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        - quay.io/community-operator-pipeline-prod/aqua@sha256:ca2b998fb0fa4e80d6f4c3be57041e8e55e43cf0b5f1fee579733fda9ea6a107
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ses-controller@sha256:0cdbea8336c2825bf1bc8dbd3f5235ff41d97fbed30106e7775c6e993f92d8b6
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        - registry.redhat.io/ansible-automation-platform/cloud-addons-operator-bundle@sha256:2cb3b1887e8eeae6708f5f46c9bfb304d7c3df1537ea433b0b23772acacf9f90
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
          createdAt: 2025/05/14
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.3.0 <1.4.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-operator-bundle@sha256:3804fcb77893d9198527b4801e2a778bf612596c9a0b125a0f0f872b23138434
        version: 1.4.1
      entries:
      - name: secondaryscheduleroperator.v1.4.1
        version: 1.4.1
      - name: secondaryscheduleroperator.v1.4.0
        version: 1.4.0
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.3
        version: 1.1.3
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      - name: secondaryscheduleroperator.v1.1.1
        version: 1.1.1
      - name: secondaryscheduleroperator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-operator-bundle@sha256:44d82d5f8434fb6e88d2da406e87df0bf4274661d9d77554b80e09d19fce03de
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - cloudoperators/ibmcloud-operator:1.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/ibmcloud-operator@sha256:b22e89daf6b7f0fd55ba6db7eb7d7a48cbaef2844dfe20e558d7b821482e9480
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudfront-controller@sha256:235ad7b4ad07644d9a638909c681620b24a85148a8655b7c903faf6aa42c2242
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator-bundle@sha256:0192000a2994a62a904ce4aabfc1ff04caf2fa9d779d27c016922b9121877fac
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.connect.redhat.com/synaxg/synaxg-operator@sha256:ff3d9ab5506c4487e5b6697ffe960584aadfe9f16408e7976d345c3d93375171
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        - registry.connect.redhat.com/anchore/engine-operator-bundle@sha256:fbbe7e6c1d75c4de2f47e2c825c930568e85f1134545e9d890a0c9f3d9187a4d
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/community-operator-pipeline-prod/windup-operator@sha256:4e2bafba54eb01b0ae77555d2ff3da2c34286c7c173756dd20a90de28be36535
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/layer7-operator:v1.2.1
        - quay.io/community-operator-pipeline-prod/layer7-operator:1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/mongodb-operator@sha256:1c06d3e936348ceeed773eb45144c9dd702e50ba293610421a93e3b3942ac9ef
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        - registry.connect.redhat.com/raftt/operator-bundle@sha256:7108686c168c35acd996e61a981bdd885b454837cf6823a48a28cdeeebbb0af9
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:d0bd55bad99bfd1ce09da5ee885a38229829570ebf80e1acc4341dff17a54e64
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:b9ffc18b2d88dfab92583af35b27e0ae36bceac7778cb448ab2ea2eae2ab3e94
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/community-operator-pipeline-prod/nfs-provisioner-operator@sha256:d7ef0fa3b5a1a6094b6fd60c6197a52cb14766865b3549806de100bb3522821a
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-kafka-controller@sha256:1a32352f30d0835299e14e8d0f0bac5585396dd6a10a63fe633e3d81c4d71b50
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        - quay.io/community-operator-pipeline-prod/postgresql@sha256:f9b7137598dcaf0774de71f4633c7e3b0359b0b40f313ad9c91dc8eca58592eb
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle-rhmp@sha256:875d81ba1b91a72d9bb4f1d86e2285992370a04ace3831b3c914127ab55a57e7
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:a657269d0212d952f5107514db63c3ea2b6c86ffed0a44f0f33eca96a242cf35
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:1ec1264dcc4ee289ada0747083ad9b072bad2e33544a3a879383694c2974a885
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:60b1d25ebb3aada94985e359bb059d4491bf0148ac75d51b68a461670659e429
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:3128cd81b56975643c1c88c6ceff6d72e918a67ec2d6a5a606b6e04753c5b0ce
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:808529c14a8873324096fbd1622fb895c4dc7f60c1cf9e3ded646f9a81d7b496
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:708291528f26c70c561745967a1200edcb2f984687f749a49a2860fee580423f
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:d1c636cc17155c4764d331e050f9bf3ec9f39f6ec6ec14e972179b185ea9f5b8
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:5339c41013727fa5dd3ae5bc1d51eb90e02e5baa85d4039b1efdde30baeae1ae
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:1532fc1351ca44f863feaaf54ba8810be7a62396f98dd292ccaf975fde31ce5b
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:fa72cc3e0b09510d71a097b92ee101e057f1c11e1cb363172bd26b2d94272417
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:b897e25e03133513f9afdb5abe38cbf0f2ac0556c27518d9d80e792eb07ecf28
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-sfn-controller@sha256:01036536877e5b236dc83774eb3b6fe9493eaa593cf5d1cb5b103798d0727bf8
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:20:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:36393d8945938325e4f330eee801ba28ab9b7604f58ea927004e17bb514f328f
        version: 4.17.13-rhodf
      entries:
      - name: odr-hub-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-hub-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-hub-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-hub-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-hub-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-hub-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-hub-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-hub-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-hub-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-hub-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-hub-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-hub-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-hub-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-hub-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-hub-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:bb352ffcfe7e79981f96fc5b5a4f2f557fc7fe39de27c36f195349bf4ad6248b
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-hub-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-hub-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-hub-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-hub-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-hub-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-hub-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-hub-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-hub-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-hub-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-hub-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-hub-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-hub-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - registry.connect.redhat.com/kong/gateway-operator-bundle@sha256:91b3d73bc75eb170035110e5c63747e4918f738841aa97b6df27d22e37b4879b
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:d7bd0e90aff6e8fc9fbc4119e5f1ec89eba93487b84356e3af95208c1250fe44
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:fc374491270b9566b0a5c80f3dd665b2227b1e8fb224eada7e5c98c82926ae08
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:411a47151609b7da29974ef3ae3b04f1bb1526f4328fd3e1d9b0165380018049
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:076c345de449353cc9837c603024ba9d790d26c5a8cebfc97db61ded1482ab78
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:f481b2e364b9fe417a1cbbe9eff70180a3b4153a14712a8935527e25f919ee0d
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:a85d4f1c8a686abc0b9032dddda75991216d44dfade214625723d7009be314d3
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:b71367012d1ed17d3967cd77fc28b38c69c76f75848fb20021dbb176dc382eed
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:32b442eb6344d3ab64d2885f53817e61f0e8b72b5c6e28845c8e056fc176b666
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:268aa90a9f6eebf6cdeb32dc47335e75e7483dcce4ea3331b1b736c2fd79bea4
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecrpublic-controller@sha256:6acdf23661d603339e998307e9d5130b21c45665f4b0c35019ea80aac3d7343e
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
          createdAt: "2024-09-09T20:32:09Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:6e80289c93ce0ef26c776beca59eec792832dc7f2f57cb76a7321eabc12405c8
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:8b8d39f64cef8a1ed37a9ee44ab3bd5b98575ff32fd3962f94cf7a9c79e623ae
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:6173ab8b48c178ea8bad0d6b98e73801aa28135fcb7455146449a66543891e26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:63ef9626a82b5d591abaff2e9fe76f52e2bac16b910da40af5e92920694a2a67
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:c88a193654c191820bee307d12ae5e721ef82740ec912de9f710481b386edde5
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:c3cbe6cf4158aec38c67582fb79358315d196de76e0da611b4b8d6183babeb58
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:33cefb8803cd212932fe2bbb6052e53562d04084403f0166c8e077957b92c58d
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:ad206c2f78dbcdec43755eca3cb92229ee77e737e741dc276efec23f6379ecd9
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:e90e8df1726698a79ec689358d323eea7a26a72290660db62443e3f0cd696a4e
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:66dfedd62d474aaed03b1fee142b91358271591b651a3766d8e0542d08da914f
        version: 1.0.17
      entries:
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:7317ed83242340597b5cdde8ee314606ddd563d0067ee85c12eb21037e2feb99
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:16b39335e97bcc7c970a8ba385c61d8333e9912faaee813ccf8520d434b7bb7a
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:51ac9aced282f62c8fb2d3696da3696b8496d8f877e9846fe3a4a743ccd63bea
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:a82288782b61ea46d068cc09ee10608d3201c136abfc767e4d37734775101390
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:828d31bf8a441091b8aed03da213a6b4ffae260070eff58c3319156f9bf04f5a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:f938cc534fc4b24bb9449f1e1e1e904dc588e6646c98474de4b8c81d083b4d96
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:095629572a40469eb651c9f81d6c10ff46254eff8b9813cadce54d6f0075c2fb
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:91456907c61952017daf07e376b565472ef86d67a2925f60e189c639ceb68595
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        - registry.connect.redhat.com/loadcore/keysight-wap-core-bundle@sha256:2f8295f8f56ef2c56e8215d4125cf08facd2f1d07fad0ef3a01eb0a2ef5f0555
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.connect.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:623a93173e8f703849238806b40aac843cb2ed016c8763a4f8ea1a0b0990eeea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/special-resource-operator@sha256:df87c400713311f2495d6d02769acf29e5251330194f992c11719063095fdaea
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.5.2
          createdAt: "2023-10-07T15:15:08Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.5.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/openshift-community-operators/kube-green@sha256:e0c9f5ecdef1f7172af098f20cce17afebbedf02c0e0bd516d4d6aae736d1b27
        version: 0.5.2
      entries:
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
          createdAt: "2025-09-19T19:13:34Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-iam-controller@sha256:2d904b6462679d578505409d9bc585795983df7a4f0dd6668253c724aaa5063b
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
        version: 1.5.1
      entries:
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        - quay.io/openshift-community-operators/edp-keycloak-operator@sha256:82ff141bb86599f12d4a516fd1b6c5252f9ff24ef2547be6c58e2f3d88219914
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        - quay.io/community-operator-pipeline-prod/edp-keycloak-operator@sha256:10151387a79d0fccb47114b56525f28f507faee96e4fe29babd2ec46c3dd9343
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/ibm-bamoe/bamoe-operator-bundle@sha256:54df7eba0292459c0491ba39c15bfcf793e1c0aef042a4b3f6c587b0d45f77af
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/cincinnati-operator-bundle@sha256:50e9c88ca31182c54f9c4bc6682642a5115fa280044e6c125f99626c69d999ad
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-03-25T14:27:42Z"
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.0 <4.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-operator-bundle@sha256:a372f2b3b61c0012618837e349d389f9f300ccd2ae0fa005e9cd7d379b8569a1
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:da231dc87dad63f4fe28e3b438a1ca9a96f04839389f7c3c19e8368bf4604283
        version: 4.18.1
      entries:
      - name: numaresources-operator.v4.18.1
        version: 4.18.1
      - name: numaresources-operator.v4.18.0
        version: 4.18.0
      name: "4.18"
    defaultChannel: "4.18"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:8ad2282afef801c794e316c93a12834d688cfa44766132231f93013c3124944f
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:d4ec9966370c3e770209042e437c53b0780401150f5385069b8476e26c343919
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle@sha256:10c0eed1e8ca3154f4c815241f6275f40113ae2ba0abdc940ad9a04682edafef
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-07-07T18:31:48Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:abe7b5029a0956943b5af20cd66192a6fa816998523c05b671dc1280a0219683
        - registry.connect.redhat.com/minio/volume-manager-operator@sha256:e3a69a9ef99d98d1cd8eb43d3acca48a6806ccae5b6747fff0ce79712a7d363d
        version: 5.0.0
      entries:
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:43bf1b9f06a7ace06e0b87921ba9b62f430ee3c8b2357c713054151e01f9cd02
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:5ebd9fb9277844967f55b985fc44069c1c0209360735c5321eea3aaef8628528
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:e668de3c5a2eed2b2b89515860f735be8ac03c2646ab7b65ae3c1270262bc1ed
        version: 8.17.30
      entries:
      - name: prometurbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchserverless-controller@sha256:e31d29d7c3389932e42c82645fd7b697aa16e03e20b17be0f355ef6b301d5ab8
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle-rhmp@sha256:d54904777c861e00858ddd8f1218cc39db47fbfb1225b8d35902ed4e13e357b1
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - quay.io/openshift-community-operators/kuadrant-operator@sha256:398c9230878feb756ad393d7de5675a7370c883a18acf73483a660e9303d807c
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        - quay.io/community-operator-pipeline-prod/kuadrant-operator@sha256:f21456392af3b37ae48d77dc7f19fa23c6946269555bb1280978b29420bbf31e
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator@sha256:6b7a1e34c33b16c464745f70658a10d6ea3d96099c1e2ab008e96589ef73e7eb
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        - quay.io/community-operator-pipeline-prod/shipwright-operator@sha256:f4d3b97812c69b396c8e6d081d8fa9b412ebe1365432658288e6ab065c220fca
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/istio-workspace-operator@sha256:7a581f048aee64701b23f1a0e94d77f2e6932ee8229fd46b1635d8df6994e174
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        - quay.io/community-operator-pipeline-prod/kubecost@sha256:ff1e00f5a04e49fe9daf06db6b60c2ca14692efd098a38ecd622ef6bc5de8dd5
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        - registry.connect.redhat.com/loadcore/keysight-wap-ui-bundle@sha256:7ca0365e849fc98950021f86f3acb9048303fad3dcb6eac0ee09dd4bcdaf8fe0
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator-bundle@sha256:3715abae1ffeea2d0bebf665b1dfaddce4686bb8a61f0697e6fd201b157798b6
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:fb7bc294096fa78f199d909f6ff8f7fcbcd771c3dc19f3f767d77dd13a2081a9
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:10e4bdca44a1a0cdef27b415765238249f81ae6ba53f94f992e37d350bc9f800
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:3e8aa02c71fc99e5eeabfce39c8fc24a9f9e8e826a58f9ff59c8b7c0e9c14e0d
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - quay.io/openshift-community-operators/datadog-operator@sha256:b1977170e5991e1f08c49037d6f34873c8543d29d3a2c3692ebb1d66734f5b3e
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        - quay.io/community-operator-pipeline-prod/datadog-operator@sha256:0a5684e5bb442c5daad6737f1560198ca380507f2da583efd0a7d66e18f2dade
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:4e12e48d0b2b884de2a0d32682da044cff419aea53bb162915538e2179fcd7f3
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:9eca42cc297a93293200a1c46e8ed949a61770550366044ef54e36bed9d7a116
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:31eda4f1888b180feb023c42a6ec93478b2e956f13420d55987d9abf7c7797c3
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:7d5f9e02f3d6c3e2ac63f1fd33240dc696ee79563c214d21d2897ee773517e9c
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:5b061126dc3adf1fc9c8c22dfded18f4652bd62537aefca88ddcc585078b94ae
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:007a3d43633fe355202002737f766a7c636f537f3daeffa1f62356c0bbf969aa
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:87594d78977ead074e53b39a52e905a2d2861d24d69d2f2e33b066077cd279a4
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:ff8e563d97f0d0e42c690ca3e61d3625e920947264ea99c396392c440cd43201
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:12a02131f13e532a26572a524ae542766092ff75624113994d304824951161b4
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-operator-bundle@sha256:006edc597eec839f960bfd191b75217fc8c6bfaea9d1694e7ac9fa34266f9b4f
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/service-telemetry-operator-bundle@sha256:c4ec01d0d27eaf93cf820d8da1f2ad7ef3e557c5b4ed651c5a1d07a1fc116f79
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/rhosdt/jaeger-operator-bundle@sha256:f250e39033d7cb1d786e5a7ec6798c25d4c5d8c6ecbcf6828915605fc4658da5
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/annotationlab@sha256:b69fde0eb341114893aba097d58455c9e316a97cb829b41c0d080987490a83d3
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:2dded4ad8e80f711fc18ada34ea825e844045da44f0813550a586371dac84b52
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:a20986fafd0a475a5a77296b3f8a18c78edb84da5cc8246b06c805dd5f693511
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        - quay.io/openshift-community-operators/hawtio-operator@sha256:38b36745547a08afd10e1620f8243e6f1ce18a3eb8883ec78cd631577177c5ee
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: helloworld-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: helloworld-operator.v4.18.0-202506160933
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.example.com/v1alpha1",
                "kind": "HelloWorld",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "helloworld-operator"
                  },
                  "name": "helloworld-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-05-18T20:30:40Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202506160933'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HelloWorld is the Schema for the helloworlds API
            displayName: Hello World
            kind: HelloWorld
            name: helloworlds.apps.example.com
            version: v1alpha1
        description: helloworld-operator
        displayName: helloworld-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - helloworld-operator
        links:
        - name: Helloworld Operator
          url: https://helloworld-operator.domain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/helloworld-operator-bundle@sha256:056744dd517acc3c12908e5ce06bfffe2c634d427b837e244dd7185c70915c00
        - registry.redhat.io/openshift4/helloworld-operator-rhel9@sha256:7c9a290bef3368d01e44bbecafdd99d2e18376221557938ab82c483bf189f370
        version: 4.18.0-202506160933
      entries:
      - name: helloworld-operator.v4.18.0-202506160933
        version: 4.18.0-202506160933
      - name: helloworld-operator.v4.18.0-202506090359
        version: 4.18.0-202506090359
      - name: helloworld-operator.v4.18.0-202506050817
        version: 4.18.0-202506050817
      name: stable
    defaultChannel: stable
    packageName: helloworld-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:13caed923cfbf9d5ff215c8f87fc64ecd753d183f5573c31ea89c9bc5915c6e1
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:c47f099ae5ddd9b25fb7aa5ecfc253fdf24df2e7929318bbcf162583df91f886
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
          createdAt: "2025-05-30T01:26:40Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/saison-technology/hulft-bundle@sha256:b7d1348b16ae564bc1092dbdb2042bc2a1ee8a3b7fdc85b7d6fe18be9bb99c19
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
        version: 10.4.0
      entries:
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:e16accb01b2ca6bf4ebd60c6ce7faab73b784828578ee74305a3a97ff9db1e26
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:ea37c68775b14e7d590b6eeec3c9b59eba8e988860fc04d40775da3957d6d040
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:1673f08529a94671f881edec6fd29b73f43c2115dc86b643f9dd79d7e36be38e
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:5eabf561c0549d81862e521ddc1f0ab91a3f2c9d99dcd83ab5a2cf648a95dd19
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:758bfd5455d0ca96f5404229a73fdcd56686564fc4c315ca3468ebe8588dd9ca
        - registry.redhat.io/openshift4/ose-tools-rhel8@sha256:486b4d2dd0d10c5ef0212714c94334e04fe8a3d36cf619881986201a50f123c7
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:0765d248fd1c4b4f8cbcbee23f35506c08c9eeb96a8b0e8ff1511319be6c0ae6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:0ad1961d6dd1c8636f5f1bcb5ffdbea51ad29b88c6539df366305ef801c79877
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:06382df99dfff9f002365a31eee0855f8032e351bc9583e42e489ab7623fe287
        - registry.redhat.io/rhtas/client-server-cg-rhel9@sha256:f0c8f5129c0f611332f7b34dfc5f0dc39c73ed8a985a0845d1fe6ce31d2ba020
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:d4ea970447f3b4c18c309d2f0090a5d02260dd5257a0d41f87fefc4f014a9526
        - registry.redhat.io/rhtas/tuf-server-rhel9@sha256:6c26d2394b1d82995a6f3d3483d2c18befd6a4ae826af2cd25aac30dfabef1cc
        - registry.access.redhat.com/ubi9/httpd-24@sha256:7874b82335a80269dcf99e5983c2330876f5fe8bdc33dc6aa4374958a2ffaaee
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:496b29b00a9831278eb3a8316bd5fb233ad73cb9a6217b1ff4a5f72a91b87414
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:58140d6b7fee762b148336b342a6745b1910ab758a701b2e8077b1fca2af1180
        - registry.redhat.io/rhtas/client-server-re-rhel9@sha256:83b9f9a5ea40473c8525a16d9965b1ed1eb2d8f7c2ec1d17ae38e5f10ac1652e
        version: 1.0.2
      deprecation:
        message: channel stable-v1.0 is no longer supported.  Please switch to channel
          'stable' or channel 'stable-v1.1'.
      entries:
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:540556453c98db3919e5f305a99c699e767e2891a95fe7527ace578f3c67a308
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:6131053778ea04e437f3005f90d1138aa11ebc58e3a9295e2a8d8ef6713a52be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:09b8259ec9e982a33bde84c7c901ad70c79f3e5f6dd7880b1b1e573dafcbd56c
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:571c48ecb2658ce70199c06dc483ea48a16e032a764a6830388ad845f508d981
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        - registry.redhat.io/ubi9/httpd-24@sha256:366cc5bc3b96069fecaa0abeb9d3510c507e69a20de1d205d881da8ae8561ab4
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:dc994a95be22b0f4bab022fc362c4f44c6a7d1887a2eb0d04870d75654ec013b
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:5a752cefdaf28bfc53847185cdd5fef1ee47e3dcff8472f8a8bf7bbdc224ef57
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:75f1049431f9e92898a4217870309cbbb3b39c8362e929c0bad3b53cad4459db
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:603a2ffa54df0bd45ff42ac710dd16331560eda2be93442232b045c959773c4d
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:329a0dad7842d4ef60430b0e3f188231bea9e579427443ecd42eb62a5c282266
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1f5a30a285a16635a7234c3c1763dfb385c8bffd605fc862b782bdb5c6c61ea3
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:3b8f49c41df15022f8ffdf3a8f8605b14c14f4e10eae754a06a86b6585d158b3
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:775b261b8f258b369213246f2da5c91218678acb5e3b8bb4670b143f487767af
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:76c24a38ac89ed632d38e44049f37e4997abfa27fa8cadbb8afb42575031296f
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:796860a3e85712c60398c36983e0ff4d45325c7a4de869da2ebf1b6ba4b19825
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:7ce611aefdfedd8b2a633def482cf41f390c95b8f8c800b6163a585f117a9e2e
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/openshift-community-operators/prometheus@sha256:1de089becfcceb3747b3a3ce310c1612861b03446906cab5e981071db10709a9
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - quay.io/openshift-community-operators/azure-service-operator@sha256:2430ebe724ead60e84f12aa485c8dae4af82af10d14400ce76356392030c8c0b
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-operator-bundle@sha256:c6bc928d220fa001e2a7b5e8661a3c8a604c21385dc1bb97b8dd121ca2347916
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/openshift-community-operators/spark-helm-operator@sha256:9bd49276e02f2fb28bb1ff116112cd7ebe0aa12bdb9f1e46aae2501fd9e4e755
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patterns-operator@sha256:99ac398b06fa9d9a7626b99cb47fb380d0cf4a4307fa11da8e486df35966debb
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        - quay.io/openshift-community-operators/varnish-operator@sha256:46de72c5abad80cef3d6426be967131c51c2c15029a6a38744975578c805a44e
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:3f7fc52c79f3e4ef7ddd6f5efc8eec1ee4836bbf2076e7dd5f68bb3b7c6e9d8f
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:34d4f950bce7863f15d5f1e7d6555a3ccf97ec31dc3ee140f68442890c4f0f96
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:96b77663961daea3a336b97e34e9412281bfb2727854dc970a69edd029b9af85
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:5170255ecd45b9bb7a70abec0f6eeb0518bbc2871de759e468dd61a80261c0ad
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:9a3f5b6c403fb84c9cb93c9d7cf902c80990196983607c8ca568d4b558938e17
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/kasten-rhmp-bundle@sha256:c6b58e5f6c9b2a1f5c4bb770d1fa1db6eaaee8ae7d9751a48338623d0d812438
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.connect.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:99e845965f7eca4b5a3361c494a385292d655e7c1906185197826c773cf9aaf6
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        - registry.connect.redhat.com/qc-qranium/aic-operator-bundle@sha256:581b21f44d6b993888681ed702ed10b9d1bb8e52dcca53a91e856febed2613a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/devspaces-operator-bundle@sha256:a48dd94d5616183843ba939146b7e4d985210cdd902038ff80ff6df3df3d74ca
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:18452ba29f4e0f52b765e8f9359e4143ebfe1203160e6f64ce46ec9d4f095a62
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:9716402247ed874f19574a0d5125d1035fd198f36081d9129c1d4d4ad2f714a6
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:397feef6fc1dffb7e2c3798bbe3bfebb86d6173c225eb61473029cc4500b0cbd
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a70fceaad3b6a23e6f247b4dff41fa64349c4bbf841b66adbd0f8309c6acce74
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a7a4d20b5acc59dbb8c99a37c19b2749f15ed0be7f841490a829796865c5a956
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        - quay.io/openshift-community-operators/loki-helm-operator@sha256:fa3662ec16e6cb3f866ad46982a50d8fdefb43c68f3caa13291239013e03af13
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/openshift-community-operators/infinispan@sha256:3ba9d20e83825f3c041a7dfd195b0800f74fa0f5cd1e6218fbd3547fedd85d64
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:f165693aa550903d7a26b429ed4d093f60194a7a04148e65e61c52d67b527fad
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:5e1a1fdb5a78a34155abe77754c774c65335ea8cbfdec2a222ee22a062c2db46
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:e43b0581415af648c0b82830377e354a951814d3eb7825bbe31b9f3e633917c5
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:4ec16d88828e580fd9223125043ca49569324c4f06b7a65bbaa7d76923de2bad
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:7f959a4076c36d76f8f1489a92d6984d1cfd2ae5c0533a78e8b16010eb6d4d3c
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:a3132caf9cddf6d6186b3e982a6f6232a208fc727b433ee6d200a3b72f50ca88
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:d89f4d647534569f401790d6e10bc42433ae791be86558b100a4027a74b92cd3
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
          createdAt: "2025-09-30T14:21:42Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/flux-operator@sha256:557ae37fa6a6942e3a9d06966df67721011ac7905e2241867c975948092f2936
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
        version: 0.30.0
      entries:
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-operator-bundle@sha256:84d65eed99300b0ba87369cbf9a9fc17a37781f8b6f9de2819d0a159f9071a6c
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:e140415e0361cabeab192ba6d22634283cd47749dd25e9cee4a775d3c409cd97
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:4b7796235a0a545db83570add7a675fdd9d82b17f931eb81d0cafbedb1fbf22c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:9f9dadba5a35c8ab98ddacee22593a247a31e96a4b6e1b4f1a946c2e34006bb9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:01ca6a65a041a47efd611673a720e69464f02424d21feadc496c356065ff2f65
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        - registry.connect.redhat.com/appian/appian-operator-bundle@sha256:2cfb2d2e43fa207c8c06df6de8d3e90123c4cd52cdb6a4c6d618ec576b79c2eb
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:ac244e06ecd2f8a5e4de7b121efb8e82e3e7d23194878cb8327ff0b2affb5f04
        - registry.redhat.io/openshift4/ose-smb-csi-driver-operator-bundle@sha256:91d04cbe0eee6c4fe0230d0d179366942e6c54afb65bb77a43a681e1a2411caf
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        version: 4.18.0-202509100344
      entries:
      - name: smb-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: smb-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: smb-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: smb-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: smb-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: smb-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: smb-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: smb-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: smb-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: smb-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: smb-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: smb-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: smb-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: smb-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: smb-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        - registry.connect.redhat.com/enterprisedb/pg4k-pgd@sha256:fb206b95690e9604c41715369c69125a906b866dc35cfc5c63638ce7a4337fa2
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.connect.redhat.com/loc-devops/operator-bundle@sha256:b9f7a678ceec2d367d55499a48a15a9dd729bc34b0e707989be23c97fc3c5f4e
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "aes_key_rotation_interval": "2160h",
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_location_id": "devspatcher",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "enable_server_session": "true",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "session_duration_seconds": "86400",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          },
                          {
                            "email": "admin@mycompany.com",
                            "hash": "$2y$10/yYwtXRrOagYO2jNtwVzGQPPEfDvX0QllV8NXbku",
                            "userID": "b7e2c2e2-2e7a-4c7b-9e2a-7e6b8c1f2a3d",
                            "username": "Admin MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
          createdAt: "2025-09-24T08:25:29Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/edb-hcp-operator@sha256:87833c7edfa714b433fb066533bd415049b4ae96588ea8457a85a62326801d5c
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
        version: 1.3.0
      entries:
      - name: edb-hcp-operator.v1.3.0
        version: 1.3.0
      - name: edb-hcp-operator.v1.2.3
        version: 1.2.3
      - name: edb-hcp-operator.v1.2.2
        version: 1.2.2
      - name: edb-hcp-operator.v1.2.1
        version: 1.2.1
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator-bundle@sha256:a26586b5f2d748f56dc29fbe3871d84de422e4103ce13c93dbe477718c582487
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        - quay.io/openshift-community-operators/yaks@sha256:0031de9a68502cb65923c4e17453195663f59e46b1cda37619a857add503c016
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:91632b89b4001e2ffa0f1aef56207eb48f0275e20c635e7f1ff7c6de57a59ec2
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:eed8048d085eef7eefaa643a09627521fc2596ebd26280e05c79dddaa0b01afe
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        version: 4.17.13-rhodf
      entries:
      - name: mcg-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: mcg-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: mcg-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: mcg-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: mcg-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: mcg-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: mcg-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: mcg-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: mcg-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: mcg-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: mcg-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: mcg-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: mcg-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: mcg-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: mcg-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{},"bucketNotifications":{"enabled":false}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:ed5c7c8cbbe455b6d31b1c55871585a0679a91800f53743d81d4cc36c6aac7cc
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:88387ba7a9e85c1bcf57966558db57c099fd301ea80a68e43afc2bca3dfe94a7
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.18.11-rhodf
      entries:
      - name: mcg-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: mcg-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: mcg-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: mcg-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: mcg-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: mcg-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: mcg-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: mcg-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: mcg-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: mcg-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: mcg-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: mcg-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-operator-bundle@sha256:b637a02d23bd8364cc8914421049eb60169c163ac70bff2f33591df1a1193002
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e9612ec99daa171c403b5ddf37788526e5375b83987e5de9b1ebe519198607b4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:34d177abb56b948eaf9ce6fa62e37427ddee9fcecca48e752d09f72c69847a46
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:932e83fe9c9b54934767a7c5af683b2ab08bd2b4f0524c191f1c766dc64d9c3d
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e1c19bbcd4c010ef2bd4f760eb24b2c96cdb82b4e025361a433e6a241eb23cf5
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:a6e02e1876fd5382ba4921dd65650b2d747666feb96c839f34e99d2972dfc323
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cb711fcb32b5ca0f34c4e3c67624bc4e035de4247617d988a9e841fd58c1c911
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:26af4e950a33755114298386a70229672a09ce45b6203025856f460a55c4c8a0
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        - registry.connect.redhat.com/amd/amd-t2-january-bundle@sha256:7c87354aa07e783a2cbf4ab29877a8ba56ea4269bc1f41bfbe08c558b3666841
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/deployment-validation-operator@sha256:2803298c7cf3eeaffe9da3153142058d94d65ff1dae90efd8819a4cd23e0518f
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        - quay.io/community-operator-pipeline-prod/eginnovations-operator@sha256:19b12c75f7a2dfab027881ed9f0b09d25e3baa936d21f7be3ec1e0177ed9fdf2
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        - quay.io/openshift-community-operators/limitador-operator@sha256:7aa0a4fe55e5e099e2e6e4c2cd44b9bfeba69d8cfade10500ccba3092fb60dd2
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        - quay.io/community-operator-pipeline-prod/limitador-operator@sha256:450f9bf43b57b83e9975a1865981af22e28c725193573354d1180f406887664a
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/flux@sha256:cd94037ff3055ab6ce40b0670f900248a5fd233cc67a2a4f92c1559cf9d2f3ad
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-framework-bundle@sha256:53d056ac37177263d79098cda775deff86a327c65106ea5a40109b08a856cf84
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/keycloak-permissions-operator@sha256:4de19e52e84cdba0963bdd852a63912f097280800d65add8d474c0953da478a0
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle@sha256:a90a748aa70ee5f1b8fd375a7877db1292792ba46f7b49f629533c82a25c43a0
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.connect.redhat.com/cpopen/ibm-zen-operator@sha256:5a9120d5977e934ca3801eab44ed48ecd4a9fee10c0b4980594d33cb02e64855
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        - registry.connect.redhat.com/airlock/microgateway@sha256:80d49a299a97bed33c28941f93312e86b949ce945e843c93da6bca49865c81d0
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        - registry.connect.redhat.com/airlock/microgateway@sha256:087d4d726f8080111d2b18cd0f9ab290f4560333fef308e553991484b6210c61
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/openshift-community-operators/nlp-server@sha256:6f80f44822301252540c6b808a20cea327f0eedeadc7643d31fc743de5c5e696
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:fdc0c228ca38db30f89812e41c09ec96d9276ed844b0801013375bbffd6c1a16
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:ea606dcfa4cbe4df7b27c10d3fc45f464845518c4896845079ca9194600ceb6f
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - registry.connect.redhat.com/loadcore/keysight-kcos-ingress-bundle@sha256:f04b8f15b069c06eac5d9a97f81a2f069704237eea998d35727b89854c793746
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        - registry.connect.redhat.com/loadcore/keysight-wap-msg-broker-bundle@sha256:a7dfca98b4e09c20317cbb486e7893de8d88faf5db2ee26549ba81332ec526cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - quay.io/community-operator-pipeline-prod/dell-csm-operator@sha256:6e2e813b7ca426a793da717d42f3620edbc3eaa84be710600f2a5262924378b4
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:0739d5035590cece6d2a573aa1e799b0b0cab8b548a41cc2488cf465486cf4a5
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:3028579f4909be6b8dcd970e778bc88d0576c76e55b98c886738b8f601cdf4c1
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:e50343879b4c3c4002c5e7ff67f4664df441cf5f239f59fa98b7fddd75516d41
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:add5d14653f57ac1e3254c2d1c70fc9232959f76ee6286fe5cb572696a5a4fd3
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:7fc4d97b5ee63085672fc9a562ebdad3d4c00d1b555e06679d1dd9ac945f1bb9
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:05e12cc4caaebb7f185f651f8ca876c442c02d79794e561b743beba0d189fa04
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:84796005b6e71efa5d783c55b1f72bcc3762944bf32740cb75498bf8b908bffb
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:424e138a17e2af91603e834ff65d529400d6793771ddaab92d3988f4e6d2f2e3
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:9d505bf561650eb9785811ba50c920480d85679eac6a611a1c5794544e175934
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3abd8c3f97c9bba127b6e8c7efcc078c90c9d7cc35d74cfcd9bc0d303f62d2ed
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:21c17448932f053f1d4845111288a317786b190154bae9da7947c315ba12f796
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:023b9cc1e4ec2542b8d68b636b21a2e8c4eb143424581c62d16598fddb47da6a
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3843fd48c57be102576db93e8483a30f8efe8cd46b5e09c704ac0bc5f18553a3
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:55bb4133abc945ce7b75e6c01c6beaf63c19c129fc8fbfb48f1b53b8e28aae97
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:43b959fcdcf107b70e4224d80b63d1c123560c3af055109bbcc7d68476546584
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:2845843f32b895dee0242e31380b44625e3f8381caee7ecb3258648311c302f7
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:c7cd3ffa88eae188e94abcc0af56d328aff43463b98a3bc3c0b574f29373944b
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: openshift
      provider-url: ""
    name: pf-status-relay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pf-status-relay-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pfstatusrelay.openshift.io/v1alpha1",
                "kind": "PFLACPMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "pf-status-relay-operator"
                  },
                  "name": "pflacpmonitor-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.19.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PFLACPMonitor is the Schema for the pflacpmonitors API
            displayName: PFLACPMonitor
            kind: PFLACPMonitor
            name: pflacpmonitors.pfstatusrelay.openshift.io
            version: v1alpha1
        description: It deploys the pf-status-relay application.
        displayName: pf-status-relay-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pf-status-relay
        - sriov
        - lacp
        links:
        - name: Pf Status Relay Operator
          url: https://pf-status-relay-operator.domain
        maintainers:
        - email: marguerr@redhat.com
          name: Marcelo Guerrero
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: openshift
        relatedImages:
        - registry.redhat.io/openshift4/pf-status-relay-rhel9@sha256:8cdc0714d02144f5b32e26c97f7678d0377e0c3826c475a784e76a42f781d3e3
        - registry.redhat.io/openshift4/pf-status-relay-operator-bundle@sha256:fb9635a614b6676fbe6a532d07b3206f12ee525b7947378785e79951d43940b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/pf-status-relay-rhel9-operator@sha256:e6719af6fdc99c652238f3c31a014782811eb46da91e65f93c083a7a001e2496
        version: 4.18.0-202509100344
      entries:
      - name: pf-status-relay-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: pf-status-relay-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: pf-status-relay-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: pf-status-relay-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: pf-status-relay-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: pf-status-relay-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: pf-status-relay-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: pf-status-relay-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: pf-status-relay-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: pf-status-relay-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: pf-status-relay-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: pf-status-relay-operator
    provider:
      name: openshift
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/rhpam-7/rhpam-operator-bundle@sha256:46e7d899d64ea8cbfc35221d05535db1023c41c1d3a7c580cb03e7e7d769144b
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-rhel9-operator-bundle@sha256:896a615487c0ffd5335b869f0dc0c0ec1c9200dbcafa52f2493003fecbc030bc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        - quay.io/openshift-community-operators/alvearie-imaging-ingestion@sha256:f65cc3cb83b1c75385970ce70c6a84b15b143e682ad0d1784148f1643df0991e
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-operator-bundle@sha256:f596dc9323da89f5e45e71192f0fa698b7bb29b8205eaa0fc1659d68fa7401f4
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:8e223a537f487bc3cef10769adc6fb619085dafe3ea797cfd52c20e3066f694f
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle-rhmp@sha256:81476f897726c85293e64e6188ed314e3e1e637c568c6b0ff8f475a876bee29c
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:4ae986761a69d25697b9b6a3a99de88d8587beba136c3de29b90624e2ea6738f
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:adbfebc73bc5a9cd12d17e9e1fbdf87c2176aae7a7217bcefcec3068e51bc904
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    defaultChannel: stable-v0.2
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/gitlab/gitlab-runner-operator-bundle@sha256:68b22c412a44d5fa09b45e204d7c8e7b6ca6afa4ed6879de608dbcb0b1929676
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/openshift-community-operators/pcc-operator@sha256:e99a466ddb52ad44c435ccf952754d685744fb5b7b7e4cf77dd542e530e4f968
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53-controller@sha256:227a306cc90998489bfb681cb29b416238b404bc2db28a09b8949d4cca2d87a4
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-operator-bundle@sha256:e83707d0b396c38cf55046d3159f1f427188404aa11dc47d7b1c1f25605182de
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.3-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:64a623d5d536c790dee04c1604ec2f4429f8b8f1973b4232b543e8cbe23d6590
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:7c9b4327d8677f48c48888230fc38eec8c20c7f646dde568dbd96b5e4c8f7967
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:8797a3be94ce6f4e4da15e4efefe41ea5b6ffb4cdc6d698638c80976bb22491c
        version: 4.17.6
      entries:
      - name: lvms-operator.v4.17.6
        version: 4.17.6
      - name: lvms-operator.v4.17.5
        version: 4.17.5
      - name: lvms-operator.v4.17.4
        version: 4.17.4
      - name: lvms-operator.v4.17.3
        version: 4.17.3
      - name: lvms-operator.v4.17.2
        version: 4.17.2
      - name: lvms-operator.v4.17.1
        version: 4.17.1
      - name: lvms-operator.v4.17.0
        version: 4.17.0
      name: stable-4.17
    - currentCSV: lvms-operator.v4.18.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:5d0534e08790e91d5d355eac2e5c187a217c1ec6eb59464ea9a771490ac040e9
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:6e809f8393e9c3004a9b8d80eb6ea708c0ab1e124083c481b48c01a359684588
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:b3b1c4836a82a7ec6c9b02e4a86c352f4ae8c45a3fe027ae7200869afe3ab028
        version: 4.18.3
      entries:
      - name: lvms-operator.v4.18.3
        version: 4.18.3
      - name: lvms-operator.v4.18.2
        version: 4.18.2
      - name: lvms-operator.v4.18.1
        version: 4.18.1
      - name: lvms-operator.v4.18.0
        version: 4.18.0
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-rhmp@sha256:1502ccda3c067d74689d0341008f05362bc23904c8e35b20e020fe0dca82aef7
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.connect.redhat.com/nec/sanstoragecsi-operator-bundle@sha256:1957586c13e3bf80bdc688f519be440ec1dcc4ffe63343f526018dcb93e21cea
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-key-manager-operator@sha256:7a998ea17ef537ec5e998e1627b065ae04ab01b7988196da1cdd1763e05a9924
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:6ef2956b297795a9d0c83119e71309549975df3ae5652e6a98cef035ace5d51a
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:8831faee6979acfb5bd031b64cbc09fb90e9ea4f48bbff15e9d39ccdf6bf3d95
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        - quay.io/openshift-community-operators/prometheus-exporter-operator@sha256:e62276f4c7e583d1fb24997fa970662f1a8d510de0baea159ba796b48cff709b
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        - quay.io/community-operator-pipeline-prod/ack-dynamodb-controller@sha256:547bdd3d1037dbc3b189f4c2f8e1d49379e1d547cd9063391d6c745e1a103b80
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:59ff4966f4f653e49d42d97bfed6351d3730128db74f666ee4ecacff11790baf
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:01f2ca529d2486f113bcefc9fedce6a6fd07bcb48aaf534394b5b04c353f8853
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.connect.redhat.com/cpopen/ds-operator@sha256:6ec2387e45d9557cea2df749831e136120945794773ed0d48e0b04f0b3d9e115
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:ca70ebce43d451d040aedcec036413f986e620234fd9a56d45983805396db0b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:4aec6aa2d2b9a24167bb6bc92db7c2c323847b7a2f3ad7f998c96bf1d2903366
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        - registry.connect.redhat.com/netapp/netapp-supported-apache-spark-on-openshift-operator@sha256:4f15a0fc375156fafdf2d504f9980a8eb4451808b5554357877345cacf8f3594
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-athena-controller@sha256:8cff9258583f8acbdd26d263a61ffa11792cf9f4c1d65410bb4c71f9f7aaf178
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - quay.io/openshift-community-operators/devopsinabox@sha256:1c2ad76102151a56ecd753d9b86fd9f155d64b912067738d1858d33a02557928
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        - quay.io/community-operator-pipeline-prod/activemq-artemis-operator@sha256:1e3c6ff2d2612c3e85d40b1974b5a3c4f603b77353752389fa588728bd4263c4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator-bundle@sha256:df0dcd31867bb1c3c4321db4412a744fd1b3ba8faddb509d191fab92407b6677
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        - quay.io/community-operator-pipeline-prod/apicurio-registry@sha256:7087ec1377c2612edf3b8a76f6e115d66c3ec2b4b20f5d8374638d6bf0037e93
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-pipes-controller@sha256:e46b540219ca42c3b6c08ff99afabf09daea2dca5a335cda6013343df2f35878
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        - registry.redhat.io/openshift-serverless-1/logic-operator-bundle@sha256:5fff2717f7b08df2c90a2be7bfb36c27e13be188d23546497ed9ce266f1c03f4
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:53d8cf8bf5c1103860ffc4110e2a422371761e76ee79edfef8680e947407c5ce
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        version: 1.3.2
      deprecation:
        message: '''builds-1.3'' has reached the end of maintenance support. Please
          upgrade to ''builds-1.5''.'
      entries:
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:aadba00eadb3ac46d321a14498fc3dffd5be9afbf698d80adddde0a7a1681fc4
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        - registry.connect.redhat.com/smilecdr/smilecdr@sha256:c44f5661a598aa18c4c10372918a45edb0f26d7d2772c47a55c66e4ff874cb5f
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - registry.connect.redhat.com/mariadb/mariadb-enterprise-operator-bundle@sha256:73e7f243a6676d559068bf8ab80cc0ffe0db58152675603ca8bd360334439f26
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
          createdAt: "2025-09-09T14:14:54Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:ca10a842900d6515f714a17916a32dfd5a2909543e4d339acefc7e3e41b8d7ac
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:31d145c7074ad6cd85684b5191a3033533bf2149124c7ead5367871da5ed0dd8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        version: 4.17.13-rhodf
      entries:
      - name: ocs-client-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-client-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-client-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-client-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-client-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-client-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-client-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-client-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-client-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-client-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-client-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-client-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-client-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-client-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-client-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
          createdAt: "2025-09-11T15:23:25Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:60a91d6e4fe7e0ab999eab2224ef09fd0681e74920b23b306c7a9893bbb29cb8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:725eabb38876b93ab7693d75013d0292756a2efc00b8608f5716f31f8c324a53
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        version: 4.18.11-rhodf
      entries:
      - name: ocs-client-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-client-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-client-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-client-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-client-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-client-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-client-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-client-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-client-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-client-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-client-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-client-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-sso-bundle@sha256:f2c5a8abd1a9ec7665edf1e6206a6212cb752082b09532e4c218bc0990e48287
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:50b7f0a5f3fdd98cfb0294d664cc01ffd1e5359be66a3da14f5552e9cb3bae2c
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:29d46db99d97fe28758c0efd2acf27203ce9958734b82cf37c3a9063b5159506
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:b83c3263e77904ce1c0f95b7bfab969605c97297a5598fd794d64daf508c52c0
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:ea88a3090291f3742ccd4641785c19a80f4d3bd485a4f9c362a7645f73896db3
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:3dbb71004362a70683a5ee9f0916bca75c0d1c351d144677fd9d85bcb2ca8bf7
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:89d480ae76480829660d4cccfede74717faeb7e62a4a48b932750e41a945783d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a71b6fe33997db924669253992436e1e02d88769bb97376c0871848c2d330fed
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:1074f45a625df6ea412a9ffef3d4f4e53f3107ca45bb1de9bb030c47ce6518d9
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:0543e615d8a48ba6aaf90a8de0b0a79d37689c64bfe71f91c340455d9a359426
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a93fdee9e8d05ebe65e9dd8682fa7bc99c54f9d9f26d2cc8519f529255805da8
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:9d825d79fffd39a1149b262d48d08616795e5cfdb4c4bc40f5f192242f71042d
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: nexus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        - quay.io/openshift-community-operators/nexus-operator@sha256:277b94985c0dc982abb76b5dd8c93b37305ccca32c000d67552d540d4918cbba
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kinesis-controller@sha256:de760dbe730e38cd9fadebe5fdb69299c04512e6a70f567ce678fc0bedc53e9c
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/jaeger@sha256:907dfe1d01f73c8a385da3661ddbe03db31e0b9ad6f5ab2336460402d0d915a0
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchservice-controller@sha256:699f483cf96822576e3b79b690f256e2dbf4ca8d4426daf4298c980aefdeba4f
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/percona-xtradb-cluster-operator@sha256:d0c9d03130774d961f62032f684c494313787354186d028e9709e2463f69431c
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:a5645cb55c3e6c22655c6c8606b9bd51a49db44aecc3fe52e0eb80cca83d17c4
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:19e133c41b8f947aa83afba0417d8db3570cc2e1368813996fa5d135fdca4e04
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        - registry.connect.redhat.com/ibm/ibm-application-gateway-operator-bundle@sha256:e4ece7113fa7056980507408ac29d60676cf51c20d71a058a1accf765a6ea7e3
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-commonui-operator@sha256:87788afbc24851deb2c9c3f14b773bab174b2a2f02ac2fa0dd656a97433ffca2
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dns-operator@sha256:d5eadbf006aade69ac18156c891614da899d0779ba9e8fbaf755a256c46f055c
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/openshift-community-operators/gitops-primer@sha256:97af48469275f1b3a669f8e24a97a80ce9897fc5b12643a4f298b82dbe26e20b
        - quay.io/openshift/origin-oauth-proxy:4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso-operators/watcher-operator-bundle@sha256:01ab69b9c33a91cb194ad644bb64884b358af78467e2bb766ef373e4c36e795d
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4478b8a78bdde23e0d40476f8ad9e4bf0628fba7603a390e8de7e6a319abc1ec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:6c2ab763239f6e5a2a72890120bf1f20f6277850cec59a7e11cea2b6a06f12e8
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:8f30d4c7c491e9a804a8abf80570d1e1924cd0d787c6abb727331541c435d5ff
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:2b6b1a0b56cfb348ed4035b13d31501ff5757c101ed21027b330a7852576b891
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4c4c98b6cecd51c6644eb79b85789b7ebbcdc662edbb66ce435ea3b5e58d9273
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:b8aa74c5b2635a6d3413eeac6fc0f6eb9a2e459089aaabbf626ea3ef92a67ce0
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:73d91324ccc995e9a6db525f194bee340c17dd624ce520c767ea8fcf2346f786
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:d29abe14abc67d8836cc6f7656556a0a7fa8f4a8eea3f78e40265b339e7430f5
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:f7c4ed28a549fe7414312bd22c2ae140f1f28d30656fb9fb0c8ade05a8f82088
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c59956d70416fab0c6d7b3ecd8c9cfa3cf9415f75cfdfe2214f44b0b4aced594
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:7cace73b03bb02202a04b6026eb9f7323610fd0ccd5ba18a6d00fd66577a96b4
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c173a99026499700c08092f8418c79f39527d9b914690b199b41ed781c279de1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        - registry.connect.redhat.com/loadcore/keysight-wap-logging-bundle@sha256:2aeeefa7493eed8960004904567894244418982209e1b3a40488b759463edcdb
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.102
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-09-03T09:23:19Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.102
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator-bundle@sha256:fdef8f42c5425b9607915439daf3e5e2b77705cd220a083a00b42ee2e122d894
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:80c6d45a2ca70f55809620a82c309abb1e8272a0515f9c1a2905b3302c1467a1
        version: 0.1.102
      entries:
      - name: hibernation-operator.v0.1.102
        version: 0.1.102
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/lib-bucket-provisioner@sha256:132ecc88db79fb8b117a7a7974ca8f704600bac489923d918b78fd00378d5a5f
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-operator-bundle@sha256:aeae57b2fc706b782a3e3905e5c1825fe831f96f5b1c9478e1f2b217307a892b
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataDirect Networks
      provider-url: https://www.ddn.com
    name: vmstore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vmstore-csi-driver-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.tintri.com/v1alpha1",
                "kind": "VmstoreCsi",
                "metadata": {
                  "name": "vmstorecsi-sample"
                },
                "spec": {
                  "config": {
                    "debug": true,
                    "vmstore_map": {
                      "vmstore2": {
                        "dataIp": "\u003cdataIp\u003e",
                        "managementAddress": "\u003cmanagementAddress\u003e",
                        "mountPoint": "/tintri-csi",
                        "password": "tintri99",
                        "username": "admin",
                        "vmstoreFS": "/tintri/csi"
                      }
                    }
                  },
                  "csi_snapshotter": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                    "resources": null,
                    "tag": "v6.3.2",
                    "timeout": "120m"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "tintri/vmstore-openshift-file-driver",
                    "tag": "v1.1.0"
                  },
                  "priorityClassName": "system-cluster-critical",
                  "provisioner": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                    "resources": null,
                    "tag": "v4.0.1",
                    "timeout": "120m"
                  },
                  "registrar": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                    "resources": null,
                    "tag": "v2.10.1"
                  },
                  "resizer": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-resizer",
                    "resources": null,
                    "tag": "v1.12.0"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "512m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "64Mi"
                    }
                  },
                  "secretName": "vmstore-csi-file-driver-config",
                  "volumeNamePrefix": "pvc-vmstore"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-13T21:26:54Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: VmstoreCsi
            name: vmstorecsis.charts.tintri.com
            version: v1alpha1
        description: "A Container Storage Interface (CSI) Driver for Tintri Vmstore
          filesystems. This operator eases the installation\nand management of the
          Vmstore CSI Driver in an OpenShift or other Operator Lifecycle Manager enabled
          cluster.\n\n\n## Prerequisites\nMake sure that the top level folder where
          the CSI volumes will be created exists on VMStore and is shared over NFS.\nFor
          example `tintri/csi`.\n## Usage\n\n### Dynamically provisioned volumes\n\nFor
          dynamic volume provisioning, the administrator needs to set up a _StorageClass_
          in the PV yaml (examples/nginx-dynamic-volume.yaml for this example) pointing
          to the driver.\nFor dynamically provisioned volumes, Kubernetes generates
          volume name automatically (for example `pvc-vmstore-cfc67950-fe3c-11e8-a3ca-005056b857f8`).\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\n```\n\n#### Example\n\nRun
          Nginx pod with dynamically provisioned volume:\n\n```bash\noc apply -f examples/nginx-dynamic-volume.yaml\n\n#
          to delete this pod:\noc delete -f examples/nginx-dynamic-volume.yaml\n```\n\n###
          Static (pre-provisioned) volumes\n\nThe driver can use already existing
          Vmstore filesystem,\nin this case, _StorageClass_, _PersistentVolume_ and
          _PersistentVolumeClaim_ should be configured.\n\n#### _StorageClass_ configuration\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc-nginx-persistent\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\nallowedTopologies:\n-
          matchLabelExpressions:\n  - key: topology.vmstore.csi.tintri.com/zone\n
          \   values:\n    - zone-1\nmountOptions:                         # list
          of options for `mount -o ...` command\n#  - noatime                          #\n```\n\n####
          _PersistentVolume_ configuration\n\n```yaml\napiVersion: v1\nkind: PersistentVolume\nmetadata:\n
          \ name: vmstore-csi-file-driver-pv-nginx-persistent\n  labels:\n    name:
          vmstore-csi-file-driver-pv-nginx-persistent\nspec:\n  storageClassName:
          vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n    - ReadWriteOnce\n
          \ capacity:\n    storage: 1Gi\n  csi:\n    driver: vmstore.csi.tintri.com\n
          \   volumeHandle: vmstore1/172.30.230.128:/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img\n```\n\n####
          _PersistentVolumeClaim_ (pointing to created _PersistentVolume_)\n\n```yaml\napiVersion:
          v1\nkind: PersistentVolumeClaim\nmetadata:\n  name: vmstore-csi-file-driver-pvc-nginx-persistent\nspec:\n
          \ storageClassName: vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n
          \   - ReadWriteOnce\n  resources:\n    requests:\n      storage: 1Gi\n  selector:\n
          \   matchLabels:\n      # to create 1-1 relationship for pod - persistent
          volume use unique labels\n      name: vmstore-csi-file-driver-pv-nginx-persistent\n```\n\n###
          Topology configuration\nIn order to configure CSI driver with kubernetes
          topology, use the `zone` parameter in driver config or storageClass parameters.
          Example config file with zones:\n  ```bash\n  vmstore_map:\n    vmstore1:\n
          \     mountPoint: /tintri-csi                            # mountpoint on
          the host where the tintri fs will be mounted\n      dataIp: 172.30.200.37\n
          \     managementAddress: tafzq23.tintri.com\n      vmstoreFS: /tintri/fs01\n
          \     zone: zone-1\n      username: admin\n      password: password\n\n
          \   vmstore2:\n      mountPoint: /tintri-csi                            #
          mountpoint on the host where the tintri fs will be mounted\n      dataIp:
          10.136.40.71\n      managementAddress: tafbm05.tintri.com\n      vmstoreFS:
          /tintri/csi\n      zone: zone-2\n      username: admin\n      password:
          password\n  ```\n\nThis will assign volumes to be created on Vmstore that
          correspond with the zones requested by allowedTopologies values.\n\n\n###
          Defaults/Configuration/Parameter options\n\n| Default  | Config               |
          \ Parameter    |  Desc                             |\n|----------|----------------------|---------------|-----------------------------------|\n|
          \  -      | dataIp               |   dataIp      | Data address used to
          mount Vmstore filesystem |\n| -    | mountPoint       |    -      | Mount
          point inside of the driver container where Vmstore filesystem will be mounted|\n|
          \  -      | username             |      -        | Vmstore API username
          |\n|   -      | password             |      -        | Vmstore API password
          |\n|   -      | zone                 |      -        | Zone to match topology.vmstore.csi.tintri.com/zone
          \           |\n|   -      | vmstoreFS            |   vmstoreFS   | Path
          to filesystem on Vmstore |\n|   -      | managementAddress    |      -        |
          FQDN for Vmstore management operations |\n|   -      | defaultMountOptions
          \ |   mountOptions| Mount options |\n|   0777   | mountPointPermissions|
          \  mountPointPermissions | File permissions to be set on mount point |\n\n####
          Example\n\nRun nginx server using PersistentVolume.\n\n**Note:** Pre-configured
          volume file should exist on the Vmstore:\n`/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img`.\n\n```bash\noc
          apply -f nginx-persistent-volume.yaml\n\n# to delete this pod and volume:\noc
          delete -f nginx-persistent-volume.yaml\n```\n\n## Snapshots\nTo use CSI
          snapshots, the snapshot CRDs along with the csi-snapshotter must be installed.\n```bash\noc
          apply -f deploy/kubernetes/snapshots/\n```\n\nAfter that the snapshot class
          for VMStore CSI must be created\n\n```bash\noc apply -f examples/snapshot-class.yaml\n```\n\nNow
          a snapshot can be created, for example\n```bash\noc apply -f examples/snapshot-from-dynamic.yaml\n```\n\nWe
          can create a volume using the snapshot\n```bash\noc apply -f examples/nginx-from-snapshot.yaml\n```\n\n##
          Troubleshooting\n### Driver logs\nTo collect all driver related logs, you
          can use the `oc logs` command.\nAll in on command:\n```bash\nmkdir vmstore-csi-logs\nfor
          name in $(oc get pod -owide | grep vmstore | awk '{print $1}'); do oc logs
          $name --all-containers > vmstore-csi-logs/$name; done\n```\n\nTo get logs
          from all containers of a single pod \n```bash\noc logs <pod_name> --all-containers\n```\n\nLogs
          from a single container of a pod\n```bash\noc logs <pod_name> -c driver\n```\n\n####
          Driver secret\n```bash\noc get secret vmstore-csi-file-driver-config -o
          json | jq '.data | map_values(@base64d)' > vmstore-csi-logs/vmstore-csi-file-driver-config\n```\n\n####
          PV/PVC/Pod data\n```bash\noc get pvc > vmstore-csi-logs/pvcs\noc get pv
          > vmstore-csi-logs/pvs\noc get pod > vmstore-csi-logs/pods\n```\n\n####
          Extended info about a PV/PVC/Pod:\n```bash\noc describe pvc <pvc_name> >
          vmstore-csi-logs/<pvc_name>\noc describe pv <pv_name> > vmstore-csi-logs/<pv_name>\noc
          describe pod <pod_name> > vmstore-csi-logs/<pod_name>\n```"
        displayName: Vmstore CSI file driver
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - CSI
        - Vmstore
        - storage
        - Tintri
        - DDN
        links:
        - name: DDN vmstore Filesystem
          url: https://www.ddn.com/products/lustre-file-system-vmstore/
        maintainers:
        - email: akhodos@tintri.com
          name: Alexey Khodos
        - email: rrethinam@tintri.com
          name: Ranjith Rethinam
        maturity: alpha
        provider:
          name: DataDirect Networks
          url: https://www.ddn.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/ddn/vmstore-csi-file-driver-operator@sha256:64df1c504a9d70bd2b8b34520c8d431a92bb54702d601a57a2a989c6dd982462
        - tintri/vmstore-csi-driver-operator-controller@sha256:749ceff2bd46c71d81338510c1cc5866d39c66345f971d3093f986ea0417311d
        version: 1.1.0
      entries:
      - name: vmstore-csi-driver-operator.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vmstore-csi-driver-operator
    provider:
      name: DataDirect Networks
      url: https://www.ddn.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:042623e1faacc9fc22222a1c3e546a561abf7fc05d41ae9c915a69795f6bec7e
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:67b2b8783a04460019057cf826a74c18747c8c64ee7fe8ce83c8d0b7cd87cb36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/resource-locker-operator@sha256:9cc08639be2af4b72a9b816420d87dfa290a168fb83edc947a04b3e06cdb675b
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator-bundle@sha256:e284d0839795f763ca53402c2b050fa9ffedebccd0f3832fe7259368c3ad32f2
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-s3-controller@sha256:058a37a9a2fe6a42bbdddabb3fc7495450f1ebbd7f46194866c1491c40ce1a9f
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ClusterPulse
      provider-url: ""
    name: clusterpulse
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterpulse.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.clusterpulse.io/v1alpha1",
                "kind": "ClusterPulse",
                "metadata": {
                  "name": "clusterpulse-prod"
                },
                "spec": {
                  "api": {
                    "enabled": true,
                    "imagePullPolicy": "Always",
                    "logging": {
                      "json": true,
                      "level": "INFO"
                    },
                    "oauth": {
                      "enabled": true,
                      "headerEmail": "X-Forwarded-Email",
                      "headerUser": "X-Forwarded-User"
                    },
                    "rbac": {
                      "auditEnabled": true,
                      "cacheTTL": 0,
                      "defaultDeny": true,
                      "enabled": true,
                      "groupSyncInterval": 300
                    },
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "security": {
                      "allowedOrigins": [
                        "*"
                      ],
                      "sessionExpireMinutes": 1440,
                      "trustedHosts": [
                        "*"
                      ]
                    },
                    "server": {
                      "host": "0.0.0.0",
                      "port": 8080,
                      "reload": false,
                      "workers": 1
                    }
                  },
                  "clusterEngine": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "timing": {
                      "cacheTTL": "600",
                      "fullScanInterval": "300",
                      "reconciliationInterval": "30"
                    }
                  },
                  "frontend": {
                    "config": {
                      "features": {
                        "devTools": false,
                        "mockData": false,
                        "nodeDetails": true,
                        "operatorDetails": true,
                        "registryStatus": true
                      },
                      "refreshInterval": 30000,
                      "theme": {
                        "defaultTheme": "auto"
                      }
                    },
                    "enabled": true,
                    "oauth": {
                      "enabled": true,
                      "image": "quay.io/openshift/origin-oauth-proxy:4.17"
                    },
                    "replicas": 3,
                    "route": {
                      "enabled": true,
                      "tls": {
                        "termination": "reencrypt"
                      }
                    }
                  },
                  "policyEngine": {
                    "batch": {
                      "cacheClearBatchSize": 500,
                      "redisPipelineBatchSize": 1000,
                      "redisScanBatchSize": 100
                    },
                    "cache": {
                      "groupCacheTTL": 300,
                      "policyCacheTTL": 300
                    },
                    "enabled": true,
                    "limits": {
                      "maxPoliciesPerUser": 100,
                      "policyValidationInterval": 300
                    },
                    "logging": {
                      "level": "INFO"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "enabled": true,
                    "volumePermissions": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "ClusterConnection",
                "metadata": {
                  "name": "example-cluster"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "cluster-credentials"
                  },
                  "displayName": "Example Production Cluster",
                  "endpoint": "https://api.example.openshift.com:6443",
                  "monitoring": {
                    "interval": 30,
                    "timeout": 10
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "MonitorAccessPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "access": {
                    "effect": "Allow",
                    "enabled": true
                  },
                  "identity": {
                    "priority": 100,
                    "subjects": {
                      "groups": [
                        "default-group"
                      ],
                      "users": [
                        "default-user"
                      ]
                    }
                  },
                  "operations": {
                    "audit": {
                      "logAccess": false,
                      "requireReason": false
                    }
                  },
                  "scope": {
                    "clusters": {
                      "default": "none",
                      "rules": [
                        {
                          "permissions": {
                            "view": true,
                            "viewAuditInfo": true,
                            "viewCosts": true,
                            "viewMetadata": true,
                            "viewMetrics": true,
                            "viewSecrets": true,
                            "viewSensitive": true
                          },
                          "resources": {
                            "namespaces": {
                              "visibility": "all"
                            },
                            "nodes": {
                              "filters": {
                                "hideMasters": false
                              },
                              "visibility": "all"
                            },
                            "operators": {
                              "visibility": "none"
                            },
                            "pods": {
                              "visibility": "all"
                            }
                          },
                          "selector": {
                            "matchPattern": ".*"
                          }
                        }
                      ]
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "RegistryConnection",
                "metadata": {
                  "name": "dockerhub",
                  "namespace": "clusterpulse"
                },
                "spec": {
                  "displayName": "Docker Hub",
                  "endpoint": "https://registry-1.docker.io",
                  "healthCheckPaths": [
                    "/v2/"
                  ],
                  "insecure": false,
                  "labels": {
                    "environment": "production",
                    "provider": "dockerhub"
                  },
                  "monitoring": {
                    "checkCatalog": false,
                    "interval": 60,
                    "maxCatalogEntries": 100,
                    "timeout": 10
                  },
                  "skipTLSVerify": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: docker.io/clusterpulse/clusterpulse-operator:0.1.0
          createdAt: "2025-09-26T21:09:27Z"
          description: Provides OpenShift cluster health monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/ClusterPulse/operator
          support: None
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a connection to a monitored Kubernetes/OpenShift
              cluster
            displayName: Cluster Connection
            kind: ClusterConnection
            name: clusterconnections.clusterpulse.io
            version: v1alpha1
          - description: Manages the deployment and configuration of ClusterPulse
              monitoring platform
            displayName: ClusterPulse Instance
            kind: ClusterPulse
            name: clusterpulses.charts.clusterpulse.io
            version: v1alpha1
          - description: Defines fine-grained access control policies for cluster
              monitoring
            displayName: Monitor Access Policy
            kind: MonitorAccessPolicy
            name: monitoraccesspolicies.clusterpulse.io
            version: v1alpha1
          - description: Defines a connection to a container registry for monitoring
            displayName: Registry Connection
            kind: RegistryConnection
            name: registryconnections.clusterpulse.io
            version: v1alpha1
        description: |
          ## About ClusterPulse
          ClusterPulse is a comprehensive multi-cluster monitoring and management platform designed for OpenShift environments. It provides centralized visibility, policy-based access control, and real-time monitoring across multiple clusters through a unified dashboard. For more information, visit [https://github.com/ClusterPulse/operator](https://github.com/ClusterPulse/operator).

          ## Features and Capabilities
          - **Multi-Cluster Monitoring**: Connect and monitor multiple OpenShift/Kubernetes clusters from a single interface
          - **Policy-Based Access Control**: Fine-grained access policies using MonitorAccessPolicy CRDs
          - **Real-Time Metrics**: Live cluster health, node status, and resource utilization tracking
          - **Registry Monitoring**: Monitor connected container registries
          - **OAuth Integration**: Seamless OpenShift OAuth integration for authentication
          - **Redis-Backed Storage**: Scalable data storage using Redis for metrics and state
        displayName: ClusterPulse
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - clusterpulse
        links:
        - name: ClusterPulse
          url: https://github.com/ClusterPulse/operator
        maintainers:
        - email: asmoore@redhat.com
          name: Asa Moore
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: ClusterPulse
        relatedImages:
        - docker.io/clusterpulse/clusterpulse-operator:0.1.0
        - quay.io/community-operator-pipeline-prod/clusterpulse:0.1.0
        version: 0.1.0
      entries:
      - name: clusterpulse.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: clusterpulse
    provider:
      name: ClusterPulse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:61685ecb519d17160149a5dc16ae28d6fd96264652a6f19fef861addfa4ec6f4
        - quay.io/rasen708/ncn-operator:v0.2.2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:a99a5ca46c5bd3065905588f4d0e3cdc6409d435abdc02793be88e77c1a8f0f5
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:7d2ad8e800339c15efa5d5954106e01151d34135b25e9a23e30f0027ee23f193
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:eaa77ae0f1a1f3d32e269e73eb29a73d11012d5e920af1011d1842bd904d7293
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:4d9d003842e3ffc386f4e07ce7a81cfa85e676031fb6fdc96aefa9654244974a
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:a7641fc5f4c7aed4b73d7d98ccb58ef64cf14ab16d72378f56dba3ad4b8714c6
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:7dffd170b9b7d021b1615e6d746cbdfcbd819a1306ecdbaf042bcaf2202f7daa
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:111d1f4ab8f006386102d1bec4cd6a41b446d9ae401a43c821b2f318c8ba74c8
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:d7a4f9b8e5eb9bf6d11e00898eb168e86636f0633959eeb2b36b2d6e5e505f71
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle-rhmp@sha256:1f3d005224eada1b71d04f21d07a7c4fde159725c7a44d967d996152c802bede
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle-marketplace@sha256:7864f9f1cf3ad1d551b698b2e803302d207704b448b15f6c0013cbd9e854593c
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
          createdAt: "2025-07-16T09:03:58Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <2.8.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk@sha256:99aec4a74ea4c9910b50cc035b4c0c85ed4711a4c9b456028690d0f0c7af2657
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/splunk/splunk-operator-bundle@sha256:433bec6334c4771e140a24ea8753e7bcbf358c4504c942658fd3364915ac86ac
        - splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
        version: 2.8.1
      entries:
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
          createdAt: "2025-09-19T19:24:03Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-mq-controller@sha256:702ffd931cb4baafb1eac49f9fda5588d8b02f3b8c696e477dd626b0efad9f54
        version: 1.1.1
      entries:
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ssm-controller@sha256:922fd6daa660d2250bad8a3c54733cceae845624a14651cb20e049e29bf90e1c
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        - quay.io/openshift-community-operators/community-trivy-operator@sha256:e3652e68e1e569d7f40432fd8cbb7bfdc6738a116ecc1011c71a36f7ca5663c6
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:3b538de0bbaa43d2368402ea1c0eaf767c8b9e5a28e112a309148a630dd3c052
        - registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:5b61484c6a2bbc377435093bd9abb31df980ece470325cae92f3106963b6482c
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:af8ccc7a075f11bf6f31310830704e3a27b23587094f56e518b1f264c92b0fe4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: kubernetes-nmstate-operator.4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: kubernetes-nmstate-operator.4.18.0-202508271223
        version: 4.18.0-202508271223
      - name: kubernetes-nmstate-operator.4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: kubernetes-nmstate-operator.4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: kubernetes-nmstate-operator.4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: kubernetes-nmstate-operator.4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: kubernetes-nmstate-operator.4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: kubernetes-nmstate-operator.4.18.0-202505280403
        version: 4.18.0-202505280403
      - name: kubernetes-nmstate-operator.4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: kubernetes-nmstate-operator.4.18.0-202505061007
        version: 4.18.0-202505061007
      - name: kubernetes-nmstate-operator.4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: kubernetes-nmstate-operator.4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: kubernetes-nmstate-operator.4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: kubernetes-nmstate-operator.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: kubernetes-nmstate-operator.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: kubernetes-nmstate-operator.4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: kubernetes-nmstate-operator.4.18.0-202503190933
        version: 4.18.0-202503190933
      - name: kubernetes-nmstate-operator.4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: kubernetes-nmstate-operator.4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: kubernetes-nmstate-operator.4.18.0-202502250802
        version: 4.18.0-202502250802
      - name: kubernetes-nmstate-operator.4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/cluster-observability-operator-bundle@sha256:8c27ac0dfae9e507601dc0a33ea19c8f757e744350a41f41b39c1cb8d60867b2
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        - registry.connect.redhat.com/loadcore/keysight-wap-diagnostics-bundle@sha256:4e8400a38a04c61dc20ece1ee7dac1ad8a56ac309621f3225e7389e84e20c870
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator-bundle@sha256:85e75ec797f70c5807ec1ea25208cf0bb1990daa05eca9fbc82bc38aa4a840ab
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        - quay.io/community-operator-pipeline-prod/percona-postgresql-operator@sha256:44e687b38a48f68bbbb919e6df0d3b87b75492aca7bd41e508a2aa0e80d024a9
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        - quay.io/community-operator-pipeline-prod/ack-elasticache-controller@sha256:a489c815e9d3777d775ccd65bbd0ecd5fe7c577de36c504b63cf1620531d4ccb
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:301c3b793d8068f9e50c855196da91b6c07111ef500b02a91fbb354c9fc1316b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:8edd2f89d7adda6c1a5aae7414f50e0d786ec8443c1db13834ea4924009b6419
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:227159c5a93966a43fa3a84cb69536ff0051cc261da068561241261d3837998a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:60635156d6b4e54529195af3bdd07329dcbe6239757db86e536dbe561aa77247
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:e4d49cf6a2fb8b3585ecfe3b3da3e9de090fb99de3dee2da0601b1f6b071173c
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:095e93f8dd31a0985d6fef6a6168da456630febe89ab9e0cb9480170f1036541
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:524d315efa226f97ff26b8668f9297e755c6f0c815b8f3385b317667553f0e62
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:f5f025c7f56643fab9148bb5de8ee34a84d02b78a6abb4b5592306cf54296654
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/smart-gateway-operator-bundle@sha256:f12b2f1dd34c73b9c8a5ad461e771c0ce56c9ea4eb88f03225be1462a42c5a78
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:0cda4be8f74da7524a72c961c8762927c66e145d82e59af92073a54179bc9936
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:4bab095c5dd6f8c95a17a8ff2bf70ff3f53faea48b56eea8baa9cf9a42a5227a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/keycloak-operator@sha256:a4da3bcbb4e99d60b8217d9e62742eeb83eef163dfefb2f7efda131f312e46e3
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        - quay.io/openshift-community-operators/keycloak-operator@sha256:65fe00f310289b70538a6dde897ef663731d0da44a7bb1d0f55a3001a50e53fa
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/keycloak-operator@sha256:70bbd44f6ecf86aa8e9710175a96dd3c1d813aec40667a09800241ce30afd93f
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
          createdAt: "2025-06-19T11:07:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-zookeeper@sha256:c8487e18318a92a64e84a5ab29f844a0973f3b8b2aa291723014731471459176
        - docker.io/confluentinc/cp-server@sha256:072719eef75ee61294027ba77f7b9cd821437158761495a1644b5687af9305a8
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:7a271565db9c500f8c5755de397e8093251673d2fc6b8a10974f927f382b4704
        - confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2520a514b848b53338c411f0e0dfa68491596761f8bbbb090d652f71eda010af
        - docker.io/confluentinc/cp-kafka-rest@sha256:9c19e7900669e555810d26676393737a7d573b39c31e284c0ad78807bfe05ce3
        - docker.io/confluentinc/cp-server-connect@sha256:05e77fe8e19747521e1f1d5630bbf384b0fe0ec9efe8dc798de6e4a3eb0bfa2d
        - docker.io/confluentinc/cp-schema-registry@sha256:5b6bf5b971dd8fe2c821e227391858f58f6ae2128f1ae1faa988c3ea59b86801
        - docker.io/confluentinc/confluent-init-container@sha256:550a6b92fc5aca81ad8e0d3d47c0063af9da877e45c1ecda1a9cc44a70e5d87a
        - docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df1f6a99ecfe8d8acc0af222848aad13349f06071fc2bf1360d6d20f99769394
        version: 2.10.2
      entries:
      - name: confluent-for-kubernetes.v2.10.2
        version: 2.10.2
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.2",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
          createdAt: "2025-06-19T11:14:20Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:b7cde5736b5275abafafbb81f4c917890abfe8f400f436c3a1f712cf200832f8
        - docker.io/confluentinc/cp-schema-registry@sha256:4f6cabeea11ae532cb47905eadaa3192d7113ea6d0a0da310cdfd5eb7142fc45
        - docker.io/confluentinc/confluent-init-container@sha256:37161a90b89dfad625bf61ba7d299707d7b2a96271e9667044c3eaba3ffc5f66
        - docker.io/confluentinc/cp-server@sha256:dfed43a1618de1fea6d202921ca4cc2168f7455c3d37fe046f3e64febf042a87
        - docker.io/confluentinc/cp-zookeeper@sha256:f1e6d3d698a6e2757e3c225b5c6f8f007251838226c46f652a4cbb4314b50277
        - confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:2baa0cb5df8017d9d14c2f0c7e02ce371ba92462d8d1e93a3ca1bc548c20a74e
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:507e1538a64c4232ffda86504594def45cb20bbcb6b13fc6196c1f41e71f93bf
        - docker.io/confluentinc/cp-ksqldb-server@sha256:744f5d844ac223c857116f3994c5be2105898ea3896d53026eb62e76c7240c4f
        - docker.io/confluentinc/cp-server-connect@sha256:73822ffff10d386b98c9ca2bb295b4d079deb210116c8e4283cdc26ce304526f
        version: 2.11.2
      entries:
      - name: confluent-for-kubernetes.v2.11.2
        version: 2.11.2
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.0",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
          createdAt: "2025-06-13T13:02:11Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:ba49d46654770994c302d586a609afdfc58a7cc05f204aa090b643f0680a5014
        - docker.io/confluentinc/cp-server-connect@sha256:e4cbc1836402497ea42b2c961c5d6b4420c25b3c6a5f6f71be3e99fcc0f09702
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:a283f2a72b6346febaf711e71dee1e0e3666d7265320f3f69307fc2c0b95d5ef
        - docker.io/confluentinc/cp-schema-registry@sha256:3dcb9652a4c83d4566a7334041c1ac3ec359665a9450a70c5d7b41adaea55c90
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:35f2799f71c308d87e6a7dada00263abb9e08e895b10f10c815e7a9edf19dd49
        - docker.io/confluentinc/confluent-init-container@sha256:82277228368a783ba84367c920a9fa15fad6844723df1e16b07c9dcc829a54ad
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-server@sha256:ffd5961412da42c81c9b20f35e334aa8bdf351dcb452610e84156ed511116ef5
        version: 3.0.0
      entries:
      - name: confluent-for-kubernetes.v3.0.0
        version: 3.0.0
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator-bundle@sha256:21b69b5eb896a37691a8e086f3a51c0f68086a0ff9370fd66667acc6c935a850
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/mta/mta-operator-bundle@sha256:65c0932bfda32b1e62be9edd0f96bb4d6e41e302a2e8080fc6cc794ee5c52276
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-operator-bundle@sha256:f3d585e24d65905b2e9f594dcdb67721aa95057ab7587ae531c3487aa2f94e53
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/mta/mta-operator-bundle@sha256:a13643117c2867351718a872f7f1b2350c67855ca73a727a1dc140754ffe6589
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/mta/mta-operator-bundle@sha256:9eeb43af2bcab84f5261d1575f7c897903a7696dba011d256abffa1fe850eba2
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-operator-bundle@sha256:6aef520983fe01351886e108d32b998b20e74116fc0280c754a914bfad03d0bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ed10e35f45a140e21d826bee71b3ab96e67f1b4dc8c2f7f1676380fa51b67fe
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        - registry.redhat.io/mta/mta-operator-bundle@sha256:68953a2a721309fd103e09f3171d0c9d50020413631b575f1a44602a5d2ac855
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ec977b0c9c26551db110dcf2d5069647b1a5495f869ff48a0319cb496b84cdc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - registry.connect.redhat.com/eginnovations/universal-agent-operator@sha256:65f851074e0e1db57a62b559c0f10f9d30543011727aec3b832f9a35468139e9
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator-bundle@sha256:8d130a5f2407931dbc73e45a53610b95ceb04cb507b8f6519f630cada1c0da45
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:9009198860c7a1e0de4f429261d442b75b5b2893a943bb77914b14d643db8829
        - registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/openshift-community-operators/sosivio@sha256:110d7f4fb993d65ff497eca57c42206faab07dc5f5b1f43b1b682387e7cd3a10
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        - quay.io/openshift-community-operators/l5-operator@sha256:e4cb13801a80753c99a47f009ca0a29416b3e04afe5314d07d97c2464a6f0d75
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:712d31214423b326ab8f107435c85431ce8d99c2e52fd8d615005a4274e0ff98
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/operator@sha256:2cb5ec186af3264cba42333245616dd01f0f9868a849129b1a31afbc51ee1410
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - registry.connect.redhat.com/harness/gitops-operator@sha256:f97cdc2245b7133a04235c55540d3a87165e080b1d5b9572a663731f6ccbe597
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        - quay.io/openshift-community-operators/tf-controller@sha256:cdaeb470d9b331a3fa08167e5c904befb00be983e1354651670e21e13560bf9b
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.18.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
          - description: L2ServiceStatus is the Schema for the l2servicestatus API
            kind: ServiceL2Status
            name: servicel2statuses.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-metallb-operator-bundle@sha256:c80c3467cf52fc4b2a55e444fba9fdc8cf3b1efedad5d1276a889d07aed44e68
        - registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:6f23dd225f2bccfe39d05428a51de990a4a9338c49da079384eb30fb470c04f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509100344
      entries:
      - name: metallb-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: metallb-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: metallb-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: metallb-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: metallb-operator.v4.18.0-202506251003
        version: 4.18.0-202506251003
      - name: metallb-operator.v4.18.0-202506070804
        version: 4.18.0-202506070804
      - name: metallb-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: metallb-operator.v4.18.0-202505081435
        version: 4.18.0-202505081435
      - name: metallb-operator.v4.18.0-202504291304
        version: 4.18.0-202504291304
      - name: metallb-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: metallb-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: metallb-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: metallb-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: metallb-operator.v4.18.0-202503181802
        version: 4.18.0-202503181802
      - name: metallb-operator.v4.18.0-202503110933
        version: 4.18.0-202503110933
      - name: metallb-operator.v4.18.0-202502261233
        version: 4.18.0-202502261233
      - name: metallb-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: metallb-operator.v4.18.0-202502041632
        version: 4.18.0-202502041632
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:latest"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:latest
          createdAt: "2025-01-03T19:50:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The eBPF Manager Operator is a Kubernetes Operator for deploying
          [bpfman](https://bpfman.netlify.app/), a system daemon\nfor managing eBPF
          programs. It deploys bpfman itself along with CRDs to make deploying\neBPF
          programs in Kubernetes much easier.\n\n## Quick Start\n\nTo get bpfman up
          and running quickly simply click 'install' to deploy the bpfman-operator
          in the bpfman namespace via operator-hub.\n## Configuration\n\nThe `bpfman-config`
          configmap is automatically created in the `bpfman` namespace and used to
          configure the bpfman deployment.\n\nTo edit the config simply run\n\n```bash\nkubectl
          edit cm bpfman-config\n```\n\nThe following fields are adjustable\n\n- `bpfman.agent.image`:
          The image used for the bpfman-agent`\n- `bpfman.image`: The image used for
          bpfman`\n- `bpfman.log.level`: the log level for bpfman, currently supports
          `debug`, `info`, `warn`, `error`, and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`:
          the log level for the bpfman-agent currently supports `info`, `debug`, and
          `trace` \n\nThe bpfman operator deploys eBPF programs via CRDs. The following
          CRDs are currently available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n
          - KprobeProgram\n- TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n-
          XdpProgram\n - BpfNsApplication\n- TcProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n\n ## More information\n\nPlease checkout the [bpfman community
          website](https://bpfman.io/) for more information."
        displayName: eBPF Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - registry.redhat.io/bpfman/bpfman-operator-bundle@sha256:b6d43f1fec6ba7336c23d1fbceab15ebb2ed9e9e2eebb08c78d7541ee47b4204
        - registry.redhat.io/bpfman/bpfman-rhel9-operator@sha256:c3d878d77091bed03d99f72943d9b78e210e2f6194ea4b5cd2aa23f8d48f3f61
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      name: latest
    defaultChannel: latest
    packageName: bpfman-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-applicationautoscaling-controller@sha256:4f1b92eac30d72a66d8fe8c2c44339c7a5fd6af65d48c0d8697459c7af290741
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Agents",
                "metadata": {
                  "name": "loadcore-agents-sample",
                  "namespace": "keysight-loadcore-agents"
                },
                "spec": {
                  "agents": [
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent1",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "replicas": 2,
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    },
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent2",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    }
                  ],
                  "image": {
                    "registry": "prod.harbor.keysight.digital",
                    "repository": "loadcore/loadcore-agent",
                    "tag": "UBI8-0.1-7d990d7254"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "regcred"
                    }
                  ],
                  "networks": [
                    {
                      "config": "{ \"cniVersion\": \"0.3.1\", \"type\": \"ipvlan\", \"master\": \"ens192\", \"mode\": \"l2\", \"ipam\": { \"type\": \"whereabouts\", \"range\": \"192.168.1.0/24\", \"routes\": [ { \"dst\": \"0.0.0.0/0\" } ] } }",
                      "name": "loadcore-agent-network"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:53:55Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:f1d1e9f2aca11505407785bc33905744b667bc8aea49df73fcc50db379c0b375
        - registry.connect.redhat.com/loadcore/keysight-loadcore-agents-bundle@sha256:f16fcea424cb7b1eb647f3e9d3c08bd93c8a1b8bb7eab71a4e756da72a7a56a0
        version: 0.1.21
      entries:
      - name: keysight-loadcore-agents.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:a58c9c8454ac337c5a359dfe5b7c6b06983b7050ad62a0d11054f06fc17e92c1
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.89.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.89.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:f6c37a0808c137f226b6afb6cac1e2ce05184d6a01b9234969ce63056571e991
        version: 42.89.0
      entries:
      - name: t8c-operator.v42.89.0
        version: 42.89.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - quay.io/community-operator-pipeline-prod/gitlab-runner-operator@sha256:3e0314a2695c071fe0a34f3dc7684b43b5a31b6705f42982b4dbaacabe360d2b
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-operator-bundle@sha256:2bab9fa63b128a3ba0a6031f70f0342f95d17dd53a3ec2930b19940b5b2e6f47
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - registry.connect.redhat.com/eweb/cloudnfv@sha256:0bb427954fb8df5de197353908a5189922f514e121ae8cdb8242d125b1b3807d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - quay.io/openshift-community-operators/odoo-operator@sha256:616e1067ebf8deb66636d57cae2ac1f9981b83419d8e7f023bc7df1c1b745536
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/docling-operator@sha256:9a3717ee195572b29d1ee31b96bb95561f5521b032dbaa09a000413b227b5f12
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
          createdAt: "2025-09-19T21:59:41Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatch-controller@sha256:254df5f0e6821a65277bc3d029e7af87b10800b192f9724997996224525a9943
        version: 1.2.1
      entries:
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        - registry.connect.redhat.com/scaleops-sh/scaleops-operator-bundle@sha256:efd483d9c826e25b0d1cd79deff4263772260c72da79bad149ef7985cc00a44b
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:54bb383ac6c5f57ae77dcd96cdc33cea5c4c1aff9f9d356c1721591aa820e82d
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:948ad5eb47b7809ba5dd78cb6bb30b1bc64163617f229dc33cec251c1d5bb20e
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        - quay.io/community-operator-pipeline-prod/lbconfig-operator@sha256:b0bb724a5d9905633962a3e8267e75d371f754bce866a29cba3ac0d0c671ebff
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        - quay.io/community-operator-pipeline-prod/apicurio-registry-3@sha256:e207a4f7f53486b78f95ad958f7e0053a8a8561c3dcb1463a7176f27eb9ea082
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:610a5e66cf1b6c7f5451c9a0f5e0a6cc9435f2548335fbd50f44b8656adadd99
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:bc0d8b62d5814407e08817ece44626400cc3d9d29b14aa35dfe711b907fddf21
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.connect.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:15e8cbce994b783fa6dac737ac28f1b713d1c75891dbce4987d2c09ff295376b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:12d6ae3f139daf33dbf104499c195d9915470b957f401b5f622cfde92350a6c1
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:965ec7875eccc6a5fc56c3b8f0da668d826d34580117ba1438a607a9d0d13b73
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:f8a8cfcebba4b17fa6cfa99b3ef430af272c5237476dba4665803a90adfd5d2a
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        version: 3.74.9
      deprecation:
        message: |
          The `latest` channel is no longer supported.  Please switch to the `stable` channel.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:7fd7595e6a61352088f9a3a345be03a6c0b9caa0bbc5ddd8c61ba1d38b2c3b8e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        version: 3.62.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:5187b04f8010a274d6bc017fc82f1b00d886124e520c1b7de8b4da8cd2f37bc0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        version: 3.64.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:d3bf33899ca8b199ca8782978618ca343164e43fd0ae2ac6a0cfbf697efa6428
        version: 3.65.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6536ee795e8aaeab9321ee4820e55933195983a768ba24fab40508d03e50a6e4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        version: 3.66.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:ad88773ab141f45fe24c61e83ec736aab27f652fb82ca2a85fee6f885568c5b9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        version: 3.67.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6a10a4859527946f61f497ff84b80f973cb7be5982d428423836b5aca81a1f5d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:297e7f863106c6b41f882bbd5b691a12c797c24d832473a04e604d4639e4a68e
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        version: 3.69.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:9b35ae6df2c19935783735d213483f3fcecead2f49f873a4f7844ef9fb5c7f1a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        version: 3.70.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e001b9f844a6c298dd74f1ab683f3dbbf0ad66044cec419c373ef4f97d47be73
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        version: 3.71.3
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:878ed77d07d3f40f049a594af44376797305786a2a264dae3d0c50401643bfa5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        version: 3.72.4
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:459a8d91beff91638872d377a8b82c279ad2905ce3f4ad4bfc00f6822418ac57
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        version: 3.73.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        version: 3.74.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:a0d9cd31d953aaf7e84f5a39f405c5aa2ea62e68d6c728a936ca1ffc4dc1b93e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        version: 4.0.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:b26133843687cc50e2c52d6067d8b3369b5cb28dfbfd15dd4b6062096673195a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        version: 4.1.6
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:08aebe6e8275d3314b6756d522ec568ccdfb9c52df98c9822f187f6c5a3af19c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        version: 4.2.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:30f9384244819da013358325cec364bb210b3b33b4cf9680698531fe50562aac
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        version: 4.3.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:123c0108bd7a190dd2accfd21786600f787147f040e57fb5f840861e2caa441d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        version: 4.4.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:41351e88688d23a443ad4f141caf0433b2c2789e6a1a85b4f5b36abeb5656bb1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        version: 4.5.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6d6b2696a12d49cf3403e338a0e028187473b6f43ef2b0bd72c7701200ac8e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:c7f894ab773d1d389656d13ed88fbf575faddd95c5b8396c6593ab3f993f331e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-bundle-rhm@sha256:5d14e0c44ffb543b7cf59817fe587c8e819542d8edfa261a44e58aea0ecdb79b
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        - registry.connect.redhat.com/randoli/app-insights-operator@sha256:21ed3d9b8b8466665c1cb784d0bc244279d0730f6319625caf46af111c25d333
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
          createdAt: "2025-05-09T05:49:47Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-operator-bundle@sha256:f49b7a2c60e79edaba1a5faabc3983f87f9123b6d84121a438d179ccbb8102e3
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
        version: 1.1.1
      entries:
      - name: multiarch-tuning-operator.v1.1.1
        version: 1.1.1
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.9-r3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
          createdAt: "2025-08-05"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:327a9899adfe89af9aebdcfac651a9eddec9ecf8a345caee1cef1b4a1fb279d5
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:df0493f7edbdedf754c70e4efc256fa4d246a24d6c8882c562e56ef3844278c1
        - registry.redhat.io/apicurio/apicurio-registry-3-operator-bundle@sha256:728354aee3729a1fba233c2aa92cacb21d27efd4869706ddd214a2d3688e33b2
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
        version: 3.0.9-r3
      entries:
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:d9930900b0d2f92ae813246fc322290234ad8bbd165b10e36671b53852228718
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:87448055a162e215a4d4982a8494fec6e7cc59fe6704494ddfdcfea076a2100c
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:a7c6f1c9260c79ebbdfb25ea996516ba1e2bcbedf25401e9e6848b955a2d3d78
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-operator-bundle@sha256:9ab3d5eb70c654270fef9e8e9d4e5b68e1ac3c86bd0e0dcd36cc6f8741be1603
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:36561e5888cd34f8fb83ceff2910141fceaa6ea290bd246125930fe183e2ebc3
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:9e8bcbff46b97f3bc1071146b8db233abd1630016e6fc0efb679df8243c6fd03
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
          createdAt: "2024-07-10 00:59:51"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:518691d157b3e3b5031a2b1e4cc3f415839536079bb8570ae8690ac92d302ba0
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:a64f67f29a4eb8ce426ed842f1a5a5fcd83fd9d91acebb064cf65c89c9d8bfb6
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/keepalived-operator@sha256:5a46fcb72f02a1f0d9cbc72f44948ef4b09149a4b3fe2e45fd31df90d20b3c56
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-keyspaces-controller@sha256:6e56400162867a4f55a70c340aeeb074726fd3dca13b6da60c395d6f35a48b48
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/automotive-infra@sha256:6d4180ff6899d4de8ba2cd4a06ece234bbd8f0ec1bd0b4b0ae4b7a0c047121da
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        - quay.io/openshift-community-operators/aws-efs-operator@sha256:6372847df81faddc23c938928157c3ac280e537a4eec9f7edd599480260bc566
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/cluster-impairment-operator@sha256:6fd114eb8f525f670e505373c2999a4e66b53beda988e99b267204d2fe0e1b4f
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:8197e4a96ac8949d43cda62ac2257a9b1e71f0b54f42283ee4925f88c369f873
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:728074151b245f5e0d65bd8956149f05ba22a1b6438120ebca2d85b0d7d0c837
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:4109698147632b858fe2787964772a1339cdebee72b75b42d9b1c5dbbbecb45c
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:088d0eebf60c8f6961ca72c268f61d12648d575aa41552ae6398fefc6d9700cb
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d6d444e642bfd567eb40c7e084f5df743c23b47935e52e35ef541c67fcc00433
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:b3469cd953ef2015fad2f7f96d960cee5ecaa6c19df631a9c882154ab5b3de98
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:f49d3eddbfcec1018fd3716cd0ec2addef9961d2d308099636541868566979b0
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        - quay.io/openshift-community-operators/t8c@sha256:403b29a6262ba0e06de117f788aea7224268b798868c3e5b738b05e96df20456
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4bfaf97ab09a1aff1316536b99957b1f82763b009dde6d08fc9eea1fdd290fbf
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        - quay.io/openshift-community-operators/t8c@sha256:61f1ea9b3851d7272703f5bc21203ff03bb35fd52ed9780d0cf996379c78a00b
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:48a5a9fbfa92b4b87b3653613467a3fcceacadbbabd274e367b40f6871039cc8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d65834d8f9946c2764d45666c90c50f428b74afe7b066d666c18e9545994c9a8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4964196aba839fde569ce58ab7e4eef438c9638ae4057194f2d7a58eaa5a98d9
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        - quay.io/openshift-community-operators/t8c@sha256:46c815fa635d58820757f41b1d8f04333638c39471f4d6c901a6500379680b57
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d8a575a84dd86d241acaaec53af1e164bc1ff1b95b5eb710851ae698bbed6bf5
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:dfdc8fbdfc0ee76b2d5c62f6fd5deac2c50756661871d8b4357e25043e5ce66a
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:081d7f4f20332599fdbbd07f22cc18eca6c5625b619efeae6e861239cde4b12a
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:cf7ccc1a60d6787ebcd1d76415cca4cc55c2c3ee18d52c0511e66ca6bf9310f8
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: loki-operator.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
          createdAt: "2025-06-11T13:21:40Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b311677919183c6c91c6da40b0ddb2529f92ed52b2c46a07e399a4bad5dea280
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:fce287241be64b7b2c20a971c4f2d468c079ae81556aaab756572ebad27a17f5
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:ff98532bd8f21ecf375062d465f2aee4cdcd70b647d258c1e09f7e2286f998c9
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:ea2d995744cd4a4bd6e7467c2d6927bfc89a47d551ef0d6b048c0c57652f81be
        version: 6.3.1
      entries:
      - name: loki-operator.v6.3.1
        version: 6.3.1
      - name: loki-operator.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b793aef88881aa75d47a22377d8932efc0b497bb01b2f3f376d64891d1dace54
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-operator-bundle@sha256:722cf3c05eb5663a737cb1385b971b6d1f0c8b2595d5ac425f7b100c25e2fd26
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-operator-bundle@sha256:ed97b11fa0f0f7aadf193810a5a599bf26015a0af8ee6fd90061110dcb0d50f4
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f19ae051041c46fe7587f903a9c6221bcbe8be2c4cc0235739f7592eb93ae164
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:28e37ec35de1307405264cd4f79aaa6ce4d4ecbd6cbc27f4fef5db415c678d64
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        - registry.redhat.io/quay/quay-operator-bundle@sha256:4a004584fa6c7187481e8c8480f17e0e4edd18c32f52a2e271c68d02f1a9655f
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f16c0a135dd7380e2a489f963b38bd76a63a973f482d06ece8d3656aff223703
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b499eebed32c0261db414e307f96f8aef9e743035544c80c048526e7ba4fbdc2
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-operator-bundle@sha256:55a5e6297109aea029bd40fac09a9fbc10c28a8c5558f9afc9b9b86c5dc7a4d4
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-operator-bundle@sha256:e3f63ea56de7ca5388a20c17ac63e16f07935ab09bcdcd38a86639c9d59c8a7d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-operator-bundle@sha256:53fcddadaa31a5f7d74c38f0bbd0d80b4c09796a580ef8494f60ff82fb6fe6c1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-bundle@sha256:9d0ed6d279800a189c1466e0b32f80af2f8bad96a65b7dc0322c64337dddf39c
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:9d21b738b0d357ef064690eae66bd4b3810664bcde57237822f1e90e9fd6ad03
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - quay.io/openshift-community-operators/reportportal-operator@sha256:67a7ffaf6a303b92541170f266d7b2871f5c2fe73f6b6c9a2b7b33dbc4555fa5
        - quay.io/waynesun09/rp5-operator:v0.0.8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/openshift-community-operators/etcd@sha256:e42581eacbc8349f40c6a4ed6d7d82d6168518e15b1a24cba36817205bd8688f
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:1d308347e5e290ac4bb7dd7aab0a1c0659d661153541d81dd30733b52d1fbf83
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:f1757deff8c4048ae42cb033f3ddd7686dec2968e5a4cd61ff8f74fda69a8a09
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - docker.io/grafana/tempo-query:2.8.2
        - docker.io/grafana/tempo:2.8.2
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - quay.io/community-operator-pipeline-prod/tempo-operator@sha256:14282a0ca95401ab3e4c11dc2b611a6f57c00c461254f52ddf03ef74c725767b
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/jboss-webserver-5/jws5-operator-bundle@sha256:ce4060138117dbc8fc21abb60f0b8ef1bfa5ebc0dd8eb803c0e79d7447341964
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9y
            dCgzMzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQ
            LmFkZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAg
            ICAgICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24i
            LAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJh
            c2gpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3Zp
            ZGVkIGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAog
            ICAgICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmls
            ZXN5c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IG1ldGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBk
            ZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2Vw
            aGZzIGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1l
            bnQoCiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92
            aWRlcyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9v
            bC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRh
            IHBvb2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBp
            biB0aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAg
            KQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1l
            bmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZh
            bHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+
            OjxQT1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCki
            LAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAg
            ICAiLS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAg
            ICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBv
            aW50IFRMUyBjZXJ0aWZpY2F0ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRf
            YXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1za2lwLXRscyIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJJ
            Z25vcmUgVExTIGNlcnRpZmljYXRpb24gdmFsaWRhdGlvbiB3aGVuIGEgc2VsZi1zaWduZWQgY2Vy
            dGlmaWNhdGUgaXMgcHJvdmlkZWQgKE5PVCBSRUNPTU1FTkRFRCIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW1vbml0b3JpbmctZW5k
            cG9pbnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgZW5k
            cG9pbnRzIChjb21tYSBzZXBhcmF0ZWQgbGlzdCBvZiAoZm9ybWF0IGA8SVB2ND5gIG9yIGA8W0lQ
            djZdPmAgb3IgYDxGUUROPmApIGVudHJpZXMgb2YgYWN0aXZlIGFuZCBzdGFuZGJ5IG1ncnMpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tbW9uaXRvcmluZy1lbmRwb2ludC1wb3J0IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAg
            ICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkNlcGggTWFuYWdlciBw
            cm9tZXRoZXVzIGV4cG9ydGVyIHBvcnQiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1za2lwLW1vbml0b3JpbmctZW5kcG9pbnQiLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBhY3Rpb249InN0b3JlX3RydWUi
            LAogICAgICAgICAgICBoZWxwPSJEbyBub3QgY2hlY2sgZm9yIGEgbW9uaXRvcmluZyBlbmRwb2lu
            dCBmb3IgdGhlIENlcGggY2x1c3RlciIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgZXJhc3VyZSBjb2RlZCBSQkQgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWRyeS1ydW4iLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAg
            ICAgICBhY3Rpb249InN0b3JlX3RydWUiLAogICAgICAgICAgICBoZWxwPSJEcnkgcnVuIHByaW50
            cyB0aGUgZXhlY3V0ZWQgY29tbWFuZHMgd2l0aG91dCBydW5uaW5nIHRoZW0iLAogICAgICAgICkK
            ICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yYWRvcy1u
            YW1lc3BhY2UiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1G
            YWxzZSwKICAgICAgICAgICAgaGVscD0iRGl2aWRlcyBhIHBvb2wgaW50byBzZXBhcmF0ZSBsb2dp
            Y2FsIG5hbWVzcGFjZXMsIHVzZWQgZm9yIGNyZWF0aW5nIFJCRCBQVkMgaW4gYSBDZXBoQmxvY2tQ
            b29sUmFkb3NOYW1lc3BhY2UgKHNob3VsZCBiZSBsb3dlciBjYXNlKSIsCiAgICAgICAgKQogICAg
            ICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXN1YnZvbHVtZS1n
            cm91cCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgc3Vidm9sdW1lIGdy
            b3VwIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tcmd3LXJlYWxtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAg
            ICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhlIG5hbWUg
            b2YgdGhlIHJndy1yZWFsbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXJndy16b25lLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlk
            ZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cC1uYW1lIiwKICAg
            ICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAg
            ICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSByZ3ctem9uZWdyb3VwIiwKICAgICAg
            ICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9w
            b2xvZ3ktcG9vbHMiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJl
            ZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdG9wb2xv
            Z3ktY29uc3RyYWluZWQgcmJkIHBvb2xzIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3Vw
            LmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9wb2xvZ3ktZmFpbHVyZS1kb21haW4tbGFi
            ZWwiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwK
            ICAgICAgICAgICAgaGVscD0iazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gbGFiZWwgKGV4YW1w
            bGU6IHpvbmUsIHJhY2ssIG9yIGhvc3QpIGZvciB0aGUgdG9wb2xvZ3ktcG9vbHMgdGhhdCBtYXRj
            aCB0aGUgY2VwaCBkb21haW4iLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdGhlIGs4cyBjbHVzdGVyIGZhaWx1
            cmUgZG9tYWluIHZhbHVlcyBjb3JyZXNwb25kaW5nIHRvIGVhY2ggb2YgdGhlIHBvb2xzIGluIHRo
            ZSBgdG9wb2xvZ3ktcG9vbHNgIGxpc3QiLAogICAgICAgICkKCiAgICAgICAgdXBncmFkZV9ncm91
            cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikKICAgICAgICB1cGdyYWRlX2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFkZSIsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAg
            IGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9yIGV4YW1wbGU6IGNsaWVudC5j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0aGNoZWNrZXIgY2VwaCB1c2Vy
            cyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBuZXcgY2x1c3RlciB2ZXJzaW9u
            IGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBwbGllZC4iCiAgICAgICAgICAg
            ICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rl
            ci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVwZ3JhZGUgYWxsIHRoZSBkZWZh
            dWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAgICAgICsgIkZvciByZXN0cmlj
            dGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVu
            c2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2luZyAtLXJlc3RyaWN0ZWQtYXV0
            aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9yeSBmbGFncyIKICAgICAgICAg
            ICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29sLW5hbWUsIC0tazhzLWNsdXN0
            ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUgdXBncmFkaW5nIgogICAgICAg
            ICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3UgaGF2ZSBwYXNzZWQgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcgdXNlciB0aGVuIHdoaWxlIHVw
            Z3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAgICAgICAgICsgIlNhbXBsZSBy
            dW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMu
            cHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxpY2Fwb29sIC0tazhzLWNsdXN0
            ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGllbnQuY3NpLXJiZC1ub2RlLXJv
            b2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAiUFM6IEFuIGV4aXN0aW5nIG5v
            bi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0byBhIHJlc3RyaWN0ZWQgdXNl
            ciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRlIGZsYWcgc2hvdWxkIG9ubHkg
            YmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVzZXJzLCBpdCBzaG91bGRuJ3Qg
            YmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxpZWQgcGVybWlzc2lvbiwgZm9y
            IGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2ggcG9vbCB1c2VyIGhhcyBhY2Nl
            c3MiLAogICAgICAgICkKCiAgICAgICAgIyBBZGQgY29tbWFuZC1saW5lIGFyZ3VtZW50cwogICAg
            ICAgIGNvbmZpZ19ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb25maWciKQogICAg
            ICAgIGNvbmZpZ19ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWNvbmZpZy1maWxl
            IiwKICAgICAgICAgICAgdHlwZT1zdHIsCiAgICAgICAgICAgIGhlbHA9IlBhdGggdG8gdGhlIGNv
            bmZpZ3VyYXRpb24gZmlsZSwgUHJpb3JpdHk6IGNvbW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmlu
            aSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgKQoKICAgICAgICBpZiBhcmdzX3Rv
            X3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAogICAgICAgICAgICAgICAgdHlwZShhcmdzX3Rv
            X3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICksICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJz
            ZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgYXJnc190b19w
            YXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJldHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190
            b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6
            CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCgogICAgICAgICAgICBpZiByYmRfcG9v
            bF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBpZiByYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0
            YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAg
            KQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQi
            OiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            IG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCdd
            fSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21z
            ZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZh
            bHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtleSBpbiBqc29uX291dFsicG9vbHMiXToKICAg
            ICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wg
            bmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBvb2wKICAgICAgICAgICAgICAgIGlmICgKICAg
            ICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAg
            ICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bv
            b2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxl
            IGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAg
            ICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29s
            X25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgICAgIHBvb2xfZXhpc3Qg
            PSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAg
            ICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9wb29sIG5hbWUsIgogICAgICAgICAgICAgICAg
            ICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBwb29sX2V4aXN0OgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            IlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwge3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBl
            eGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihzZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIg
            KyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoOgogICAgICAg
            ICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rp
            bmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAgICAgY29udGVudHMgPSBmLnJlYWQoKQogICAg
            ICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJzdHJpcCgpCgogICAgZGVmIF9jaGVja19jb25m
            bGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJF
            aXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRhdGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBz
            aG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAgICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBv
            aW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAgICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdl
            dHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRlbGltaXRlcgogICAgICAgIHRyeToKICAgICAg
            ICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEp
            CiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikK
            CiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgog
            ICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4o
            ZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAgICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAg
            ICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNzKGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRB
            ZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAgICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJv
            cjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUUROIgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGln
            aXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQg
            bm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGludFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBp
            ZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioqMTYgLSAxOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9y
            dH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoKICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYs
            IGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMg
            aWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0
            IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBvaW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShz
            ZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gIiIsICIiLCAiIgog
            ICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJz
            cGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            Ik5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSAhPSAiWyI6CiAgICAgICAgICAgICAg
            ICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRwb2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAg
            ICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2VuZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3Ry
            X3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBbImh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJl
            c3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZvciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlcCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRf
            c3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9IE5vbmUKICAgICAgICAgICAgICAgICMgSWYg
            dmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBkaXJlY3RvcnksCiAgICAgICAgICAgICAgICAj
            IHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4gcHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFz
            aCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNTTC4KICAgICAgICAgICAgICAgIGlmIHByZWZp
            eCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RsczoKICAgICAgICAg
            ICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1GYWxzZSkKICAgICAgICAgICAgICAg
            IGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNlcnQ6CiAgICAgICAgICAgICAgICAgICAgdmVy
            aWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1l
            b3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAgICAgICAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10aW1lb3V0KQogICAgICAg
            ICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAyMDA6CiAgICAgICAgICAgICAgICAgICAgcmV0
            dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAgICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVy
            cjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vycm9yID0gZXJyCiAgICAgICAgICAgICAgICBj
            b250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRv
            IGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNw
            b25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICIiLAog
            ICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCItMSIpLAogICAgICAgICkKCiAgICBkZWYgcGFy
            c2VfY29uZmlnX2ZpbGUoc2VsZiwgY29uZmlnX2ZpbGUpOgogICAgICAgIGNvbmZpZyA9IGNvbmZp
            Z3BhcnNlci5Db25maWdQYXJzZXIoKQogICAgICAgIGNvbmZpZy5yZWFkKGNvbmZpZ19maWxlKQog
            ICAgICAgIGZvciBhcmcgaW4gbGlzdCh2YXJzKHNlbGYuX2FyZ19wYXJzZXIpKToKICAgICAgICAg
            ICAgIyBweXRob24gdHJlYXRzIGZsYWctbmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28g
            Y29udmVydGluZyBiYWNrIHRvIGZsYWctbmFtZSwKICAgICAgICAgICAgIyBzbyB3ZSBjYW4gZ2V0
            IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAgIGFyZ3VtZW50ID0gYXJn
            LnJlcGxhY2UoIl8iLCAiLSIpCiAgICAgICAgICAgIGFyZ3VtZW50VmFsdWUgPSBzdHIoZ2V0YXR0
            cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcpKQogICAgICAgICAgICBjb25maWdfdmFsdWUgPSBjb25m
            aWcuZ2V0KCJDb25maWd1cmF0aW9ucyIsIGFyZ3VtZW50LCBmYWxsYmFjaz1Ob25lKQogICAgICAg
            ICAgICAjIGdpdmUgcHJpb3JpdHkgdG8gY29tbWFuZCBsaW5lIGFyZ3VtZW50LCBpZiBjb21tYW5k
            IGxpbmUgYXJndW1lbnQgaXMgbm90IHByZXNlbnQgdXNlIGNvbmZpZy5pbmkgdmFsdWUsCiAgICAg
            ICAgICAgIGlmIChzdHIoc3lzLmFyZ3YpLmZpbmQoYXJndW1lbnQpID09IC0xKSBhbmQgKAogICAg
            ICAgICAgICAgICAgKGNvbmZpZ192YWx1ZSAhPSBOb25lKSBhbmQgKGNvbmZpZ192YWx1ZSAhPSAi
            IikKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuX19zZXRh
            dHRyX18oYXJnLCBjb25maWdfdmFsdWUpCgogICAgICAgIHJldHVybiBjb25maWcKCiAgICBkZWYg
            X19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAgICAgICAgc2VsZi5vdXRfbWFwID0ge30K
            ICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0KCkKICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3RvX3BhcnNlPWFyZ19saXN0KQogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGU6CiAgICAgICAgICAgIHNlbGYuY29uZmln
            ID0gc2VsZi5wYXJzZV9jb25maWdfZmlsZShzZWxmLl9hcmdfcGFyc2VyLmNvbmZpZ19maWxlKQog
            ICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoKQogICAgICAgIHNlbGYucnVu
            X2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRw
            dXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAgICAgc2VsZi5jZXBoX2NvbmYg
            PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNlbGYuY2VwaF9rZXlyaW5nID0g
            c2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1c2VyIG5vdCBwcm92aWRlZCwg
            Z2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxmLnJ1bl9hc191c2VyIGFuZCBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLnJ1bl9hc191c2Vy
            ID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGU6CiAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gc2VsZi5ERUZBVUxUX1JH
            V19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25mOgogICAgICAgICAgICBrd2Fy
            Z3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmluZzoKICAgICAgICAgICAgICAg
            IGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBoX2tleXJpbmd9CiAgICAgICAg
            ICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxlPXNlbGYuY2VwaF9jb25mLCAq
            Kmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIgPSByYWRvcy5S
            YWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3JlYWRfZmlsZSgpCiAgICAgICAg
            c2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToKICAgICAgICBp
            ZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAgICAgICAgICAgIHNlbGYuY2x1
            c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2Vw
            aCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rlci5nZXRfZnNpZCgpKQoKICAg
            IGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNvbik6CiAgICAgICAgY21kID0g
            anNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAgICAgICAgcmV0X3ZhbCwgY21k
            X291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFuZChjbWQsIGIiIikKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAgICAgIHByaW50KGYiQ29tbWFu
            ZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgIGYiUmV0
            dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2NtZF9vdXR9XG4iCiAgICAgICAg
            ICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0tLS0tLS0tLVxuIgogICAgICAg
            ICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJy
            b3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQnIGlzIG5vdCBlbXB0eQogICAg
            ICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24gb3V0cHV0CiAgICAgICAgaWYg
            cmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBqc29uX291dCA9IGpzb24ubG9h
            ZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cKCiAg
            ICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6CiAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9y
            dW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            IyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9ydW1fc3RhdHVzJyBjb21tYW5k
            IGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwg
            IT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICApCiAgICAgICAg
            cV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVyX25hbWUiXQogICAgICAgIHFf
            bGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21hdGNoaW5nX2xpc3QgPSBbCiAg
            ICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJdWyJtb25zIl0gaWYgbFsibmFt
            ZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQiKQogICAgICAgIHFfbGVhZGVy
            X2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAgICAgICAgIyBnZXQgdGhlIGFk
            ZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAgICAgcV9sZWFkZXJfYWRkcnZl
            YyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMiLCB7fSkuZ2V0KCJhZGRydmVj
            IiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9kZXRhaWxzWyJwdWJsaWNfYWRk
            ciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudjJfcG9ydF9l
            bmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIi
            OgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bImFkZHIiXQog
            ICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+IDEgYW5kIHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVh
            ZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAg
            c3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUg
            bm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAg
            ICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9h
            ZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBw
            b3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15
            IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRo
            b2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNp
            bnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYg
            aXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91
            bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUg
            aG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlw
            CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1h
            dCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlz
            IGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBm
            YWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3Zh
            bCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBh
            ZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28g
            dmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRw
            b2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAg
            ICAgICAgaXAgPSBbXQogICAgICAgICAgICAjIGV4YW1wbGUgb3V0cHV0IFsoPEFkZHJlc3NGYW1p
            bHkuQUZfSU5FVDogMj4sIDxTb2NrZXRLaW5kLlNPQ0tfU1RSRUFNOiAxPiwgNiwgJycsICgnOTMu
            MTg0LjIxNi4zNCcsIDgwKSksIC4uLl0KICAgICAgICAgICAgIyB3ZSBuZWVkIHRvIGdldCA5My4x
            ODQuMjE2LjM0IHNvIGl0IHdvdWxkIGJlIGlwWzBdWzRdWzBdCiAgICAgICAgICAgIGlmIGhvc3Rf
            aXBfdHlwZSA9PSAiSVB2NiI6CiAgICAgICAgICAgICAgICBpcCA9IHNvY2tldC5nZXRhZGRyaW5m
            bygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZhbWlseT1zb2NrZXQuQUZf
            SU5FVDYsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbGlmIGhvc3RfaXBfdHlwZSA9PSAiSVB2NCI6CiAgICAgICAgICAgICAgICBpcCA9IHNv
            Y2tldC5nZXRhZGRyaW5mbygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZh
            bWlseT1zb2NrZXQuQUZfSU5FVCwgcHJvdG89c29ja2V0LklQUFJPVE9fVENQCiAgICAgICAgICAg
            ICAgICApCiAgICAgICAgICAgIGRlbCBzb2NrZXQKICAgICAgICAgICAgcmV0dXJuIGlwWzBdWzRd
            WzBdCiAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQoKICAgIGRlZiBnZXRfYWN0aXZlX2FuZF9zdGFu
            ZGJ5X21ncnMoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAg
            ICAgICAgICByZXR1cm4gIiIsIHNlbGYuZHJ5X3J1bigiY2VwaCBzdGF0dXMiKQogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gc2VsZi5fYXJn
            X3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50CiAgICAgICAgc3RhbmRieV9tZ3JzID0gW10KICAg
            ICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0OgogICAgICAgICAgICBjbWRf
            anNvbiA9IHsicHJlZml4IjogInN0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1w
            dCwKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAg
            ICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICAgICAgIidtZ3Igc2VydmljZXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAg
            ICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlf
            T1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtb25pdG9yaW5nX2Vu
            ZHBvaW50ID0gKAogICAgICAgICAgICAgICAganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0
            KCJzZXJ2aWNlcyIsIHt9KS5nZXQoInByb21ldGhldXMiLCAiIikKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgImNhbid0IGZp
            bmQgbW9uaXRvcmluZ19lbmRwb2ludCwgcHJvbWV0aGV1cyBtb2R1bGUgbWlnaHQgbm90IGJlIGVu
            YWJsZWQsICIKICAgICAgICAgICAgICAgICAgICAiZW5hYmxlIHRoZSBtb2R1bGUgYnkgcnVubmlu
            ZyAnY2VwaCBtZ3IgbW9kdWxlIGVuYWJsZSBwcm9tZXRoZXVzJyIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgIyBub3cgY2hlY2sgdGhlIHN0YW5kLWJ5IG1nci1zCiAgICAgICAgICAgIHN0
            YW5kYnlfYXJyID0ganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0KCJzdGFuZGJ5cyIsIFtd
            KQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5IGluIHN0YW5kYnlfYXJyOgogICAgICAgICAg
            ICAgICAgaWYgIm5hbWUiIGluIGVhY2hfc3RhbmRieS5rZXlzKCk6CiAgICAgICAgICAgICAgICAg
            ICAgc3RhbmRieV9tZ3JzLmFwcGVuZChlYWNoX3N0YW5kYnlbIm5hbWUiXSkKICAgICAgICAgICAg
            dHJ5OgogICAgICAgICAgICAgICAgcGFyc2VkX2VuZHBvaW50ID0gdXJscGFyc2UobW9uaXRvcmlu
            Z19lbmRwb2ludCkKICAgICAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYi
            aW52YWxpZCBlbmRwb2ludDoge21vbml0b3JpbmdfZW5kcG9pbnR9IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBwYXJzZWRfZW5kcG9p
            bnQuaG9zdG5hbWUKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoK
            ICAgICAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHN0cihwYXJzZWRfZW5k
            cG9pbnQucG9ydCkKCiAgICAgICAgIyBpZiBtb25pdG9yaW5nIGVuZHBvaW50IHBvcnQgaXMgbm90
            IHNldCwgcHV0IGEgZGVmYXVsdCBtb24gcG9ydAogICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQ6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYu
            REVGQVVMVF9NT05JVE9SSU5HX0VORFBPSU5UX1BPUlQKCiAgICAgICAgIyB1c2VyIGNvdWxkIGdp
            dmUgY29tbWEgYW5kIHNwYWNlIHNlcGFyYXRlZCBpbnB1dHMgKGxpa2UgLS1tb25pdG9yaW5nLWVu
            ZHBvaW50PSI8aXAxPiwgPGlwMj4iKQogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlz
            dCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdC5yZXBsYWNlKCIsIiwgIiAiKQogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCA9IG1vbml0b3JpbmdfZW5kcG9pbnRf
            aXBfbGlzdC5zcGxpdCgpCiAgICAgICAgIyBpZiBtb25pdG9yaW5nLWVuZHBvaW50IGNvdWxkIG5v
            dCBiZSBmb3VuZCwgcmFpc2UgYW4gZXJyb3IKICAgICAgICBpZiBsZW4obW9uaXRvcmluZ19lbmRw
            b2ludF9pcF9saXN0X3NwbGl0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKCJObyAnbW9uaXRvcmluZy1lbmRwb2ludCcgZm91bmQiKQogICAgICAgICMg
            Zmlyc3QgaXAgaXMgdHJlYXRlZCBhcyB0aGUgbWFpbiBtb25pdG9yaW5nLWVuZHBvaW50CiAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9z
            cGxpdFswXQogICAgICAgICMgcmVzdCBvZiB0aGUgaXAtcyBhcmUgYWRkZWQgdG8gdGhlICdzdGFu
            ZGJ5X21ncnMnIGxpc3QKICAgICAgICBzdGFuZGJ5X21ncnMuZXh0ZW5kKG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXBfbGlzdF9zcGxpdFsxOl0pCiAgICAgICAgZmFpbGVkX2lwID0gbW9uaXRvcmluZ19l
            bmRwb2ludF9pcAoKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50ID0gIjoiLmpvaW4oCiAgICAg
            ICAgICAgIFttb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnRd
            CiAgICAgICAgKQogICAgICAgIGlwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KG1vbml0
            b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBv
            aW50X2lwID0gc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQog
            ICAgICAgICAgICApCiAgICAgICAgICAgICMgY29sbGVjdCBhbGwgdGhlICdzdGFuZC1ieScgbWdy
            IGlwcwogICAgICAgICAgICBtZ3JfaXBzID0gW10KICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRi
            eV9tZ3IgaW4gc3RhbmRieV9tZ3JzOgogICAgICAgICAgICAgICAgZmFpbGVkX2lwID0gZWFjaF9z
            dGFuZGJ5X21ncgogICAgICAgICAgICAgICAgbWdyX2lwcy5hcHBlbmQoCiAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgICAgICAg
            ICAgZWFjaF9zdGFuZGJ5X21nciwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAg
            ICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiQ29udmVyc2lvbiBvZiBob3N0OiB7ZmFpbGVkX2lwfSB0byBJUCBmYWlsZWQuICIKICAgICAg
            ICAgICAgICAgICJQbGVhc2UgZW50ZXIgdGhlIElQIGFkZHJlc3NlcyBvZiBhbGwgdGhlIGNlcGgt
            bWdycyB3aXRoIHRoZSAnLS1tb25pdG9yaW5nLWVuZHBvaW50JyBmbGFnIgogICAgICAgICAgICAp
            CgogICAgICAgIF8sIF8sIGVyciA9IHNlbGYuZW5kcG9pbnRfZGlhbChtb25pdG9yaW5nX2VuZHBv
            aW50LCBpcF90eXBlKQogICAgICAgIGlmIGVyciA9PSAiLTEiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGVycikKICAgICAgICAjIGFkZCB0aGUgdmFsaWRhdGVk
            IGFjdGl2ZSBtZ3IgSVAgaW50byB0aGUgZmlyc3QgaW5kZXgKICAgICAgICBtZ3JfaXBzLmluc2Vy
            dCgwLCBtb25pdG9yaW5nX2VuZHBvaW50X2lwKQogICAgICAgIGFsbF9tZ3JfaXBzX3N0ciA9ICIs
            Ii5qb2luKG1ncl9pcHMpCiAgICAgICAgcmV0dXJuIGFsbF9tZ3JfaXBzX3N0ciwgbW9uaXRvcmlu
            Z19lbmRwb2ludF9wb3J0CgogICAgZGVmIGNoZWNrX3VzZXJfZXhpc3Qoc2VsZiwgdXNlcik6CiAg
            ICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJhdXRoIGdldCIsICJlbnRpdHkiOiBmInt1c2Vy
            fSIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgIGlmIHJldF92YWwgIT0gMCBv
            ciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIHJldHVy
            biBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaGZzX3Byb3Zpc2lvbmVy
            X2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAi
            YWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRh
            dGE9KiIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFk
            YXRhPXtjZXBoZnNfZmlsZXN5c3RlbX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAg
            ICBkZWYgZ2V0X2NlcGhmc19ub2RlX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRp
            dHkgPSAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIKICAgICAgICBjYXBzID0gewogICAgICAgICAg
            ICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAg
            ICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAiYWxsb3cgcncgdGFn
            IGNlcGhmcyAqPSoiLAogICAgICAgICAgICAibWRzIjogImFsbG93IHJ3IiwKICAgICAgICB9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAi
            azhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVy
            LW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lz
            dGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAg
            IGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7
            ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVz
            eXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgKj17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5Cgog
            ICAgZGVmIGdldF9lbnRpdHkoCiAgICAgICAgc2VsZiwKICAgICAgICBlbnRpdHksCiAgICAgICAg
            cmJkX3Bvb2xfbmFtZSwKICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgIGs4c19j
            bHVzdGVyX25hbWUsCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgKToKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgIG9y
            IHJiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgIG9yIGFsaWFzX3JiZF9w
            b29sX25hbWUgIT0gIiIKICAgICAgICAgICAgIyBjaGVja2luZyBhbGlhc19yYmRfcG9vbF9uYW1l
            IGlzIG5vdCBlbXB0eSBhcyB0aGVyZSBtYXliZSBhIHNwZWNpYWwgY2hhcmFjdGVyIHVzZWQgb3Ro
            ZXIgdGhhbiAuIG9yIF8KICAgICAgICApOgogICAgICAgICAgICBpZiBhbGlhc19yYmRfcG9vbF9u
            YW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICAgICAicGxlYXNlIHNldCB0aGUgJy0tYWxpYXMtcmJkLWRhdGEt
            cG9vbC1uYW1lJyBmbGFnIGFzIHRoZSByYmQgZGF0YSBwb29sIG5hbWUgY29udGFpbnMgJy4nIG9y
            ICdfJyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAg
            YWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgICAgIG9yIGFs
            aWFzX3JiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgICk6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICInLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgdmFsdWUgc2hvdWxkIG5vdCBj
            b250YWluICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9
            IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXthbGlhc19yYmRfcG9vbF9uYW1lfSIKICAg
            ICAgICBlbHNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfS17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZToKICAgICAg
            ICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICByZXR1
            cm4gZW50aXR5CgogICAgZGVmIGdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNf
            Y2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJz
            ZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJi
            ZC1kYXRhLXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYg
            azhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAg
            ICAgICAgZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAg
            ICAgIGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1l
            LAogICAgICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmIHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFt
            ZX0gbmFtZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBv
            b2w9e3JiZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVm
            IGdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktcmJkLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            b3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9
            IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xfbmFtZQog
            ICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rl
            cl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFk
            b3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgazhzX2Ns
            dXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBmb3VuZCwg
            cGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAgICAgICAg
            ZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAg
            ICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAgICAgIGlm
            IHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gKAog
            ICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0gbmFt
            ZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBvb2w9e3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9k
            ZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gc2VsZi5y
            dW5fYXNfdXNlcgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLAogICAg
            ICAgICAgICAibWdyIjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwKICAgICAgICAgICAgIm9zZCI6
            IGYicHJvZmlsZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD17c2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5yb290LCBhbGxv
            dyBydyBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2ws
            IGFsbG93IHJ4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubG9n
            LCBhbGxvdyB4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuYnVj
            a2V0cy5pbmRleCIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQoKICAg
            ICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jYXBzX2FuZF9lbnRpdHkoc2Vs
            ZiwgdXNlcl9uYW1lKToKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIi
            IGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lv
            bmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3Ry
            aWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRf
            Y2VwaGZzX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBz
            ZWxmLmdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            Y3NpLXJiZC1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50
            LmNzaS1yYmQtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KCkKICAgICAg
            ICBpZiAiY2xpZW50LmNzaS1yYmQtbm9kZSIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAi
            Y2xpZW50LmNzaS1yYmQtbm9kZSIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAg
            cmV0dXJuIHNlbGYuZ2V0X3JiZF9ub2RlX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2VudGl0eSgpCgogICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgIGYibm8gdXNlciBmb3VuZCB3aXRo
            IHVzZXJfbmFtZToge3VzZXJfbmFtZX0sICIKICAgICAgICAgICAgImdldF9jYXBzX2FuZF9lbnRp
            dHkgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcihzZWxmLCB1c2VyKToKICAgICAgICAiIiIKICAgICAgICBjb21tYW5kOiBjZXBo
            IGF1dGggZ2V0LW9yLWNyZWF0ZSBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciBtb24gJ2Fs
            bG93IHInIG1nciAnYWxsb3cgcncnIG9zZCAnYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            JwogICAgICAgICIiIgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2Vu
            dGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1
            dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAg
            ICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBj
            YXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0K
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4g
            KAogICAgICAgICAgICAgICAgc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgICAgICJjZXBo
            ICIKICAgICAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAg
            ICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50aXR5Il0KICAg
            ICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21k
            X2pzb25bImNhcHMiXSkKICAgICAgICAgICAgICAgICksCiAgICAgICAgICAgICAgICAiIiwKICAg
            ICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAg
            dXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJf
            a2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXksIGYie2VudGl0eS5zcGxpdCgn
            LicsIDEpWzFdfSIKICAgICAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4nLDEpWzFdIHRvIHJlbmFt
            ZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1ub2RlCgogICAgICAgIHJl
            dF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRf
            anNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAg
            ICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGdl
            dC1vci1jcmVhdGUge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYi
            RXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJ
            U1R9IgogICAgICAgICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0p
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4n
            LDEpWzFdIHRvIHJlbmFtZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1u
            b2RlCgogICAgZGVmIGdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoc2VsZik6CiAgICAgICAg
            Y21kX2pzb24gPSB7InByZWZpeCI6ICJmcyBscyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            aWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlf
            cnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsIHJlcG9ydCBhbiBlcnJvcgog
            ICAgICAgIGlmIHJldF92YWwgIT0gMDoKICAgICAgICAgICAgIyBpZiBmcyBhbmQgZGF0YV9wb29s
            IGFyZ3VtZW50cyBhcmUgbm90IHNldCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSA9
            PSAiIgogICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9v
            bF9uYW1lID09ICIiCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICByZXR1cm4KICAgICAg
            ICAgICAgIyBpZiB1c2VyIGhhcyBwcm92aWRlZCBhbnkgb2YgdGhlCiAgICAgICAgICAgICMgJy0t
            Y2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgb3IgJy0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lJyBhcmd1
            bWVudHMsCiAgICAgICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIGFzIHdlIGFyZSB1bmFibGUg
            dG8gdmVyaWZ5IHRoZSBhcmdzCiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIidmcyBscycgY2VwaCBjYWxsIGZhaWxlZCB3aXRoIGVy
            cm9yOiB7ZXJyX21zZ30iCiAgICAgICAgICAgICkKCiAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQg
            PSB7fQogICAgICAgICMgaWYgJy0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgYXJndW1lbnQgaXMg
            cHJvdmlkZWQsCiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBwcm92aWRlZCBmaWxlc3lzdGVt
            LW5hbWUgZXhpc3RzIG9yIG5vdAogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZToKICAgICAgICAgICAgIyBnZXQgdGhlIG1hdGNoaW5nIGxpc3QKICAgICAg
            ICAgICAgbWF0Y2hpbmdfanNvbl9vdXRfbGlzdCA9IFsKICAgICAgICAgICAgICAgIG1hdGNoZWQK
            ICAgICAgICAgICAgICAgIGZvciBtYXRjaGVkIGluIGpzb25fb3V0CiAgICAgICAgICAgICAgICBp
            ZiBzdHIobWF0Y2hlZFsibmFtZSJdKSA9PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWUKICAgICAgICAgICAgXQogICAgICAgICAgICAjIHVuYWJsZSB0byBmaW5kIGEgbWF0
            Y2hpbmcgZnMtbmFtZSwgcmFpc2UgYW4gZXJyb3IKICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5n
            X2pzb25fb3V0X2xpc3QpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiRmlsZXN5c3RlbSBwcm92aWRlZCwg
            J3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWV9JywgIgogICAgICAgICAg
            ICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBmcy1saXN0OiB7W3N0cih4WyduYW1lJ10p
            IGZvciB4IGluIGpzb25fb3V0XX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1hdGNo
            aW5nX2pzb25fb3V0ID0gbWF0Y2hpbmdfanNvbl9vdXRfbGlzdFswXQogICAgICAgICMgaWYgY2Vw
            aGZzIGZpbGVzeXN0ZW0gbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgIyB0cnkgdG8gZ2V0
            IGEgZGVmYXVsdCBmcyBuYW1lIGJ5IGRvaW5nIHRoZSBmb2xsb3dpbmcKICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAjIGEuIGNoZWNrIGlmIHRoZXJlIGlzIG9ubHkgb25lIGZpbGVzeXN0ZW0gaXMg
            cHJlc2VudAogICAgICAgICAgICBpZiBsZW4oanNvbl9vdXQpID09IDE6CiAgICAgICAgICAgICAg
            ICBtYXRjaGluZ19qc29uX291dCA9IGpzb25fb3V0WzBdCiAgICAgICAgICAgICMgYi4gb3IgZWxz
            ZSwgY2hlY2sgaWYgZGF0YV9wb29sIG5hbWUgaXMgcHJvdmlkZWQKICAgICAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICMg
            YW5kIGlmIHByZXNlbnQsIGNoZWNrIHdoZXRoZXIgdGhlcmUgZXhpc3RzIGEgZnMgd2hpY2ggaGFz
            IHRoZSBkYXRhX3Bvb2wKICAgICAgICAgICAgICAgIGZvciBlYWNoSiBpbiBqc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSBpbiBlYWNoSlsiZGF0YV9wb29scyJdOgogICAgICAgICAgICAgICAgICAgICAgICBtYXRjaGlu
            Z19qc29uX291dCA9IGVhY2hKCiAgICAgICAgICAgICAgICAgICAgICAgIGJyZWFrCiAgICAgICAg
            ICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIG1hdGNoaW5nIGZzIGV4aXN0cywgdGhhdCBtZWFucyBw
            cm92aWRlZCBkYXRhX3Bvb2wgbmFtZSBpcyBpbnZhbGlkCiAgICAgICAgICAgICAgICBpZiBub3Qg
            bWF0Y2hpbmdfanNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhX3Bv
            b2wgbmFtZSwge3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfSwiCiAgICAg
            ICAgICAgICAgICAgICAgICAgICIgZG9lcyBub3QgZXhpc3RzIgogICAgICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgIyBjLiBpZiBub3RoaW5nIGlzIHNldCBhbmQgY291bGRuJ3QgZmluZCBh
            IGRlZmF1bHQsCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGp1c3QgcmV0dXJu
            IHNpbGVudGx5CiAgICAgICAgICAgICAgICByZXR1cm4KCiAgICAgICAgaWYgbWF0Y2hpbmdfanNv
            bl9vdXQ6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSA9IHN0cihtYXRjaGluZ19qc29uX291dFsibmFtZSJdKQogICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19tZXRhZGF0YV9wb29sX25hbWUgPSBzdHIoCiAgICAgICAgICAgICAgICBt
            YXRjaGluZ19qc29uX291dFsibWV0YWRhdGFfcG9vbCJdCiAgICAgICAgICAgICkKCiAgICAgICAg
            aWYgaXNpbnN0YW5jZShtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdLCBsaXN0KToKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBoYXMgYWxyZWFkeSBwcm92aWRlZCBkYXRhLXBvb2wtbmFt
            ZSwKICAgICAgICAgICAgIyB0aHJvdWdoIC0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lCiAgICAgICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAg
            ICAgICAgIyBpZiB0aGUgcHJvdmlkZWQgbmFtZSBpcyBub3QgbWF0Y2hpbmcgd2l0aCB0aGUgb25l
            IGluIHRoZSBsaXN0CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgICAgICAgICBu
            b3QgaW4gbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXQogICAgICAgICAgICAgICAgKToK
            ICAgICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lOiAne3NlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfScsICIKICAgICAgICAgICAgICAgICAgICAg
            ICAgImRvZXNuJ3QgbWF0Y2ggZnJvbSB0aGUgZGF0YS1wb29scyBsaXN0OiAiCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGYie1tzdHIoeCkgZm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFf
            cG9vbHMnXV19IgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiBkYXRhX3Bv
            b2wgbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgICAgICMgdGhlbiB0cnkgdG8gZmluZCBh
            IGRlZmF1bHQgZGF0YSBwb29sIG5hbWUKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAg
            ICMgaWYgbm8gZGF0YV9wb29scyBleGlzdCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgICAg
            ICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPT0gMDoKICAgICAgICAg
            ICAgICAgICAgICByZXR1cm4KICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25f
            b3V0WyJkYXRhX3Bvb2xzIl1bMF0KICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiB0
            aGVyZSBhcmUgbW9yZSB0aGFuIG9uZSAnZGF0YV9wb29scycgZXhpc3QsCiAgICAgICAgICAgICMg
            dGhlbiB3YXJuIHRoZSB1c2VyIHRoYXQgd2UgYXJlIHVzaW5nIHRoZSBzZWxlY3RlZCBuYW1lCiAg
            ICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdKSA+IDE6CiAg
            ICAgICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgICAgICAiV0FSTklORzogTXVsdGlw
            bGUgZGF0YSBwb29scyBkZXRlY3RlZDogIgogICAgICAgICAgICAgICAgICAgIGYie1tzdHIoeCkg
            Zm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFfcG9vbHMnXV19XG4iCiAgICAgICAgICAg
            ICAgICAgICAgZiJVc2luZyB0aGUgZGF0YS1wb29sOiAne3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lfSdcbiIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2No
            ZWNrZXJLZXkoc2VsZiwgdXNlcik6CiAgICAgICAgY2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fw
            c19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21kX2pzb24gPSB7CiAgICAgICAgICAgICJwcmVm
            aXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAgICAgICAgICAgImVudGl0eSI6IGVudGl0eSwK
            ICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBjYXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMo
            KSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAg
            ICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oCiAgICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAg
            ICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAg
            ICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAgICAgICApCiAgICAg
            ICAgIyBjaGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYu
            Y2hlY2tfdXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAg
            ICAgICAgIHJldHVybiB1c2VyX2tleQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21z
            ZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVy
            ZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHtzZWxmLnJ1bl9h
            c191c2VyfScgY29tbWFuZCBmYWlsZWRcbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJf
            bXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAg
            ICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0pCgogICAgZGVmIGdl
            dF9jZXBoX2Rhc2hib2FyZF9saW5rKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgi
            OiAibWdyIHNlcnZpY2VzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiAr
            IGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1
            bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25f
            b3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIGlmICJkYXNoYm9hcmQi
            IG5vdCBpbiBqc29uX291dDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICByZXR1cm4g
            anNvbl9vdXRbImRhc2hib2FyZCJdCgogICAgZGVmIGNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIo
            c2VsZik6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAg
            ICAgICAgICJ1c2VyIiwKICAgICAgICAgICAgImNyZWF0ZSIsCiAgICAgICAgICAgICItLXVpZCIs
            CiAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAg
            ICAgICAgICItLWRpc3BsYXktbmFtZSIsCiAgICAgICAgICAgICJSb29rIFJHVyBBZG1pbiBPcHMg
            dXNlciIsCiAgICAgICAgICAgICItLWNhcHMiLAogICAgICAgICAgICAiYnVja2V0cz0qO3VzZXJz
            PSo7dXNhZ2U9cmVhZDttZXRhZGF0YT1yZWFkO3pvbmU9cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgIiAiLmpvaW4oY21kKSkKICAgICAgICB0cnk6CiAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBhbHJlYWR5IGV4aXN0cywgd2UganVzdCBxdWVyeSBpdAog
            ICAgICAgICAgICBpZiBleGVjRXJyLnJldHVybmNvZGUgPT0gZXJybm8uRUVYSVNUOgogICAgICAg
            ICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAg
            ICAgICAgICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICAgICAgICAgImluZm8iLAogICAg
            ICAgICAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5FWFRFUk5B
            TF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUs
            CiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICAgICAg
            ICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICAgICAgICAgIF0KICAgICAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVy
            cj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxs
            ZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgICAgICBlcnJfbXNnID0g
            KAogICAgICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2Nt
            ZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShl
            cnJfbXNnKQogICAgICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0x
            IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4
            ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0
            dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0
            dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5m
            bz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19j
            YXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAg
            ICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAg
            ICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Jl
            YWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUi
            LAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQog
            ICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQo
            Y21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNh
            bGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZl
            cnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRh
            dGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2Fw
            czogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4
            ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2Fw
            X3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJf
            bXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7
            Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJD
            b2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAg
            ICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNv
            bm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAg
            IGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0
            cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9y
            dGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0
            cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3twb29sX25hbWV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYgaW5pdF9yYmRfcG9vbChzZWxmLCBy
            YmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlS
            YWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9w
            ZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAg
            ICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkKCiAgICBkZWYgdmFsaWRhdGVfcmFk
            b3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmJkX21l
            dGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQogICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYgcmFkb3NfbmFt
            ZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRvc19uYW1lc3Bh
            Y2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNl
            LCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRhaW5zIHVwcGVy
            IGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAgICAg
            IGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICBp
            ZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3BhY2UpIGlzIEZh
            bHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFj
            ZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9vbCAne3JiZF9w
            b29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRlX3N1YnZvbHVt
            ZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUpOgog
            ICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAg
            ICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIsCiAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19v
            dXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9j
            ZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAg
            ICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAgICJzdWJ2b2x1
            bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgICAgIGNlcGhm
            c19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3Mu
            Y2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmInN1YnZvbHVt
            ZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3JlYXRlZCIKICAg
            ICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBzZWxmLCBzdWJ2
            b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBwaW5fc2V0dGlu
            ZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAgICAgICAgICAg
            ImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgInBpbiIsCiAg
            ICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9n
            cm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0aW5nLAogICAg
            ICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1
            dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3Mu
            Q2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZvbHVtZV9ncm91
            cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRf
            cmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNzX2tleSA9IHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAgICAgIHNlY3Jl
            dF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0KICAg
            ICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAg
            IGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIvYWRtaW4vaW5m
            bz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmVxdWVzdF91
            cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToKICAgICAgICAg
            ICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJsLAogICAgICAg
            ICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3X2VuZHBvaW50
            KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkKICAgICAgICBl
            eGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90IGFibGUgdG8g
            Y2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAgIGlmIHIxIGlz
            IE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAgICAgICAgKQog
            ICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICAgICAi
            LTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJzdG9yYWdlX2Jh
            Y2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X2VuZHBv
            aW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXIn
            IGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91
            dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVt
            bXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJndyBlbmRwb2lu
            dCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9IHNlbGYuX2lu
            dmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cg
            ZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBv
            aW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd190
            bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2VsZi5lbmRwb2lu
            dF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAgIGlmIGVyciAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBpZiB0aGUgcmd3
            IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMgb25seSBjaGVj
            ayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2FwX3N1cHBvcnRl
            ZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAgICByZ3dfZnNp
            ZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAgICAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAgICAgICAgICBp
            ZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAg
            ICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7c2VsZi5fYXJn
            X3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlkYXRpbmcgYnkg
            Y2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFuZCB2YWxpZGF0
            aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUgY2VwaCBjbHVz
            dGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9vbCBleGlzdAog
            ICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFzc2VkIGVsc2Ug
            aXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0aGVzZSBkZWZh
            dWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggIT0g
            ImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBbCiAgICAgICAg
            ICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3Lm1ldGEiLAog
            ICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAgICAgICAgZiJ7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAgICAgICAgICBd
            CiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bvb2xzX3RvX3Zh
            bGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0cyhf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2wsICd7X3Jnd19w
            b29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4gIiIKCiAgICBk
            ZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1l
            LCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0ZV9jb25maWcg
            IT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWlu
            IiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAgICAgICAgICAg
            ImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVfY29uZmlnLAog
            ICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAgICAgICAgXQog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgIGVycl9t
            c2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtj
            bWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNv
            ZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAg
            ICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAgZGVmIGNvbnZl
            cnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAgICByZXR1cm4g
            dmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5
            X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAgICgKICAgICAg
            ICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIKICAgICAgICAg
            ICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAgIG9yIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIiCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICAgICAg
            ICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxhZ3MgLS10b3Bv
            bG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10b3BvbG9neS1m
            YWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfdG9w
            b2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6CiAgICAgICAg
            aWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJv
            dmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIKICAgICAgICAg
            ICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9JywiCiAgICAg
            ICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3BvbG9neV9wb29s
            cyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAgICAgICAgICkK
            ICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYs
            IHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bv
            b2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgogICAgZGVmIGlu
            aXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAg
            Zm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLmluaXRfcmJk
            X3Bvb2wocG9vbCkKCiAgICAjIHRoaXMgd2lsbCByZXR1cm4gdGhlIGZpbmFsIGFyZ3MgdGhhdCBz
            Y3JpcHQgdXNlcyB0byBwcm9jZXNzCiAgICAjIHRoZSBwcmlvcml0eSB0byBzZXQgYSBwYXJ0aWN1
            bGFyIHZhbHVlIGlzLAogICAgIyBjb21tYW5kLWxpbmUtYXJncyA+IGNvbmZpZy5pbmkgZmlsZSB2
            YWx1ZXMgPiBkZWZhdWx0IHZhbHVlcwogICAgZGVmIGdldEZpbmFsVXNlZEFyZ3Moc2VsZik6CiAg
            ICAgICAgYXJndW1lbnQgPSBmIltDb25maWd1cmF0aW9uc11cbiIKICAgICAgICBmb3IgYXJnIGlu
            IHZhcnMoc2VsZi5fYXJnX3BhcnNlcik6CiAgICAgICAgICAgIGlmIHN0cihnZXRhdHRyKHNlbGYu
            X2FyZ19wYXJzZXIsIGFyZykpOgogICAgICAgICAgICAgICAgIyBweXRob24gdHJlYXRzIGZsYWct
            bmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28gY29udmVydGluZyBiYWNrIHRvIGZsYWct
            bmFtZSwKICAgICAgICAgICAgICAgICMgc28gd2UgY2FuIGdldCB0aG9zZSB2YWx1ZXMgZnJvbSBj
            b25maWcgZmlsZQogICAgICAgICAgICAgICAgYXJnVmFsdWUgPSBhcmcucmVwbGFjZSgiXyIsICIt
            IikKICAgICAgICAgICAgICAgICMgZG8gbm90IGFkZCB0aGUgY29uZmlnLWZpbGUgZmxhZyBhbmQg
            YWxzbyBub3QgYWRkIHRoZSBib29sZWFuIGZsYWdzIHdoaWNoIGFyZSBzZXQgdG8gRmFsc2UKICAg
            ICAgICAgICAgICAgICMgYmVjYXVzZSBjb25maWcuaW5pIGZpbGUgdHJlYXRzIGJvb2xlYW4gZmxh
            Z3MgYXMgVHJ1ZSBhbHdheXMKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBhcmdWYWx1ZSAhPSAiY29uZmlnLWZpbGUiCiAgICAgICAgICAgICAgICAgICAgYW5kIGdldGF0
            dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSAhPSBGYWxzZQogICAgICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgICAgICBhcmd1bWVudCArPSBmInthcmdWYWx1ZX0gPSB7c3RyKGdldGF0dHIo
            c2VsZi5fYXJnX3BhcnNlciwgYXJnKSl9XG4iCiAgICAgICAgcmV0dXJuIGFyZ3VtZW50CgogICAg
            ZGVmIF9nZW5fb3V0cHV0X21hcChzZWxmKToKICAgICAgICBpZiBzZWxmLm91dF9tYXA6CiAgICAg
            ICAgICAgIHJldHVybgogICAgICAgICMgc3VwcG9ydCBsZWdhY3kgZmxhZyB3aXRoIHVwZ3JhZGVz
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWU6CiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuY2x1c3Rl
            cl9uYW1lCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gKAogICAg
            ICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUubG93ZXIoKQogICAgICAg
            ICkgICMgYWx3YXlzIGNvbnZlcnQgY2x1c3RlciBuYW1lIHRvIGxvd2VyY2FzZSBjaGFyYWN0ZXJz
            CiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZSkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2woc2VsZi5fYXJnX3BhcnNlci5y
            YmRfZGF0YV9wb29sX25hbWUpCiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uo
            KQogICAgICAgIHNlbGYuX2V4Y2x1ZGVkX2tleXMuYWRkKCJLOFNfQ0xVU1RFUl9OQU1FIikKICAg
            ICAgICBzZWxmLmdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoKQogICAgICAgICMgZG91Ymxl
            IHN0cmluZyBuZWVkZWQgZm9yIHVwc3RyZWFtIGV4cG9ydHMgb2YgZmxhZ3MKICAgICAgICBzZWxm
            Lm91dF9tYXBbIkFSR1MiXSA9IGYnIntzZWxmLmdldEZpbmFsVXNlZEFyZ3MoKX0iJwogICAgICAg
            IHNlbGYub3V0X21hcFsiTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFyc2VyLm5hbWVzcGFjZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiSzhTX0NMVVNURVJfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0ZT
            SUQiXSA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSTkFNRSJdID0gc2VsZi5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0gPSBzZWxmLmdldF9jZXBoX2V4dGVybmFsX21vbl9k
            YXRhKCkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSA9
            IHNlbGYuY3JlYXRlX2NoZWNrZXJLZXkoCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FS
            RF9MSU5LIl0gPSBzZWxmLmdldF9jZXBoX2Rhc2hib2FyZF9saW5rKCkKICAgICAgICAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICBz
            ZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxm
            LmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1ub2RlIikKICAgICAg
            ICAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVu
            dC5jc2ktcmJkLXByb3Zpc2lvbmVyIikKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09M
            X05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDRVBIRlNfTUVUQURBVEFfUE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAg
            IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRhX3Bvb2xfbmFtZQogICAgICAgICkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVU
            SF9QRVJNSVNTSU9OIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3Rl
            ZF9hdXRoX3Blcm1pc3Npb24KICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQURPU19O
            QU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXSA9IHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1l
            X2dyb3VwCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0gPSAi
            IgogICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXSA9
            ICIiCiAgICAgICAgIyBjcmVhdGUgQ2VwaEZTIG5vZGUgYW5kIHByb3Zpc2lvbmVyIGtleXJpbmcg
            b25seSB3aGVuIE1EUyBleGlzdHMKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19O
            QU1FIl0gYW5kIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAo
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwK
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1F
            Il0sCiAgICAgICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGll
            bnQuY3NpLWNlcGhmcy1ub2RlIikKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLWNl
            cGhmcy1wcm92aXNpb25lciIpCiAgICAgICAgICAgICMgY3JlYXRlIHRoZSBkZWZhdWx0ICJjc2ki
            IHN1YnZvbHVtZWdyb3VwCiAgICAgICAgICAgIHNlbGYuZ2V0X29yX2NyZWF0ZV9zdWJ2b2x1bWVf
            Z3JvdXAoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmls
            ZXN5c3RlbV9uYW1lCiAgICAgICAgICAgICkKICAgICAgICAgICAgIyBwaW4gdGhlIGRlZmF1bHQg
            ImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5waW5fc3Vidm9sdW1lKAogICAg
            ICAgICAgICAgICAgImNzaSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSwgImRpc3RyaWJ1dGVkIiwgIjEiCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXToKICAgICAgICAgICAgICAgIHNlbGYuZ2V0X29yX2Ny
            ZWF0ZV9zdWJ2b2x1bWVfZ3JvdXAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAg
            c2VsZi5waW5fc3Vidm9sdW1lKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            c3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAiZGlzdHJpYnV0ZWQiLAog
            ICAgICAgICAgICAgICAgICAgICIxIiwKICAgICAgICAgICAgICAgICkKICAgICAgICBzZWxmLm91
            dF9tYXBbIlJHV19UTFNfQ0VSVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJ
            TkdfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBP
            SU5UX1BPUlQiXSA9ICIiCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIuc2tpcF9tb25p
            dG9yaW5nX2VuZHBvaW50OgogICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIk1PTklUT1JJTkdfRU5EUE9JTlQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsi
            TU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICkgPSBzZWxmLmdldF9hY3Rp
            dmVfYW5kX3N0YW5kYnlfbWdycygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1F
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoKQogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzCiAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJd
            ID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWlu
            X2xhYmVsCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lf
            ZmFpbHVyZV9kb21haW5fdmFsdWVzCiAgICAgICAgKQogICAgICAgIGlmICgKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICBhbmQgc2Vs
            Zi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAgICAg
            ICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMg
            IT0gIiIKICAgICAgICApOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3ZhbHVl
            cygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXko
            c2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKSwKICAgICAgICAgICAgICAgIHNlbGYuY29u
            dmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXQogICAgICAgICAgICAgICAg
            KSwKICAgICAgICAgICAgKQogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3JiZF9w
            b29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJy
            YXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIHNlbGYuaW5pdF90b3BvbG9neV9yYmRfcG9vbHMoCiAgICAgICAgICAgICAgICBzZWxmLmNv
            bnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSkKICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHNlbGYucmFpc2Vf
            ZXhjZXB0aW9uX2lmX2FueV90b3BvbG9neV9mbGFnX2lzX21pc3NpbmcoKQoKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJHV19QT09MX1BSRUZJWCJdID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXgKICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gIiIKICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludDoKICAgICAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICAgICAgc2VsZi5jcmVhdGVfcmd3X2FkbWluX29w
            c191c2VyKCkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAh
            PSAiIgogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZh
            bGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUsICJyZWFsbSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwgInpv
            bmVncm91cCIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgZXJyID0g
            c2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsICJ6b25lIgogICAgICAgICAgICAgICAgICAgICkK
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgICkgb3Ig
            KAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgIT0g
            IiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3Vw
            X25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgICgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19B
            RE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIGluZm9f
            Y2FwX3N1cHBvcnRlZCwKICAgICAgICAgICAgICAgICAgICAgICAgZXJyLAogICAgICAgICAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICAg
            ICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50KGluZm9fY2FwX3N1cHBvcnRl
            ZCkKICAgICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9w
            YXRoOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9p
            bnRfdGxzX2NlcnQoKQogICAgICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciwgc2V0IHRoZSBSR1dfRU5EUE9JTlQKICAgICAgICAg
            ICAgICAgICAgICBpZiBlcnIgIT0gIi0xIjoKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJSR1dfRU5EUE9JTlQiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIGVyciA9ICJQbGVhc2UgcHJv
            dmlkZSBhbGwgdGhlIFJHVyBtdWx0aXNpdGUgcGFyYW1ldGVycyBvciBub25lIG9mIHRoZW0iCiAg
            ICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnIpCgogICAgZGVmIGdlbl9zaGVs
            bF9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAgICAgIHNoT3V0
            SU8gPSBTdHJpbmdJTygpCiAgICAgICAgZm9yIGssIHYgaW4gc2VsZi5vdXRfbWFwLml0ZW1zKCk6
            CiAgICAgICAgICAgIGlmIHYgYW5kIGsgbm90IGluIHNlbGYuX2V4Y2x1ZGVkX2tleXM6CiAgICAg
            ICAgICAgICAgICBzaE91dElPLndyaXRlKGYiZXhwb3J0IHtrfT17dn17TElORVNFUH0iKQogICAg
            ICAgIHNoT3V0ID0gc2hPdXRJTy5nZXR2YWx1ZSgpCiAgICAgICAgc2hPdXRJTy5jbG9zZSgpCiAg
            ICAgICAgcmV0dXJuIHNoT3V0CgogICAgZGVmIGdlbl9qc29uX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICBqc29uX291dCA9IFsKICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgIm5hbWUiOiAiZXh0ZXJuYWwtY2x1c3Rlci11c2VyLWNvbW1hbmQi
            LAogICAgICAgICAgICAgICAgImtpbmQiOiAiQ29uZmlnTWFwIiwKICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICJhcmdzIjogc2VsZi5vdXRfbWFwWyJBUkdTIl0s
            CiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAg
            ICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uLWVuZHBvaW50cyIsCiAgICAgICAgICAgICAg
            ICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAg
            ICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05f
            REFUQSJdLAogICAgICAgICAgICAgICAgICAgICJtYXhNb25JZCI6ICIwIiwKICAgICAgICAgICAg
            ICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9
            LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwK
            ICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4tc2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAg
            ICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJ
            RCJdLAogICAgICAgICAgICAgICAgICAgICJtb24tc2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9yLWNyZWRzIiwKICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxf
            VVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAg
            XQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBh
            ZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENsdXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAg
            ICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBz
            ZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9pbnQiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdFbmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAgICAgICAiTW9uaXRvcmluZ1Bv
            cnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgIyBp
            ZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1y
            YmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lf
            UkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJLZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAg
            ICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JC
            RF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJk
            LXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNS
            RVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJJRCI6
            IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQog
            ICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQn
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNy
            ZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSdd
            fSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAg
            ICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lP
            TkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAg
            ICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfQ0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMs
            IHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhmcy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAg
            ICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewog
            ICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBI
            RlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2Vj
            cmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAg
            ICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAg
            ICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRf
            TElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNl
            Y3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElO
            SyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxp
            bmsiLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0Un
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJBRE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAg
            ICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAg
            ICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0K
            ICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAg
            ICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJyYWRv
            c05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAg
            ICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09M
            X05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92
            aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9y
            YWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5v
            dXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            IGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xf
            TkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6
            IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXIt
            c2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9T
            RUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9t
            YXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICAgICAgfQogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAog
            ICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2Vw
            aC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1l
            IjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUn
            XX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgICAgICkKCiAgICAg
            ICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwn
            LCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhl
            biBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VDbGFzcwogICAgICAgIGlmICgKICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91
            dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTIl0KICAgICAgICApOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9wb2xvZ3kiLAogICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYu
            b3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RP
            TUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAgICAgICAgXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZhbHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAg
            ICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAg
            ICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhw
            YW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycgU3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2Vs
            Zi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwK
            ICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5v
            dXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6
            IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAg
            ICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtz
            ZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAg
            ICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5k
            LXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBz
            ZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ci
            LAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJlbmRwb2ludCI6IHNl
            bGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            UHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZ
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJzZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2VwaC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0
            LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJj
            ZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImNl
            cnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdLAogICAgICAgICAgICAgICAgICAgIH0s
            CiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVt
            cHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRlZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25z
            KHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMt
            bm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAg
            ICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwKICAgICAgICAgICAgImNsaWVudC5jc2ktcmJk
            LXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAg
            ICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNlciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNl
            ciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVzZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2Vy
            KQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgogICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNl
            cl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBnZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBn
            cmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAgICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRo
            Y2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBhZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBw
            b29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAgICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBu
            YW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIgdXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcg
            Y29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdldCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMg
            cmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAg
            ICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0oLio/KVwucmd3XC5tZXRhIgogICAgICAgICAg
            ICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVybiwgY2FwcykKICAgICAgICAgICAgICAgIGlm
            IG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wgbmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAg
            ICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2Vy
            KToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhlIGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAog
            ICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNl
            cn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNn
            ID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBwcmludChmInVzZXIge3Vz
            ZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIpCiAgICAgICAgICAgIHJldHVybgogICAgICAg
            IGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsiY2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3
            X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2VyLCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAg
            ICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNh
            cF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIsICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQog
            ICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlzOgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVl
            cyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBuZXdfY2FwX3Zh
            bHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBjdXJfY2FwX3Blcm1f
            bGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVl
            cy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBdCiAgICAgICAgICAgIG5ld19j
            YXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAgICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19j
            YXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJpcCgpCiAgICAgICAgICAgIF0KICAgICAgICAg
            ICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUg
            b3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2Nh
            cF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxpbWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1
            c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCByZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3Qg
            YW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIKICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0g
            W10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5hcHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9w
            ZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9saXN0XQogICAgICAgICAgICBleGlzdGluZ19j
            YXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19jYXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhp
            c3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAgICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXAp
            CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChleGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAg
            ICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1dGggY2FwcyIsCiAgICAg
            ICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAgICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAg
            ICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQogICAgICAgIHJldF92YWwsIGpzb25fb3V0
            LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBp
            ZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNhcHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5c
            biBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVk
            IHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoKICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAg
            IGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFk
            ZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBl
            bGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJh
            dGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19w
            YXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNl
            bGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0
            OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBzZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0
            ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0gb3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2Rl
            PSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAgICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRf
            b3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3NlKCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19tYWluX18iOgogICAgcmpPYmogPSBSYWRvc0pT
            T04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1haW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAgIHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDog
            e2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAgZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAg
            ICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9z
            RXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hpbGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0
            YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAgICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:35e88716ace3576683a9c4aa686df5f7a8b5bb6ee7a144fd2d87e3c9f5bf3961
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        version: 4.17.13-rhodf
      entries:
      - name: rook-ceph-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: rook-ceph-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: rook-ceph-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: rook-ceph-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: rook-ceph-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: rook-ceph-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: rook-ceph-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: rook-ceph-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: rook-ceph-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: rook-ceph-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: rook-ceph-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: rook-ceph-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: rook-ceph-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: rook-ceph-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: rook-ceph-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9VHJ1ZSwgICMgZGVmYXVsdCBpcyBUcnVlIHRvIGVuYWJsZSB2MiBtb24g
            cG9ydCBpbiBkb3duc3RyZWFtCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9ydCgz
            MzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQLmFk
            ZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAgICAg
            ICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24iLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJhc2gp
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3ZpZGVk
            IGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmlsZXN5
            c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAg
            ICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZzIG1l
            dGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50
            KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRl
            cyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9vbC1u
            YW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2Us
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRhIHBv
            b2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBpbiB0
            aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAgKQog
            ICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1lbmRw
            b2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+OjxQ
            T1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAg
            IHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50
            IFRMUyBjZXJ0aWZpY2F0ZSAob3IgaW50ZXJtZWRpYXRlIHNpZ25pbmcgY2VydGlmaWNhdGUpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAg
            IGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBUTFMgY2VydGlmaWNhdGlv
            biB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0ZSBpcyBwcm92aWRlZCAo
            Tk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRl
            ZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJD
            ZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMgKGNvbW1hIHNlcGFyYXRl
            ZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBvciBgPEZRRE4+YCkgZW50
            cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAgICkKICAgICAgICBvdXRw
            dXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25pdG9yaW5nLWVuZHBvaW50
            LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgcG9y
            dCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFs
            c2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRv
            IG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0aGUgQ2VwaCBjbHVzdGVy
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0
            aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29sIiwKICAgICAgICApCiAg
            ICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tZHJ5LXJ1biIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1
            ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBleGVjdXRlZCBjb21tYW5k
            cyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAg
            IGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxw
            PSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFtZXNwYWNlcywgdXNlZCBm
            b3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSAoc2hv
            dWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwKICAgICAgICAgICAgZGVm
            YXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9InBy
            b3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAogICAgICAgICkKICAgICAg
            ICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0tbmFt
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3LXJlYWxtIiwKICAg
            ICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0t
            cmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3
            LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhl
            IG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1wb29scyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25zdHJhaW5lZCByYmQgcG9v
            bHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAg
            ICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJrOHMg
            Y2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9uZSwgcmFjaywgb3IgaG9z
            dCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBjZXBoIGRvbWFpbiIsCiAg
            ICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICIt
            LXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBh
            cmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gdmFsdWVzIGNvcnJl
            c3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3BvbG9neS1wb29sc2AgbGlz
            dCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJnUC5hZGRfYXJndW1lbnRf
            Z3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAg
            ICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iVXBncmFkZXMgdGhlIGNl
            cGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIp
            IGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGggbmV3IHBlcm1pc3Npb25z
            IG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9sZGVyIHBlcm1pc3Npb24g
            d2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhv
            bjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3Jh
            ZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3NpIHVzZXJzKG5vbi1yZXN0
            cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNlcnMoRm9yIGV4YW1wbGU6
            IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMpLCB1
            c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24gZmxhZyBuZWVk
            IHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJtYW5kYXRvcnkgZmxhZ3M6
            ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1lIGFuZCAtLXJ1bi1hcy11
            c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsgImluIGNhc2Ugb2YgY2Vw
            aGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUgZmxh
            ZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5nIGl0IHdpbGwgYmUgbWFu
            ZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBo
            L2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3JhZGUgLS1yYmQtZGF0
            YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlICAt
            LXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3JhZ2UtcmVwbGljYXBvb2xg
            IgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3RyaWN0ZWQgdXNlciBjYW5u
            b3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVwZ3JhZGluZy4iCiAgICAg
            ICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2VkIHRvIGFwcGVuZCBuZXcg
            cGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2VkIGZvciBjaGFuZ2luZyB1
            c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBsZSB5b3Ugc2hvdWxkbid0
            IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAgICAgICAgKQoKICAgICAg
            ICAjIEFkZCBjb21tYW5kLWxpbmUgYXJndW1lbnRzCiAgICAgICAgY29uZmlnX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoImNvbmZpZyIpCiAgICAgICAgY29uZmlnX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY29uZmlnLWZpbGUiLAogICAgICAgICAgICB0eXBlPXN0
            ciwKICAgICAgICAgICAgaGVscD0iUGF0aCB0byB0aGUgY29uZmlndXJhdGlvbiBmaWxlLCBQcmlv
            cml0eTogY29tbWFuZC1saW5lLWFyZ3MgPiBjb25maWcuaW5pIGZpbGUgdmFsdWVzID4gZGVmYXVs
            dCB2YWx1ZXMiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICApCgogICAgICAgIGlmIGFyZ3NfdG9fcGFyc2U6CiAgICAgICAgICAgIGFz
            c2VydCAoCiAgICAgICAgICAgICAgICB0eXBlKGFyZ3NfdG9fcGFyc2UpID09IGxpc3QKICAgICAg
            ICAgICAgKSwgIkFyZ3VtZW50IHRvICdnZW5fYXJnX3BhcnNlcicgc2hvdWxkIGJlIGEgbGlzdCIK
            ICAgICAgICBlbHNlOgogICAgICAgICAgICBhcmdzX3RvX3BhcnNlID0gc3lzLmFyZ3ZbMTpdCiAg
            ICAgICAgcmV0dXJuIGFyZ1AucGFyc2VfYXJncyhhcmdzX3RvX3BhcnNlKQoKICAgIGRlZiB2YWxp
            ZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xf
            bmFtZQogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0
            YV9wb29sX25hbWUKCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJGbGFnICctLXJiZC1kYXRhLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAg
            ICAgICAgICAgICAgICkKCiAgICAgICAgICAgIGlmIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUg
            PT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAog
            ICAgICAgICAgICAgICAgICAgICJGbGFnICctLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBjbWRfanNv
            biA9IHsicHJlZml4IjogIm9zZCBkdW1wIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAg
            cmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNt
            ZF9qc29uKQogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAw
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICBmIntjbWRfanNvblsncHJlZml4J119IGNvbW1hbmQgZmFpbGVkLlxuIgog
            ICAgICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNl
            IHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBt
            ZXRhZGF0YV9wb29sX2V4aXN0LCBwb29sX2V4aXN0ID0gRmFsc2UsIEZhbHNlCgogICAgICAgICAg
            ICBmb3Iga2V5IGluIGpzb25fb3V0WyJwb29scyJdOgogICAgICAgICAgICAgICAgIyBpZiBlcmFz
            dXJlX2NvZGVfcHJvZmlsZSBpcyBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IHJl
            cGxpY2EgcG9vbAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGtleVsi
            ZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBrZXlb
            InBvb2xfbmFtZSJdID09IHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgICAg
            ICk6CiAgICAgICAgICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCA9IFRydWUKICAgICAg
            ICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgbm90IGVtcHR5IGFuZCBwb29s
            IG5hbWUgZXhpc3RzIHRoZW4gaXQgaXMgZWMgcG9vbAogICAgICAgICAgICAgICAgaWYga2V5WyJl
            cmFzdXJlX2NvZGVfcHJvZmlsZSJdIGFuZCBrZXlbInBvb2xfbmFtZSJdID09IHJiZF9wb29sX25h
            bWU6CiAgICAgICAgICAgICAgICAgICAgcG9vbF9leGlzdCA9IFRydWUKCiAgICAgICAgICAgIGlm
            IG5vdCBtZXRhZGF0YV9wb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9u
            RmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiUHJvdmlkZWQgcmJkX2VjX21l
            dGFkYXRhX3Bvb2wgbmFtZSwiCiAgICAgICAgICAgICAgICAgICAgZiIge3JiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgaWYgbm90IHBvb2xfZXhpc3Q6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgcmJkX2RhdGFfcG9v
            bCBuYW1lLCB7cmJkX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICByZXR1cm4gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQoKICAgIGRlZiBk
            cnlfcnVuKHNlbGYsIG1zZyk6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICBwcmludCgiRXhlY3V0ZTogIiArICInIiArIG1zZyArICInIikKCiAgICBkZWYg
            dmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgIHdpdGggb3BlbihzZWxmLl9h
            cmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoLCBlbmNvZGluZz0idXRmOCIpIGFzIGY6CiAgICAg
            ICAgICAgICAgICBjb250ZW50cyA9IGYucmVhZCgpCiAgICAgICAgICAgICAgICByZXR1cm4gY29u
            dGVudHMucnN0cmlwKCkKCiAgICBkZWYgX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoc2VsZik6
            CiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZSBhbmQgbm90IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIkVpdGhlciAnLS11cGdyYWRlJyBvciAn
            LS1yYmQtZGF0YS1wb29sLW5hbWUgPHBvb2xfbmFtZT4nIHNob3VsZCBiZSBzcGVjaWZpZWQiCiAg
            ICAgICAgICAgICkKCiAgICBkZWYgX2ludmFsaWRfZW5kcG9pbnQoc2VsZiwgZW5kcG9pbnRfc3Ry
            KToKICAgICAgICAjIGV4dHJhY3QgdGhlIHBvcnQgYnkgZ2V0dGluZyB0aGUgbGFzdCBzcGxpdCBv
            biBgOmAgZGVsaW1pdGVyCiAgICAgICAgdHJ5OgogICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAs
            IHBvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICBleGNlcHQgVmFsdWVF
            cnJvcjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk5vdCBh
            IHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iKQoKICAgICAgICB0cnk6CiAgICAgICAg
            ICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSA9PSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2lu
            dF9zdHJfaXAgPSBlbmRwb2ludF9zdHJfaXBbMSA6IGxlbihlbmRwb2ludF9zdHJfaXApIC0gMV0K
            ICAgICAgICAgICAgaXBfdHlwZSA9ICgKICAgICAgICAgICAgICAgICJJUHY0IiBpZiB0eXBlKGlw
            X2FkZHJlc3MoZW5kcG9pbnRfc3RyX2lwKSkgaXMgSVB2NEFkZHJlc3MgZWxzZSAiSVB2NiIKICAg
            ICAgICAgICAgKQogICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICBpcF90eXBl
            ID0gIkZRRE4iCiAgICAgICAgaWYgbm90IHBvcnQuaXNkaWdpdCgpOgogICAgICAgICAgICByYWlz
            ZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiUG9ydCBub3QgdmFsaWQ6IHtwb3J0fSIpCiAg
            ICAgICAgaW50UG9ydCA9IGludChwb3J0KQogICAgICAgIGlmIGludFBvcnQgPCAxIG9yIGludFBv
            cnQgPiAyKioxNiAtIDE6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZiJPdXQgb2YgcmFuZ2UgcG9ydCBudW1iZXI6IHtwb3J0fSIpCgogICAgICAgIHJldHVybiBp
            cF90eXBlCgogICAgZGVmIGVuZHBvaW50X2RpYWwoc2VsZiwgZW5kcG9pbnRfc3RyLCBpcF90eXBl
            LCB0aW1lb3V0PTMsIGNlcnQ9Tm9uZSk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXInIGluc3Rh
            bmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91dCB0byB0
            aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlSYWRv
            cyk6CiAgICAgICAgICAgIHJldHVybiAiIiwgIiIsICIiCiAgICAgICAgaWYgaXBfdHlwZSA9PSAi
            SVB2NiI6CiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCwg
            ZW5kcG9pbnRfc3RyX3BvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICAg
            ICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiTm90IGEgcHJvcGVyIGVuZHBvaW50
            OiB7ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgZW5kcG9p
            bnRfc3RyX2lwWzBdICE9ICJbIjoKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9ICJb
            IiArIGVuZHBvaW50X3N0cl9pcCArICJdIgogICAgICAgICAgICBlbmRwb2ludF9zdHIgPSAiOiIu
            am9pbihbZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydF0pCgogICAgICAgIHByb3Rv
            Y29scyA9IFsiaHR0cCIsICJodHRwcyJdCiAgICAgICAgcmVzcG9uc2VfZXJyb3IgPSBOb25lCiAg
            ICAgICAgZm9yIHByZWZpeCBpbiBwcm90b2NvbHM6CiAgICAgICAgICAgIHRyeToKICAgICAgICAg
            ICAgICAgIGVwID0gZiJ7cHJlZml4fTovL3tlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAg
            dmVyaWZ5ID0gTm9uZQogICAgICAgICAgICAgICAgIyBJZiB2ZXJpZnkgaXMgc2V0IHRvIGEgcGF0
            aCB0byBhIGRpcmVjdG9yeSwKICAgICAgICAgICAgICAgICMgdGhlIGRpcmVjdG9yeSBtdXN0IGhh
            dmUgYmVlbiBwcm9jZXNzZWQgdXNpbmcgdGhlIGNfcmVoYXNoIHV0aWxpdHkgc3VwcGxpZWQgd2l0
            aCBPcGVuU1NMLgogICAgICAgICAgICAgICAgaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3NraXBfdGxzOgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9IEZh
            bHNlCiAgICAgICAgICAgICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGlt
            ZW91dCwgdmVyaWZ5PUZhbHNlKQogICAgICAgICAgICAgICAgZWxpZiBwcmVmaXggPT0gImh0dHBz
            IiBhbmQgY2VydDoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBjZXJ0CiAgICAgICAgICAg
            ICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCwgdmVyaWZ5PWNl
            cnQpCiAgICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQpCiAgICAgICAgICAgICAgICBpZiByLnN0YXR1c19j
            b2RlID09IDIwMDoKICAgICAgICAgICAgICAgICAgICByZXR1cm4gcHJlZml4LCB2ZXJpZnksICIi
            CiAgICAgICAgICAgIGV4Y2VwdCBFeGNlcHRpb24gYXMgZXJyOgogICAgICAgICAgICAgICAgcmVz
            cG9uc2VfZXJyb3IgPSBlcnIKICAgICAgICAgICAgICAgIGNvbnRpbnVlCiAgICAgICAgc3lzLnN0
            ZGVyci53cml0ZSgKICAgICAgICAgICAgZiJ1bmFibGUgdG8gY29ubmVjdCB0byBlbmRwb2ludDog
            e2VuZHBvaW50X3N0cn0sIGZhaWxlZCBlcnJvcjoge3Jlc3BvbnNlX2Vycm9yfSIKICAgICAgICAp
            CiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICIiLAogICAgICAg
            ICAgICAoIi0xIiksCiAgICAgICAgKQoKICAgIGRlZiBwYXJzZV9jb25maWdfZmlsZShzZWxmLCBj
            b25maWdfZmlsZSk6CiAgICAgICAgY29uZmlnID0gY29uZmlncGFyc2VyLkNvbmZpZ1BhcnNlcigp
            CiAgICAgICAgY29uZmlnLnJlYWQoY29uZmlnX2ZpbGUpCiAgICAgICAgZm9yIGFyZyBpbiBsaXN0
            KHZhcnMoc2VsZi5fYXJnX3BhcnNlcikpOgogICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxh
            Zy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRlcm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxh
            Zy1uYW1lLAogICAgICAgICAgICAjIHNvIHdlIGNhbiBnZXQgdGhvc2UgdmFsdWVzIGZyb20gY29u
            ZmlnIGZpbGUKICAgICAgICAgICAgYXJndW1lbnQgPSBhcmcucmVwbGFjZSgiXyIsICItIikKICAg
            ICAgICAgICAgYXJndW1lbnRWYWx1ZSA9IHN0cihnZXRhdHRyKHNlbGYuX2FyZ19wYXJzZXIsIGFy
            ZykpCiAgICAgICAgICAgIGNvbmZpZ192YWx1ZSA9IGNvbmZpZy5nZXQoIkNvbmZpZ3VyYXRpb25z
            IiwgYXJndW1lbnQsIGZhbGxiYWNrPU5vbmUpCiAgICAgICAgICAgICMgZ2l2ZSBwcmlvcml0eSB0
            byBjb21tYW5kIGxpbmUgYXJndW1lbnQsIGlmIGNvbW1hbmQgbGluZSBhcmd1bWVudCBpcyBub3Qg
            cHJlc2VudCB1c2UgY29uZmlnLmluaSB2YWx1ZSwKICAgICAgICAgICAgaWYgKHN0cihzeXMuYXJn
            dikuZmluZChhcmd1bWVudCkgPT0gLTEpIGFuZCAoCiAgICAgICAgICAgICAgICAoY29uZmlnX3Zh
            bHVlICE9IE5vbmUpIGFuZCAoY29uZmlnX3ZhbHVlICE9ICIiKQogICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5fX3NldGF0dHJfXyhhcmcsIGNvbmZpZ192YWx1
            ZSkKCiAgICAgICAgcmV0dXJuIGNvbmZpZwoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhcmdfbGlz
            dD1Ob25lKToKICAgICAgICBzZWxmLm91dF9tYXAgPSB7fQogICAgICAgIHNlbGYuX2V4Y2x1ZGVk
            X2tleXMgPSBzZXQoKQogICAgICAgIHNlbGYuX2FyZ19wYXJzZXIgPSBzZWxmLmdlbl9hcmdfcGFy
            c2VyKGFyZ3NfdG9fcGFyc2U9YXJnX2xpc3QpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            b25maWdfZmlsZToKICAgICAgICAgICAgc2VsZi5jb25maWcgPSBzZWxmLnBhcnNlX2NvbmZpZ19m
            aWxlKHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGUpCiAgICAgICAgc2VsZi5fY2hlY2tfY29u
            ZmxpY3Rpbmdfb3B0aW9ucygpCiAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuX2FyZ19w
            YXJzZXIucnVuX2FzX3VzZXIKICAgICAgICBzZWxmLm91dHB1dF9maWxlID0gc2VsZi5fYXJnX3Bh
            cnNlci5vdXRwdXQKICAgICAgICBzZWxmLmNlcGhfY29uZiA9IHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aF9jb25mCiAgICAgICAgc2VsZi5jZXBoX2tleXJpbmcgPSBzZWxmLl9hcmdfcGFyc2VyLmtleXJp
            bmcKICAgICAgICAjIGlmIHVzZXIgbm90IHByb3ZpZGVkLCBnaXZlIGEgZGVmYXVsdCB1c2VyCiAg
            ICAgICAgaWYgbm90IHNlbGYucnVuX2FzX3VzZXIgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGU6CiAgICAgICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLkVYVEVSTkFMX1VTRVJf
            TkFNRQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCBhbmQg
            bm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggPSBzZWxmLkRFRkFVTFRfUkdXX1BPT0xfUFJFRklYCiAgICAgICAg
            aWYgc2VsZi5jZXBoX2NvbmY6CiAgICAgICAgICAgIGt3YXJncyA9IHt9CiAgICAgICAgICAgIGlm
            IHNlbGYuY2VwaF9rZXlyaW5nOgogICAgICAgICAgICAgICAga3dhcmdzWyJjb25mIl0gPSB7Imtl
            eXJpbmciOiBzZWxmLmNlcGhfa2V5cmluZ30KICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFk
            b3MuUmFkb3MoY29uZmZpbGU9c2VsZi5jZXBoX2NvbmYsICoqa3dhcmdzKQogICAgICAgIGVsc2U6
            CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKCkKICAgICAgICAgICAgc2Vs
            Zi5jbHVzdGVyLmNvbmZfcmVhZF9maWxlKCkKICAgICAgICBzZWxmLmNsdXN0ZXIuY29ubmVjdCgp
            CgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIGlmIHNlbGYuY2x1c3Rlci5zdGF0ZSA9
            PSAiY29ubmVjdGVkIjoKICAgICAgICAgICAgc2VsZi5jbHVzdGVyLnNodXRkb3duKCkKCiAgICBk
            ZWYgZ2V0X2ZzaWQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoIGZzaWQiKQogICAgICAgIHJldHVy
            biBzdHIoc2VsZi5jbHVzdGVyLmdldF9mc2lkKCkpCgogICAgZGVmIF9jb21tb25fY21kX2pzb25f
            Z2VuKHNlbGYsIGNtZF9qc29uKToKICAgICAgICBjbWQgPSBqc29uLmR1bXBzKGNtZF9qc29uLCBz
            b3J0X2tleXM9VHJ1ZSkKICAgICAgICByZXRfdmFsLCBjbWRfb3V0LCBlcnJfbXNnID0gc2VsZi5j
            bHVzdGVyLm1vbl9jb21tYW5kKGNtZCwgYiIiKQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            dmVyYm9zZToKICAgICAgICAgICAgcHJpbnQoZiJDb21tYW5kIElucHV0OiB7Y21kfSIpCiAgICAg
            ICAgICAgIHByaW50KAogICAgICAgICAgICAgICAgZiJSZXR1cm4gVmFsOiB7cmV0X3ZhbH1cbkNv
            bW1hbmQgT3V0cHV0OiB7Y21kX291dH1cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3IgTWVzc2Fn
            ZToge2Vycl9tc2d9XG4tLS0tLS0tLS0tXG4iCiAgICAgICAgICAgICkKICAgICAgICBqc29uX291
            dCA9IHt9CiAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciAoaS5lOyByZXRfdmFsIGlzIFpF
            Uk8pIGFuZCAnY21kX291dCcgaXMgbm90IGVtcHR5CiAgICAgICAgIyB0aGVuIGNvbnZlcnQgJ2Nt
            ZF9vdXQnIHRvIGEganNvbiBvdXRwdXQKICAgICAgICBpZiByZXRfdmFsID09IDAgYW5kIGNtZF9v
            dXQ6CiAgICAgICAgICAgIGpzb25fb3V0ID0ganNvbi5sb2FkcyhjbWRfb3V0KQogICAgICAgIHJl
            dHVybiByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZwoKICAgIGRlZiBnZXRfY2VwaF9leHRlcm5h
            bF9tb25fZGF0YShzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogInF1b3J1bV9z
            dGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5
            X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25b
            InByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29t
            bW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3Vj
            Y2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQp
            ID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAiJ3F1b3J1bV9zdGF0dXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAg
            ICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9P
            VVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICBxX2xlYWRlcl9uYW1lID0ganNvbl9v
            dXRbInF1b3J1bV9sZWFkZXJfbmFtZSJdCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHt9CiAg
            ICAgICAgcV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCA9IFsKICAgICAgICAgICAgbCBmb3IgbCBpbiBq
            c29uX291dFsibW9ubWFwIl1bIm1vbnMiXSBpZiBsWyJuYW1lIl0gPT0gcV9sZWFkZXJfbmFtZQog
            ICAgICAgIF0KICAgICAgICBpZiBsZW4ocV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCkgPT0gMDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gbWF0Y2hpbmcgJ21v
            bicgZGV0YWlscyBmb3VuZCIpCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3RbMF0KICAgICAgICAjIGdldCB0aGUgYWRkcmVzcyB2ZWN0b3Igb2YgdGhlIHF1
            b3J1bS1sZWFkZXIKICAgICAgICBxX2xlYWRlcl9hZGRydmVjID0gcV9sZWFkZXJfZGV0YWlscy5n
            ZXQoInB1YmxpY19hZGRycyIsIHt9KS5nZXQoImFkZHJ2ZWMiLCBbXSkKICAgICAgICBpcF9hZGRy
            ID0gc3RyKHFfbGVhZGVyX2RldGFpbHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYg
            cV9sZWFkZXJfYWRkcnZlY1swXVsidHlwZSJdID09ICJ2MiI6CiAgICAgICAgICAgICAgICBpcF9h
            ZGRyID0gcV9sZWFkZXJfYWRkcnZlY1swXVsiYWRkciJdCiAgICAgICAgICAgIGVsaWYgbGVuKHFf
            bGVhZGVyX2FkZHJ2ZWMpID4gMSBhbmQgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09ICJ2
            MiI6CiAgICAgICAgICAgICAgICBpcF9hZGRyID0gcV9sZWFkZXJfYWRkcnZlY1sxXVsiYWRkciJd
            CiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAg
            ICAgICAgICAgICAgICAgICIndjInIGFkZHJlc3MgdHlwZSBub3QgcHJlc2VudCwgYW5kICd2Mi1w
            b3J0LWVuYWJsZScgZmxhZyBpcyBwcm92aWRlZCIKICAgICAgICAgICAgICAgICkKCiAgICAgICAg
            cmV0dXJuIGYie3N0cihxX2xlYWRlcl9uYW1lKX09e2lwX2FkZHJ9IgoKICAgIGRlZiBfY29udmVy
            dF9ob3N0bmFtZV90b19pcChzZWxmLCBob3N0X25hbWUsIHBvcnQsIGlwX3R5cGUpOgogICAgICAg
            ICMgaWYgJ2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgdHlwZSwKICAgICAgICAjIGNhbGwg
            dGhlIGR1bW15IGluc3RhbmNlJ3MgImNvbnZlcnQiIG1ldGhvZAogICAgICAgIGlmIG5vdCBob3N0
            X25hbWU6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIkVtcHR5
            IGhvc3RuYW1lIHByb3ZpZGVkIikKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3Rlciwg
            RHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybiBzZWxmLmNsdXN0ZXIuX2NvbnZlcnRfaG9z
            dG5hbWVfdG9faXAoaG9zdF9uYW1lKQoKICAgICAgICBpZiBpcF90eXBlID09ICJGUUROIjoKICAg
            ICAgICAgICAgIyBjaGVjayB3aGljaCBpcCBGUUROIHNob3VsZCBiZSBjb252ZXJ0ZWQgdG8sIElQ
            djQgb3IgSVB2NgogICAgICAgICAgICAjIGNoZWNrIHRoZSBob3N0IGlwLCB0aGUgZW5kcG9pbnQg
            aXAgdHlwZSB3b3VsZCBiZSBzaW1pbGFyIHRvIGhvc3QgaXAKICAgICAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJvcmNoIGhvc3QgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVt
            cHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICInb3JjaCBob3N0IGxzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAg
            ICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZ
            X09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaG9zdF9hZGRyID0g
            anNvbl9vdXRbMF1bImFkZHIiXQogICAgICAgICAgICAjIGFkZCA6ODAgc2FtcGxlIHBvcnQgaW4g
            aXBfdHlwZSwgYXMgX2ludmFsaWRfZW5kcG9pbnQgYWxzbyB2ZXJpZnkgcG9ydAogICAgICAgICAg
            ICBob3N0X2lwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KGhvc3RfYWRkciArICI6ODAi
            KQogICAgICAgICAgICBpbXBvcnQgc29ja2V0CgogICAgICAgICAgICBpcCA9IFtdCiAgICAgICAg
            ICAgICMgZXhhbXBsZSBvdXRwdXQgWyg8QWRkcmVzc0ZhbWlseS5BRl9JTkVUOiAyPiwgPFNvY2tl
            dEtpbmQuU09DS19TVFJFQU06IDE+LCA2LCAnJywgKCc5My4xODQuMjE2LjM0JywgODApKSwgLi4u
            XQogICAgICAgICAgICAjIHdlIG5lZWQgdG8gZ2V0IDkzLjE4NC4yMTYuMzQgc28gaXQgd291bGQg
            YmUgaXBbMF1bNF1bMF0KICAgICAgICAgICAgaWYgaG9zdF9pcF90eXBlID09ICJJUHY2IjoKICAg
            ICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAgICAgICAgICAgICAgICAg
            IGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVUNiwgcHJvdG89c29ja2V0LklQ
            UFJPVE9fVENQCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsaWYgaG9zdF9pcF90eXBl
            ID09ICJJUHY0IjoKICAgICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAg
            ICAgICAgICAgICAgICAgIGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVULCBw
            cm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZGVs
            IHNvY2tldAogICAgICAgICAgICByZXR1cm4gaXBbMF1bNF1bMF0KICAgICAgICByZXR1cm4gaG9z
            dF9uYW1lCgogICAgZGVmIGdldF9hY3RpdmVfYW5kX3N0YW5kYnlfbWdycyhzZWxmKToKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIiwgc2Vs
            Zi5kcnlfcnVuKCJjZXBoIHN0YXR1cyIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0
            ID0gc2VsZi5fYXJnX3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5k
            cG9pbnQKICAgICAgICBzdGFuZGJ5X21ncnMgPSBbXQogICAgICAgIGlmIG5vdCBtb25pdG9yaW5n
            X2VuZHBvaW50X2lwX2xpc3Q6CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAic3Rh
            dHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVy
            cl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAj
            IGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRf
            dmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ21nciBzZXJ2aWNl
            cycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9t
            c2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnQgPSAoCiAgICAgICAgICAg
            ICAgICBqc29uX291dC5nZXQoIm1ncm1hcCIsIHt9KS5nZXQoInNlcnZpY2VzIiwge30pLmdldCgi
            cHJvbWV0aGV1cyIsICIiKQogICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9y
            aW5nX2VuZHBvaW50OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiY2FuJ3QgZmluZCBtb25pdG9yaW5nX2VuZHBvaW50
            LCBwcm9tZXRoZXVzIG1vZHVsZSBtaWdodCBub3QgYmUgZW5hYmxlZCwgIgogICAgICAgICAgICAg
            ICAgICAgICJlbmFibGUgdGhlIG1vZHVsZSBieSBydW5uaW5nICdjZXBoIG1nciBtb2R1bGUgZW5h
            YmxlIHByb21ldGhldXMnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIG5vdyBjaGVj
            ayB0aGUgc3RhbmQtYnkgbWdyLXMKICAgICAgICAgICAgc3RhbmRieV9hcnIgPSBqc29uX291dC5n
            ZXQoIm1ncm1hcCIsIHt9KS5nZXQoInN0YW5kYnlzIiwgW10pCiAgICAgICAgICAgIGZvciBlYWNo
            X3N0YW5kYnkgaW4gc3RhbmRieV9hcnI6CiAgICAgICAgICAgICAgICBpZiAibmFtZSIgaW4gZWFj
            aF9zdGFuZGJ5LmtleXMoKToKICAgICAgICAgICAgICAgICAgICBzdGFuZGJ5X21ncnMuYXBwZW5k
            KGVhY2hfc3RhbmRieVsibmFtZSJdKQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBw
            YXJzZWRfZW5kcG9pbnQgPSB1cmxwYXJzZShtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgICAg
            ICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50OiB7bW9u
            aXRvcmluZ19lbmRwb2ludH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3Jp
            bmdfZW5kcG9pbnRfaXBfbGlzdCA9IHBhcnNlZF9lbmRwb2ludC5ob3N0bmFtZQogICAgICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICAgICAgbW9uaXRv
            cmluZ19lbmRwb2ludF9wb3J0ID0gc3RyKHBhcnNlZF9lbmRwb2ludC5wb3J0KQoKICAgICAgICAj
            IGlmIG1vbml0b3JpbmcgZW5kcG9pbnQgcG9ydCBpcyBub3Qgc2V0LCBwdXQgYSBkZWZhdWx0IG1v
            biBwb3J0CiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoKICAgICAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0ID0gc2VsZi5ERUZBVUxUX01PTklUT1JJTkdfRU5E
            UE9JTlRfUE9SVAoKICAgICAgICAjIHVzZXIgY291bGQgZ2l2ZSBjb21tYSBhbmQgc3BhY2Ugc2Vw
            YXJhdGVkIGlucHV0cyAobGlrZSAtLW1vbml0b3JpbmctZW5kcG9pbnQ9IjxpcDE+LCA8aXAyPiIp
            CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcF9saXN0LnJlcGxhY2UoIiwiLCAiICIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0X3NwbGl0ID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0LnNwbGl0KCkKICAgICAg
            ICAjIGlmIG1vbml0b3JpbmctZW5kcG9pbnQgY291bGQgbm90IGJlIGZvdW5kLCByYWlzZSBhbiBl
            cnJvcgogICAgICAgIGlmIGxlbihtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQpID09
            IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vICdtb25p
            dG9yaW5nLWVuZHBvaW50JyBmb3VuZCIpCiAgICAgICAgIyBmaXJzdCBpcCBpcyB0cmVhdGVkIGFz
            IHRoZSBtYWluIG1vbml0b3JpbmctZW5kcG9pbnQKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X2lwID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzBdCiAgICAgICAgIyByZXN0
            IG9mIHRoZSBpcC1zIGFyZSBhZGRlZCB0byB0aGUgJ3N0YW5kYnlfbWdycycgbGlzdAogICAgICAg
            IHN0YW5kYnlfbWdycy5leHRlbmQobW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzE6
            XSkKICAgICAgICBmYWlsZWRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwCgogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnQgPSAiOiIuam9pbigKICAgICAgICAgICAgW21vbml0b3JpbmdfZW5k
            cG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydF0KICAgICAgICApCiAgICAgICAgaXBf
            dHlwZSA9IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQobW9uaXRvcmluZ19lbmRwb2ludCkKICAgICAg
            ICB0cnk6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBzZWxmLl9jb252ZXJ0
            X2hvc3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCwg
            bW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgIyBjb2xsZWN0IGFsbCB0aGUgJ3N0YW5kLWJ5JyBtZ3IgaXBzCiAgICAgICAgICAgIG1ncl9p
            cHMgPSBbXQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5X21nciBpbiBzdGFuZGJ5X21ncnM6
            CiAgICAgICAgICAgICAgICBmYWlsZWRfaXAgPSBlYWNoX3N0YW5kYnlfbWdyCiAgICAgICAgICAg
            ICAgICBtZ3JfaXBzLmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9jb252ZXJ0X2hv
            c3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgICAgICAgICBlYWNoX3N0YW5kYnlfbWdyLCBt
            b25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgZXhjZXB0OgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJDb252ZXJzaW9uIG9mIGhvc3Q6
            IHtmYWlsZWRfaXB9IHRvIElQIGZhaWxlZC4gIgogICAgICAgICAgICAgICAgIlBsZWFzZSBlbnRl
            ciB0aGUgSVAgYWRkcmVzc2VzIG9mIGFsbCB0aGUgY2VwaC1tZ3JzIHdpdGggdGhlICctLW1vbml0
            b3JpbmctZW5kcG9pbnQnIGZsYWciCiAgICAgICAgICAgICkKCiAgICAgICAgXywgXywgZXJyID0g
            c2VsZi5lbmRwb2ludF9kaWFsKG1vbml0b3JpbmdfZW5kcG9pbnQsIGlwX3R5cGUpCiAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZXJyKQogICAgICAgICMgYWRkIHRoZSB2YWxpZGF0ZWQgYWN0aXZlIG1nciBJUCBpbnRvIHRo
            ZSBmaXJzdCBpbmRleAogICAgICAgIG1ncl9pcHMuaW5zZXJ0KDAsIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXApCiAgICAgICAgYWxsX21ncl9pcHNfc3RyID0gIiwiLmpvaW4obWdyX2lwcykKICAgICAg
            ICByZXR1cm4gYWxsX21ncl9pcHNfc3RyLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKCiAgICBk
            ZWYgY2hlY2tfdXNlcl9leGlzdChzZWxmLCB1c2VyKToKICAgICAgICBjbWRfanNvbiA9IHsicHJl
            Zml4IjogImF1dGggZ2V0IiwgImVudGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0K
            ICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoK
            ICAgICAgICAgICAgcmV0dXJuICIiCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5
            Il0pCgogICAgZGVmIGdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAg
            ICJvc2QiOiAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwKICAgICAgICAgICAgIm1k
            cyI6ICJhbGxvdyAqIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0
            cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rl
            cl9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiazhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNf
            ZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgog
            ICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAg
            ICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNf
            Y2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVzeXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1si
            b3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9e2NlcGhmc19maWxlc3lzdGVt
            fSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2VwaGZzX25vZGVf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3
            IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsCiAgICAgICAg
            ICAgICJtZHMiOiAiYWxsb3cgcnciLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4
            c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0g
            PT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0
            eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAg
            ICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyAqPXtjZXBoZnNfZmlsZXN5c3Rl
            bX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2VudGl0eSgKICAg
            ICAgICBzZWxmLAogICAgICAgIGVudGl0eSwKICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAg
            IGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICBy
            YWRvc19uYW1lc3BhY2UsCiAgICApOgogICAgICAgIGlmICgKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgb3IgcmJkX3Bvb2xfbmFtZS5jb3VudCgi
            XyIpICE9IDAKICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZSAhPSAiIgogICAgICAg
            ICAgICAjIGNoZWNraW5nIGFsaWFzX3JiZF9wb29sX25hbWUgaXMgbm90IGVtcHR5IGFzIHRoZXJl
            IG1heWJlIGEgc3BlY2lhbCBjaGFyYWN0ZXIgdXNlZCBvdGhlciB0aGFuIC4gb3IgXwogICAgICAg
            ICk6CiAgICAgICAgICAgIGlmIGFsaWFzX3JiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAg
            ICJwbGVhc2Ugc2V0IHRoZSAnLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgYXMgdGhl
            IHJiZCBkYXRhIHBvb2wgbmFtZSBjb250YWlucyAnLicgb3IgJ18nIgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLmNv
            dW50KCIuIikgIT0gMAogICAgICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3Vu
            dCgiXyIpICE9IDAKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIictLWFsaWFzLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyB2YWx1ZSBzaG91bGQgbm90IGNvbnRhaW4gJy4nIG9yICdfJyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0
            ZXJfbmFtZX0te2FsaWFzX3JiZF9wb29sX25hbWV9IgogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtyYmRfcG9vbF9uYW1lfSIK
            CiAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgIHJldHVybiBlbnRpdHkKCiAgICBkZWYgZ2V0
            X3JiZF9wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0g
            ImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAg
            ICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAg
            ICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSBy
            YmQiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNl
            ci5yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19j
            bHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAg
            ICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAg
            ICAgICAgICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBm
            bGFnIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFn
            IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIi
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAn
            LS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bnRpdHkgPSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAg
            ICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFt
            ZSwKICAgICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRv
            c19uYW1lc3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNl
            ICE9ICIiOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25h
            bWVzcGFjZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBjYXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgog
            ICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X3JiZF9ub2RlX2NhcHNfYW5k
            X2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtbm9kZSIKICAg
            ICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSByYmQiLAog
            ICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJt
            aXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19jbHVzdGVy
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgcmFk
            b3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICAg
            ICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5v
            dCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIgogICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAg
            ICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAg
            ICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMt
            Y2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkg
            PSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAgICAgICAg
            ICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAg
            ICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRvc19uYW1l
            c3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlICE9ICIi
            OgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgZiJw
            cm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25hbWVzcGFj
            ZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBj
            YXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAg
            IHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2Vu
            dGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAgY2Fw
            cyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9z
            dGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLAogICAgICAgICAgICAib3NkIjogZiJwcm9maWxlIHJiZC1yZWFkLW9u
            bHksIGFsbG93IHJ3eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9e3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2csIGFsbG93IHggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5idWNrZXRzLmluZGV4IiwKICAgICAgICAg
            ICAgIm1kcyI6ICJhbGxvdyAqIiwKICAgICAgICB9CgogICAgICAgIHJldHVybiBjYXBzLCBlbnRp
            dHkKCiAgICBkZWYgZ2V0X2NhcHNfYW5kX2VudGl0eShzZWxmLCB1c2VyX25hbWUpOgogICAgICAg
            IGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAg
            ICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fw
            c19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgaW4gdXNl
            cl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgIT0gdXNlcl9u
            YW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVy
            bWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19ub2RlX2Nh
            cHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiBp
            biB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X3JiZF9w
            cm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1u
            b2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1ub2RlIiAh
            PSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRf
            YXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX25v
            ZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiIGlu
            IHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5oZWFsdGhjaGVja2VyIiAhPSB1c2Vy
            X25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9w
            ZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfZGVmYXVsdFVzZXJf
            Y2Fwc19hbmRfZW50aXR5KCkKCiAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgZiJubyB1c2VyIGZvdW5kIHdpdGggdXNlcl9uYW1lOiB7dXNlcl9uYW1l
            fSwgIgogICAgICAgICAgICAiZ2V0X2NhcHNfYW5kX2VudGl0eSBjb21tYW5kIGZhaWxlZC5cbiIK
            ICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKHNlbGYsIHVzZXIp
            OgogICAgICAgICIiIgogICAgICAgIGNvbW1hbmQ6IGNlcGggYXV0aCBnZXQtb3ItY3JlYXRlIGNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIG1vbiAnYWxsb3cgcicgbWdyICdhbGxvdyBydycg
            b3NkICdhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSonCiAgICAgICAgIiIiCiAgICAgICAg
            Y2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fwc19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21k
            X2pzb24gPSB7CiAgICAgICAgICAgICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAg
            ICAgICAgICAgImVudGl0eSI6IGVudGl0eSwKICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBj
            YXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMoKSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAg
            ICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICBzZWxm
            LmRyeV9ydW4oCiAgICAgICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAgICAg
            ICsgY21kX2pzb25bInByZWZpeCJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgICAgICsgIiAi
            CiAgICAgICAgICAgICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAg
            ICAgICAgICAgKSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCiAgICAgICAgIyBj
            aGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYuY2hlY2tf
            dXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAgICAgICAg
            IHJldHVybiB1c2VyX2tleSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICAg
            ICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0eShjbGllbnQuY3NpLXJi
            ZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9t
            c2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhl
            cmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9y
            IGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7dXNlcn0nIGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0
            X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAg
            ICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSksIGYie2VudGl0eS5zcGxpdCgnLics
            IDEpWzFdfSIKICAgICAgICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0
            eShjbGllbnQuY3NpLXJiZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICBkZWYgZ2V0X2NlcGhm
            c19kYXRhX3Bvb2xfZGV0YWlscyhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4Ijog
            ImZzIGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29u
            WyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2Nv
            bW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1
            Y2Nlc3NmdWwgYXR0ZW1wdCwgcmVwb3J0IGFuIGVycm9yCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            OgogICAgICAgICAgICAjIGlmIGZzIGFuZCBkYXRhX3Bvb2wgYXJndW1lbnRzIGFyZSBub3Qgc2V0
            LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPT0gIiIKICAgICAgICAg
            ICAgKToKICAgICAgICAgICAgICAgIHJldHVybgogICAgICAgICAgICAjIGlmIHVzZXIgaGFzIHBy
            b3ZpZGVkIGFueSBvZiB0aGUKICAgICAgICAgICAgIyAnLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1l
            JyBvciAnLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUnIGFyZ3VtZW50cywKICAgICAgICAgICAgIyBy
            YWlzZSBhbiBleGNlcHRpb24gYXMgd2UgYXJlIHVuYWJsZSB0byB2ZXJpZnkgdGhlIGFyZ3MKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiJ2ZzIGxzJyBjZXBoIGNhbGwgZmFpbGVkIHdpdGggZXJyb3I6IHtlcnJfbXNnfSIKICAgICAg
            ICAgICAgKQoKICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IHt9CiAgICAgICAgIyBpZiAnLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lJyBhcmd1bWVudCBpcyBwcm92aWRlZCwKICAgICAgICAjIGNo
            ZWNrIHdoZXRoZXIgdGhlIHByb3ZpZGVkIGZpbGVzeXN0ZW0tbmFtZSBleGlzdHMgb3Igbm90CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lOgogICAgICAg
            ICAgICAjIGdldCB0aGUgbWF0Y2hpbmcgbGlzdAogICAgICAgICAgICBtYXRjaGluZ19qc29uX291
            dF9saXN0ID0gWwogICAgICAgICAgICAgICAgbWF0Y2hlZAogICAgICAgICAgICAgICAgZm9yIG1h
            dGNoZWQgaW4ganNvbl9vdXQKICAgICAgICAgICAgICAgIGlmIHN0cihtYXRjaGVkWyJuYW1lIl0p
            ID09IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgdW5hYmxlIHRvIGZpbmQgYSBtYXRjaGluZyBmcy1uYW1lLCByYWlzZSBh
            biBlcnJvcgogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRfbGlzdCkgPT0gMDoK
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgZiJGaWxlc3lzdGVtIHByb3ZpZGVkLCAne3NlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgZiJpcyBub3QgZm91
            bmQgaW4gdGhlIGZzLWxpc3Q6IHtbc3RyKHhbJ25hbWUnXSkgZm9yIHggaW4ganNvbl9vdXRdfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBtYXRjaGlu
            Z19qc29uX291dF9saXN0WzBdCiAgICAgICAgIyBpZiBjZXBoZnMgZmlsZXN5c3RlbSBuYW1lIGlz
            IG5vdCBwcm92aWRlZCwKICAgICAgICAjIHRyeSB0byBnZXQgYSBkZWZhdWx0IGZzIG5hbWUgYnkg
            ZG9pbmcgdGhlIGZvbGxvd2luZwogICAgICAgIGVsc2U6CiAgICAgICAgICAgICMgYS4gY2hlY2sg
            aWYgdGhlcmUgaXMgb25seSBvbmUgZmlsZXN5c3RlbSBpcyBwcmVzZW50CiAgICAgICAgICAgIGlm
            IGxlbihqc29uX291dCkgPT0gMToKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0g
            anNvbl9vdXRbMF0KICAgICAgICAgICAgIyBiLiBvciBlbHNlLCBjaGVjayBpZiBkYXRhX3Bvb2wg
            bmFtZSBpcyBwcm92aWRlZAogICAgICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgIyBhbmQgaWYgcHJlc2VudCwgY2hlY2sg
            d2hldGhlciB0aGVyZSBleGlzdHMgYSBmcyB3aGljaCBoYXMgdGhlIGRhdGFfcG9vbAogICAgICAg
            ICAgICAgICAgZm9yIGVhY2hKIGluIGpzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lIGluIGVhY2hKWyJkYXRhX3Bvb2xz
            Il06CiAgICAgICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0gZWFjaEoKICAg
            ICAgICAgICAgICAgICAgICAgICAgYnJlYWsKICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            bm8gbWF0Y2hpbmcgZnMgZXhpc3RzLCB0aGF0IG1lYW5zIHByb3ZpZGVkIGRhdGFfcG9vbCBuYW1l
            IGlzIGludmFsaWQKICAgICAgICAgICAgICAgIGlmIG5vdCBtYXRjaGluZ19qc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGFfcG9vbCBuYW1lLCB7c2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9LCIKICAgICAgICAgICAgICAgICAgICAgICAgIiBk
            b2VzIG5vdCBleGlzdHMiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGMuIGlm
            IG5vdGhpbmcgaXMgc2V0IGFuZCBjb3VsZG4ndCBmaW5kIGEgZGVmYXVsdCwKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgICMganVzdCByZXR1cm4gc2lsZW50bHkKICAgICAgICAgICAg
            ICAgIHJldHVybgoKICAgICAgICBpZiBtYXRjaGluZ19qc29uX291dDoKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID0gc3RyKG1hdGNoaW5nX2pzb25f
            b3V0WyJuYW1lIl0pCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0WyJtZXRh
            ZGF0YV9wb29sIl0KICAgICAgICAgICAgKQoKICAgICAgICBpZiBpc2luc3RhbmNlKG1hdGNoaW5n
            X2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0sIGxpc3QpOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGhhcyBhbHJlYWR5IHByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lLAogICAgICAgICAgICAjIHRocm91
            Z2ggLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUKICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGlmIHRoZSBwcm92aWRl
            ZCBuYW1lIGlzIG5vdCBtYXRjaGluZyB3aXRoIHRoZSBvbmUgaW4gdGhlIGxpc3QKICAgICAgICAg
            ICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19k
            YXRhX3Bvb2xfbmFtZQogICAgICAgICAgICAgICAgICAgIG5vdCBpbiBtYXRjaGluZ19qc29uX291
            dFsiZGF0YV9wb29scyJdCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYi
            UHJvdmlkZWQgZGF0YS1wb29sLW5hbWU6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWV9JywgIgogICAgICAgICAgICAgICAgICAgICAgICAiZG9lc24ndCBtYXRjaCBmcm9t
            IHRoZSBkYXRhLXBvb2xzIGxpc3Q6ICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4
            KSBmb3IgeCBpbiBtYXRjaGluZ19qc29uX291dFsnZGF0YV9wb29scyddXX0iCiAgICAgICAgICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGlmIGRhdGFfcG9vbCBuYW1lIGlzIG5vdCBwcm92aWRl
            ZCwKICAgICAgICAgICAgIyB0aGVuIHRyeSB0byBmaW5kIGEgZGVmYXVsdCBkYXRhIHBvb2wgbmFt
            ZQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBpZiBubyBkYXRhX3Bvb2xzIGV4
            aXN0LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dFsiZGF0YV9wb29scyJdKSA9PSAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPSBzdHIo
            CiAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXVswXQog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGlmIHRoZXJlIGFyZSBtb3JlIHRoYW4gb25l
            ICdkYXRhX3Bvb2xzJyBleGlzdCwKICAgICAgICAgICAgIyB0aGVuIHdhcm4gdGhlIHVzZXIgdGhh
            dCB3ZSBhcmUgdXNpbmcgdGhlIHNlbGVjdGVkIG5hbWUKICAgICAgICAgICAgaWYgbGVuKG1hdGNo
            aW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID4gMToKICAgICAgICAgICAgICAgIHByaW50KAog
            ICAgICAgICAgICAgICAgICAgICJXQVJOSU5HOiBNdWx0aXBsZSBkYXRhIHBvb2xzIGRldGVjdGVk
            OiAiCiAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4KSBmb3IgeCBpbiBtYXRjaGluZ19qc29u
            X291dFsnZGF0YV9wb29scyddXX1cbiIKICAgICAgICAgICAgICAgICAgICBmIlVzaW5nIHRoZSBk
            YXRhLXBvb2w6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9J1xuIgog
            ICAgICAgICAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2hlY2tlcktleShzZWxmLCB1c2VyKToK
            ICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAg
            ICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVh
            dGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtj
            YXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0
            XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigK
            ICAgICAgICAgICAgICAgICJjZXBoICIKICAgICAgICAgICAgICAgICsgY21kX2pzb25bInByZWZp
            eCJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50
            aXR5Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICArICIgIi5qb2luKGNt
            ZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            CgogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBh
            dHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmIidhdXRoIGdldC1vci1jcmVhdGUge3NlbGYucnVuX2FzX3VzZXJ9JyBjb21tYW5kIGZhaWxl
            ZFxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVs
            c2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            c2VsZik6CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJtZ3Igc2VydmljZXMiLCAiZm9y
            bWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQog
            ICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAog
            ICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAg
            IHJldHVybiBOb25lCiAgICAgICAgaWYgImRhc2hib2FyZCIgbm90IGluIGpzb25fb3V0OgogICAg
            ICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIHJldHVybiBqc29uX291dFsiZGFzaGJvYXJkIl0K
            CiAgICBkZWYgY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcihzZWxmKToKICAgICAgICBjbWQgPSBb
            CiAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAgICAgICAgICAgInVzZXIiLAogICAgICAg
            ICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgIi0tdWlkIiwKICAgICAgICAgICAgc2VsZi5FWFRF
            Uk5BTF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgIi0tZGlzcGxheS1uYW1l
            IiwKICAgICAgICAgICAgIlJvb2sgUkdXIEFkbWluIE9wcyB1c2VyIiwKICAgICAgICAgICAgIi0t
            Y2FwcyIsCiAgICAgICAgICAgICJidWNrZXRzPSo7dXNlcnM9Kjt1c2FnZT1yZWFkO21ldGFkYXRh
            PXJlYWQ7em9uZT1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmVn
            cm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAog
            ICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyAiICIuam9pbihjbWQp
            KQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNz
            LkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGFscmVhZHkgZXhpc3RzLCB3ZSBqdXN0IHF1ZXJ5IGl0CiAgICAgICAgICAgIGlmIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSBlcnJuby5FRVhJU1Q6CiAgICAgICAgICAgICAgICBjbWQgPSBbCiAgICAg
            ICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgICAgICJ1c2Vy
            IiwKICAgICAgICAgICAgICAgICAgICAiaW5mbyIsCiAgICAgICAgICAgICAgICAgICAgIi0tdWlk
            IiwKICAgICAgICAgICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAi
            LS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgICAg
            ICAgICAgXQogICAgICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVj
            RXJyOgogICAgICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAg
            ICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5v
            dXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJu
            Y29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCiAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRv
            IGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAg
            ICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVj
            RXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2Us
            ICItMSIKCiAgICAgICAgIyBzZXBhcmF0ZWx5IGFkZCBpbmZvPXJlYWQgY2FwcyBmb3Igcmd3LWVu
            ZHBvaW50IGlwIHZhbGlkYXRpb24KICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQgPSBUcnVlCiAg
            ICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICJj
            YXBzIiwKICAgICAgICAgICAgImFkZCIsCiAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAg
            IHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICItLWNh
            cHMiLAogICAgICAgICAgICAiaW5mbz1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAg
            Ii0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVn
            cm91cF9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAg
            ICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNz
            LlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4
            ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIGNlcGggdmVyc2lvbiBub3Qgc3VwcG9ydGVkIGZv
            ciBhZGRpbmcgYGluZm89cmVhZGAgY2FwKHJnd192YWxpZGF0aW9uKQogICAgICAgICAgICBpZiAo
            CiAgICAgICAgICAgICAgICAiY291bGQgbm90IGFkZCBjYXBzOiB1bmFibGUgdG8gYWRkIGNhcHM6
            IGluZm89cmVhZFxuIgogICAgICAgICAgICAgICAgaW4gZXhlY0Vyci5zdGRlcnIuZGVjb2RlKCJ1
            dGYtOCIpCiAgICAgICAgICAgICAgICBhbmQgZXhlY0Vyci5yZXR1cm5jb2RlID09IDI0NAogICAg
            ICAgICAgICApOgogICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gRmFsc2UKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICBqc29ub3V0cHV0ID0ganNvbi5sb2Fkcyhv
            dXRwdXQpCiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJhY2Nlc3Nfa2V5Il0sCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsic2VjcmV0
            X2tleSJdLAogICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAgICIiLAog
            ICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX3Bvb2woc2VsZiwgcG9vbF9uYW1lKToKICAg
            ICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKHBvb2xfbmFtZSk6CiAgICAgICAg
            ICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRo
            ZSBwcm92aWRlZCBwb29sLCAne3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAg
            ICAgKQoKICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYsIHJiZF9wb29sX25hbWUpOgogICAgICAg
            IGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0
            dXJuCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1l
            KQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0
            KGlvY3R4LCBUcnVlKQoKICAgIGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAg
            ICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgog
            ICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFf
            ZWNfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5y
            YWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAg
            ICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScs
            ICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMgdXBwZXIgY2FzZSIKICAgICAgICAgICAgKQog
            ICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIu
            b3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9l
            eGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkgaXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRl
            ZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAg
            IGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAg
            ICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVt
            ZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSk6CiAgICAgICAgY21kID0gWwogICAgICAg
            ICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIsCiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91
            cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1f
            bmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgIF0KICAgICAgICB0cnk6
            CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJw
            cm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9y
            OgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            ICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgXQogICAgICAgICAgICB0
            cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRl
            cnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQ
            cm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vidm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3Jv
            dXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVkIgogICAgICAgICAgICAgICAgKQoKICAgIGRl
            ZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9zZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9
            IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAic3Vi
            dm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGluIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVz
            eXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICBwaW5f
            dHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcsCiAgICAgICAgXQogICAgICAgIHRyeToKICAg
            ICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nl
            c3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQg
            cGlubmVkIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3Vy
            bCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5f
            T1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAgc2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsi
            UkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXQogICAgICAgIHJnd19lbmRwb2ludCA9IHNl
            bGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAgICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCAr
            ICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1pbi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7
            ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5j
            b2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAgICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAog
            ICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAgICAgICAgICAgICAgICBhdXRoPVMzQXV0aChh
            Y2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5kcG9pbnQpLAogICAgICAgICAgICAgICAgdmVy
            aWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRp
            b25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgICAg
            ICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJsZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3
            X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAgICAgICAgICByZXR1cm4gIiIsICItMSIKICAg
            ICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYgcjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8i
            KSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAne3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiAoCiAg
            ICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAgICAgICItMSIsCiAgICAgICAgICAgICkKCiAg
            ICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3JhZ2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9p
            ZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3Vw
            cG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBv
            bmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAg
            ICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAg
            cmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVuZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBp
            cyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5k
            cG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUK
            ICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19z
            a2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAg
            ICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBi
            YXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBp
            cF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYgZXJyICE9ICIiOgogICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0
            aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3Vw
            cG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3VwcG9ydGVkOgogICAgICAgICAgICBmc2lkID0g
            c2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJnd19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3df
            ZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAg
            ICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAg
            ICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgICAgIGYiVGhl
            IHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0n
            LCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGluZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBh
            cGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZhbGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3ty
            Z3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBoIGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgICMgY2hl
            Y2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRl
            IGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQgZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2
            YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNlIGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAg
            IHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSIsCiAgICAgICAgICAgICAgICAiLnJndy5y
            b290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19w
            b29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3df
            cG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNfdG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAg
            ICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6
            CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tfcmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHJl
            dHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlz
            aXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmlnX25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmln
            KToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2NvbmZpZyAhPSAiIjoKICAgICAgICAgICAgY21k
            ID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgcmd3
            X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICAiZ2V0IiwKICAgICAgICAgICAgICAg
            ICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICByZ3dfbXVs
            dGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAgICBdCiAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9j
            ZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJv
            ciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAg
            ICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIu
            b3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29k
            ZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0dXJuICIt
            MSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYgY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9f
            YXJyYXkoc2VsZiwgdmFsdWUpOgogICAgICAgIHJldHVybiB2YWx1ZS5zcGxpdCgiLCIpCgogICAg
            ZGVmIHJhaXNlX2V4Y2VwdGlvbl9pZl9hbnlfdG9wb2xvZ3lfZmxhZ19pc19taXNzaW5nKHNlbGYp
            OgogICAgICAgIGlmICgKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9m
            YWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAg
            KQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9t
            YWluX3ZhbHVlcyA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIG9yICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVy
            ZV9kb21haW5fdmFsdWVzICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAgICAgICAgICAg
            ICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICApOgogICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgInBy
            b3ZpZGUgYWxsIHRoZSB0b3BvbG9neSBmbGFncyAtLXRvcG9sb2d5LXBvb2xzLCAtLXRvcG9sb2d5
            LWZhaWx1cmUtZG9tYWluLWxhYmVsLCAtLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIK
            ICAgICAgICAgICAgKQoKICAgIGRlZiB2YWxpZGF0ZV90b3BvbG9neV92YWx1ZXMoc2VsZiwgdG9w
            b2xvZ3lfcG9vbHMsIHRvcG9sb2d5X2ZkKToKICAgICAgICBpZiBsZW4odG9wb2xvZ3lfcG9vbHMp
            ICE9IGxlbih0b3BvbG9neV9mZCk6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCB0b3BvbG9neSBwb29scywg
            J3t0b3BvbG9neV9wb29sc30nLCBhbmQgIgogICAgICAgICAgICAgICAgZiJ0b3BvbG9neSBmYWls
            dXJlIGRvbWFpbiwgJ3t0b3BvbG9neV9mZH0nLCIKICAgICAgICAgICAgICAgIGYiYXJlIG9mIGRp
            ZmZlcmVudCBsZW5ndGgsICd7bGVuKHRvcG9sb2d5X3Bvb2xzKX0nIGFuZCAne2xlbih0b3BvbG9n
            eV9mZCl9JyByZXNwY3RpdmVseSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybgoKICAgIGRl
            ZiB2YWxpZGF0ZV90b3BvbG9neV9yYmRfcG9vbHMoc2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToK
            ICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9yYmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYu
            dmFsaWRhdGVfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgaW5pdF90b3BvbG9neV9yYmRfcG9vbHMo
            c2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToKICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9y
            YmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYuaW5pdF9yYmRfcG9vbChwb29sKQoKICAgICMgdGhp
            cyB3aWxsIHJldHVybiB0aGUgZmluYWwgYXJncyB0aGF0IHNjcmlwdCB1c2VzIHRvIHByb2Nlc3MK
            ICAgICMgdGhlIHByaW9yaXR5IHRvIHNldCBhIHBhcnRpY3VsYXIgdmFsdWUgaXMsCiAgICAjIGNv
            bW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmluaSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVz
            CiAgICBkZWYgZ2V0RmluYWxVc2VkQXJncyhzZWxmKToKICAgICAgICBhcmd1bWVudCA9IGYiW0Nv
            bmZpZ3VyYXRpb25zXVxuIgogICAgICAgIGZvciBhcmcgaW4gdmFycyhzZWxmLl9hcmdfcGFyc2Vy
            KToKICAgICAgICAgICAgaWYgc3RyKGdldGF0dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSk6CiAg
            ICAgICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxhZy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRl
            cm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxhZy1uYW1lLAogICAgICAgICAgICAgICAg
            IyBzbyB3ZSBjYW4gZ2V0IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAg
            ICAgICBhcmdWYWx1ZSA9IGFyZy5yZXBsYWNlKCJfIiwgIi0iKQogICAgICAgICAgICAgICAgIyBk
            byBub3QgYWRkIHRoZSBjb25maWctZmlsZSBmbGFnIGFuZCBhbHNvIG5vdCBhZGQgdGhlIGJvb2xl
            YW4gZmxhZ3Mgd2hpY2ggYXJlIHNldCB0byBGYWxzZQogICAgICAgICAgICAgICAgIyBiZWNhdXNl
            IGNvbmZpZy5pbmkgZmlsZSB0cmVhdHMgYm9vbGVhbiBmbGFncyBhcyBUcnVlIGFsd2F5cwogICAg
            ICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGFyZ1ZhbHVlICE9ICJjb25maWct
            ZmlsZSIKICAgICAgICAgICAgICAgICAgICBhbmQgZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBh
            cmcpICE9IEZhbHNlCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIGFyZ3Vt
            ZW50ICs9IGYie2FyZ1ZhbHVlfSA9IHtzdHIoZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcp
            KX1cbiIKICAgICAgICByZXR1cm4gYXJndW1lbnQKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAg
            IyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBj
            bHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRl
            X3JiZF9wb29sKHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNl
            bGYuaW5pdF9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVk
            ZWRfa2V5cy5hZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19k
            YXRhX3Bvb2xfZGV0YWlscygpCiAgICAgICAgIyBkb3VibGUgc3RyaW5nIG5lZWRlZCBmb3IgdXBz
            dHJlYW0gZXhwb3J0cyBvZiBmbGFncwogICAgICAgIHNlbGYub3V0X21hcFsiQVJHUyJdID0gZici
            e3NlbGYuZ2V0RmluYWxVc2VkQXJncygpfSInCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BB
            Q0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJL
            OFNfQ0xVU1RFUl9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgp
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1
            bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RB
            VEEiXSA9IHNlbGYuZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0
            X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktl
            eSgKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZGFzaGJvYXJkX2xpbmsoKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoImNsaWVudC5jc2ktcmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2Vs
            Zi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIi
            KQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19N
            RVRBREFUQV9QT09MX05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfbWV0YWRhdGFfcG9vbF9uYW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQ
            SEZTX0ZTX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAg
            ICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJn
            X3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9H
            Uk9VUCJdID0gc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBD
            ZXBoRlMgbm9kZSBhbmQgcHJvdmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFw
            WyJDRVBIRlNfUE9PTF9OQU1FIl06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNl
            bGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQog
            ICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJP
            VklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBI
            RlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAg
            ICAgICAgICAgIyBjcmVhdGUgdGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAg
            ICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAg
            ICJjc2kiLCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAg
            ICAgKQogICAgICAgICAgICAjIHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAog
            ICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9V
            UCJdOgogICAgICAgICAgICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUs
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAg
            ICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAog
            ICAgICAgICAgICAgICAgICAgICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEi
            LAogICAgICAgICAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0g
            PSAiIgogICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAg
            ICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lO
            VCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BP
            UlQiXSwKICAgICAgICAgICAgKSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkK
            ICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNf
            UE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2Vj
            X3Bvb2xfbmFtZSgpCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSA9IHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9t
            YXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMK
            ICAgICAgICApCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X3Bvb2xzICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5
            X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2Vs
            Zi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZ
            X1BPT0xTIl0pLAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90
            b19hcnJheSgKICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX1ZBTFVFUyJdCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAg
            c2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9M
            T0dZX1BPT0xTIl0pCiAgICAgICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5
            X3JiZF9wb29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRf
            dG9fYXJyYXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAg
            ICAgICAgZWxzZToKICAgICAgICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9s
            b2d5X2ZsYWdfaXNfbWlzc2luZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJF
            RklYIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0
            X21hcFsiUkdXX0VORFBPSU5UIl0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X2VuZHBvaW50OgogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAg
            ICAgICAgICAgICBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5f
            YXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAp
            OgogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwg
            InJlYWxtIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBz
            ZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0
            aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVf
            bmFtZSwgInpvbmUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAg
            IGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAg
            ICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAg
            ICBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAg
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9L
            RVkiXSwKICAgICAgICAgICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAgICAgICAgICAgICBlcnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxp
            ZGF0ZV9yZ3dfZW5kcG9pbnQoaW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVy
            cm9yLCBzZXQgdGhlIFJHV19FTkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAi
            LTEiOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRp
            c2l0ZSBwYXJhbWV0ZXJzIG9yIG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMu
            c3RkZXJyLndyaXRlKGVycikKCiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAg
            ICBmb3IgaywgdiBpbiBzZWxmLm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQg
            ayBub3QgaW4gc2VsZi5fZXhjbHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3Jp
            dGUoZiJleHBvcnQge2t9PXt2fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdl
            dHZhbHVlKCkKICAgICAgICBzaE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAg
            ICBkZWYgZ2VuX2pzb25fb3V0KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkK
            ICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAi
            IgogICAgICAgIGpzb25fb3V0ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJleHRlcm5hbC1jbHVzdGVyLXVzZXItY29tbWFuZCIsCiAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgImFyZ3MiOiBzZWxmLm91dF9tYXBbIkFSR1MiXSwKICAgICAgICAgICAgICAgIH0sCiAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2st
            Y2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZpZ01hcCIs
            CiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHNl
            bGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAgICAgICAg
            ICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5nIjogInt9
            IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAg
            ICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZnNp
            ZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAgICAgICAg
            ICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgt
            b3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91
            dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAgICAidXNl
            cktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYgJ01PTklU
            T1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1lbmRwb2lu
            dCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05J
            VE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9OSVRPUklO
            R19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9uaXRvcmlu
            Zy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0ZXIiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiTW9u
            aXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAg
            ICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVfU0VDUkVU
            JyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicgU2VjcmV0
            CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNS
            RVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JC
            RF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAog
            ICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgK
            ICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAg
            ICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ1c2Vy
            S2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAj
            IGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            ICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAg
            ICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMg
            aWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBq
            c29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5h
            bWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119
            IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAg
            ICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0
            X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRoZW4gb25s
            eSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9M
            SU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAg
            ICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            IHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAgICAgICAg
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0
            X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQgbm90IHNl
            bGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6IHNlbGYu
            b3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            Ijogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmImNlcGgt
            cmJkLXJhZG9zLW5hbWVzcGFjZS17c2VsZi5vdXRfbWFwWydSQURPU19OQU1FU1BBQ0UnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAg
            ICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRf
            bWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsn
            Q1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIs
            CiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAg
            ICAgICkKICAgICAgICBlbHNlOgogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJCRF9NRVRB
            REFUQV9FQ19QT09MX05BTUUiXToKICAgICAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAg
            ICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgt
            cmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICAiZGF0YVBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9P
            TF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlv
            L3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0K
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJuYW1lIjogImNlcGgtcmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3Rv
            cmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25l
            ci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICMgaWYgJ1RPUE9MT0dZX1BPT0xTJywgJ1RPUE9MT0dZX0ZBSUxVUkVf
            RE9NQUlOX0xBQkVMJywgJ1RPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUycgIGV4aXN0cywK
            ICAgICAgICAjIHRoZW4gb25seSBhZGQgJ3RvcG9sb2d5JyBTdG9yYWdlQ2xhc3MKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXQogICAgICAgICAg
            ICBhbmQgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXRvcG9sb2d5IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAg
            ICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidG9wb2xvZ3lGYWlsdXJlRG9tYWlu
            TGFiZWwiOiBzZWxmLm91dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xP
            R1lfRkFJTFVSRV9ET01BSU5fTEFCRUwiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5WYWx1ZXMiOiBzZWxmLm91
            dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xPR1lfRkFJTFVSRV9ET01B
            SU5fVkFMVUVTIgogICAgICAgICAgICAgICAgICAgICAgICBdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAidG9wb2xvZ3lQb29scyI6IHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNy
            ZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9j
            b250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lf
            UkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBI
            RlNfRlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZz
            TmFtZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAg
            ICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFt
            ZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQ
            SEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JH
            V19FTkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNz
            CiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3Mi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            ZW5kcG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAgICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAi
            a2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAg
            ICAgICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VT
            RVJfQUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2Vs
            Zi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAg
            ICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdS
            R1dfVExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0
            IiBzZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5k
            IjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJl
            dHVybiBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vy
            c19wZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlz
            aW9uZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNl
            bGYucnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2Vs
            Zi5ydW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2Vs
            Zi51cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25h
            bWVfZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAi
            Y2xpZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBw
            cm92aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0
            aGUgcmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2Fz
            IHVzZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1
            bHQnIHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0
            YSIKICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAg
            ICAgICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3Jh
            ZGUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lv
            bnMoc2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4
            aXN0cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVu
            dGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29u
            X291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAg
            ICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJp
            bnQoZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICBy
            ZXR1cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAg
            IHNlbGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5n
            X2NhcHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNl
            cikKICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAg
            ICBjYXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAg
            Y3VyX2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAg
            ICAgbmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            Y3VyX2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4g
            Y3VyX2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAg
            ICAgICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBm
            b3IgeCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAg
            ICBdCiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8g
            bWFpbnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3Qu
            ZXh0ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNh
            dGVzIHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1z
            IGluIHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5l
            d19jYXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4
            IGluIGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAg
            ICAgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAg
            ICAgICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFw
            cGVuZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tl
            YWNoQ2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRo
            IGNhcHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBj
            YXBzLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRf
            dmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pz
            b24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29t
            bWFuZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHBy
            aW50KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihz
            ZWxmKToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9u
            cygpCiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAg
            ICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRl
            ZF9vdXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3Vw
            cG9ydGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQog
            ICAgICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmls
            ZSBhbmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRw
            dXRfZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3Jp
            dGUoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJq
            T2JqID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2Vw
            dCBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1
            dGlvbiBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJv
            ciBhcyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQg
            T1NFcnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBv
            dXRwdXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0
            ZG93bigpCg==
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:62cd98508ab41f9526f10c55f6c8054475a1b029230f255fe7461fb750fd1def
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        version: 4.18.11-rhodf
      entries:
      - name: rook-ceph-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: rook-ceph-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: rook-ceph-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: rook-ceph-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: rook-ceph-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: rook-ceph-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: rook-ceph-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: rook-ceph-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: rook-ceph-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: rook-ceph-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: rook-ceph-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: rook-ceph-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:bbfddd7cb4a6cb7a131a0cef174054890fd39956d63e737c92fe8233ad8f98e4
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5a35593357fbf280561a1d4a64ecc0d5c2044409acff3ce6cb490a4a61462a4e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:791ac512370af5cde1e089de3def760e7dbce69badc03137051520ce316d63e2
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5dc66b7c486d1604c4bbb04c85544147c43264840e597958e6c5ca83e7e34c0e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:ddfdf850c9238a6988889e1840efe3175f03dad2489a5651e21cac62ac44e23e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/application-services-metering-operator@sha256:c3600cc8cd789df1f3c8f48bb3b1bc4c9cbf0285b6f6ce06f1d563c52f50915e
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.connect.redhat.com/cpopen/common-service-operator@sha256:13f5b6699ab505d02550b09e100f7e205c7fde3f2bc36317025bc795be216761
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - registry.connect.redhat.com/isovalent/clife-bundle@sha256:6d2ba514e40559fe3568dd0107f5e6648d43b543b3dbe4319e4114ca05fe20a6
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/vault-config-operator@sha256:696bfe3769605870286e228f463dc2582202d6aeb8f199d78d87f153c68d7401
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-emrcontainers-controller@sha256:cde2c51ca18d0969be0549e37ecd03cc430be3b4746d53b8744cb552a2c475df
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:29ec87ce725d2c277824312607915bb7de92e18015c388543173dbdadd5834a5
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:c906c822fbab57c1eb95880b0dfb9cd04100c86222a5d5cfe6342298e3bb1b9e
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:e26bc4b183f69a448552f9fe740292d4e803165d341d7d66047688d01fcd060c
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
        version: 4.17.13-rhodf
      entries:
      - name: ocs-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:a4b5197db1ccda9934f30bfd63e41b0bef430ea9bd224cd3dc022466742b82b5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:0fcc80541d8966f3ea3502da1c4fdda24de277d8d73d0752b46e7d69fbcf7038
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
        - registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: ocs-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:9fcdfa4615b91a5bb919aff71fc4df046365b2ee6d540b5f5f47f9ccaeb9877c
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-003
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:9f5b63888cbf4b1c90e3803d0017a01ad9c0c22693b978c9ee8299bfc5337832
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:212ab9249707ca8ff835186d88e7888af18118fcbe98d948a3156b9700307a79
        version: 7.6.12-opr-003
      entries:
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/ipm-bundle@sha256:68c38ac59cf0d6cc145eb694e5b5a724da733caddb82c5585ef132e880dde544
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.connect.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:19b163f055da43946643057c07500784210d02db868d2aa10f25a6501d767d3d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/community-operator-pipeline-prod/cluster-manager@sha256:a4c8a54148538b371e5cea3b2f2d4109f7b6bc4a7f2e5b52a840c9340fbf5e47
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/openshift-community-operators/oadp-operator@sha256:f8c5c6e59f13db8847080ea6d964da6e50e56f7d373bcc4e4278b1a8a80395e4
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/openshift-community-operators/oadp-operator@sha256:b7b4d340157513da2ebef6bc7bc262967d0e825e2baa1f2bc2ab2e170457e015
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a6399dfd6fe0e8dc9f59fac4382320d23f9e1709e48ed6da7db69e76354b2646
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:0a86a1808d9906eb57886d3d3cd5c04cf9578a80860342132d36ecfc411bf807
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:b02bab3af52a2469ced96618f1d122464568ba21ebc0a3a945b78ec4e6bdc80d
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a614ec94c2468e34e129e9ecd4bc0d1895eb67b96e77eb5ad94cd8bd507860f5
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Net AI
      provider-url: https://netai.tech
    name: netai-demo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netai-demo-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deployments.netai.tech/v1alpha1",
                "kind": "AntennaDeployment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netai-demo-operator"
                  },
                  "name": "antennadeployment-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Networking
          com.redhat.openshift.versions: v4.18
          containerImage: ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
          createdAt: "2025-07-09T11:33:08Z"
          description: Net AI Operator that uses EnergAIze engine forecasts to scale
            Red Hat OpenShift resources.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Net AI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntennaDeployment is the Schema for the antennadeployments
              API
            displayName: Antenna Deployment
            kind: AntennaDeployment
            name: antennadeployments.deployments.netai.tech
            version: v1alpha1
          - description: Demo is the Schema for the Demo API
            displayName: Demo
            kind: Demo
            name: demos.deployments.netai.tech
            version: v1alpha1
        description: POC Operator that uses saved EnergAize forecasts to manage OpenShift
          Custom Resources
        displayName: netai-demo-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - energaize
        - scaling
        - scale
        - manage
        - forecast
        - cells
        - 5g
        - networking
        - net
        - ai
        - network
        - telecommunication
        - xapp
        links:
        - name: Netai Demo Operator
          url: https://netai-demo-operator.domain
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Net AI
          url: https://netai.tech
        relatedImages:
        - registry.access.redhat.com/ubi8/ubi@sha256:68fecea0d255ee253acbf0c860eaebb7017ef5ef007c25bee9eeffd29ce85b29
        - registry.connect.redhat.com/net-ai/demo-operator-bundle@sha256:b1a30de1e984da0487a7d042bd940876bedb5a2f43dd73c7895e684a3a35d81e
        - ghcr.io/net-ai-tech/netai-demo-operator-bundle@sha256:0783aba7b039a569b49de2a61ef6dc6ff1d1f301d93c26d623b7f652321871c5
        - ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
        version: 0.0.1
      entries:
      - name: netai-demo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: netai-demo-operator
    provider:
      name: Net AI
      url: https://netai.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:42102a2e462e27dd1de28ca3f7db02d227c44736f2d57c05037a83ecbff6dd9f
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:b405d6148aa680bf31bdc6185c1702916c3219983d8482f27d89ed2f93659aa7
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-recyclebin-controller@sha256:9d916ed0af37d60d90f9b4711b1544b9373bd247476278c9c4041c9dbce6acc9
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:c60664e5da8eb23bbd918b1c13b83565b1f33a1aa8961b425232fa1ef4d1a6f1
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:82cafd48db8107dd141064ce7530468b11804c8c54235a317eb450a7606bac3e
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:5d9b56300a3b17178467954b533bee19622f8e96fe9d4ac9745ead2c59b9ec5f
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-efs-controller@sha256:338a28600b08d1ead95ed409e3326e03f3c5e02ed840d3be56a950db6b1de061
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/ipfs-operator@sha256:03a61ef328feb626408d42904cdfb9d45f66e022d5a1e1088a15b74455487743
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.connect.redhat.com/yugabytedb/yugabyte-platform-operator-bundle@sha256:4cee86e973206b5300e88af088aca700cde3b2e2e5e1c70b9a20289e9f592349
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3e0fde49a8e8c5f780a76087bbc42507a42fea0e1fea572ae05ba66d27338365
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3b36dcea36ebf4b2e95fac6691f6fc8f1242f7e5c38195a9af28bfad514c28cb
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:36dc3710ac11d86e1cf0a3c6994ae0e6feb6c8ff1e8110997f27f4e573c259bc
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:64cadb49077eff7863dd678c4aec70f4101f17ee8e0bf7ca64ce7010c9bf4b9b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:4981d96081cc3d10cd26f20aab792e9c195b05bdcd609367011b7e9c7e6759b1
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-apigateway-controller@sha256:44d34788b43d103530115fc53b0252bfdbe6e07a692847a3d32aa21e45892ff9
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        - registry.connect.redhat.com/tigera/operator@sha256:4d0d814fcfeb628aae8a05cb7a0a416ae85395197950085c907b6cdcf2fb4317
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        - registry.connect.redhat.com/tigera/operator@sha256:b1a8b56ff5e1bb5326e1bde4931303358410da9768f0ebf9c08e2c2d8661d742
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        - registry.connect.redhat.com/tigera/operator@sha256:cb7f3367ff064cd01dc636a43fd89d5f6c4e6a7004cb7c94b1758b2398e36d71
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        - registry.connect.redhat.com/tigera/operator@sha256:de9724b84ca493d3b14c8f801c1049ac4713166b00d9bf3ea0b1986aac5040fa
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        - registry.connect.redhat.com/tigera/operator@sha256:8e0c026f369f80e8b36d0c94f1fd60f7f53656e57dc5be0b8b12189546874909
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        - registry.connect.redhat.com/tigera/operator@sha256:1fb1007f7aad744f0e1b0f6a32022c2571ee8558e9db53c85976bbd2a891a6b0
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - registry.connect.redhat.com/tigera/operator@sha256:42a4d0d0e24752dbb51daebcfca4d4850d04e79578bfb8b41cdc2a2a289df4a2
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        - registry.connect.redhat.com/tigera/operator@sha256:7894b770bb02b2271e1414d67f300218cc646534c9f2a6917e51dc8bff8ba731
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/openshift-community-operators/awss3-operator-registry@sha256:9ab3a2cd00b8704a8bae8e9b75f2a95d0a3e60fa3cb52589db8a4e7641b73027
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/openshift-community-operators/syndesis@sha256:9e8c08793aebc8857dcd451fa2fde2b3f05b9ff4f3d3cc604f9f752193f58d42
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/verticadb-operator@sha256:b7691da906474a8ab829966bb3da4a846eaf12909c97bd0d9cb6a1e5e7ff6c03
        - docker.io/vertica/verticadb-operator:1.11.2
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        - quay.io/community-operator-pipeline-prod/verticadb-operator@sha256:be706a582a4b8cae487835e920288b41229834cd3477b7389adb363db4b4f2ca
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              },
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "ServiceFunctionChain",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "servicefunctionchain-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "servicefunctionchain",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "servicefunctionchain-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-10-15T19:06:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-dpu-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
          - description: ServiceFunctionChain is the Schema for the servicefunctionchains
              API
            displayName: Service Function Chain
            kind: ServiceFunctionChain
            name: servicefunctionchains.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-dpu-operator-bundle@sha256:4e98c0f629e938a0c9d011e6a675d45589b044a02b3f9551fc0e8f7cd172c8a0
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:acb5ab931f4d7a46bb195c59d156441af5e9092b2739bcb0379b89d47ccf4e81
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:5dedc4733081fc21bcb77dbf4f4c032bf418f00bc4371c29c7ce028f51833c0b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: dpu-operator.v4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: dpu-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: dpu-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: dpu-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: dpu-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: dpu-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: dpu-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: dpu-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: dpu-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: dpu-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: dpu-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: dpu-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: dpu-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: dpu-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: dpu-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:4ab17b92016288f1ebfb2106c826a17e75d041f0dba7eaaf7545193c3305a408
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - registry.connect.redhat.com/streamnative/function-mesh-bundle@sha256:e4377187c20a70d768a253bc0794f3152aa6fcee22ddbf2fc88f45756e9681fe
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - registry.connect.redhat.com/loadcore/loadcore@sha256:7fca6883b092ac7955bb42f893f163403960c81b83e360cfd1eac802e57ea839
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.connect.redhat.com/cpopen/odlm@sha256:064b597625550ca75dbc9f3c8fd6b2de6e0b9cb7f42d604e871bfc668ad05ac1
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:eccacdf9ac2b0b1f6061aea457c705319f268bfa30631ed9919ae393f0ef7f71
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:cf330a9e405a9641c213a74baf28307a6e247c4f03cc0523d071dad8f5f9e2e9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:412403b096566965b2c92a0a55beed8ebd6d8113a7e1fdf5eb3334769929ca41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:c89d8be470c2fd052c5be7a7dbfb133c9e3528bfa5b3681ffb8b3223c4c6bac2
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:6c9e04df3f5ddc45788333a85f6aeff8953b008a4b165e9cc2d5a079c6c73ece
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:5611a2fd53859d55d5dfdfdd117e80a0f30be0a4489ad838d571e872d3b5158a
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        - quay.io/openshift-community-operators/horreum-operator@sha256:a9d82ce368b015e45b93f917d4069978a0ba323f3dd684d470cdfd262e0c9f05
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-wafv2-controller@sha256:cb9a426baec724225c0132973d5b90a83672f9b0b75451adc6d7fd858c79ab58
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-glue-controller@sha256:2e1e23de186f3c61be8c7a30f3b295fe3b8b101509cf30bd6b7bff055ab74c82
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.109.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
          createdAt: "2025-09-18T09:13:25Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/eclipse/che-dashboard@sha256:8e2285991a725e87d2a9131ea23e08ce97cd5bdf1118f53a55a140c2c97ca25d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:1df1b75f702952b650e21717b04de5f39c84ca8f55f4afa0f32644ef8638d03a
        - quay.io/che-incubator/che-code@sha256:3b5c7e1ae664c82ddb3eaedc8161eb72a3a1a22c6c5e3f342f4aad972a6282a5
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/community-operator-pipeline-prod/eclipse-che@sha256:545d98c512c99bc7bf4ccd02c1eb554bbc4bca7da8cc66e0e26934edbfd50c59
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/che-idea-dev-server@sha256:0344c9574a6544c4c0ef93a798a9705ca3a4624508a55c8f94c2bae7196f914a
        - quay.io/eclipse/che-plugin-registry@sha256:bd71aa6817eeda0449056e78a6e832b82dd37b65ec6cba2bac3ffb9defa4048e
        - quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/che-incubator/configbump@sha256:77bd0fc269ff3724583b43ec68e158fcd300e825e1b44e24e2b9e67ed28df6da
        - quay.io/devfile/universal-developer-image@sha256:f76ea662cfdb9e80b93bcb58a393aed8b52aa72e7da73762d30a3ad22c3f06d9
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/eclipse/che-server@sha256:e5a542da0484eb2223ac292e6005cf7e116bec15d377557f92949728e1f5c8bf
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        version: 7.109.0
      entries:
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/group-sync-operator@sha256:fe6fda2de67fc4996d3f388e11faa0130fee19453111c829ba0aaa5be5410d01
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:e6d5ef5583ae26a8312dc331bd6003fda6c7c248de15b6790a128d0cd03965dd
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:35e0ee6b3ec5cb654b709c5e4a76a1f08182d826b47f077031d9139d4c2cc308
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b876ec88e1e64c17572e956b1c7e4020f5a423c1275f7f57025823ff7f749c61
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b3a45189ae8fc5302fb8a9e15d5747bacce044537923621edf21e9d3b267a27e
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-bundle@sha256:6cd4cba02a9e5c22062108363f6a052734c398019a19f84afc18a7b05a71d578
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        - registry.connect.redhat.com/loadcore/keysight-wap-stats-dashboards-bundle@sha256:f69f61eb818a1f0f4035ae79487402237b739c55eb104fc4643a7d3ef8d7d0c4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:6d9eabdb4f645aa1b8b6c08ca94f9ec89466cdc35b3cb02aaba22e160d8fd9ea
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:5484f8ecf079ee126d24c05e1bcb06467146b3b8e751e90e0fd68f3e933125f0
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:11ce1c35f9c3d9b5ef9539b4402eb8358df843b29ffd94af7a453e5d765e4767
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecs-controller@sha256:280d5ec57ad4deb45c145e702242468a4cc04d23112c33f4e72722a269c900e8
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-object-store-operator@sha256:c5af6adc5e7f9bb2d5bf6025b1b8c3e33ce5b9349a0246df1a26b7ec91b2ea2f
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lightbits Labs
      provider-url: https://www.lightbitslabs.com
    name: lightbits-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightbits-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "CSIDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "csidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "csidriver",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "csi.lightbitslabs.com",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "controller": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "tag": "v1.21.0-ubi9"
                  },
                  "node": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "sidecars": {
                      "lb-nvme-discovery-client": {
                        "registry": "docker.lightbitslabs.com",
                        "repository": "lightbits-operator/lb-nvme-discovery-client",
                        "tag": "v1.21.0-ubi9"
                      }
                    },
                    "tag": "v1.21.0-ubi9"
                  }
                }
              },
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "cluster1",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cluster",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "cluster1",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "addresses": "192.168.40.46,192.168.40.47,192.168.40.48",
                  "apiService": "192.168.40.46:443,192.168.40.47:443,192.168.40.48:443",
                  "clusterAdminSecretName": "lb-secret",
                  "clusterName": "cluster1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          com.redhat.openshift.versions: v4.17
          containerImage: docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
          createdAt: "2025-08-20T11:35:15Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/supported-versions: v4.17
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '[]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/suggested-namespace: lightbits-operator
          repository: https://github.com/LightBitsLabs/lightbits-operator
          support: Lightbits Labs <support@lightbitslabs.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster is the Schema for the clusters API
            displayName: Lightbits Cluster
            kind: Cluster
            name: clusters.storage.lightbitslabs.com
            version: v1alpha1
          - description: CSIDriver is the Schema for the csidrivers API
            displayName: Lightbits CSI Driver
            kind: CSIDriver
            name: csidrivers.storage.lightbitslabs.com
            version: v1alpha1
        description: |-
          Operator for deploying and managing Lightbits CSI driver.

          By using the operator users can quickly and easily deploy and upgrade Lightbits CSI Driver.

          ## Requirements

          - Kernel support for NVMe-TCP
          - Kernel support for NVMe Asymmetric Namespace Access (ANA)

          ## CSI driver Supported Features

          - Filesystem Volumes (xfs/ext4)
          - Raw Block Volumes
          - Dynamic/Static Provisioning
          - Volume Snapshots
          - Volume Clones
          - Multi-Tenancy & Credentials
          - Volume Expansion (Online/Offline)
          - Storage Capacity Tracking
          - Host-Side Encryption (LUKS)
          - Publish ReadOnly volumes

          ## How to install

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/install_lightbits_operator.html#installation)

          ## Documentation

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/)
        displayName: Lightbits Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - storage
        - lightbits
        links:
        - name: Lightbits Operator
          url: https://github.com/lightbitslabs/lightbits-operator
        maintainers:
        - email: k8s@lightbitslabs.com
          name: Lightbits Labs
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Lightbits Labs
          url: https://www.lightbitslabs.com
        relatedImages:
        - docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - registry.connect.redhat.com/lightbitslabs/lightbits-operator@sha256:53db496f7a254632a25acf8974aca4041cbddb109499871622ffc2182e7aaa83
        version: 1.2.0
      entries:
      - name: lightbits-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: lightbits-operator
    provider:
      name: Lightbits Labs
      url: https://www.lightbitslabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ram-controller@sha256:a90b07270f0cfa8929e01569895ab9a209ee44dcc452a3e1f6e0e9a52c5f13b9
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-09-30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
          createdAt: "2025-09-30T03:27:25Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.41429998

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - quay.io/community-operator-pipeline-prod/sailoperator:1.28.0-nightly-2025-09-30
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-testing/proxyv2@sha256:28f058629e7969e8441b677b308a09bf5d4968dbd5024d54df8472d1f43e1a8e
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-testing/ztunnel@sha256:f4895a66a99ca0c87e54260d6d5784068a48fe7fbfa25144debad420ea7223fa
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-testing/pilot@sha256:42d64978872f6be8540027002a361933e4fe69590458353e39d2643b2459bb3c
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-testing/install-cni@sha256:a8c8f6fb3048461e08b86056000a3034f7945df067d5ebf43cca9bcb0ffa3916
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.28.0-nightly-2025-09-30
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      - name: sailoperator.v1.28.0-nightly-2025-09-21
        version: 1.28.0-nightly-2025-09-21
      - name: sailoperator.v1.28.0-nightly-2025-09-19
        version: 1.28.0-nightly-2025-09-19
      - name: sailoperator.v1.28.0-nightly-2025-09-18
        version: 1.28.0-nightly-2025-09-18
      - name: sailoperator.v1.28.0-nightly-2025-09-17
        version: 1.28.0-nightly-2025-09-17
      - name: sailoperator.v1.28.0-nightly-2025-09-16
        version: 1.28.0-nightly-2025-09-16
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/install-cni:1.23.0
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/pilot:1.23.0
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:a5698068f57fb27c6e4b8c385878c4651967ae5f17412c6ea6cf8bd34bfe76a1
        - docker.io/istio/proxyv2:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/pilot:1.24.1
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:5416c5fc62c7d9498349ed771bf93bfc0f712ad8da0b8ee109dc49d1e684fa56
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.3
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/proxyv2:1.24.2
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:9cc4d5a9f46e4d05f8adb5232459002a073c101cc813330325afb6682e2138ae
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.4
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - quay.io/community-operator-pipeline-prod/sailoperator:1.26.3
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-sagemaker-controller@sha256:c962ee90891a05197edc08b4cf015fce1e2431cced90bd274e89bb87708420f3
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-bedrockagent-controller@sha256:6688b26f71500af0e2ffbd0ccc53926b0c41adaaf817a88f0fdfd6fb398075e7
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        - quay.io/community-operator-pipeline-prod/trident-operator:25.6.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.32.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90",
                  "kernelVfDrivers": [
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
          description: This operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Intel device
            plugins for Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              GpuDevicePlugin is the Schema for the gpudeviceplugins API. It represents
              the GPU device plugin responsible for advertising Intel GPU hardware resources to
              the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              QatDevicePlugin is the Schema for the qatdeviceplugins API. It represents the QAT device
              plugin responsible for advertising Intel QuickAssist Technology hardware resources
              to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              SgxDevicePlugin is the Schema for the sgxdeviceplugins API. It represents
              the SGX device plugin responsible for advertising SGX device nodes to
              the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fpga
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4
        - registry.connect.redhat.com/intel/intel-device-plugins-operator-bundle@sha256:4a422ca897caeabb9fc3be3dfef3746819868de1674ea61c657cad257fbddf12
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
        version: 0.32.1
      entries:
      - name: intel-device-plugins-operator.v0.32.1
        version: 0.32.1
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        - quay.io/community-operator-pipeline-prod/flink-kubernetes-operator@sha256:7c919847b88a5969d0c0830e4154512bbca28015cf6fb7fa70cd20824a726685
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:13013220d7376d23df4780fe65c3bea325bdeb7b7a89d5704aa52e948ed7a2d5
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-operator-bundle@sha256:3fb13020c6a34b329825d268dd70c3c9310075780768312dc52cac1e7134a030
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:aa2267c1c871699db37763e938125f290ace4aff2aca6f668b2ce01f7df92b8d
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        version: 4.18.0-202509100344
      entries:
      - name: aws-efs-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: aws-efs-csi-driver-operator.v4.18.0-202508260754
        version: 4.18.0-202508260754
      - name: aws-efs-csi-driver-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: aws-efs-csi-driver-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: aws-efs-csi-driver-operator.v4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: aws-efs-csi-driver-operator.v4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: aws-efs-csi-driver-operator.v4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: aws-efs-csi-driver-operator.v4.18.0-202505280604
        version: 4.18.0-202505280604
      - name: aws-efs-csi-driver-operator.v4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: aws-efs-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: aws-efs-csi-driver-operator.v4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: aws-efs-csi-driver-operator.v4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: aws-efs-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: aws-efs-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: aws-efs-csi-driver-operator.v4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: aws-efs-csi-driver-operator.v4.18.0-202503190402
        version: 4.18.0-202503190402
      - name: aws-efs-csi-driver-operator.v4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: aws-efs-csi-driver-operator.v4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: aws-efs-csi-driver-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: aws-efs-csi-driver-operator.v4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-operator-bundle@sha256:7f3d41672fa8c60725304f0d45e468a4f01defcab7352501737ffddfd32edded
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-operator-bundle@sha256:75cb78f7a40d3daedbc11cbbb5505e029328b4120f0de11d4f6423de943e92d8
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:279f5d817f9dc90b202924fc2b2ab59e5303e9dddc1d78d71017d65cdd66d565
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
          createdAt: "2025-08-26T06:53:59Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/hpestorage/xspc-csi-operator-bundle@sha256:658dd12fe1b92bf38dd24601043cb84e27a27b0d987a36a0ec7d99b0514574e2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:4c289e5023fd3e59b2cf40335eba9eb53caf562472f16dd97b370773a0ba608b
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: xspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle@sha256:8cba51274389030d632c8b82255399217d485146f4797907134e3c4ee55cb4d2
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:3bc76999ece42fcb3acd913b32be3fa0e0da056832922819b81119dcfb5f1582
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:2aeba259864bbddaa8cdd46ab6d8c3495450a0081469a27b9acde53008d91bed
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:f58f60283f6cf9ea745f4a25378ef6690d98af8fda0eb63f663066631e71450b
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:48b37c9f9417ef442d896e79851b176505045b487bddf77bb4e6d91fd5ad3dd1
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        - quay.io/community-operator-pipeline-prod/node-discovery-operator@sha256:899baace103c58e7337ae5cd75e3764e6de26966aab7340037c612c26ca7cf87
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:53777788253eaa1199e1854d14782619aeec0df9844d86b83742fe0005671cde
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:e3a13ea8b84b9e1abd158d394e11113e5f6c04d879547fa1d3e8638f1ac979d8
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:31258400bb37f4b8837ef294820f8a728ae83f8031d4e01ff0392690a315a30f
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        - quay.io/openshift-community-operators/move2kube-operator@sha256:cdf288be5ea779e03900af9b57c4d922b8760b88da4fcbf88437f8031d679aef
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:236a53d6fad949285b6abafcdd1a4541a7748f952c0c96fa2e71500079579ba4
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:38d03dacca046dda3ba08c39943ec2e7db33f1c7bb897e84fe1308a0982c63f1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        - registry.redhat.io/rhcl-1/limitador-operator-bundle@sha256:4c56cd8462531cf8ba39316a0c6f7493240bfa4a6ace77352fe7ac1597e2fd90
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        - registry.connect.redhat.com/netapp/trident-operator@sha256:53a4889e1036af9f858c9fdb1bd58f59b758a301bc773379572cf279f930f5c6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - quay.io/openshift-community-operators/ocm-operator@sha256:ca6a3b24ca4ecfcab52ee0402e007d6f14b23fa7f3408d740923806989b170e4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/community-windows-machine-config-operator@sha256:2b79f0e6b8bc069fac7d6cb5c86783fdc98ccae30d279ca0492f1eca96b6ea3f
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-eventbridge-controller@sha256:3a284fb7884e50ed9976368f312c75fc1f1cb2ec361297b38ee72d0ac6b66fc7
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:1f4fd5ed8e398b713e733fb552e7294f22b3927d4e017f2ac15ea20024d3de32
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:759dc7d6836f8c8830ec145e3088bddc1b63bd40ff3047079fa8f0e986e74c30
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/openshift-community-operators/konveyor-operator@sha256:04e63481c5104bd034283d5d4b0567167f62480d2c61154b1d16dbd41c344dfb
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/openshift-community-operators/konveyor-operator@sha256:d4c590033b1a2bcd67375465298e5cb6d136fe091abb12b8d009aa683c23ef62
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:848db711d9e8a329b689b40003293fce19c43faedd31afd80d9ae9d5c11cda49
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:e29f175a799597c84f0663a66767bba24feb78f802516056b35b5beb14d9eb41
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:2d58f12044b0951c87c851527240196db148acb9797e6ac49f9474a6e6039d71
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:7b8bdecbe3b3def1ec810ad53f604c7a71daebcfa7433cbb3a0db937580d087c
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:16c4f165143b8bab7f3eb08c815c2cae0dbf7fffd9ce7a625357be5806c24c5c
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:c28522543b039454d2d97b48198d099ac0b8211cbc58f13468d5f70b9f69b037
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:18ba5c587dc613f9c1a8bc9e66081e52f3a37eedeea5e75881e628e9c35ef460
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - registry.connect.redhat.com/nvidia/nim-operator-bundle@sha256:3dbd307d32a39e37a02ac8fa9c9fb15a0c2fd6dc5aea12d55abb061e86365124
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-directory-operator@sha256:9d93e3aeef3cfbba0b655de7783ba9fd1f4cb4ee5de4389005b7c31d971d4947
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/global-load-balancer-operator@sha256:50771c320c8b7faea144f765535e89979f9f81e61d34c8dd2e86139fa68e222e
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/aerospike-kubernetes-operator@sha256:bfa83ff729ef8b3a4ebca27b02e223327d3e1351d8bc78a06c18873d288dfb02
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        - quay.io/community-operator-pipeline-prod/aerospike-kubernetes-operator@sha256:63cdb7ea6cd6bda4c6ab2c1f3467489c2cef6c8a95836763b9c88b20c43e8168
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - quay.io/openshift-community-operators/sap-data-intelligence-observer-operator@sha256:5217360266f62fc8ba087fbc730074c14aacf62dbd124fdc3aa46650d51a243b
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sap-data-intelligence-observer-operator@sha256:79aa33cb7e4b570fc9c962fc6bd59708c7e28e81e2b188608780fa085db91442
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/kueue-operator
    name: kueue-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kueue-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "policy": "Disabled"
                    },
                    "integrations": {
                      "frameworks": [
                        "AppWrapper"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "FairSharing"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameOptional"
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob",
                        "Pod",
                        "Deployment",
                        "StatefulSet"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "byWorkload": "Parallel",
                      "policy": "ByWorkload"
                    },
                    "integrations": {
                      "frameworks": [
                        "RayJob",
                        "RayCluster",
                        "AppWrapper",
                        "PyTorchJob"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "Classical"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameRequired"
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-05-01T20:58:29Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.operator.openshift.io
            version: v1alpha1
        description: Kueue Operator description. TODO.
        displayName: Kueue Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:de644d640f355815f3aa40ced5456bb5b0e958aeba8bfbdc4ab360fcdb90f26d
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:0c4d9cd97f7579adbf2afd238c953ca5f170408cf27e3b2bf4aa16c108a76881
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:dd765c446564d4397570f4bc056b55989f6c0bae0d70514874e67765c9c77889
        version: 0.1.0
      entries:
      - name: kueue-operator.v0.1.0
        version: 0.1.0
      name: stable-v0.1
    - currentCSV: kueue-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-06-24T02:48:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:30659b42f5053dd756190211454c01f049bbc7689dddda3ecd8c406b157873a7
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:9a0b85b0752135694193befe00d10f7abdc0c99c6b9f295d59b237fa0fe3e6d2
        - registry.redhat.io/kueue-tech-preview/kueue-must-gather-rhel9@sha256:e5306c95bf6c5f55be6e791fcb66fe9b34f4f982536f4f90f8ed74acf3f6c31d
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:6f0b5f62f6e549c6f6769dd9a358a14668d7244f9c20753d1690d3e09948588c
        version: 0.2.1
      entries:
      - name: kueue-operator.v0.2.1
        version: 0.2.1
      - name: kueue-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: kueue-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-07-15T01:52:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue/kueue-rhel9-operator@sha256:5db0af8088cce6e9e7a88faf80855b6eaa1dd8679c6a56dd5ee98c50a3a8e2e4
        - registry.redhat.io/kueue/kueue-rhel9@sha256:16007bd38c480b607b63c82b3ee8721441cc374ee31c04b9e329426314590195
        - registry.redhat.io/kueue/kueue-must-gather-rhel9@sha256:f5f624ef8dbb3676ed4e37335e226e76e5f886c1990c6e0e8c0e4762740ed98d
        - registry.redhat.io/kueue/kueue-operator-bundle@sha256:1bafac48c00fb47972f512ccd5d846efbbccccd2259395de17537b39834d89df
        version: 1.0.1
      entries:
      - name: kueue-operator.v1.0.1
        version: 1.0.1
      - name: kueue-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: kueue-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/kueue-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-bundle-certified@sha256:a9b8102ab0ec81e5bea84332ef20b023d06660b410e7e2deadb00904d2098613
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opentelemetry-operator@sha256:5dc347a5085c10823a122fab0bd42afd5d9b9aa19ba17e8992e9492000886629
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        - quay.io/openshift-community-operators/snyk-operator@sha256:97908144406700ae9c038316661dc7904f3e0f4e01b1bbdf68259b59dfb86d6c
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cadee58c1c1040fad5a46fc9be5fae4a1944845c5e736c50fa192192d5c17a36
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cf782e9e2a62a2fc71c81032e6bf40f9ba15d37c33feff2df551edb111c4e17c
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:60a9e4b05e7eb17ebfe58dd55db96ab2c68580d9282117f74f21869cb63e4a73
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:1461576c2e06c2ee8d84fd40836184566a346ed596b99cb81fe6a08d0ad99096
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        - quay.io/openshift-community-operators/camel-k@sha256:2b8369f729c18384f45a458e88c9f551eb00f26792b40dbf286e6ea2837e9400
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:6154ff0f72c9c49f1b7ea0edbe93e0a3b04f6e63abbd851d0e6706796050d672
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:21ce00e81ce495c142cbe9611f494cfb26b52b3c09a507bee0572c1e1e3333d7
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        - quay.io/openshift-community-operators/camel-k@sha256:26df9d918e942f7fc637212df87a114b5edd52570da18cc212b3499354cdb993
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:11dac0881f5ed66b5d42b2d69b13aba31cd56e6b32157cb3ce977be197189b79
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        - quay.io/openshift-community-operators/camel-k@sha256:cba8cde86a870b31f7dae9a554f4c496ed81731f91a8e64cd62b0caa88efaa38
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:06d31bb777b37f8f7c8981efc24ca26a2f9c56a1c579243d0c7bdddb34c76a98
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-dependencies
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-dependencies.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          createdAt: "2025-09-11T15:25:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          deploys dependencies.
        displayName: Data Foundation Dependencies
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-dependencies-operator-bundle@sha256:6c2223357b672b1a59d22eeab742d3c241d87d5524935974876c9371ad650afa
        version: 4.18.11-rhodf
      entries:
      - name: odf-dependencies.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-dependencies.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-dependencies.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-dependencies.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-dependencies.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-dependencies.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-dependencies.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-dependencies.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-dependencies.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-dependencies.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-dependencies.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-dependencies.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-dependencies
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "debug": false,
                  "ebpfProgramManagerMode": false,
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
          createdAt: "2024-09-13T12:46:10Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:577ab5d4780ba01e130ca5713fb7b3764244cdcb937b6b00e6bb17638f15fea2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ingress-node-firewall-operator-bundle@sha256:6c6a76287353a02c05164f9a0a3e07d27b75f0e5d403eb1b47f1e9dc7c9883d8
        version: 4.18.0-202509100344
      entries:
      - name: ingress-node-firewall.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ingress-node-firewall.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ingress-node-firewall.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ingress-node-firewall.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ingress-node-firewall.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: ingress-node-firewall.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: ingress-node-firewall.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ingress-node-firewall.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ingress-node-firewall.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: ingress-node-firewall.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ingress-node-firewall.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ingress-node-firewall.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: ingress-node-firewall.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ingress-node-firewall.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ingress-node-firewall.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:e39a0b57e73d92c5db330ac9b8a184eeb400397f14c05c1ef75c952d91e4f5b2
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:5024be9a3190288983e3c42e8a2dbb49aa44cd596237ad9ed0e72de5090dc9b9
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:ece58f1e360eccf7d970dca2f9da0c4d99a692648f165f677f11ff080649cf62
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:6bd423056b555d39337c0e3a9c31e7732e80f34d0c01f0766a6aadb2aff586f8
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle@sha256:261b77cfb315c608505e5659b2d01dc4e1ea96cd37fae25f417cd7df148f4ce4
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        - quay.io/community-operator-pipeline-prod/beegfs-csi-driver-operator@sha256:1a9f2aab026c2d1549be5b4624c56d65e813c619fdd5beefb062d5c9fb8981b8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        - quay.io/openshift-community-operators/nexus-operator-m88i@sha256:46cd06d477b4d09dbfca7ca4434ea58f881314138ed744c6390033bd6fb7dc0f
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-secretsmanager-controller@sha256:c92ea9b492b6585a816668bd2a77d65471e339e02e25e44c9558a34b352db593
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:151e7d9bc500b7252af46d0856e3e16559ff310a1d0efedcc5d914164bbdef55
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:309ba89dece3ee9901ed66c40bd28d71f20c7d91a7feff552e28ea0c772b9178
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:63bd0faf0b64e4672014205f9dcfeec6c436edd7dc2aaf83cd053d6e4569d90a
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:1b464c682cd29c8fb7bbc566a1575b9002ad61163c2b6e94138e0e4b0ce51140
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:8223bda16ac5ffac4b7407c849c4c30ee006ddafd34cad8a03d2af8314d56f58
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:d7333e377a5aace0d42e42ca936fa7a84d288b181eda4747f9367a9539f2bedc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-bundle@sha256:0064fcdc252a04010e605706be377417de4c9be4289358ec0f331d6a37befb0b
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        - registry.connect.redhat.com/loadcore/keysight-nats-operator-bundle@sha256:89bde716c2aa4dba0738d2481a2c4707c59a748f84349f9fb01cb430e69d7beb
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-rhmp-bundle@sha256:b66e3cfad24cf28e0a738179cab062667b0103c52cf2e84abb88d5e018e12b2c
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:27c5d91d62371f3cf0dea4d50044d8a6e0a6015a453c4216a030683ea2474c64
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        - quay.io/community-operator-pipeline-prod/percona-server-mongodb-operator@sha256:8659c9f25245e541609079970c8193fc6a58a7d323c249a81d743de696dc15d4
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:9fe87e1b0c49dd00a6ab0c35b68ebf641f347ac5b405d31f3c5ace02a7a30540
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:35d2c7c2849f89474fd5648a0d2781dde28131303cc8a597ce243a3e314fe0af
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:dd0ab94059f782aa31696b15e1b25787cd58bab4905fe64faec7e49bbae3d289
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator-bundle@sha256:0f20db33b1eb069a767089862412846aacb723dc22b44adeb9793495ad20ea50
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle@sha256:506afdc804dc30e0a720e8e6f151b75a32148844de42d90d8c0648e9108d9ea7
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        - registry.connect.redhat.com/elastic/eck@sha256:063a1f405b87751c5102693451ae5c2e9ad203c6b8a31b062a5b403bb8b2febc
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/fence-agents-remediation@sha256:6ab21048f86979bcc34a84965dc2a97467a79507ab03469fe8c6deed224881f7
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:18:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:1a98d55c28a0679134e24b6aafb3c4dc9db0098094c8d2bcfaa41a63c33308e1
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: odr-cluster-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-cluster-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-cluster-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-cluster-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-cluster-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-cluster-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-cluster-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-cluster-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-cluster-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-cluster-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-cluster-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-cluster-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-cluster-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-cluster-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-cluster-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:16f820575acdca39c0b7e85edff7b910809f6d184a9a479de2bb2f687c94a6f2
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-cluster-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-cluster-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-cluster-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-cluster-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-cluster-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-cluster-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-cluster-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-cluster-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-cluster-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-cluster-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-cluster-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-cluster-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a93b382258a34e0b56ef6616a8490e5f44ef1c6aa5d29c0d08d2147ae5b440e6
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3170da029869d15a246ad31d0debd4b9c03ba13855fb877a754cfdffb2286b80
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a7ce92f7f4aadca2781c44839bc357615395587bc8b68a735642967cf8694362
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:093c6a66d4faa1d980e4319a048db8b2869eba8f13f40f02615d6696257e6719
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:5a19032dda9a0f584b41cbe9e729ed23e6003d90ebbb026eb9e7bb9769d4a4e4
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3b2cf972e1ced0a46d3559da4eb1a1b32c50457293ede4958a13538a84b45316
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/proactive-node-scaling-operator@sha256:0ae01e9adc67ff7341fa3faf51a67c02206c41342374cab0bb2e6079d84fef4a
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        - quay.io/openshift-community-operators/oci-ccm-operator@sha256:eaf63c6f9ee9e9c3562d6ced4cff091c1e401daa826ee3a92b321a5cc8f13da2
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.connect.redhat.com/vastdataorg/csi-operator-bundle@sha256:925cc412650ec3a32ea4a55ffe65d16b76916d51f0b57ff62182b4ce12c5d2d8
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator-bundle@sha256:e0e7fc3a69575199a6fcf93985451dd66a5eb7d1ab8caecf059d7d52f3953dc4
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle-rhmp@sha256:d90f6bce87b783a80b1b81275cb61162c7c2a7898aa67ea85e3223deb1dfcb4b
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-operator-bundle@sha256:1a950c41ce2ae204d6bcba959b2d1a3df2d994bf1fefce4c7095c4ca20a5fd6d
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-iam-operator@sha256:d7e2507481b280118997ec9fe8062a9d56ed0867265732ecb5c89ef6fccd98fc
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator-metadata@sha256:7fb2750941193d64f8cca278ff7d81d3ca5e29a2ad0568dff7ddae4396548bb3
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:57e796822bcd705974c0efcd6c190ce39eac2e70b9d6a0c27f69d85dfe966db2
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:1e7781ff09c5cfb87ede98b96d31e58b760bedbf7a23f41778654faba3507591
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:f234e8af232f2d3a9472307ae82834f96756841b8d60b8c42ac4f9ca3fd17545
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-lambda-controller@sha256:9b2943aa770ebd5638927b19f05f629575e1f8fef129f8520e01c2238771d7dc
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-operator-bundle@sha256:bcc3521d20a1662d4032901f46e3bd4744c9910b060f52bdf10c6a8dfea1276b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/sap-commerce-operator@sha256:9b6a3c9c7a3af1a91578fa06c202d128fc7bfff3ae21300a4382cffb11def17a
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.connect.redhat.com/ogaye/ogaye@sha256:f35a21fb3936b259aabe27b453612616fca4e6327c9042361b4d00ef332905c0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - registry.connect.redhat.com/hpestorage/csi-driver-operator-bundle@sha256:39fedc5650ff2fef6ed1d47a5745851101b7dd8f6ce1f8b72ac3a78bf3c1fd6a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        - quay.io/community-operator-pipeline-prod/cass-operator-community@sha256:428effade29fcb3e8981e84f7df945f9bc49fe1a9b56217573db8b387706c26b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - quay.io/openshift-community-operators/seldon-operator@sha256:864e87ef21e81624010a1cbcee6b17c8c816dcb0a699c66badd32eec3baf39fd
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
          createdAt: "2025-09-23 17:13:29"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.16-3,hyperconverged-cluster-operator-test-rhel9:v4.18.16-1,kubevirt-common-instancetypes-test-rhel9:v4.18.16-1,kubevirt-tekton-tasks-test-rhel9:v4.18.16-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:86d21c945452fb380a895aa96630726fdba73b4e10549aa8ae353c7cac37fe44
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:aa24c4a937c9c1e52c30f229af5b26ce413b684375801d0f5f1ef2a50e8502e2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:6542fcf51604d3ba95ec0d42aa7bc19bdef940e1959808efc98f170a68a7d01c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:82dc19865ed64c129b5214b413f5f8da16d90da532c40d3e02b28f767d72f3c5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:a67782645abd4a869ca8db132953f5f4da72c853fe16f9521c77f241a0703bfa
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:134cb74301a9f9f704d35dea38ffa73aa3ee0ecf43b7fb3e73a89c9b432af9d4
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:6cbb8785d06e58af9b8fc1a66d0d5989a1b490b04d0191252cbb16ace14a3cdd
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:424ea484cd63823b4585e630ffc4053c32309e3f7dc24a212582afc60064e503
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:01082e3cdbac6fa50bec206401660dedd832a71f1f9e52c27281587f1f0122ba
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:a4b7120550d915cb74f03dc3530082f3a0d6850809062469f59934d8518479b9
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:0db7542ab7d67866046b1dde2f3ec8a04c47caa9c622ea9832dc59d127b4e919
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:fb9321f06c842dcb628f9082094fb697146e8740130c889bee81ba6941177303
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:cfba9b634c5a8871511ae6c1644e63560fa51d8b2729eb0db207b89ea9a1b3e9
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:0384141b5c99df83759f69c88895bc23f946da385d2c6db0fc75f80c0853a77c
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:2cffcd87d0a42748ffdbc63707751eb8212ce34ac33d9225690f04ce696178cb
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:42e5ad70d114edd15c6e4d13d48399f572ce71ead1f45a9d94de2262a77754db
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:3da5b2baa67e468fe72efcdb4caeeb4f3384365719d5edb5fff6656edcd74141
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:521b8f630c112e9d8eb4d5686ec7a6907fff8cf491628b99d6aa4a10253b3d78
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:350bc99751e2cac633b98c633b6c83e60bfd55dde4301282ff8eb61f85ea8d49
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:e21e94b2778c786d2b6aa0c107df267055b1f7d36c589a5900d313b3fe3404c6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8ab2c1ea24981ca684505c2f9e050cd938efbcbdedabd74897e0c6c0d5c5a433
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:68b318dd7fed09f60f6df215c9812b38791707b4e9af1de27acf91365766808f
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:c3a694c846fc85709e756a617d2581c06062d482b0bba88983192f1efbf8cb21
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:7e60c4e1fb9a0e10b039887747287cc06ca156f813d88c7421f3950eaf07bf31
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:710d24725c6aa3b8282d32a310d3e107b6a2e57708425d5e3acde9e82e33591f
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:deaf31fb3ac137556fedec55c44abb71d01c6927aaa21e5d23627ce139f12107
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:8c3f276434dfb7462c556f7be3f93a0c857a463e2fa2befcaef3a7c0859c7368
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:e51dfca6233c67b1b5372c682a2fc6e0fe54f2beef86e4c9451aabc80017a5fb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e3af0269b0a3225b2b5d43cce57a40aef6788830e8af8f4330bd36e4aaf8ec97
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:1c4bcdb3e2074b3f99c4dd0c57a264330c7989b8564efcc50bfe0e086d487f5c
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:27d1ea63ca9dc4fe073a28793f574da3a991a4c36e8da55fc2c7889cede187ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:9f8ab3e7e76ccad5a794de310ed0ba39e070e00647e42cf2b9a1bcdd8c377f3f
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:e0d239237d8052c36aff1509d65c00909b8563ce14341b7eb7fb79e9b84deace
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:af0b3d4305dcc5ccf00569deddf122aabf444bda0d3f0c2a5c4fad7e0c4117e2
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:071b2bf8f6470a6f87c3804eac3a10a947376469e42a55e4f19d3b2fc98c7e95
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:8339ad1cbefafbb91be533f1980ffb0e22114d52fd5c11c1b06cc55d9554c9f4
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:f514eb6d468fdbc6f5bc2b60cafdaff19ea55cb9a9c60f22b9db466f5b8f9000
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:c839c0eef642ac854e04129b007bdd529f9b380840e052fc306ecc0d62e4d055
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:d661d52528a89d87764e16b5d1dcb6ed328075477453e69c6cb9814e2a2d0ea9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:21c02925a871875a8eb70efda72f79d4c287da5cb9c0a4fc19de9ebfcb06e496
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:ae92990aa7ee3cf0bb98549ba78882c032afc8ed323fa83d1e8de24951b09d11
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:ecbdb919859e032ba4f2fe3af4dc79e87a5ad28545942d9d75c734aa814b7b23
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:e6507ebc8d71561772e573691b160f60ee6819b3f6404d5f0db123f0e4bc17e4
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:250cb53bbe119c5609a99ad32e97fc23572af9c08490e0f5212cf0977f555827
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:7eb05621f7cfe8162b08bff8a7411d6b05959affd7dc0b7729b4ccf1d90bb005
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:55996948c6d66abfafc630d7f6fe4b67201af0b61613a39e1747218393fb5e01
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:7f61b66615c9b22ac0e7e6ded50fb5a4f0a71ca228d34340977d82c8eb8322d4
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:09abeaf5bd5a7ac9f272759932e8442828efe6935dcd3b365eb42b7655c1de18
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:d32a745d70478945bb25f04ca85c17dbc1fefa0a791522e42591a6876f085029
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:db86fa7625d770a7eb8a4af977499f4569023992d33caac0caf8e3b71871f574
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:f5ebfeb2d0bcb0e86999d4efdd08d03c34c90daa2523b1382e02f9d4930ded67
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:e3f9e0c1b53a54354246d9c934ea48c038a547845d6a75c7bfac5e82f780ee8a
        version: 4.18.16
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.16
        version: 4.18.16
      - name: kubevirt-hyperconverged-operator.v4.18.15
        version: 4.18.15
      - name: kubevirt-hyperconverged-operator.v4.18.14
        version: 4.18.14
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.12
        version: 4.18.12
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.10
        version: 4.18.10
      - name: kubevirt-hyperconverged-operator.v4.18.9
        version: 4.18.9
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.7
        version: 4.18.7
      - name: kubevirt-hyperconverged-operator.v4.18.6
        version: 4.18.6
      - name: kubevirt-hyperconverged-operator.v4.18.5
        version: 4.18.5
      - name: kubevirt-hyperconverged-operator.v4.18.4
        version: 4.18.4
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:a32214d166ee692a4cae917fe97e0535a4a696155e1c878ae5335c77dadef0df
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.13
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
          createdAt: "2025-08-17 16:56:34"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.13-6,hyperconverged-cluster-operator-test-rhel9:v4.18.13-4,kubevirt-common-instancetypes-test-rhel9:v4.18.13-3,kubevirt-tekton-tasks-test-rhel9:v4.18.13-2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:a4d0418e49929e186e37f20f64c46f8eca20e314f6da4deb55f0b5407686c2e2
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:9fbf91cd986fe4064b99d6eb4bde7a79f58966763765a5001d9ae75ed22655f5
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:710dd1bcf4e171943d5260bd16121abd75058bbeb80bc8c237d24aadca99585c
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:396eec2f40ad14e9348b229498313e478903125dbb8c6f7c5f3dc3dad359d18f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:36a46082ca4ce5fd4784a9613d49c30585883bccf7433dc2ecd8de787eace937
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c73b90403d587fa6993e736fd84c08ed39381ac506502c3711da99dd4ea39b4b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:47d3263ddc6d80e844f8240502df81c0b71c2ed85a29a0d4afac7b5c98379091
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:6c39b35d0142e12757091831e2f9900c5f8ebcb1498e0eb75b62c86ee2e7238c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:8a69cfd1bcf5745779a019499632141ecdbc4b2d9a4858c66d28df11087359f3
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:b7771fb216f69af28f49c615fa050bde00d27067acdb505578d104ce1337febd
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:21ad0c18f57b964b86ced3498ac0c36a6daf8a60bdeece77a0b41b168dbd6586
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:877fe3539fe1a4d9f8cf25bfe2c1bb847ff9413693848ad539c26442472c1d3e
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:6ceaacb96b29fbd5a07dac9e8aeb038b76d53f6dc30188b88068755f35da68a9
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:4f806c32d077c6021033733674ed5701bd45086f9fdbfe90985813e2efc27ee9
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:1e73a912388b8b3de05e669128d30dd3081d5747932dc10287d613e19f59a49b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ca5969eb64a112ae5c12ae3b6752b6858dfb999d78d61daffd61db5e07a7cdb9
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:b0a4d68244205f839c0bb2c5cd4f4f305537ce57b42d24d0b46fa4fe2fe5cbba
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:960ab26612067d50aae973b63ffe482649c90ff8f42c2dde698ef548422460cd
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9600fd87e8331989db167870e83618bfb5f273e34e28c2006fa72b49aceb86e0
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:d90907bbc0d2810d592316f609bc8fd0f551de4c37ac02c1c03ccc014cdef056
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:c78c9ee196047ada7ecb614a42cbbfe3dec934534942ee5cacf34cf030e50378
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:f54b97303ba8579a2ef38d971035117efbab46d0eece22a759bc2fba98d11511
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:1d984a34589472f1c554bb4a53b6df470104affacd341eccd5fa1e334a7d8854
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:28d40c603ff685d922149ae30dc77ba1d9bc9af0ce2ab6d9038ff8272940ca08
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:c399069fc6a0d3b3e7093ab6cc2d540d05a04a68213a24a7a500a83d084246df
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:3c2706973b2bca2fbbf88d5af1de0b3c53ef1c7e5029b9a3ef42f961795ad65c
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:725286e0c9265a66e3f4b2a86fcfa7cbdda4e963863af9fb2af9e1226e10f34f
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:3e546a40d6a3e2cce2567750f93260b3db0978477e8ce34f63cd5d195c58c195
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:3765263eb3b193d860699d3e348346b762eabca51ecf31f8f97db69f3ca5e4bd
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:8985f049a49cae810ffb59dd4c94afca632808d3c8ba40809c5447a382b63bad
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:7095f50cb41030a9f49af2ad9a99d81407b5d8491c421c34934f788f02eb7528
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:f501a890dd87d37eb24cdd86d5889eadfad1e454fcc782e8ca2454150e8dc494
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:04b135c11a1e4518febbad5ee56e118ca711ad3454db19f1824cda0d4d497742
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ebc17f70a654181527a8a1d39e069076af82343cab55531fd7a86eaa85b12fd2
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:a9788ba6756715f286ede2fa04e460fd021b24f2a0df068c2aa626c89814a7b5
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:5794ed8b4af6cc37d278cd91f08f0418b89550269d69c9196772fcca1a4933e7
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:0c2c54d18dddf8ae7f2993dd42b3d07217f63266218c329119dda13e0c431fb6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:f15c753ca6687605d956d5200ec9104067c9bfe66e987dce7c32bc45ed79c57d
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:ee7ae60e4ba4b1d13b0aecb053e3ccd22b62d2277d28413f394d32e6a3ffe139
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:b5a6c60e663b81cef29e9622af8f0eb62838be99ff51f576ca1b275c6e5a204a
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:4c58b6e5356866deb394d55e2f96e4d83489d24c988e8016ccddd8738b895b7b
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:88d6db3caa017e4c2d8114af1abd28df4d0751b01f291d42b84586378e0c9484
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:0838f54f7ab4fbc04346ef67501fbca0b6c2097558bbd9c6fec2632ba0ce7bc8
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:9e46d9a0ca2e1678a02a52d14e2d8f0eb007b21ae105100106532cd12f317a41
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:a0821002f32c485c28d43dcc3afe5f447ecec4a567ef1b5f07b162f96245cbe6
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:fd17e015ad32d54f8464bc337c874b9131e0edca750adb4165161f0d382b8c3a
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:45c7a05260790a6ea78c809316dc58ed98b32cd94a884e82f24650051a27a163
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:2e6b6d7d04690a616eb210fe4b142372807bb9c59b6af6e7d6fa196f8dadff3e
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:8e69000f409c115b6af634c057106dbe894395708a694c3a4cb030b06e77407b
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:498282a9a9afceec8587eef445214f5db6ed6c09c2c3362049c4097a210e5aa5
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:11031e69287b4fafa77e2d54dc40d8c7de64e7ed8528abfca7a798b5aa8970b9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:faf31a311d34ea8bad50c83918d5ac1dc32277b9d8e356a866d166f97209e40d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:61f96cac775c4f636f0685900c419a6771fe3f3b44648d93a176b4ab5efa9092
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:70219973c00c12519a1e5abd9fa3061669de8e7273ea1dcf9f72d40ae683d244
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:f0eba5e54eced4a4478d395b098b3dca201b7b374f9e6b7f4a5c1669953c0991
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:be1b6c7e71d25c45b0f4bc2ac8507a1c215e30e39f51bc9f907436f6bcc96b4b
        version: 4.18.13
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      - name: kubevirt-hyperconverged-operator.v4.18.2
        version: 4.18.2
      - name: kubevirt-hyperconverged-operator.v4.18.1
        version: 4.18.1
      - name: kubevirt-hyperconverged-operator.v4.18.0
        version: 4.18.0
      - name: kubevirt-hyperconverged-operator.v4.17.4
        version: 4.17.4
      - name: kubevirt-hyperconverged-operator.v4.17.3
        version: 4.17.3
      - name: kubevirt-hyperconverged-operator.v4.17.2
        version: 4.17.2
      - name: kubevirt-hyperconverged-operator.v4.17.1
        version: 4.17.1
      - name: kubevirt-hyperconverged-operator.v4.17.0
        version: 4.17.0
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        - registry.connect.redhat.com/fortanix/ccm-node-agent-operator-bundle@sha256:12746020eefd383963e08a86e3542608aed1999b4e520435e366a307a6a9eebc
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - registry.connect.redhat.com/microsafe/operador-aims-mobile-bundle@sha256:1572f715c13d852c2a824b726912c4e5cc49aea43d0d91b6fd673f7198b0d038
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - quay.io/openshift-community-operators/ibm-security-verify-access-operator@sha256:ec3970c7a6a5ee3ac863dfe6f16234fe19cd7ea4a541fd0cbae9a8183d0763a2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.18.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-05-14T20:50:15Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.17.0 <10.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.18/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.18/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.18 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.18/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.18/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.18/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:f391499545a2981a3c3492766bbaae6a72a82db07d88251e85fda5d672a108e0
        - registry.redhat.io/openshift4-wincw/windows-machine-config-operator-bundle@sha256:4ebf4ace7dcaf0a935b83314b2a514baffca704b2f36cec5cd61dcf13e459946
        version: 10.18.1
      entries:
      - name: windows-machine-config-operator.v10.18.1
        version: 10.18.1
      - name: windows-machine-config-operator.v10.18.0
        version: 10.18.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:6e58f2d6a1619a06be05807ff5407924087b7b27611628f9baf999733b67645b
        - registry.redhat.io/web-terminal/web-terminal-operator-bundle@sha256:fb969c8b8ab53ab6639795086efad6870dd212f04c94934b75910598efcf3ca3
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:0b133afa920b5180a3c3abe3dd5c73a9cfc252a71346978f94adcb659d683404
        version: 1.13.0
      entries:
      - name: web-terminal.v1.13.0
        version: 1.13.0
      - name: web-terminal.v1.12.1
        version: 1.12.1
      - name: web-terminal.v1.12.1-0.1745393748.p
        version: 1.12.1+0.1745393748.p
      - name: web-terminal.v1.12.1-0.1743657317.p
        version: 1.12.1+0.1743657317.p
      - name: web-terminal.v1.12.1-0.1742484227.p
        version: 1.12.1+0.1742484227.p
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso-operators/openstack-operator-bundle@sha256:945271aec89541d6d8a749c41af2fff3fe24e41fd50e935e3944451c940bf940
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/customized-user-remediation@sha256:dc517e2349fe993e3417b858865fb73b98b7b163c0b4231d4d7be1f11992bb78
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pubsubplus-eventbroker-operator@sha256:2761ebb89dace09a1c1ccad92c28c877f3a6f75a674191ffc3788be7fa549926
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/ember-csi-community-operator@sha256:5dce9b3018cfb300cd651c7980353cedc6617383d5c525e6a4a39ffb8e105774
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator.2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v2",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "600m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "baseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c",
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "image": "registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733",
                    "resources": {
                      "limits": {
                        "cpu": "400m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "10Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "50Gi",
                    "analyticsvol": "50Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-02-15T09:21:28Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v2
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: ABot-Operator-v2.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-operator-bundle@sha256:35f90c08438e4e5aadd01de20bc82e59454c394c0e1d24ade6c56d6e02053bd4
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c
        - registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:e3b3c32d12223097a9d0eb466c3f49fbdab519a675123c7e69597b53ab4c79fe
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:50f1b4c65541b56b85f42bc9b3127bc7bc9faad1974067ce11481de7596e4c6e
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        version: 2.0.0
      entries:
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:4603bf04e53ce948bbf0ee5d239d5e265b40598044ef9e1f4539f4e466e59cb2
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:a178716e62ab8a752bcb9023ad9631af2dbc72357266cd13b43398e21f230828
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:c5288fea5b2ef8125f7b1ff7f7633eccd4e52d14ac9cfc5e473c2e317018ca3d
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:704925128f81d4ffadc4dcfaa69fc8011cf3d0f431d24222a8d7a03bad675331
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:8da482ee4c94b2546bf2b83345b1e7df6e70dcc63f68d8a2e839f10138951d2e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/openshift-community-operators/ibm-spectrum-scale-csi-operator@sha256:01d24dfa0d6b6146230459f8f7e74f2749d7fff3c4d75c40de386e014f9367d6
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        - quay.io/openshift-community-operators/openebs@sha256:ff179bdd6e63b8e2cb1ce3b96d5e03d20c1a2410296fed6fb331bf5bc2e6b14d
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:895fcf14b337398585f51e4b6e1f46a4a17c01a153ca889bd4374d691d89b9c8
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:dbf68e729484ba11aaafbcf1a590aca7f233de10e71b01d78c319dc5b5c6c278
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:e9824dd4ca39accfc674b88f7a2b76fba3006ff856e78402348c720f2aa8a6b7
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:767d55dc7343406637dd16f1b3a73bfaec353e0b058143fd68df2474271e3b76
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:43c90bbe3b20e95e1f352cdad097d63820a1282deedeb769352cb100da43ecc3
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:2f853ad4fece4510e847e96fe8085b53c56ee342e1f387677ff8cef345298e72
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:70c5237a251adb5f44a085d091ca6c24d5b313aa73f6545389a2b938740a9a20
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/socmmd@sha256:5adf9706daf49b35cfcf74b6fb07b056f87714b87047803b797a56f138c77486
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        - quay.io/openshift-community-operators/must-gather-operator@sha256:a71e5e25aaa92f03334776fee87316809bf2a46dfb34c497d63b228a194982f2
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: nbde-tang-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nbde-tang-server.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
          createdAt: "2025-04-24T18:06:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/nbde-tang-server
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tang Server allows definition of the NBDE Tang Server deployment
            displayName: Tang Server
            kind: TangServer
            name: nbde.openshift.io
            version: v1alpha1
        description: NBDE Tang Server operator provides Network-Bound Disk Encryption
          in OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NBDE
        - PBD
        - Installation
        - Network
        links:
        - name: NBDE Tang Server
          url: https://github.com/openshift/nbde-tang-server
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c15928ecc2f0ff370f567a2ca909b7a71e2becaff0af649275396e59232b7b8c
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-operator-bundle@sha256:79bf610b144cbdf4404a1153f1c395179ce1743274eec7102c862aae4f45890a
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
        version: 1.1.1
      entries:
      - name: nbde-tang-server.v1.1.1
        version: 1.1.1
      - name: nbde-tang-server.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: nbde-tang-server
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:6f0e07eb583201703f50ce9b3bf47e7aa63630074afcb588902caf05ff3ba2bf
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:35fae43ffc92770e13919f081195b35d8dd293f223fb9e577d75ae53f156305a
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:e135baa6b7eefcaf70780b6833acb689c5aef076b6280ff29c02949fbb33b06e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:6a447f455b8a3a3a8390787b36fe1d59a7d8bbf834f66a63fb0023f0557796e4
        version: 4.17.13-rhodf
      entries:
      - name: odf-prometheus-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-prometheus-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-prometheus-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-prometheus-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-prometheus-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-prometheus-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-prometheus-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-prometheus-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-prometheus-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-prometheus-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-prometheus-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-prometheus-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-prometheus-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-prometheus-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-prometheus-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:64c4cc3155e8ea6630bc7fecd392a41d71193cfb20f832b6334cd4795cf0735e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7916e81066baac027fb575a1edaab3c98a00505c73c4ac3c2530c111c11f0030
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:4352d0aecaac2efc3a15bf87925c08f7f85ec7baac9eebc29787f05af047514e
        version: 4.18.11-rhodf
      entries:
      - name: odf-prometheus-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-prometheus-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-prometheus-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-prometheus-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-prometheus-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-prometheus-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-prometheus-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-prometheus-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-prometheus-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-prometheus-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-prometheus-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-prometheus-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-operator-bundle@sha256:bd5bdc7085f3e9ce8494488f29b4d83a29796c47d536106cd6e14a7b67ab07ee
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/3scale-amp2/3scale-operator-bundle@sha256:934fd9123d1da778b9f8375854f2a61167f0efbc76371edfdc9962f7626e694a
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.11.8-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:9c2aba45971c924fd55b2aab50ee8e6861a9defd3808b38ff30316727af31f9a
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:4f807ee6e885e966853f70e1c40bbc179dc1f994b401ac868b7f2b921f9d17a0
        - registry.redhat.io/rhel8/mysql-80@sha256:7c05e61c0f4f5dad6aeda136e9a1af2f04f4de3c51004e997d4eacae3778da17
        - registry.redhat.io/3scale-mas/3scale-operator-bundle@sha256:a6c2eff926feb11da27eb9972cb86ee0ef9ca0b013afb6fe6af4f72439d557ef
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:419eb09d1b165363d5b90df01c1fa10f1f91f03649a8758620ffa49276685feb
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:a03a1ae7f01259192b8d7b4ecee6ddfcdc0cd9e2d24a8c2306accb93ee767135
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:7d07caf7b6c5a195ba4706e82f3d28c3793afab99b1a6d5a7020547c48534cdc
        - registry.redhat.io/3scale-mas/searchd-rhel7@sha256:1426e4bd4aa0eb28b479900ffc02621203886280d24997cd11818387ff30d9df
        - registry.redhat.io/3scale-mas/system-rhel7@sha256:b0fd85a4994502a3a2ae9ba6a7610c38af64b6cc897efac12feb4b115a484f77
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:22431acc2c991b6ee6b4de0219366dee6ea745122a49a0175d5fb58ed7d9205f
        version: 0.11.8-mas
      entries:
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator-bundle@sha256:4629bd129addb09478e48964a62acee88f5dd2746875801bd19d2d55cadd9da2
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        - registry.connect.redhat.com/loadcore/keysight-wap-test-core-bundle@sha256:16c938699ab0d71138693641ba7dcce2f4304da1d357a67a057c4effb4300401
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-cluster-operator@sha256:681e51da33e3cd9d6a4b617dca4338827237b70d1e8065322bc4135846039c56
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/cluster-aas-operator@sha256:aad6afc2f6f1f4e3b2c0d14b6c178d6f71ef318b0d3efc97faa5fdb736eaa721
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        - registry.redhat.io/openshift4/compliance-rhel8-operator-metadata@sha256:1c06868bce1f5d7145ef658bfcc284c44011aa24a17166aa124547b38f2259e4
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:8a2194d8d26dff03ec53ac0ab93c62c268e48675079a88db18bb14a977a391be
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:06e4abf4a94d44728b6bfa685b6ed027962b60fc2a0060a7256297857fd5f6fa
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:b1dc39979be1c347b123b2dd8070e4509179b7b499d4cf3fc7feb4552c6e2413
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:14fb2aa627ea7f38fba738cc4302cf1962660b10d3806c71161d8c7d64a8d060
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
          createdAt: "2025-08-26T06:53:32Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.1.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.1
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:498bcfc3c56bf642aaa70aec4db61b8610af0a22e9e378dfba02356770566198
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hitachi/hspc-operator-bundle@sha256:2d5e1ed3782a992034bec29d85d24befed31ad5eaf7a21ea44acdd775f90cb73
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: hspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        - registry.connect.redhat.com/opdev/operator-certification-operator-bundle@sha256:c390f32b40449fb09477e288fe5347319b691d3e89e020163e32007e6679c685
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/openshift-community-operators/tagger@sha256:a528169d305dbc2bc69794437665e4f940ea708dfd957c58d69bc79316b68612
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - registry.connect.redhat.com/amd/gpu-operator-v1@sha256:8a92a0d18a5ae6a837ca6590c2beb65d889cd744a098c22d3f01d2fffdc4dc57
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/community-operator-pipeline-prod/mongodb-atlas-kubernetes@sha256:f12108323fc990f869394e09b4ff6d9751d0a07f3c152fbd45b0f738746a468b
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
          createdAt: "2025-04-23T07:03:43Z"
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - kind: NodeFeatureGroup
            name: nodefeaturegroups.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.18/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-nfd-operator-bundle@sha256:ffb2d53dfdc63bfc9320bc4e927d037244111ab0cdbe5721ad69ce06792d9753
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:a1a60b53db81b32de0127dbb63f8c3a9ef7f70c95968aa53ffe8b49eced7dc43
        version: 4.18.0-202509100344
      entries:
      - name: nfd.4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: nfd.4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: nfd.4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: nfd.4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: nfd.4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: nfd.4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: nfd.4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: nfd.4.18.0-202505052133
        version: 4.18.0-202505052133
      - name: nfd.4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: nfd.4.18.0-202504232132
        version: 4.18.0-202504232132
      - name: nfd.4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: nfd.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: nfd.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: nfd.4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: nfd.4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: nfd.4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: nfd.4.18.0-202502101733
        version: 4.18.0-202502101733
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:f6de1136ced78ddf1ff48fdadc259c251919dd3a70819fe0e9247411999d51f3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
          createdAt: "2025-05-15 07:58:52"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.20.1'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:0885c82cd09105a078d4308b4ec1047b96f3cbf1fd555a1c90a8b053befee27e
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5783a00c386858d1eb45f224431d07d40bb9be746de965e7b129ee30969d4014
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5e822b12b5aadaa08a2f95b879be911455607a9c04922a641625515f955dc9ca
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:2217d9128fe50d61d1dff48b588113b5267002f72d025b1b6fbf87a1c2e670d1
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:90505e5ffaee7af330306c4e045d3755e0c74d30b9f45e1e7739c205d945872d
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:2e18b858f09e7595a89c4dc6ea5a85f8c5952f581b769093bfaab8156bab0813
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:6b4d34e7a40b21cf464e2f42cef21ef72a69ae8a3f4e7600a7bd85aa689117d9
        version: 0.20.1
      entries:
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:e533406d832461cd8889f5e95569ea9f8dbbd57b51dcd08b56050ba1e3775ad0
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:4530ed52522a74051aaedabd62b99d5c5274a0ab72eb28d4e75207d9e9d3391b
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:02e6baf8619cc2b782cbb1f3072e554a7638eaf3fd804abe1ba5f43f692e511b
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:df59ef548b82e9d72fecd06aa6b9c569f96485693c8a4ecb8ec278ae92df436f
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:e68814481ca33a422ed5c2b55b8a369e8b48092fbc71b138cb938a119d39e689
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:8ce258f36d3baab3175b29d5ed6df89588f128d249285bb1e7cb45539933c511
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/korrel8r@sha256:f00f8a7e2011bcc97d87fc7f5bb40a19d6b5737d1f5d7f3b2322ec39009c0618
        - quay.io/korrel8r/operator:0.1.8
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53resolver-controller@sha256:542229539932458d76180d23f1d5c5795e2de89169d8054f7b95be6f3700881b
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:3a96ab2db995d24a9ca064b5566821828d6e73ea20df77c9932d8972cb8e5780
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:c2b996e8dac2f4c81a3e48e4203499f48e4e15d0692e98d61dc8c97192e0a776
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:68eb666168d3a41cc98f91a0f2d74656af73cf76ec2cbe2cc308ae2ffab84875
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        - registry.redhat.io/rhosdt/opentelemetry-operator-bundle@sha256:1844430afea89707b70bba22cb9a2db98161cfc62b53b1086b63f109058028c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - registry.connect.redhat.com/citrix/admagent-operator-bundle@sha256:2a32d9a8cc3bd67ba7ebfe3f96b59e42009121d630918d417717d8ac9f2e40f4
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        - quay.io/openshift-community-operators/rhoas-operator@sha256:5299cd4c88012692b31d560ef981f80f994a0468552f43e03623adbf89b49ff6
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-networkfirewall-controller@sha256:96c35a47f58c48532fa3a5750e52a6d7bc2d530b54e807d57547f8780e9d7aea
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:20:34Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:85ba83a60fece2cc1c39ff245193712f918fbd55e1e283e940425fbb1b1f0e29
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: recipe.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: recipe.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: recipe.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: recipe.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: recipe.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: recipe.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: recipe.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: recipe.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: recipe.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: recipe.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: recipe.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: recipe.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: recipe.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: recipe.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: recipe.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:27:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:3c7033c68ff5e65533d2c3d0f2032cf34aecc153fd34f33ad3239331442d370e
        version: 4.18.11-rhodf
      entries:
      - name: recipe.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: recipe.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: recipe.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: recipe.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: recipe.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: recipe.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: recipe.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: recipe.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: recipe.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: recipe.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: recipe.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: recipe.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        - registry.connect.redhat.com/sonatype/nxiq-operator-bundle@sha256:48bccf9c8cc07127153af7974b949b8cd8429adf8714dcc866f026806e38fb08
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:9d24507aabd06856c7c35d73417b6ce065164f9c966053b494d68f14f0e260b7
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:00e6945419be8ac83aa7c888fe243c2fb67ee4ae5030cae117c111fdf80f5fe4
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-organizations-controller@sha256:76ea247239dde4e52e5016944a41908de8fdcc4173cad8dfff04413a0e3cdf48
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - quay.io/openshift-community-operators/universal-crossplane@sha256:b3104adb471083de96d49fa53f663dcf0e3e7ff5e3573bb1422aa6045277bea0
        - upbound/crossplane:v1.5.1-up.1
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
          createdAt: "2024-06-12"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:87c29bad5a920ea676a5aa8f6e582492547e391ea166694f84cbb75044718a19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:a29fc3bfed3add2768499e94435e8993bfd704a8a46268e6e63a8d20d2c4ca6c
        - registry.redhat.io/openshift4/ose-ptp-operator-bundle@sha256:7202697ea75c3503d4165b1d970802ac776c2977755a8b240094046f74c162c0
        version: 4.18.0-202509100344
      entries:
      - name: ptp-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ptp-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ptp-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ptp-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ptp-operator.v4.18.0-202506260833
        version: 4.18.0-202506260833
      - name: ptp-operator.v4.18.0-202506121732
        version: 4.18.0-202506121732
      - name: ptp-operator.v4.18.0-202505291136
        version: 4.18.0-202505291136
      - name: ptp-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ptp-operator.v4.18.0-202505071704
        version: 4.18.0-202505071704
      - name: ptp-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ptp-operator.v4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: ptp-operator.v4.18.0-202504170406
        version: 4.18.0-202504170406
      - name: ptp-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ptp-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ptp-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: ptp-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ptp-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ptp-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator-bundle@sha256:b473fba287414d3ccb09aaabc64f463af2c912c322ca2c41723020b216d98d14
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
          createdAt: 2025/06/23
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.0.0 <5.1.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kube-descheduler-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * LongLifecycle
          * CompactAndScale
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `mode` - Configures the descheduler to either evict pods or to simulate the eviction.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
          * `evictionLimits` - Restricts the number of evictions during each descheduling run.
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:4c653bee02c0135c8625fea6a01e18a046013ee330c5253ff66a7cd705ba08c1
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-operator-bundle@sha256:37afe091e3b9656c107196d7120ce171d8b0bf895264a9791c4559e97c81a00a
        version: 5.1.3
      entries:
      - name: clusterkubedescheduleroperator.v5.1.3
        version: 5.1.3
      - name: clusterkubedescheduleroperator.v5.1.2
        version: 5.1.2
      - name: clusterkubedescheduleroperator.v5.1.1
        version: 5.1.1
      - name: clusterkubedescheduleroperator.v5.1.0
        version: 5.1.0
      - name: clusterkubedescheduleroperator.v5.0.1
        version: 5.0.1
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:afbe40e3dd6894210741fa65a3f0bb6e4a3d00ca261eb6bc3adac6a22251a74e
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.connect.redhat.com/vfunction/vfunction-server-operator-bundle@sha256:f1784afe63b2e834a6b0bf6da254339ffe3597d02dc2d389f9162d806bedbe74
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator-bundle@sha256:b2bc3e161311f8f5d2f5fc967815259697ce30bf94e980214bc05aa0e1529d55
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/odf-node-recovery-operator@sha256:77b0ef6110a71da56216719ceed0f448257dec9633c98d7dfabc1038b23b6e29
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/hyperfoil-bundle@sha256:1d3625d12d95deb4f48660459fd0fc8b250b52b46d68136927d0f12465aaadf1
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        - quay.io/community-operator-pipeline-prod/wandb-operator@sha256:9bebaf0e72f0b3ce20dabf7ace62cb5ecd1015243438248e8101a238988cd60e
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:bf79fd57e1df791fcd45dd4745c923046dd7759ff5b3762cd7a329c5bf611316
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:b03f4e88ab1fed4744356a584d673fb2fdb15de7e1de7786bdd222b853690670
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:aa261c79b421bd49f7b46974a3bb5a790d5a1a73dd8dea4e1d8a4634fb68a51f
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:123488f0877c054a348f5b4135e68d77c17523bc4e5ea14dd73168372592f4ce
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:e66501276e3a10ae6d08dad3c4658fdb8b0b635e4b2fe72fa943b0e5300d7f2d
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle@sha256:47ad4efd37ad85bccf915a6d9448d404b7ce3164d7afb4c944648eeba9f0b04b
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-bundle@sha256:3e7ab287f64398b2aa434d9cfd5d99e2e9cf2d40c2e9fa1b1418537807b7aeea
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle@sha256:6c09391edf49d71b558f8ddbaf4e947dee17490202521539910d78b5ddcb3531
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        - quay.io/community-operator-pipeline-prod/ack-prometheusservice-controller@sha256:dd9c62f7493d43c6f28055e6f1aa7d586e25b2d6d359418782cf18de45da32e8
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/datatrucker-operator@sha256:4ab1d683d95d3eeba944b977594aff2e3b3e7203f56598b2fb2f7cff75354be3
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        - quay.io/community-operator-pipeline-prod/neuvector-community-operator@sha256:788f742c9f2def80d0b768eb96c5138f2cebdad767ea4ebb8cc26451771e23f5
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/community-operator-pipeline-prod/codeflare-operator@sha256:3dbd2b65d7841cc97c37db59ae8aff9991c4d0f0324c4391afd41c4de20c232a
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        - quay.io/community-operator-pipeline-prod/trustify-operator@sha256:49def1c9612621f90d2398cbe7385488c167cc8ccdcd951f098433eb42cb9ac5
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/susql-operator@sha256:4a47a8448610e54b8dc9700828c9250b84ee19126d6d4edbb62b35597e40bf37
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/grafana-operator@sha256:ae9aa1d05ab9dadcc9961ca4293d0878fd7656be79c869c7fdbee718f259ad86
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/orchestrator-operator-bundle@sha256:b82fd43d2dad2e525464177081869c923e239dcb09319458162f33a96370a6e6
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle-marketplace@sha256:b7f67aec4e911d89b3e6217e8718c349c20a859149b829a0390807c113e27196
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - registry.connect.redhat.com/kedify/kedify-agent-helm-operator@sha256:d176136a44e790298158ea3c4232e2aa7d0ad88ff18b6a5b39238162ed41226a
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        - quay.io/openshift-community-operators/tidb-operator@sha256:4f73b39de6129b906b5eb7bb0b7485f3da406cfad8616f81e375cc8ffcb156ed
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.829
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.829'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/aquasec-operator-bundle@sha256:59e62e5cbd479eed9ddd083db881469056c35253d76aef6bda9cbdb578e05b3f
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/console@sha256:05bd172efdb4729027ea6a491c874a1a3368156a506b7b5d228837b12282ee1d
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/gateway@sha256:cc8a3e950ae45fb40866b7369aa9393e92d1ab91dc0795ee682217fc421e9ec2
        - registry.connect.redhat.com/aquasec/enforcer@sha256:b48f026d7876656c0b365a90bcaaca7f4544f38b85400c110cf20e57c67ac8f6
        - registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:ce827db4a2b01b054e33f1091ee7046a14757b8781444bc9868783fe1efbf641
        - registry.connect.redhat.com/aquasec/scanner@sha256:d0ab5e0dad4e042324874de02ee36585a93b53ad1c3e905416af2fcb98f15e2f
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.829
      entries:
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:ad6b8e10abfff1cdac1ea35b44c456e3d59b4dd1372cd47c2e4221ed93e1f71f
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:167b9063a2dec7f9a5f1c5f3c02daf481c45c5e6a2f1b25353102e85921b6fd7
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:83c5e966addca5197f25c5f9c81a52648db44c2fd9e4969bf30b82c4880f0a40
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:bfbb9d0d5556eb067543f4e139a935000319b439f700240adaf393c68bd8244b
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:e0b1cff317bd4e35334429d1cc7c548dbda54065716aea96ef30610f0b91901b
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:eef6e7cf8f76ccd5e9ed16a5e4bdeb4398859121a6b0e86284b9766408cc2b30
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/openshift-community-operators/dbaas-operator@sha256:b2474b569406fb0f47a7042042429bbc5cd2edbbd119e701c6f440a139e1b3ac
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
          createdAt: "2025-09-09T14:18:23Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.18"}]'
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5189547a19ea81eb18142ad41e6412577eb9ac64d3987ba15b6665a1117a2e55
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:63d8b19d73e5cddcdce67c94e677666d3385ab8dfa1a534b9860fe1db33a2f91
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 4.17.13-rhodf
      entries:
      - name: odf-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
          createdAt: "2025-09-11T15:26:46Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:d45bf7fc74c58bc192a0c9d30157bd745f7d48b2a47820e3b768e8cf29b3c6a0
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:e76c59039b39b3caeabca73353a91501e8374e832c45eaf4005b1c08244e00b1
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odf-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:b4e4d262801b8ff616710075ff05689a8956e975e8ac563e50204a26d1725458
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:afc1e5bdade3ff0e3918fffb8fafb2700dfba7c473f99a2f347cfc7f889c8678
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: cluster-logging.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing, Observability
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
          createdAt: "2025-05-14T12:47:05Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.

              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:255a73d34b85f25b3655fb2b63954aa6b635b89080a8c249fc41568fd8365b10
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:2495de935f503808bf8700d9d66b8412d0f15fc754cf569d02abfbd74b5a6ee8
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:8c6decb25972dffb37917fb5f5a91ea05fb21b87823eb75807d8c601c12fc1bb
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
        version: 6.3.1
      entries:
      - name: cluster-logging.v6.3.1
        version: 6.3.1
      - name: cluster-logging.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - registry.connect.redhat.com/nginx/nginx-ingress-operator-bundle@sha256:4e032549f2c36c5a660994b5ec460ef214bb875a53f6f7e9b814d57049683a7a
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-bundle@sha256:31fd95c5232105d3354515102f099b9b662a8659f08da5d67d4e147d4f934fc7
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:17:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dd20cf6cd0b69404d2ddf5554e7f89870429c117740d2df51f3be5a8163b8db9
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:26c47d0498f8b8a61338cda3fffb10724c7e0f78d74d14537580763614130256
        version: 4.17.13-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-csi-addons-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-csi-addons-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-csi-addons-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-csi-addons-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-csi-addons-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-csi-addons-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-csi-addons-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-csi-addons-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-csi-addons-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-csi-addons-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-csi-addons-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-csi-addons-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-csi-addons-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-csi-addons-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:25:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationCronJob is the Schema for the encryptionkeyrotationcronjobs
              API
            displayName: EncryptionKeyRotation CronJob
            kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationJob is the Schema for the encryptionkeyrotationjobs
              API
            displayName: EncryptionKeyRotation Job
            kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFenceClass is the Schema for the networkfenceclasses
              API
            displayName: Network Fence Class
            kind: NetworkFenceClass
            name: networkfenceclasses.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationClass is the Schema for the volumegroupreplicationclasses
              API
            displayName: Volume Group Replication Class
            kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationContent is the Schema for the volumegroupreplicationcontents
              API
            displayName: Volume Group Replication Content
            kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dc3cbef1d890b1be5cd59d4e2fcb91f4f88e2aa53bf59c78a2cbe9af136ec3cd
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:c3e9d2deb1454fbd423ff9eeca34225df86f92b536c962300a75158505341678
        version: 4.18.11-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-csi-addons-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-csi-addons-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-csi-addons-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-csi-addons-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-csi-addons-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-csi-addons-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-csi-addons-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-csi-addons-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-csi-addons-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-csi-addons-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-csi-addons-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:a7b60f1c65a6fb4ee0109c7b2b0bd12fb9487e2f6edb059bd18cfd2337ea9116
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        - quay.io/community-operator-pipeline-prod/apicurio-api-controller@sha256:c783cd6db1ae5a6152848f5e4258e89b37077a177199be905d0cfe451a48e5ef
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        - quay.io/openshift-community-operators/fossul-operator@sha256:95c043ddc818f44a303e27508e0acdcbbcb3d9b1329712c85a8e3fd4cb647e9a
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        - quay.io/community-operator-pipeline-prod/victoriametrics-operator:0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        - registry.connect.redhat.com/flux/flux-operator@sha256:13c2f3b8092fc7b84548dd6383edbb8e41b7a8f3acefe63620f07f7a6454d8b9
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.connect.redhat.com/ibm/modelbuilder-bundle@sha256:4940d853ba477aeeec99550c35c132b69bab8bd7b6fbc1a2fe118c828c7ba088
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-elbv2-controller@sha256:f40f12efb734af5930c225925e99f31eb8db5fecece704a27ea50bd8661658c2
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - rasen708/cxl-operator:latest
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/ncn-community-operator@sha256:15edc1898fe9f8f56708da7d6de3548d06f1df3bdbb3f66a9cf4f7990fe9fad6
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - quay.io/observatorium/opa-openshift:latest
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - docker.io/grafana/loki-operator:0.8.0
        - docker.io/grafana/loki:3.4.2
        - quay.io/community-operator-pipeline-prod/loki-operator@sha256:b7e52fc9b9a988eed8b8338a2b2eba34b020cb24a189c2fa4db63110118dbbc1
        - quay.io/observatorium/api:latest
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
          createdAt: "2025-09-19T19:13:23Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sns-controller@sha256:f535d7861f6ef1488ccf95d0c5629a924033acb133d9452dc86e4c80352e69d2
        version: 1.2.1
      entries:
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
          createdAt: 2025/06/24
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-run-once-duration-override-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-operator-bundle@sha256:8ff7f1c707b40448a719137f37f6a3e09860812f7f07c3e61b7f73d1216399dc
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:42f73634df2c06216fac7b3742b2ac21ef5f12ee98d8694c06f7c9073773acd0
        version: 1.2.2
      entries:
      - name: runoncedurationoverrideoperator.v1.2.2
        version: 1.2.2
      - name: runoncedurationoverrideoperator.v1.2.1
        version: 1.2.1
      - name: runoncedurationoverrideoperator.v1.2.0
        version: 1.2.0
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-paygo-bundle-certified@sha256:e0a8c4d0de20c91b73957d6ea09fb94005e680a092c54389e6e411edbeb59296
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-access-operator@sha256:6e37f565564de371b5797895f3e96e0efd4f592d7d3e899ea8507438bdb60e78
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        - quay.io/community-operator-pipeline-prod/ack-ec2-controller@sha256:58d3249744b82a7f75ea2b1461adb03259350930da791da62841f05d91e94d01
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/egressip-ipam-operator@sha256:1e2443488e0aad361cccece51689e2b8da882862cb62c1208a14ab1b5ec2922d
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.connect.redhat.com/intel/sriov-fec-operator-bundle@sha256:c30849cbf451eb016786bc367e3716389b1d45676ad298c4de6a492c7bb00f4f
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:2338eaab863304250b3b1d267e9f2261c48eef1f97c280b7fa0af0ebc3ba208b
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:40802ba8bf531637f19adbbaeba3631e0b0557168ecf61f566fb5f2da25d1d4f
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        - registry.connect.redhat.com/loadcore/keysight-wap-db-bundle@sha256:3940b8cbf7579d9ce888f176f96c4d191a7cdfaec4be9251e8da4942e7897e97
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:2647a21dfb60364cfb7c4d1b81508adc13d859e97553bb2f710901e6ad59f4a3
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:dfaeb7ac2ccfed03042febea709f0ab66a4c57d62fa041b7ea887a0783d60a36
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:8d20352c0091985cf93e39cd18b2bb092bc15fc5e9900fe46ed5c6fc3c0af885
        version: 8.17.30
      entries:
      - name: kubeturbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - quay.io/openshift-community-operators/airflow-helm-operator@sha256:b3d05d2e2b4ff8ae64e4fc0338fcfa1ffdc345c6fdda8eaefed4e9108750fe58
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/kubernetes-imagepuller-operator@sha256:a72e3728dcf395667820c10c4c5ccb3f3d040ae5d04f50680e193398d3b72b88
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:5d4e4e62b03888bcbf6336934f2ddf5c5cb3d91bc8482a3aa1697e09625c6865
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-operator:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:df0939534c6aaa988714e64728ecdb387ff7dc4b32a5d97d56703ab143787171
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:ef65de2143df2c93aa73f099fa66d39e3239c30a9d547ed7bf258da0a19e8e35
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        - quay.io/openshift-community-operators/prometurbo@sha256:83444ab464bce4fb5021a2fca6d4d6ff72b40e90a6cfc9c5ac6c74357c775ad5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:032f053f21e273714e9f8a8f6854c83bc71cb84f5b32174704b490eca09eebc8
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:40b961a28ea9761c052a5cd17d2452fb72ea4786a92e13a0a1b589c7de31bf26
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:e4d902722d83f82372021d49c80f9880dfdb8bb312d4635e1d97980b60ca1fb6
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:961aa224525116fd575cff3a37a6465e9c652ee446927dd2bd03249b29469472
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:ad94f962d7aa064dc2b5b12ac2ea66ae5973909c82e6a96829165eaeff7ea952
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:504672b2db7fd25d1c85c9614f8242f98b4d2d04184a947dbaaada41ec205b45
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-codeartifact-controller@sha256:9bc95a0db7be356d2f657ae61c1c3098ff6e7e8bf1a675810dfb1b18805c6c8b
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:e80f87f289ea944a78b5c38df600c0a22fa470a2ce90f0b1e9453cd63586d75d
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:bddfd4ebd973089f7446cb16ed9668d418f91c3907a5ede22052be4ccbea6d0c
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:b31ebfab28ec75c0d7110d6f01e3b26fd973c36013308e12f69584ff38ab3779
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:7edf59729c2ae8b4506d5361eda01a98f33243987ac082abf758719c76a4804a
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:94bade4574c3679a425628f68067f9aead57bdd78e60f635041f421ed30341de
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:f5dda99364d07df3a3680a2d7031c27df071c8e80746b1ececb5487240e9bbb0
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:a4704a3977fccceb145e8972d2d7de074455bc228f2e925e5b427ffe3dcc623d
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:4f8f9dae947e9d34e86dc4237aec9167a9a25f83cecfae4982dc952494502486
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:6c0899610a78d2737c2096552e5146d9d5602b5568fab24b51b409dbffcf1961
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:84c6c826473b844293bd2048a498c5e0e8a558224a09f8e2e0cc85a2f10fcacf
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:c3913778f986ae4f3a7d94da2a9d0157233dbb676e3c1eecacbdbc02a864685a
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-s3control-controller@sha256:114ea0669459ae148ab32f5366f239da9cfde23cd6bd6a67c14e5844856837a4
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/openshift-community-operators/nsm-operator@sha256:330a930f885dacf65537beb600fe7f428e02687e83778e47dcda3dc2df15ae24
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:30375ef2cab46b967f6d2cd46862377c83e58bc6489eb0e75b495a173ebfd84f
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:466b9c30c02a3dfcd9cc94fe8eb432dca7076b8c79eb97ebe5802d716fd963f8
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:d5121af03b32a4b2286ff4f39499931a1ec04e05af45dccc6599dcc1b49e27ad
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:a0739ee7420b64e6195ec53057b032c356faf0ea4b8c1943725052f925eaeb19
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.connect.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:02d101a5929b60e612bf64a944e282ad9fd353c47d4ff28d2c9d5c0b84704ce5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        - registry.connect.redhat.com/coroot/coroot-operator@sha256:44aa4ea98d468f1d1ced671bd2e2bd5bf79ce467f8c572c2248abdce2b7c5acf
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acmpca-controller@sha256:c3929eb9a46af8feb2c830fd0b9225b6e3ad2559a9b5e04fda9239ef4cc07d9a
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        - quay.io/openshift-community-operators/machine-deletion-remediation@sha256:34d1e6cbfaf248a7b28c2a6e6a39f3fe46c183b9c38ac5fbb98eb3f421ba903c
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/machine-deletion-remediation@sha256:99b0b439d37ecdac8ec74433acf1a0ef787d0d6d1d7ef7cfe5f0a29f9dfe6071
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:90f512ce90433a33b50c58613d9b7136e00f116c48e179c4a1be89d6a35f9abc
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:53e73c53c31e0e06c990d3d966c2b4053d559a58a17a781155a3ceff3efac831
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:246712ce23cdd590ff83d84e9800b4ed761886ed5bab9f9cda4125a5d930e7e5
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:68e766235bdcd8f3335e2ab6e09942d98b183b3e498abb65acab6481664d46ed
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:91afe3cccd5eabbe4933459e63fd64a6bd87e399ea1aa81ae30dc3e8341c83e4
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:56db2fd54dbe982dde6a4558b664b19a676a8ac99bfbc8082d097cd6d71a03ea
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:d17af97462de060dc21f997dac973d9f895ce91f80821f1a810d93dbcbda31e3
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:783380361431859301477dfe63b6915240e934ffd07a2af2651094579be47a15
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:46f8c0ea6e383b742ab9cdaed1c8f624c2cf405bb1ce39c0ea6977940d52de07
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:e4ea95a86e77aa64da0b1800f00c2e0e598776cde7cfc7bb9941a60bac207ed9
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:aca3d91b7b9f8c1e8a1e0d045c42ba5f27444e0c1c0bc4a87c697b428c2fff78
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:5af7a32da67d39e7a011672de333792e63604cbfee997c19a495fc759d2fd739
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:74f0cf91e2feb60fe794f71b9f56fa8afc60e53379a0aef104701dd6692272e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:72724358b360ab74cc0857e2f18749d3d5b4c76bb6e930f31c5b04c7716e4c2d
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.connect.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:61f935c5985994994eb2f454fb12d74312f7d7d1627fc2b55f3d6ec9bdeb4bc7
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - registry.connect.redhat.com/mongodb/mongodb-kubernetes-bundle@sha256:354b16ea43cfcb9bff704b6265fb10d94037beb231138e1fb44963f3b348bb40
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sqs-controller@sha256:a22009a5d2037aacf52d223e45b74e6aa08020d4cfd1836ff137254b73e19bb9
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/namespace-configuration-operator@sha256:971991d2bf717001a98b175c50580cd7db9d230492e72670606c2ed50c53feb0
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ecr-controller@sha256:03ca7f898cfbf14f18a06cf3b510cddade3863275df934be891734b8c177c1e6
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        - quay.io/openshift-community-operators/eventing-kogito@sha256:abe8d5803d2d96ebbb9fe4cda0feaa0217e92def6a669404362bcf741d934690
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.21.0
          createdAt: "2025-09-15T11:39:08Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        - quay.io/community-operator-pipeline-prod/kepler-operator@sha256:37fe88e7234e5eb07c493a6412f27bb8df33896f7eab0a94b0e396fcdf870715
        - quay.io/sustainable_computing_io/kepler-operator:0.21.0
        - quay.io/sustainable_computing_io/kepler:v0.11.0
        version: 0.21.0
      entries:
      - name: kepler-operator.v0.21.0
        version: 0.21.0
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - quay.io/openshift-community-operators/eclipse-amlen-operator@sha256:bb89e2a7cb1c9172b080c72ad3227017d135fc30076d2a68b4e215fa8fa55846
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/eclipse-amlen-operator@sha256:5fbfb0bd4f764643cdc1021e99e489514a1f035d85131b13e56b11a3045883c2
        - controller:latest
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/microcks@sha256:1313bb08037e57cde02d9d48c4d8c9663eed08651a3c29f872b5ef85ac764112
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-community-operators/openshift-nfd-operator@sha256:880517267f12e0ca4dd9621aa196c901eb1f754e5ec990a1459d0869a8c17451
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:e5e0c3b3a25c13ea916d70fa6f6ddc20f2c730bd9909aa2654672280c27afda2
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:8c65e65bbaf0ebb10dba03655c5854c69c82ddc0e6b2a29293eee029222e5efa
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
          createdAt: "2025-03-03 15:18:52"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virtio-container-disk@sha256:3cff6989c7d7f72d2f0ac27e512d99873768766ba420d65a6b7794676d0c1e88
        - quay.io/kubevirt/virt-operator@sha256:d5fce85dd4232745408daebfbf70c2bc879a06f8340d1f7448b141d5ba3f1821
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-uploadserver@sha256:2b59d3c0a7c154fb6915cf3439b42eb4aaf48f94025ddb75fa9b73b16148fc13
        - quay.io/kubevirt/ssp-operator@sha256:61ac0c182528a9b54d84755fc0931ad8d408ae74e7e4b644316a5e9e630ab360
        - quay.io/kubevirt/libguestfs-tools@sha256:d655b9abbe8cce45bce3934e11d01715e170ddda4da78d7c51f356e7ee1e3263
        - quay.io/kubevirt/bridge-marker@sha256:e492ca4a6d1234781928aedefb096941d95babee4116baaba4d2a3834813826a
        - quay.io/kubevirt/virt-exportserver@sha256:ec1693cc6e8a14ab83f2feea87427643a968d65832c44b25a1bb911482403ff2
        - ghcr.io/k8snetworkplumbingwg/ovs-cni-plugin@sha256:516791acf430bc747f01366e2b748ac76c0b5e39ad8592be2b3dcc809429fa1e
        - ghcr.io/kubevirt/kubesecondarydns@sha256:8273cdbc438e06864eaa8e47947bea18fa5118a97cdaddc41b5dfa6e13474c79
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:66c55f5ffa4b44f2e58990baa7c672ec9776612a5f1c57fee5a5ab1c080f17bb
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:902af624e05eca764c17137c27813b1618f388e4b1a53b27b54665e8735251a2
        - quay.io/kubevirt/aaq-operator@sha256:5e53b755adaf6c66f59812b590a2eae2d84554fa658e93621ac977316843d918
        - quay.io/kubevirt/hostpath-csi-driver@sha256:f8d44aabe0024058b4380e1af42a57a90b7f8c66c8627091bf425fc384c0880f
        - quay.io/kubevirt/virt-exportproxy@sha256:55e38274e59bfe5a2cb5671534b8b6e75d90e43107abecc95fd463f7be8f0026
        - quay.io/kubevirt/cdi-apiserver@sha256:1849bc14687449d36561eb1c768fdbf0cd14ababf9c254b33a4ddb14abfba01f
        - registry.k8s.io/sig-storage/livenessprobe@sha256:33692aed26aaf105b4d6e66280cceca9e0463f500c81b5d8c955428a75438f32
        - quay.io/kubevirt/network-passt-binding@sha256:c794cc1b9173267f1e3d34e666d9e2f782b88d93b47fb4449675346b467f802c
        - quay.io/kubevirt/virt-controller@sha256:68a4967ef5925d4c24fb5eab0aa5acf0a82946beb7f1139bd45c85efe3653cec
        - quay.io/kubevirt/cdi-controller@sha256:94d16a6bd0d9776eddedf64165f39e63e4dfe4f8a26e858bd099db5d09136047
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:92f2ac10f28a7e80699bd984dbc4de897011b1f68c2be95f3fcb90df410b119d
        - quay.io/kubevirt/virt-artifacts-server@sha256:789f68f4e27cf947cdd4f1c0732042176e67b7faa5e37e1b503b8dace3a46125
        - quay.io/kubevirt/hostpath-provisioner@sha256:156ae46a03e2ff5550967773100122f9abc82e25bd7acaa55dbdec22238dd2ca
        - quay.io/kubevirt/sidecar-shim@sha256:cb4025f7275f8de2891c2195c5faf128ac1b68595cdb3eff776a2e5360b8f034
        - quay.io/kubevirt/cdi-importer@sha256:f9fff83c8307bca61a22e8296cd103b32e63d7c8b2a170c6781c6cd962211d26
        - quay.io/kubevirt/pr-helper@sha256:fd434b4ea1ca5682fc5b05f23de01130c0c10f77a0a2d9d089d8f3cb5b545aa6
        - quay.io/kubevirt/kubevirt-template-validator@sha256:b227679f8ccc7c53469574ee5f58c9cc39eb6c54ec01377e4b3df3205a2f4ab4
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:89ddc40700d52cda23104978498fb4d74a643e807450b630beaa71ffdfa1fc57
        - ghcr.io/kubevirt/ipam-controller@sha256:fd7c38e129ea4ed8a86d1801c9e1ca387802376fc4abc0b0d228d4278f448ecd
        - quay.io/kubevirt/aaq-controller@sha256:b9fd0818e4019cccfb38f0923db058166eafdaea8bb35a720aa22b5ce55daf26
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:16030e8320088cf74dc5fbc7dccfea40169f09722dfad15765fa28bceb8de439
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:ac955c4ad03c40d74341623b22f99e8f796a6ed2ef5cc539fba3c1504d159955
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:8061bd1276ff022fe52a0b07bc6fa8d27e5f6f20cf3bf764e76d347d2e3c929b
        - quay.io/kubevirt/cdi-uploadproxy@sha256:6a98ef32c3778e46812c698970e14736f178fd2a11b1f74ccc36c34ce2e34466
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/kubevirt/passt-binding-cni@sha256:f3fea2ec66a26850489197edf7c85d074eb78bbac52916b10016ac0de8ef7c14
        - quay.io/kubevirt/cdi-cloner@sha256:feefb3933780bf352ba03b62b900907d02c64e73f7e74b6d336f89daf2752f62
        - quay.io/kubevirt/cdi-operator@sha256:8297771144f2f358d893771a9dde208a6bedf6bfd67ab2927b16c27424ca3f26
        - quay.io/kubevirt/kubemacpool@sha256:6cc8ca860960c003b36754c6dde249f26f78d32eb22159a25b5647e197dbc9f9
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/aaq-server@sha256:18c23eb4e32ed083047c8808b35fa81a22294a7e8bbabc74e2612fd2eba97dd7
        - quay.io/kubevirt/virt-launcher@sha256:18508d8767f8530df4d6aca4ffe467c40f9891fbe052548b3b2316f8bfb1bc38
        - quay.io/kubevirt/virt-api@sha256:d92c78b3048756fd11d3c6f1a48f4ce8555dd39e866039313dc23bfa70668443
        - quay.io/kubevirt/virt-handler@sha256:4ddd39539282e79cf3602700b791eb91a6d79991bf2e4dc4cba24e3da9db3d1c
        version: 1.14.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.14.0
        version: 1.14.0
      - name: kubevirt-hyperconverged-operator.v1.13.0
        version: 1.13.0
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        - quay.io/community-operator-pipeline-prod/aqua@sha256:ca2b998fb0fa4e80d6f4c3be57041e8e55e43cf0b5f1fee579733fda9ea6a107
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ses-controller@sha256:0cdbea8336c2825bf1bc8dbd3f5235ff41d97fbed30106e7775c6e993f92d8b6
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        - registry.redhat.io/ansible-automation-platform/cloud-addons-operator-bundle@sha256:2cb3b1887e8eeae6708f5f46c9bfb304d7c3df1537ea433b0b23772acacf9f90
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
          createdAt: 2025/05/14
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.3.0 <1.4.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-operator-bundle@sha256:3804fcb77893d9198527b4801e2a778bf612596c9a0b125a0f0f872b23138434
        version: 1.4.1
      entries:
      - name: secondaryscheduleroperator.v1.4.1
        version: 1.4.1
      - name: secondaryscheduleroperator.v1.4.0
        version: 1.4.0
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.3
        version: 1.1.3
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      - name: secondaryscheduleroperator.v1.1.1
        version: 1.1.1
      - name: secondaryscheduleroperator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-operator-bundle@sha256:44d82d5f8434fb6e88d2da406e87df0bf4274661d9d77554b80e09d19fce03de
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - cloudoperators/ibmcloud-operator:1.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/ibmcloud-operator@sha256:b22e89daf6b7f0fd55ba6db7eb7d7a48cbaef2844dfe20e558d7b821482e9480
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudfront-controller@sha256:235ad7b4ad07644d9a638909c681620b24a85148a8655b7c903faf6aa42c2242
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator-bundle@sha256:0192000a2994a62a904ce4aabfc1ff04caf2fa9d779d27c016922b9121877fac
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.connect.redhat.com/synaxg/synaxg-operator@sha256:ff3d9ab5506c4487e5b6697ffe960584aadfe9f16408e7976d345c3d93375171
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        - registry.connect.redhat.com/anchore/engine-operator-bundle@sha256:fbbe7e6c1d75c4de2f47e2c825c930568e85f1134545e9d890a0c9f3d9187a4d
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/community-operator-pipeline-prod/windup-operator@sha256:4e2bafba54eb01b0ae77555d2ff3da2c34286c7c173756dd20a90de28be36535
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/layer7-operator:v1.2.1
        - quay.io/community-operator-pipeline-prod/layer7-operator:1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/mongodb-operator@sha256:1c06d3e936348ceeed773eb45144c9dd702e50ba293610421a93e3b3942ac9ef
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        - registry.connect.redhat.com/raftt/operator-bundle@sha256:7108686c168c35acd996e61a981bdd885b454837cf6823a48a28cdeeebbb0af9
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:d0bd55bad99bfd1ce09da5ee885a38229829570ebf80e1acc4341dff17a54e64
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:b9ffc18b2d88dfab92583af35b27e0ae36bceac7778cb448ab2ea2eae2ab3e94
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/community-operator-pipeline-prod/nfs-provisioner-operator@sha256:d7ef0fa3b5a1a6094b6fd60c6197a52cb14766865b3549806de100bb3522821a
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-kafka-controller@sha256:1a32352f30d0835299e14e8d0f0bac5585396dd6a10a63fe633e3d81c4d71b50
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        - quay.io/community-operator-pipeline-prod/postgresql@sha256:f9b7137598dcaf0774de71f4633c7e3b0359b0b40f313ad9c91dc8eca58592eb
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle-rhmp@sha256:875d81ba1b91a72d9bb4f1d86e2285992370a04ace3831b3c914127ab55a57e7
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:a657269d0212d952f5107514db63c3ea2b6c86ffed0a44f0f33eca96a242cf35
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:1ec1264dcc4ee289ada0747083ad9b072bad2e33544a3a879383694c2974a885
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:60b1d25ebb3aada94985e359bb059d4491bf0148ac75d51b68a461670659e429
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:3128cd81b56975643c1c88c6ceff6d72e918a67ec2d6a5a606b6e04753c5b0ce
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:808529c14a8873324096fbd1622fb895c4dc7f60c1cf9e3ded646f9a81d7b496
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:708291528f26c70c561745967a1200edcb2f984687f749a49a2860fee580423f
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:d1c636cc17155c4764d331e050f9bf3ec9f39f6ec6ec14e972179b185ea9f5b8
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:5339c41013727fa5dd3ae5bc1d51eb90e02e5baa85d4039b1efdde30baeae1ae
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:1532fc1351ca44f863feaaf54ba8810be7a62396f98dd292ccaf975fde31ce5b
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:fa72cc3e0b09510d71a097b92ee101e057f1c11e1cb363172bd26b2d94272417
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:b897e25e03133513f9afdb5abe38cbf0f2ac0556c27518d9d80e792eb07ecf28
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-sfn-controller@sha256:01036536877e5b236dc83774eb3b6fe9493eaa593cf5d1cb5b103798d0727bf8
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:20:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:36393d8945938325e4f330eee801ba28ab9b7604f58ea927004e17bb514f328f
        version: 4.17.13-rhodf
      entries:
      - name: odr-hub-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-hub-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-hub-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-hub-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-hub-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-hub-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-hub-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-hub-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-hub-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-hub-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-hub-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-hub-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-hub-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-hub-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-hub-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:bb352ffcfe7e79981f96fc5b5a4f2f557fc7fe39de27c36f195349bf4ad6248b
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-hub-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-hub-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-hub-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-hub-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-hub-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-hub-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-hub-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-hub-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-hub-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-hub-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-hub-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-hub-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - registry.connect.redhat.com/kong/gateway-operator-bundle@sha256:91b3d73bc75eb170035110e5c63747e4918f738841aa97b6df27d22e37b4879b
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:d7bd0e90aff6e8fc9fbc4119e5f1ec89eba93487b84356e3af95208c1250fe44
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:fc374491270b9566b0a5c80f3dd665b2227b1e8fb224eada7e5c98c82926ae08
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:411a47151609b7da29974ef3ae3b04f1bb1526f4328fd3e1d9b0165380018049
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:076c345de449353cc9837c603024ba9d790d26c5a8cebfc97db61ded1482ab78
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:f481b2e364b9fe417a1cbbe9eff70180a3b4153a14712a8935527e25f919ee0d
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:a85d4f1c8a686abc0b9032dddda75991216d44dfade214625723d7009be314d3
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:b71367012d1ed17d3967cd77fc28b38c69c76f75848fb20021dbb176dc382eed
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:32b442eb6344d3ab64d2885f53817e61f0e8b72b5c6e28845c8e056fc176b666
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:268aa90a9f6eebf6cdeb32dc47335e75e7483dcce4ea3331b1b736c2fd79bea4
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecrpublic-controller@sha256:6acdf23661d603339e998307e9d5130b21c45665f4b0c35019ea80aac3d7343e
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
          createdAt: "2024-09-09T20:32:09Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:6e80289c93ce0ef26c776beca59eec792832dc7f2f57cb76a7321eabc12405c8
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:8b8d39f64cef8a1ed37a9ee44ab3bd5b98575ff32fd3962f94cf7a9c79e623ae
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:6173ab8b48c178ea8bad0d6b98e73801aa28135fcb7455146449a66543891e26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:63ef9626a82b5d591abaff2e9fe76f52e2bac16b910da40af5e92920694a2a67
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:c88a193654c191820bee307d12ae5e721ef82740ec912de9f710481b386edde5
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:c3cbe6cf4158aec38c67582fb79358315d196de76e0da611b4b8d6183babeb58
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:33cefb8803cd212932fe2bbb6052e53562d04084403f0166c8e077957b92c58d
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:ad206c2f78dbcdec43755eca3cb92229ee77e737e741dc276efec23f6379ecd9
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:e90e8df1726698a79ec689358d323eea7a26a72290660db62443e3f0cd696a4e
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:66dfedd62d474aaed03b1fee142b91358271591b651a3766d8e0542d08da914f
        version: 1.0.17
      entries:
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:7317ed83242340597b5cdde8ee314606ddd563d0067ee85c12eb21037e2feb99
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:16b39335e97bcc7c970a8ba385c61d8333e9912faaee813ccf8520d434b7bb7a
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:51ac9aced282f62c8fb2d3696da3696b8496d8f877e9846fe3a4a743ccd63bea
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:a82288782b61ea46d068cc09ee10608d3201c136abfc767e4d37734775101390
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:828d31bf8a441091b8aed03da213a6b4ffae260070eff58c3319156f9bf04f5a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:f938cc534fc4b24bb9449f1e1e1e904dc588e6646c98474de4b8c81d083b4d96
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:095629572a40469eb651c9f81d6c10ff46254eff8b9813cadce54d6f0075c2fb
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:91456907c61952017daf07e376b565472ef86d67a2925f60e189c639ceb68595
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        - registry.connect.redhat.com/loadcore/keysight-wap-core-bundle@sha256:2f8295f8f56ef2c56e8215d4125cf08facd2f1d07fad0ef3a01eb0a2ef5f0555
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.connect.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:623a93173e8f703849238806b40aac843cb2ed016c8763a4f8ea1a0b0990eeea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/special-resource-operator@sha256:df87c400713311f2495d6d02769acf29e5251330194f992c11719063095fdaea
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.5.2
          createdAt: "2023-10-07T15:15:08Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.5.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/openshift-community-operators/kube-green@sha256:e0c9f5ecdef1f7172af098f20cce17afebbedf02c0e0bd516d4d6aae736d1b27
        version: 0.5.2
      entries:
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
          createdAt: "2025-09-19T19:13:34Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-iam-controller@sha256:2d904b6462679d578505409d9bc585795983df7a4f0dd6668253c724aaa5063b
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
        version: 1.5.1
      entries:
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        - quay.io/openshift-community-operators/edp-keycloak-operator@sha256:82ff141bb86599f12d4a516fd1b6c5252f9ff24ef2547be6c58e2f3d88219914
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        - quay.io/community-operator-pipeline-prod/edp-keycloak-operator@sha256:10151387a79d0fccb47114b56525f28f507faee96e4fe29babd2ec46c3dd9343
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/ibm-bamoe/bamoe-operator-bundle@sha256:54df7eba0292459c0491ba39c15bfcf793e1c0aef042a4b3f6c587b0d45f77af
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/cincinnati-operator-bundle@sha256:50e9c88ca31182c54f9c4bc6682642a5115fa280044e6c125f99626c69d999ad
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-03-25T14:27:42Z"
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.0 <4.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-operator-bundle@sha256:a372f2b3b61c0012618837e349d389f9f300ccd2ae0fa005e9cd7d379b8569a1
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:da231dc87dad63f4fe28e3b438a1ca9a96f04839389f7c3c19e8368bf4604283
        version: 4.18.1
      entries:
      - name: numaresources-operator.v4.18.1
        version: 4.18.1
      - name: numaresources-operator.v4.18.0
        version: 4.18.0
      name: "4.18"
    defaultChannel: "4.18"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:8ad2282afef801c794e316c93a12834d688cfa44766132231f93013c3124944f
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:d4ec9966370c3e770209042e437c53b0780401150f5385069b8476e26c343919
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle@sha256:10c0eed1e8ca3154f4c815241f6275f40113ae2ba0abdc940ad9a04682edafef
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-07-07T18:31:48Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:abe7b5029a0956943b5af20cd66192a6fa816998523c05b671dc1280a0219683
        - registry.connect.redhat.com/minio/volume-manager-operator@sha256:e3a69a9ef99d98d1cd8eb43d3acca48a6806ccae5b6747fff0ce79712a7d363d
        version: 5.0.0
      entries:
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:43bf1b9f06a7ace06e0b87921ba9b62f430ee3c8b2357c713054151e01f9cd02
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:5ebd9fb9277844967f55b985fc44069c1c0209360735c5321eea3aaef8628528
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:e668de3c5a2eed2b2b89515860f735be8ac03c2646ab7b65ae3c1270262bc1ed
        version: 8.17.30
      entries:
      - name: prometurbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchserverless-controller@sha256:e31d29d7c3389932e42c82645fd7b697aa16e03e20b17be0f355ef6b301d5ab8
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle-rhmp@sha256:d54904777c861e00858ddd8f1218cc39db47fbfb1225b8d35902ed4e13e357b1
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - quay.io/openshift-community-operators/kuadrant-operator@sha256:398c9230878feb756ad393d7de5675a7370c883a18acf73483a660e9303d807c
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        - quay.io/community-operator-pipeline-prod/kuadrant-operator@sha256:f21456392af3b37ae48d77dc7f19fa23c6946269555bb1280978b29420bbf31e
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator@sha256:6b7a1e34c33b16c464745f70658a10d6ea3d96099c1e2ab008e96589ef73e7eb
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        - quay.io/community-operator-pipeline-prod/shipwright-operator@sha256:f4d3b97812c69b396c8e6d081d8fa9b412ebe1365432658288e6ab065c220fca
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/istio-workspace-operator@sha256:7a581f048aee64701b23f1a0e94d77f2e6932ee8229fd46b1635d8df6994e174
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        - quay.io/community-operator-pipeline-prod/kubecost@sha256:ff1e00f5a04e49fe9daf06db6b60c2ca14692efd098a38ecd622ef6bc5de8dd5
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        - registry.connect.redhat.com/loadcore/keysight-wap-ui-bundle@sha256:7ca0365e849fc98950021f86f3acb9048303fad3dcb6eac0ee09dd4bcdaf8fe0
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator-bundle@sha256:3715abae1ffeea2d0bebf665b1dfaddce4686bb8a61f0697e6fd201b157798b6
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:fb7bc294096fa78f199d909f6ff8f7fcbcd771c3dc19f3f767d77dd13a2081a9
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:10e4bdca44a1a0cdef27b415765238249f81ae6ba53f94f992e37d350bc9f800
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:3e8aa02c71fc99e5eeabfce39c8fc24a9f9e8e826a58f9ff59c8b7c0e9c14e0d
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - quay.io/openshift-community-operators/datadog-operator@sha256:b1977170e5991e1f08c49037d6f34873c8543d29d3a2c3692ebb1d66734f5b3e
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        - quay.io/community-operator-pipeline-prod/datadog-operator@sha256:0a5684e5bb442c5daad6737f1560198ca380507f2da583efd0a7d66e18f2dade
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:4e12e48d0b2b884de2a0d32682da044cff419aea53bb162915538e2179fcd7f3
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:9eca42cc297a93293200a1c46e8ed949a61770550366044ef54e36bed9d7a116
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:31eda4f1888b180feb023c42a6ec93478b2e956f13420d55987d9abf7c7797c3
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:7d5f9e02f3d6c3e2ac63f1fd33240dc696ee79563c214d21d2897ee773517e9c
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:5b061126dc3adf1fc9c8c22dfded18f4652bd62537aefca88ddcc585078b94ae
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:007a3d43633fe355202002737f766a7c636f537f3daeffa1f62356c0bbf969aa
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:87594d78977ead074e53b39a52e905a2d2861d24d69d2f2e33b066077cd279a4
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:ff8e563d97f0d0e42c690ca3e61d3625e920947264ea99c396392c440cd43201
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:12a02131f13e532a26572a524ae542766092ff75624113994d304824951161b4
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-operator-bundle@sha256:006edc597eec839f960bfd191b75217fc8c6bfaea9d1694e7ac9fa34266f9b4f
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/service-telemetry-operator-bundle@sha256:c4ec01d0d27eaf93cf820d8da1f2ad7ef3e557c5b4ed651c5a1d07a1fc116f79
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/rhosdt/jaeger-operator-bundle@sha256:f250e39033d7cb1d786e5a7ec6798c25d4c5d8c6ecbcf6828915605fc4658da5
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/annotationlab@sha256:b69fde0eb341114893aba097d58455c9e316a97cb829b41c0d080987490a83d3
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:2dded4ad8e80f711fc18ada34ea825e844045da44f0813550a586371dac84b52
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:a20986fafd0a475a5a77296b3f8a18c78edb84da5cc8246b06c805dd5f693511
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        - quay.io/openshift-community-operators/hawtio-operator@sha256:38b36745547a08afd10e1620f8243e6f1ce18a3eb8883ec78cd631577177c5ee
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: helloworld-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: helloworld-operator.v4.18.0-202506160933
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.example.com/v1alpha1",
                "kind": "HelloWorld",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "helloworld-operator"
                  },
                  "name": "helloworld-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-05-18T20:30:40Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202506160933'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HelloWorld is the Schema for the helloworlds API
            displayName: Hello World
            kind: HelloWorld
            name: helloworlds.apps.example.com
            version: v1alpha1
        description: helloworld-operator
        displayName: helloworld-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - helloworld-operator
        links:
        - name: Helloworld Operator
          url: https://helloworld-operator.domain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/helloworld-operator-bundle@sha256:056744dd517acc3c12908e5ce06bfffe2c634d427b837e244dd7185c70915c00
        - registry.redhat.io/openshift4/helloworld-operator-rhel9@sha256:7c9a290bef3368d01e44bbecafdd99d2e18376221557938ab82c483bf189f370
        version: 4.18.0-202506160933
      entries:
      - name: helloworld-operator.v4.18.0-202506160933
        version: 4.18.0-202506160933
      - name: helloworld-operator.v4.18.0-202506090359
        version: 4.18.0-202506090359
      - name: helloworld-operator.v4.18.0-202506050817
        version: 4.18.0-202506050817
      name: stable
    defaultChannel: stable
    packageName: helloworld-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:13caed923cfbf9d5ff215c8f87fc64ecd753d183f5573c31ea89c9bc5915c6e1
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:c47f099ae5ddd9b25fb7aa5ecfc253fdf24df2e7929318bbcf162583df91f886
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
          createdAt: "2025-05-30T01:26:40Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/saison-technology/hulft-bundle@sha256:b7d1348b16ae564bc1092dbdb2042bc2a1ee8a3b7fdc85b7d6fe18be9bb99c19
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
        version: 10.4.0
      entries:
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:e16accb01b2ca6bf4ebd60c6ce7faab73b784828578ee74305a3a97ff9db1e26
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:ea37c68775b14e7d590b6eeec3c9b59eba8e988860fc04d40775da3957d6d040
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:1673f08529a94671f881edec6fd29b73f43c2115dc86b643f9dd79d7e36be38e
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:5eabf561c0549d81862e521ddc1f0ab91a3f2c9d99dcd83ab5a2cf648a95dd19
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:758bfd5455d0ca96f5404229a73fdcd56686564fc4c315ca3468ebe8588dd9ca
        - registry.redhat.io/openshift4/ose-tools-rhel8@sha256:486b4d2dd0d10c5ef0212714c94334e04fe8a3d36cf619881986201a50f123c7
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:0765d248fd1c4b4f8cbcbee23f35506c08c9eeb96a8b0e8ff1511319be6c0ae6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:0ad1961d6dd1c8636f5f1bcb5ffdbea51ad29b88c6539df366305ef801c79877
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:06382df99dfff9f002365a31eee0855f8032e351bc9583e42e489ab7623fe287
        - registry.redhat.io/rhtas/client-server-cg-rhel9@sha256:f0c8f5129c0f611332f7b34dfc5f0dc39c73ed8a985a0845d1fe6ce31d2ba020
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:d4ea970447f3b4c18c309d2f0090a5d02260dd5257a0d41f87fefc4f014a9526
        - registry.redhat.io/rhtas/tuf-server-rhel9@sha256:6c26d2394b1d82995a6f3d3483d2c18befd6a4ae826af2cd25aac30dfabef1cc
        - registry.access.redhat.com/ubi9/httpd-24@sha256:7874b82335a80269dcf99e5983c2330876f5fe8bdc33dc6aa4374958a2ffaaee
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:496b29b00a9831278eb3a8316bd5fb233ad73cb9a6217b1ff4a5f72a91b87414
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:58140d6b7fee762b148336b342a6745b1910ab758a701b2e8077b1fca2af1180
        - registry.redhat.io/rhtas/client-server-re-rhel9@sha256:83b9f9a5ea40473c8525a16d9965b1ed1eb2d8f7c2ec1d17ae38e5f10ac1652e
        version: 1.0.2
      deprecation:
        message: channel stable-v1.0 is no longer supported.  Please switch to channel
          'stable' or channel 'stable-v1.1'.
      entries:
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:540556453c98db3919e5f305a99c699e767e2891a95fe7527ace578f3c67a308
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:6131053778ea04e437f3005f90d1138aa11ebc58e3a9295e2a8d8ef6713a52be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:09b8259ec9e982a33bde84c7c901ad70c79f3e5f6dd7880b1b1e573dafcbd56c
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:571c48ecb2658ce70199c06dc483ea48a16e032a764a6830388ad845f508d981
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        - registry.redhat.io/ubi9/httpd-24@sha256:366cc5bc3b96069fecaa0abeb9d3510c507e69a20de1d205d881da8ae8561ab4
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:dc994a95be22b0f4bab022fc362c4f44c6a7d1887a2eb0d04870d75654ec013b
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:5a752cefdaf28bfc53847185cdd5fef1ee47e3dcff8472f8a8bf7bbdc224ef57
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:75f1049431f9e92898a4217870309cbbb3b39c8362e929c0bad3b53cad4459db
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:603a2ffa54df0bd45ff42ac710dd16331560eda2be93442232b045c959773c4d
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:329a0dad7842d4ef60430b0e3f188231bea9e579427443ecd42eb62a5c282266
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1f5a30a285a16635a7234c3c1763dfb385c8bffd605fc862b782bdb5c6c61ea3
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:3b8f49c41df15022f8ffdf3a8f8605b14c14f4e10eae754a06a86b6585d158b3
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:775b261b8f258b369213246f2da5c91218678acb5e3b8bb4670b143f487767af
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:76c24a38ac89ed632d38e44049f37e4997abfa27fa8cadbb8afb42575031296f
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:796860a3e85712c60398c36983e0ff4d45325c7a4de869da2ebf1b6ba4b19825
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:7ce611aefdfedd8b2a633def482cf41f390c95b8f8c800b6163a585f117a9e2e
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/openshift-community-operators/prometheus@sha256:1de089becfcceb3747b3a3ce310c1612861b03446906cab5e981071db10709a9
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - quay.io/openshift-community-operators/azure-service-operator@sha256:2430ebe724ead60e84f12aa485c8dae4af82af10d14400ce76356392030c8c0b
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-operator-bundle@sha256:c6bc928d220fa001e2a7b5e8661a3c8a604c21385dc1bb97b8dd121ca2347916
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/openshift-community-operators/spark-helm-operator@sha256:9bd49276e02f2fb28bb1ff116112cd7ebe0aa12bdb9f1e46aae2501fd9e4e755
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patterns-operator@sha256:99ac398b06fa9d9a7626b99cb47fb380d0cf4a4307fa11da8e486df35966debb
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        - quay.io/openshift-community-operators/varnish-operator@sha256:46de72c5abad80cef3d6426be967131c51c2c15029a6a38744975578c805a44e
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:3f7fc52c79f3e4ef7ddd6f5efc8eec1ee4836bbf2076e7dd5f68bb3b7c6e9d8f
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:34d4f950bce7863f15d5f1e7d6555a3ccf97ec31dc3ee140f68442890c4f0f96
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:96b77663961daea3a336b97e34e9412281bfb2727854dc970a69edd029b9af85
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:5170255ecd45b9bb7a70abec0f6eeb0518bbc2871de759e468dd61a80261c0ad
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:9a3f5b6c403fb84c9cb93c9d7cf902c80990196983607c8ca568d4b558938e17
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/kasten-rhmp-bundle@sha256:c6b58e5f6c9b2a1f5c4bb770d1fa1db6eaaee8ae7d9751a48338623d0d812438
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.connect.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:99e845965f7eca4b5a3361c494a385292d655e7c1906185197826c773cf9aaf6
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        - registry.connect.redhat.com/qc-qranium/aic-operator-bundle@sha256:581b21f44d6b993888681ed702ed10b9d1bb8e52dcca53a91e856febed2613a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/devspaces-operator-bundle@sha256:a48dd94d5616183843ba939146b7e4d985210cdd902038ff80ff6df3df3d74ca
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:18452ba29f4e0f52b765e8f9359e4143ebfe1203160e6f64ce46ec9d4f095a62
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:9716402247ed874f19574a0d5125d1035fd198f36081d9129c1d4d4ad2f714a6
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:397feef6fc1dffb7e2c3798bbe3bfebb86d6173c225eb61473029cc4500b0cbd
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a70fceaad3b6a23e6f247b4dff41fa64349c4bbf841b66adbd0f8309c6acce74
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a7a4d20b5acc59dbb8c99a37c19b2749f15ed0be7f841490a829796865c5a956
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        - quay.io/openshift-community-operators/loki-helm-operator@sha256:fa3662ec16e6cb3f866ad46982a50d8fdefb43c68f3caa13291239013e03af13
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/openshift-community-operators/infinispan@sha256:3ba9d20e83825f3c041a7dfd195b0800f74fa0f5cd1e6218fbd3547fedd85d64
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:f165693aa550903d7a26b429ed4d093f60194a7a04148e65e61c52d67b527fad
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:5e1a1fdb5a78a34155abe77754c774c65335ea8cbfdec2a222ee22a062c2db46
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:e43b0581415af648c0b82830377e354a951814d3eb7825bbe31b9f3e633917c5
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:4ec16d88828e580fd9223125043ca49569324c4f06b7a65bbaa7d76923de2bad
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:7f959a4076c36d76f8f1489a92d6984d1cfd2ae5c0533a78e8b16010eb6d4d3c
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:a3132caf9cddf6d6186b3e982a6f6232a208fc727b433ee6d200a3b72f50ca88
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:d89f4d647534569f401790d6e10bc42433ae791be86558b100a4027a74b92cd3
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
          createdAt: "2025-09-30T14:21:42Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/flux-operator@sha256:557ae37fa6a6942e3a9d06966df67721011ac7905e2241867c975948092f2936
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
        version: 0.30.0
      entries:
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-operator-bundle@sha256:84d65eed99300b0ba87369cbf9a9fc17a37781f8b6f9de2819d0a159f9071a6c
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:e140415e0361cabeab192ba6d22634283cd47749dd25e9cee4a775d3c409cd97
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:4b7796235a0a545db83570add7a675fdd9d82b17f931eb81d0cafbedb1fbf22c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:9f9dadba5a35c8ab98ddacee22593a247a31e96a4b6e1b4f1a946c2e34006bb9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:01ca6a65a041a47efd611673a720e69464f02424d21feadc496c356065ff2f65
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        - registry.connect.redhat.com/appian/appian-operator-bundle@sha256:2cfb2d2e43fa207c8c06df6de8d3e90123c4cd52cdb6a4c6d618ec576b79c2eb
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:ac244e06ecd2f8a5e4de7b121efb8e82e3e7d23194878cb8327ff0b2affb5f04
        - registry.redhat.io/openshift4/ose-smb-csi-driver-operator-bundle@sha256:91d04cbe0eee6c4fe0230d0d179366942e6c54afb65bb77a43a681e1a2411caf
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        version: 4.18.0-202509100344
      entries:
      - name: smb-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: smb-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: smb-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: smb-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: smb-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: smb-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: smb-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: smb-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: smb-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: smb-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: smb-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: smb-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: smb-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: smb-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: smb-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        - registry.connect.redhat.com/enterprisedb/pg4k-pgd@sha256:fb206b95690e9604c41715369c69125a906b866dc35cfc5c63638ce7a4337fa2
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.connect.redhat.com/loc-devops/operator-bundle@sha256:b9f7a678ceec2d367d55499a48a15a9dd729bc34b0e707989be23c97fc3c5f4e
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "aes_key_rotation_interval": "2160h",
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_location_id": "devspatcher",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "enable_server_session": "true",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "session_duration_seconds": "86400",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          },
                          {
                            "email": "admin@mycompany.com",
                            "hash": "$2y$10/yYwtXRrOagYO2jNtwVzGQPPEfDvX0QllV8NXbku",
                            "userID": "b7e2c2e2-2e7a-4c7b-9e2a-7e6b8c1f2a3d",
                            "username": "Admin MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
          createdAt: "2025-09-24T08:25:29Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/edb-hcp-operator@sha256:87833c7edfa714b433fb066533bd415049b4ae96588ea8457a85a62326801d5c
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
        version: 1.3.0
      entries:
      - name: edb-hcp-operator.v1.3.0
        version: 1.3.0
      - name: edb-hcp-operator.v1.2.3
        version: 1.2.3
      - name: edb-hcp-operator.v1.2.2
        version: 1.2.2
      - name: edb-hcp-operator.v1.2.1
        version: 1.2.1
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator-bundle@sha256:a26586b5f2d748f56dc29fbe3871d84de422e4103ce13c93dbe477718c582487
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        - quay.io/openshift-community-operators/yaks@sha256:0031de9a68502cb65923c4e17453195663f59e46b1cda37619a857add503c016
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:91632b89b4001e2ffa0f1aef56207eb48f0275e20c635e7f1ff7c6de57a59ec2
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:eed8048d085eef7eefaa643a09627521fc2596ebd26280e05c79dddaa0b01afe
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        version: 4.17.13-rhodf
      entries:
      - name: mcg-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: mcg-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: mcg-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: mcg-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: mcg-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: mcg-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: mcg-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: mcg-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: mcg-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: mcg-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: mcg-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: mcg-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: mcg-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: mcg-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: mcg-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{},"bucketNotifications":{"enabled":false}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:ed5c7c8cbbe455b6d31b1c55871585a0679a91800f53743d81d4cc36c6aac7cc
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:88387ba7a9e85c1bcf57966558db57c099fd301ea80a68e43afc2bca3dfe94a7
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.18.11-rhodf
      entries:
      - name: mcg-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: mcg-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: mcg-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: mcg-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: mcg-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: mcg-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: mcg-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: mcg-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: mcg-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: mcg-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: mcg-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: mcg-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-operator-bundle@sha256:b637a02d23bd8364cc8914421049eb60169c163ac70bff2f33591df1a1193002
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e9612ec99daa171c403b5ddf37788526e5375b83987e5de9b1ebe519198607b4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:34d177abb56b948eaf9ce6fa62e37427ddee9fcecca48e752d09f72c69847a46
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:932e83fe9c9b54934767a7c5af683b2ab08bd2b4f0524c191f1c766dc64d9c3d
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e1c19bbcd4c010ef2bd4f760eb24b2c96cdb82b4e025361a433e6a241eb23cf5
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:a6e02e1876fd5382ba4921dd65650b2d747666feb96c839f34e99d2972dfc323
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cb711fcb32b5ca0f34c4e3c67624bc4e035de4247617d988a9e841fd58c1c911
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:26af4e950a33755114298386a70229672a09ce45b6203025856f460a55c4c8a0
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        - registry.connect.redhat.com/amd/amd-t2-january-bundle@sha256:7c87354aa07e783a2cbf4ab29877a8ba56ea4269bc1f41bfbe08c558b3666841
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/deployment-validation-operator@sha256:2803298c7cf3eeaffe9da3153142058d94d65ff1dae90efd8819a4cd23e0518f
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        - quay.io/community-operator-pipeline-prod/eginnovations-operator@sha256:19b12c75f7a2dfab027881ed9f0b09d25e3baa936d21f7be3ec1e0177ed9fdf2
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        - quay.io/openshift-community-operators/limitador-operator@sha256:7aa0a4fe55e5e099e2e6e4c2cd44b9bfeba69d8cfade10500ccba3092fb60dd2
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        - quay.io/community-operator-pipeline-prod/limitador-operator@sha256:450f9bf43b57b83e9975a1865981af22e28c725193573354d1180f406887664a
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/flux@sha256:cd94037ff3055ab6ce40b0670f900248a5fd233cc67a2a4f92c1559cf9d2f3ad
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-framework-bundle@sha256:53d056ac37177263d79098cda775deff86a327c65106ea5a40109b08a856cf84
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/keycloak-permissions-operator@sha256:4de19e52e84cdba0963bdd852a63912f097280800d65add8d474c0953da478a0
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle@sha256:a90a748aa70ee5f1b8fd375a7877db1292792ba46f7b49f629533c82a25c43a0
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.connect.redhat.com/cpopen/ibm-zen-operator@sha256:5a9120d5977e934ca3801eab44ed48ecd4a9fee10c0b4980594d33cb02e64855
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        - registry.connect.redhat.com/airlock/microgateway@sha256:80d49a299a97bed33c28941f93312e86b949ce945e843c93da6bca49865c81d0
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        - registry.connect.redhat.com/airlock/microgateway@sha256:087d4d726f8080111d2b18cd0f9ab290f4560333fef308e553991484b6210c61
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/openshift-community-operators/nlp-server@sha256:6f80f44822301252540c6b808a20cea327f0eedeadc7643d31fc743de5c5e696
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:fdc0c228ca38db30f89812e41c09ec96d9276ed844b0801013375bbffd6c1a16
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:ea606dcfa4cbe4df7b27c10d3fc45f464845518c4896845079ca9194600ceb6f
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - registry.connect.redhat.com/loadcore/keysight-kcos-ingress-bundle@sha256:f04b8f15b069c06eac5d9a97f81a2f069704237eea998d35727b89854c793746
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        - registry.connect.redhat.com/loadcore/keysight-wap-msg-broker-bundle@sha256:a7dfca98b4e09c20317cbb486e7893de8d88faf5db2ee26549ba81332ec526cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - quay.io/community-operator-pipeline-prod/dell-csm-operator@sha256:6e2e813b7ca426a793da717d42f3620edbc3eaa84be710600f2a5262924378b4
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:0739d5035590cece6d2a573aa1e799b0b0cab8b548a41cc2488cf465486cf4a5
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:3028579f4909be6b8dcd970e778bc88d0576c76e55b98c886738b8f601cdf4c1
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:e50343879b4c3c4002c5e7ff67f4664df441cf5f239f59fa98b7fddd75516d41
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:add5d14653f57ac1e3254c2d1c70fc9232959f76ee6286fe5cb572696a5a4fd3
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:7fc4d97b5ee63085672fc9a562ebdad3d4c00d1b555e06679d1dd9ac945f1bb9
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:05e12cc4caaebb7f185f651f8ca876c442c02d79794e561b743beba0d189fa04
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:84796005b6e71efa5d783c55b1f72bcc3762944bf32740cb75498bf8b908bffb
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:424e138a17e2af91603e834ff65d529400d6793771ddaab92d3988f4e6d2f2e3
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:9d505bf561650eb9785811ba50c920480d85679eac6a611a1c5794544e175934
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3abd8c3f97c9bba127b6e8c7efcc078c90c9d7cc35d74cfcd9bc0d303f62d2ed
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:21c17448932f053f1d4845111288a317786b190154bae9da7947c315ba12f796
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:023b9cc1e4ec2542b8d68b636b21a2e8c4eb143424581c62d16598fddb47da6a
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3843fd48c57be102576db93e8483a30f8efe8cd46b5e09c704ac0bc5f18553a3
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:55bb4133abc945ce7b75e6c01c6beaf63c19c129fc8fbfb48f1b53b8e28aae97
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:43b959fcdcf107b70e4224d80b63d1c123560c3af055109bbcc7d68476546584
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:2845843f32b895dee0242e31380b44625e3f8381caee7ecb3258648311c302f7
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:c7cd3ffa88eae188e94abcc0af56d328aff43463b98a3bc3c0b574f29373944b
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: openshift
      provider-url: ""
    name: pf-status-relay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pf-status-relay-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pfstatusrelay.openshift.io/v1alpha1",
                "kind": "PFLACPMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "pf-status-relay-operator"
                  },
                  "name": "pflacpmonitor-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.19.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PFLACPMonitor is the Schema for the pflacpmonitors API
            displayName: PFLACPMonitor
            kind: PFLACPMonitor
            name: pflacpmonitors.pfstatusrelay.openshift.io
            version: v1alpha1
        description: It deploys the pf-status-relay application.
        displayName: pf-status-relay-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pf-status-relay
        - sriov
        - lacp
        links:
        - name: Pf Status Relay Operator
          url: https://pf-status-relay-operator.domain
        maintainers:
        - email: marguerr@redhat.com
          name: Marcelo Guerrero
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: openshift
        relatedImages:
        - registry.redhat.io/openshift4/pf-status-relay-rhel9@sha256:8cdc0714d02144f5b32e26c97f7678d0377e0c3826c475a784e76a42f781d3e3
        - registry.redhat.io/openshift4/pf-status-relay-operator-bundle@sha256:fb9635a614b6676fbe6a532d07b3206f12ee525b7947378785e79951d43940b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/pf-status-relay-rhel9-operator@sha256:e6719af6fdc99c652238f3c31a014782811eb46da91e65f93c083a7a001e2496
        version: 4.18.0-202509100344
      entries:
      - name: pf-status-relay-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: pf-status-relay-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: pf-status-relay-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: pf-status-relay-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: pf-status-relay-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: pf-status-relay-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: pf-status-relay-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: pf-status-relay-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: pf-status-relay-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: pf-status-relay-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: pf-status-relay-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: pf-status-relay-operator
    provider:
      name: openshift
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/rhpam-7/rhpam-operator-bundle@sha256:46e7d899d64ea8cbfc35221d05535db1023c41c1d3a7c580cb03e7e7d769144b
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-rhel9-operator-bundle@sha256:896a615487c0ffd5335b869f0dc0c0ec1c9200dbcafa52f2493003fecbc030bc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        - quay.io/openshift-community-operators/alvearie-imaging-ingestion@sha256:f65cc3cb83b1c75385970ce70c6a84b15b143e682ad0d1784148f1643df0991e
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-operator-bundle@sha256:f596dc9323da89f5e45e71192f0fa698b7bb29b8205eaa0fc1659d68fa7401f4
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:8e223a537f487bc3cef10769adc6fb619085dafe3ea797cfd52c20e3066f694f
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle-rhmp@sha256:81476f897726c85293e64e6188ed314e3e1e637c568c6b0ff8f475a876bee29c
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:4ae986761a69d25697b9b6a3a99de88d8587beba136c3de29b90624e2ea6738f
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:adbfebc73bc5a9cd12d17e9e1fbdf87c2176aae7a7217bcefcec3068e51bc904
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    defaultChannel: stable-v0.2
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/gitlab/gitlab-runner-operator-bundle@sha256:68b22c412a44d5fa09b45e204d7c8e7b6ca6afa4ed6879de608dbcb0b1929676
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/openshift-community-operators/pcc-operator@sha256:e99a466ddb52ad44c435ccf952754d685744fb5b7b7e4cf77dd542e530e4f968
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53-controller@sha256:227a306cc90998489bfb681cb29b416238b404bc2db28a09b8949d4cca2d87a4
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-operator-bundle@sha256:e83707d0b396c38cf55046d3159f1f427188404aa11dc47d7b1c1f25605182de
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.3-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:64a623d5d536c790dee04c1604ec2f4429f8b8f1973b4232b543e8cbe23d6590
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:7c9b4327d8677f48c48888230fc38eec8c20c7f646dde568dbd96b5e4c8f7967
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:8797a3be94ce6f4e4da15e4efefe41ea5b6ffb4cdc6d698638c80976bb22491c
        version: 4.17.6
      entries:
      - name: lvms-operator.v4.17.6
        version: 4.17.6
      - name: lvms-operator.v4.17.5
        version: 4.17.5
      - name: lvms-operator.v4.17.4
        version: 4.17.4
      - name: lvms-operator.v4.17.3
        version: 4.17.3
      - name: lvms-operator.v4.17.2
        version: 4.17.2
      - name: lvms-operator.v4.17.1
        version: 4.17.1
      - name: lvms-operator.v4.17.0
        version: 4.17.0
      name: stable-4.17
    - currentCSV: lvms-operator.v4.18.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:5d0534e08790e91d5d355eac2e5c187a217c1ec6eb59464ea9a771490ac040e9
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:6e809f8393e9c3004a9b8d80eb6ea708c0ab1e124083c481b48c01a359684588
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:b3b1c4836a82a7ec6c9b02e4a86c352f4ae8c45a3fe027ae7200869afe3ab028
        version: 4.18.3
      entries:
      - name: lvms-operator.v4.18.3
        version: 4.18.3
      - name: lvms-operator.v4.18.2
        version: 4.18.2
      - name: lvms-operator.v4.18.1
        version: 4.18.1
      - name: lvms-operator.v4.18.0
        version: 4.18.0
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-rhmp@sha256:1502ccda3c067d74689d0341008f05362bc23904c8e35b20e020fe0dca82aef7
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.connect.redhat.com/nec/sanstoragecsi-operator-bundle@sha256:1957586c13e3bf80bdc688f519be440ec1dcc4ffe63343f526018dcb93e21cea
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-key-manager-operator@sha256:7a998ea17ef537ec5e998e1627b065ae04ab01b7988196da1cdd1763e05a9924
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:6ef2956b297795a9d0c83119e71309549975df3ae5652e6a98cef035ace5d51a
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:8831faee6979acfb5bd031b64cbc09fb90e9ea4f48bbff15e9d39ccdf6bf3d95
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        - quay.io/openshift-community-operators/prometheus-exporter-operator@sha256:e62276f4c7e583d1fb24997fa970662f1a8d510de0baea159ba796b48cff709b
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        - quay.io/community-operator-pipeline-prod/ack-dynamodb-controller@sha256:547bdd3d1037dbc3b189f4c2f8e1d49379e1d547cd9063391d6c745e1a103b80
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:59ff4966f4f653e49d42d97bfed6351d3730128db74f666ee4ecacff11790baf
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:01f2ca529d2486f113bcefc9fedce6a6fd07bcb48aaf534394b5b04c353f8853
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.connect.redhat.com/cpopen/ds-operator@sha256:6ec2387e45d9557cea2df749831e136120945794773ed0d48e0b04f0b3d9e115
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:ca70ebce43d451d040aedcec036413f986e620234fd9a56d45983805396db0b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:4aec6aa2d2b9a24167bb6bc92db7c2c323847b7a2f3ad7f998c96bf1d2903366
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        - registry.connect.redhat.com/netapp/netapp-supported-apache-spark-on-openshift-operator@sha256:4f15a0fc375156fafdf2d504f9980a8eb4451808b5554357877345cacf8f3594
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-athena-controller@sha256:8cff9258583f8acbdd26d263a61ffa11792cf9f4c1d65410bb4c71f9f7aaf178
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - quay.io/openshift-community-operators/devopsinabox@sha256:1c2ad76102151a56ecd753d9b86fd9f155d64b912067738d1858d33a02557928
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        - quay.io/community-operator-pipeline-prod/activemq-artemis-operator@sha256:1e3c6ff2d2612c3e85d40b1974b5a3c4f603b77353752389fa588728bd4263c4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator-bundle@sha256:df0dcd31867bb1c3c4321db4412a744fd1b3ba8faddb509d191fab92407b6677
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        - quay.io/community-operator-pipeline-prod/apicurio-registry@sha256:7087ec1377c2612edf3b8a76f6e115d66c3ec2b4b20f5d8374638d6bf0037e93
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-pipes-controller@sha256:e46b540219ca42c3b6c08ff99afabf09daea2dca5a335cda6013343df2f35878
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        - registry.redhat.io/openshift-serverless-1/logic-operator-bundle@sha256:5fff2717f7b08df2c90a2be7bfb36c27e13be188d23546497ed9ce266f1c03f4
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:53d8cf8bf5c1103860ffc4110e2a422371761e76ee79edfef8680e947407c5ce
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        version: 1.3.2
      deprecation:
        message: '''builds-1.3'' has reached the end of maintenance support. Please
          upgrade to ''builds-1.5''.'
      entries:
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:aadba00eadb3ac46d321a14498fc3dffd5be9afbf698d80adddde0a7a1681fc4
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        - registry.connect.redhat.com/smilecdr/smilecdr@sha256:c44f5661a598aa18c4c10372918a45edb0f26d7d2772c47a55c66e4ff874cb5f
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - registry.connect.redhat.com/mariadb/mariadb-enterprise-operator-bundle@sha256:73e7f243a6676d559068bf8ab80cc0ffe0db58152675603ca8bd360334439f26
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
          createdAt: "2025-09-09T14:14:54Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:ca10a842900d6515f714a17916a32dfd5a2909543e4d339acefc7e3e41b8d7ac
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:31d145c7074ad6cd85684b5191a3033533bf2149124c7ead5367871da5ed0dd8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        version: 4.17.13-rhodf
      entries:
      - name: ocs-client-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-client-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-client-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-client-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-client-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-client-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-client-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-client-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-client-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-client-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-client-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-client-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-client-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-client-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-client-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
          createdAt: "2025-09-11T15:23:25Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:60a91d6e4fe7e0ab999eab2224ef09fd0681e74920b23b306c7a9893bbb29cb8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:725eabb38876b93ab7693d75013d0292756a2efc00b8608f5716f31f8c324a53
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        version: 4.18.11-rhodf
      entries:
      - name: ocs-client-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-client-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-client-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-client-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-client-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-client-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-client-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-client-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-client-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-client-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-client-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-client-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-sso-bundle@sha256:f2c5a8abd1a9ec7665edf1e6206a6212cb752082b09532e4c218bc0990e48287
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:50b7f0a5f3fdd98cfb0294d664cc01ffd1e5359be66a3da14f5552e9cb3bae2c
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:29d46db99d97fe28758c0efd2acf27203ce9958734b82cf37c3a9063b5159506
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:b83c3263e77904ce1c0f95b7bfab969605c97297a5598fd794d64daf508c52c0
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:ea88a3090291f3742ccd4641785c19a80f4d3bd485a4f9c362a7645f73896db3
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:3dbb71004362a70683a5ee9f0916bca75c0d1c351d144677fd9d85bcb2ca8bf7
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:89d480ae76480829660d4cccfede74717faeb7e62a4a48b932750e41a945783d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a71b6fe33997db924669253992436e1e02d88769bb97376c0871848c2d330fed
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:1074f45a625df6ea412a9ffef3d4f4e53f3107ca45bb1de9bb030c47ce6518d9
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:0543e615d8a48ba6aaf90a8de0b0a79d37689c64bfe71f91c340455d9a359426
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a93fdee9e8d05ebe65e9dd8682fa7bc99c54f9d9f26d2cc8519f529255805da8
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:9d825d79fffd39a1149b262d48d08616795e5cfdb4c4bc40f5f192242f71042d
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: nexus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        - quay.io/openshift-community-operators/nexus-operator@sha256:277b94985c0dc982abb76b5dd8c93b37305ccca32c000d67552d540d4918cbba
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kinesis-controller@sha256:de760dbe730e38cd9fadebe5fdb69299c04512e6a70f567ce678fc0bedc53e9c
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/jaeger@sha256:907dfe1d01f73c8a385da3661ddbe03db31e0b9ad6f5ab2336460402d0d915a0
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchservice-controller@sha256:699f483cf96822576e3b79b690f256e2dbf4ca8d4426daf4298c980aefdeba4f
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/percona-xtradb-cluster-operator@sha256:d0c9d03130774d961f62032f684c494313787354186d028e9709e2463f69431c
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:a5645cb55c3e6c22655c6c8606b9bd51a49db44aecc3fe52e0eb80cca83d17c4
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:19e133c41b8f947aa83afba0417d8db3570cc2e1368813996fa5d135fdca4e04
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        - registry.connect.redhat.com/ibm/ibm-application-gateway-operator-bundle@sha256:e4ece7113fa7056980507408ac29d60676cf51c20d71a058a1accf765a6ea7e3
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-commonui-operator@sha256:87788afbc24851deb2c9c3f14b773bab174b2a2f02ac2fa0dd656a97433ffca2
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dns-operator@sha256:d5eadbf006aade69ac18156c891614da899d0779ba9e8fbaf755a256c46f055c
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/openshift-community-operators/gitops-primer@sha256:97af48469275f1b3a669f8e24a97a80ce9897fc5b12643a4f298b82dbe26e20b
        - quay.io/openshift/origin-oauth-proxy:4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso-operators/watcher-operator-bundle@sha256:01ab69b9c33a91cb194ad644bb64884b358af78467e2bb766ef373e4c36e795d
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4478b8a78bdde23e0d40476f8ad9e4bf0628fba7603a390e8de7e6a319abc1ec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:6c2ab763239f6e5a2a72890120bf1f20f6277850cec59a7e11cea2b6a06f12e8
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:8f30d4c7c491e9a804a8abf80570d1e1924cd0d787c6abb727331541c435d5ff
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:2b6b1a0b56cfb348ed4035b13d31501ff5757c101ed21027b330a7852576b891
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4c4c98b6cecd51c6644eb79b85789b7ebbcdc662edbb66ce435ea3b5e58d9273
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:b8aa74c5b2635a6d3413eeac6fc0f6eb9a2e459089aaabbf626ea3ef92a67ce0
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:73d91324ccc995e9a6db525f194bee340c17dd624ce520c767ea8fcf2346f786
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:d29abe14abc67d8836cc6f7656556a0a7fa8f4a8eea3f78e40265b339e7430f5
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:f7c4ed28a549fe7414312bd22c2ae140f1f28d30656fb9fb0c8ade05a8f82088
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c59956d70416fab0c6d7b3ecd8c9cfa3cf9415f75cfdfe2214f44b0b4aced594
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:7cace73b03bb02202a04b6026eb9f7323610fd0ccd5ba18a6d00fd66577a96b4
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c173a99026499700c08092f8418c79f39527d9b914690b199b41ed781c279de1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        - registry.connect.redhat.com/loadcore/keysight-wap-logging-bundle@sha256:2aeeefa7493eed8960004904567894244418982209e1b3a40488b759463edcdb
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.102
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-09-03T09:23:19Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.102
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator-bundle@sha256:fdef8f42c5425b9607915439daf3e5e2b77705cd220a083a00b42ee2e122d894
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:80c6d45a2ca70f55809620a82c309abb1e8272a0515f9c1a2905b3302c1467a1
        version: 0.1.102
      entries:
      - name: hibernation-operator.v0.1.102
        version: 0.1.102
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/lib-bucket-provisioner@sha256:132ecc88db79fb8b117a7a7974ca8f704600bac489923d918b78fd00378d5a5f
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-operator-bundle@sha256:aeae57b2fc706b782a3e3905e5c1825fe831f96f5b1c9478e1f2b217307a892b
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataDirect Networks
      provider-url: https://www.ddn.com
    name: vmstore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vmstore-csi-driver-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.tintri.com/v1alpha1",
                "kind": "VmstoreCsi",
                "metadata": {
                  "name": "vmstorecsi-sample"
                },
                "spec": {
                  "config": {
                    "debug": true,
                    "vmstore_map": {
                      "vmstore2": {
                        "dataIp": "\u003cdataIp\u003e",
                        "managementAddress": "\u003cmanagementAddress\u003e",
                        "mountPoint": "/tintri-csi",
                        "password": "tintri99",
                        "username": "admin",
                        "vmstoreFS": "/tintri/csi"
                      }
                    }
                  },
                  "csi_snapshotter": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                    "resources": null,
                    "tag": "v6.3.2",
                    "timeout": "120m"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "tintri/vmstore-openshift-file-driver",
                    "tag": "v1.1.0"
                  },
                  "priorityClassName": "system-cluster-critical",
                  "provisioner": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                    "resources": null,
                    "tag": "v4.0.1",
                    "timeout": "120m"
                  },
                  "registrar": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                    "resources": null,
                    "tag": "v2.10.1"
                  },
                  "resizer": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-resizer",
                    "resources": null,
                    "tag": "v1.12.0"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "512m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "64Mi"
                    }
                  },
                  "secretName": "vmstore-csi-file-driver-config",
                  "volumeNamePrefix": "pvc-vmstore"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-13T21:26:54Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: VmstoreCsi
            name: vmstorecsis.charts.tintri.com
            version: v1alpha1
        description: "A Container Storage Interface (CSI) Driver for Tintri Vmstore
          filesystems. This operator eases the installation\nand management of the
          Vmstore CSI Driver in an OpenShift or other Operator Lifecycle Manager enabled
          cluster.\n\n\n## Prerequisites\nMake sure that the top level folder where
          the CSI volumes will be created exists on VMStore and is shared over NFS.\nFor
          example `tintri/csi`.\n## Usage\n\n### Dynamically provisioned volumes\n\nFor
          dynamic volume provisioning, the administrator needs to set up a _StorageClass_
          in the PV yaml (examples/nginx-dynamic-volume.yaml for this example) pointing
          to the driver.\nFor dynamically provisioned volumes, Kubernetes generates
          volume name automatically (for example `pvc-vmstore-cfc67950-fe3c-11e8-a3ca-005056b857f8`).\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\n```\n\n#### Example\n\nRun
          Nginx pod with dynamically provisioned volume:\n\n```bash\noc apply -f examples/nginx-dynamic-volume.yaml\n\n#
          to delete this pod:\noc delete -f examples/nginx-dynamic-volume.yaml\n```\n\n###
          Static (pre-provisioned) volumes\n\nThe driver can use already existing
          Vmstore filesystem,\nin this case, _StorageClass_, _PersistentVolume_ and
          _PersistentVolumeClaim_ should be configured.\n\n#### _StorageClass_ configuration\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc-nginx-persistent\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\nallowedTopologies:\n-
          matchLabelExpressions:\n  - key: topology.vmstore.csi.tintri.com/zone\n
          \   values:\n    - zone-1\nmountOptions:                         # list
          of options for `mount -o ...` command\n#  - noatime                          #\n```\n\n####
          _PersistentVolume_ configuration\n\n```yaml\napiVersion: v1\nkind: PersistentVolume\nmetadata:\n
          \ name: vmstore-csi-file-driver-pv-nginx-persistent\n  labels:\n    name:
          vmstore-csi-file-driver-pv-nginx-persistent\nspec:\n  storageClassName:
          vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n    - ReadWriteOnce\n
          \ capacity:\n    storage: 1Gi\n  csi:\n    driver: vmstore.csi.tintri.com\n
          \   volumeHandle: vmstore1/172.30.230.128:/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img\n```\n\n####
          _PersistentVolumeClaim_ (pointing to created _PersistentVolume_)\n\n```yaml\napiVersion:
          v1\nkind: PersistentVolumeClaim\nmetadata:\n  name: vmstore-csi-file-driver-pvc-nginx-persistent\nspec:\n
          \ storageClassName: vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n
          \   - ReadWriteOnce\n  resources:\n    requests:\n      storage: 1Gi\n  selector:\n
          \   matchLabels:\n      # to create 1-1 relationship for pod - persistent
          volume use unique labels\n      name: vmstore-csi-file-driver-pv-nginx-persistent\n```\n\n###
          Topology configuration\nIn order to configure CSI driver with kubernetes
          topology, use the `zone` parameter in driver config or storageClass parameters.
          Example config file with zones:\n  ```bash\n  vmstore_map:\n    vmstore1:\n
          \     mountPoint: /tintri-csi                            # mountpoint on
          the host where the tintri fs will be mounted\n      dataIp: 172.30.200.37\n
          \     managementAddress: tafzq23.tintri.com\n      vmstoreFS: /tintri/fs01\n
          \     zone: zone-1\n      username: admin\n      password: password\n\n
          \   vmstore2:\n      mountPoint: /tintri-csi                            #
          mountpoint on the host where the tintri fs will be mounted\n      dataIp:
          10.136.40.71\n      managementAddress: tafbm05.tintri.com\n      vmstoreFS:
          /tintri/csi\n      zone: zone-2\n      username: admin\n      password:
          password\n  ```\n\nThis will assign volumes to be created on Vmstore that
          correspond with the zones requested by allowedTopologies values.\n\n\n###
          Defaults/Configuration/Parameter options\n\n| Default  | Config               |
          \ Parameter    |  Desc                             |\n|----------|----------------------|---------------|-----------------------------------|\n|
          \  -      | dataIp               |   dataIp      | Data address used to
          mount Vmstore filesystem |\n| -    | mountPoint       |    -      | Mount
          point inside of the driver container where Vmstore filesystem will be mounted|\n|
          \  -      | username             |      -        | Vmstore API username
          |\n|   -      | password             |      -        | Vmstore API password
          |\n|   -      | zone                 |      -        | Zone to match topology.vmstore.csi.tintri.com/zone
          \           |\n|   -      | vmstoreFS            |   vmstoreFS   | Path
          to filesystem on Vmstore |\n|   -      | managementAddress    |      -        |
          FQDN for Vmstore management operations |\n|   -      | defaultMountOptions
          \ |   mountOptions| Mount options |\n|   0777   | mountPointPermissions|
          \  mountPointPermissions | File permissions to be set on mount point |\n\n####
          Example\n\nRun nginx server using PersistentVolume.\n\n**Note:** Pre-configured
          volume file should exist on the Vmstore:\n`/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img`.\n\n```bash\noc
          apply -f nginx-persistent-volume.yaml\n\n# to delete this pod and volume:\noc
          delete -f nginx-persistent-volume.yaml\n```\n\n## Snapshots\nTo use CSI
          snapshots, the snapshot CRDs along with the csi-snapshotter must be installed.\n```bash\noc
          apply -f deploy/kubernetes/snapshots/\n```\n\nAfter that the snapshot class
          for VMStore CSI must be created\n\n```bash\noc apply -f examples/snapshot-class.yaml\n```\n\nNow
          a snapshot can be created, for example\n```bash\noc apply -f examples/snapshot-from-dynamic.yaml\n```\n\nWe
          can create a volume using the snapshot\n```bash\noc apply -f examples/nginx-from-snapshot.yaml\n```\n\n##
          Troubleshooting\n### Driver logs\nTo collect all driver related logs, you
          can use the `oc logs` command.\nAll in on command:\n```bash\nmkdir vmstore-csi-logs\nfor
          name in $(oc get pod -owide | grep vmstore | awk '{print $1}'); do oc logs
          $name --all-containers > vmstore-csi-logs/$name; done\n```\n\nTo get logs
          from all containers of a single pod \n```bash\noc logs <pod_name> --all-containers\n```\n\nLogs
          from a single container of a pod\n```bash\noc logs <pod_name> -c driver\n```\n\n####
          Driver secret\n```bash\noc get secret vmstore-csi-file-driver-config -o
          json | jq '.data | map_values(@base64d)' > vmstore-csi-logs/vmstore-csi-file-driver-config\n```\n\n####
          PV/PVC/Pod data\n```bash\noc get pvc > vmstore-csi-logs/pvcs\noc get pv
          > vmstore-csi-logs/pvs\noc get pod > vmstore-csi-logs/pods\n```\n\n####
          Extended info about a PV/PVC/Pod:\n```bash\noc describe pvc <pvc_name> >
          vmstore-csi-logs/<pvc_name>\noc describe pv <pv_name> > vmstore-csi-logs/<pv_name>\noc
          describe pod <pod_name> > vmstore-csi-logs/<pod_name>\n```"
        displayName: Vmstore CSI file driver
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - CSI
        - Vmstore
        - storage
        - Tintri
        - DDN
        links:
        - name: DDN vmstore Filesystem
          url: https://www.ddn.com/products/lustre-file-system-vmstore/
        maintainers:
        - email: akhodos@tintri.com
          name: Alexey Khodos
        - email: rrethinam@tintri.com
          name: Ranjith Rethinam
        maturity: alpha
        provider:
          name: DataDirect Networks
          url: https://www.ddn.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/ddn/vmstore-csi-file-driver-operator@sha256:64df1c504a9d70bd2b8b34520c8d431a92bb54702d601a57a2a989c6dd982462
        - tintri/vmstore-csi-driver-operator-controller@sha256:749ceff2bd46c71d81338510c1cc5866d39c66345f971d3093f986ea0417311d
        version: 1.1.0
      entries:
      - name: vmstore-csi-driver-operator.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vmstore-csi-driver-operator
    provider:
      name: DataDirect Networks
      url: https://www.ddn.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:042623e1faacc9fc22222a1c3e546a561abf7fc05d41ae9c915a69795f6bec7e
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:67b2b8783a04460019057cf826a74c18747c8c64ee7fe8ce83c8d0b7cd87cb36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/resource-locker-operator@sha256:9cc08639be2af4b72a9b816420d87dfa290a168fb83edc947a04b3e06cdb675b
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator-bundle@sha256:e284d0839795f763ca53402c2b050fa9ffedebccd0f3832fe7259368c3ad32f2
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-s3-controller@sha256:058a37a9a2fe6a42bbdddabb3fc7495450f1ebbd7f46194866c1491c40ce1a9f
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ClusterPulse
      provider-url: ""
    name: clusterpulse
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterpulse.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.clusterpulse.io/v1alpha1",
                "kind": "ClusterPulse",
                "metadata": {
                  "name": "clusterpulse-prod"
                },
                "spec": {
                  "api": {
                    "enabled": true,
                    "imagePullPolicy": "Always",
                    "logging": {
                      "json": true,
                      "level": "INFO"
                    },
                    "oauth": {
                      "enabled": true,
                      "headerEmail": "X-Forwarded-Email",
                      "headerUser": "X-Forwarded-User"
                    },
                    "rbac": {
                      "auditEnabled": true,
                      "cacheTTL": 0,
                      "defaultDeny": true,
                      "enabled": true,
                      "groupSyncInterval": 300
                    },
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "security": {
                      "allowedOrigins": [
                        "*"
                      ],
                      "sessionExpireMinutes": 1440,
                      "trustedHosts": [
                        "*"
                      ]
                    },
                    "server": {
                      "host": "0.0.0.0",
                      "port": 8080,
                      "reload": false,
                      "workers": 1
                    }
                  },
                  "clusterEngine": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "timing": {
                      "cacheTTL": "600",
                      "fullScanInterval": "300",
                      "reconciliationInterval": "30"
                    }
                  },
                  "frontend": {
                    "config": {
                      "features": {
                        "devTools": false,
                        "mockData": false,
                        "nodeDetails": true,
                        "operatorDetails": true,
                        "registryStatus": true
                      },
                      "refreshInterval": 30000,
                      "theme": {
                        "defaultTheme": "auto"
                      }
                    },
                    "enabled": true,
                    "oauth": {
                      "enabled": true,
                      "image": "quay.io/openshift/origin-oauth-proxy:4.17"
                    },
                    "replicas": 3,
                    "route": {
                      "enabled": true,
                      "tls": {
                        "termination": "reencrypt"
                      }
                    }
                  },
                  "policyEngine": {
                    "batch": {
                      "cacheClearBatchSize": 500,
                      "redisPipelineBatchSize": 1000,
                      "redisScanBatchSize": 100
                    },
                    "cache": {
                      "groupCacheTTL": 300,
                      "policyCacheTTL": 300
                    },
                    "enabled": true,
                    "limits": {
                      "maxPoliciesPerUser": 100,
                      "policyValidationInterval": 300
                    },
                    "logging": {
                      "level": "INFO"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "enabled": true,
                    "volumePermissions": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "ClusterConnection",
                "metadata": {
                  "name": "example-cluster"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "cluster-credentials"
                  },
                  "displayName": "Example Production Cluster",
                  "endpoint": "https://api.example.openshift.com:6443",
                  "monitoring": {
                    "interval": 30,
                    "timeout": 10
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "MonitorAccessPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "access": {
                    "effect": "Allow",
                    "enabled": true
                  },
                  "identity": {
                    "priority": 100,
                    "subjects": {
                      "groups": [
                        "default-group"
                      ],
                      "users": [
                        "default-user"
                      ]
                    }
                  },
                  "operations": {
                    "audit": {
                      "logAccess": false,
                      "requireReason": false
                    }
                  },
                  "scope": {
                    "clusters": {
                      "default": "none",
                      "rules": [
                        {
                          "permissions": {
                            "view": true,
                            "viewAuditInfo": true,
                            "viewCosts": true,
                            "viewMetadata": true,
                            "viewMetrics": true,
                            "viewSecrets": true,
                            "viewSensitive": true
                          },
                          "resources": {
                            "namespaces": {
                              "visibility": "all"
                            },
                            "nodes": {
                              "filters": {
                                "hideMasters": false
                              },
                              "visibility": "all"
                            },
                            "operators": {
                              "visibility": "none"
                            },
                            "pods": {
                              "visibility": "all"
                            }
                          },
                          "selector": {
                            "matchPattern": ".*"
                          }
                        }
                      ]
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "RegistryConnection",
                "metadata": {
                  "name": "dockerhub",
                  "namespace": "clusterpulse"
                },
                "spec": {
                  "displayName": "Docker Hub",
                  "endpoint": "https://registry-1.docker.io",
                  "healthCheckPaths": [
                    "/v2/"
                  ],
                  "insecure": false,
                  "labels": {
                    "environment": "production",
                    "provider": "dockerhub"
                  },
                  "monitoring": {
                    "checkCatalog": false,
                    "interval": 60,
                    "maxCatalogEntries": 100,
                    "timeout": 10
                  },
                  "skipTLSVerify": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: docker.io/clusterpulse/clusterpulse-operator:0.1.0
          createdAt: "2025-09-26T21:09:27Z"
          description: Provides OpenShift cluster health monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/ClusterPulse/operator
          support: None
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a connection to a monitored Kubernetes/OpenShift
              cluster
            displayName: Cluster Connection
            kind: ClusterConnection
            name: clusterconnections.clusterpulse.io
            version: v1alpha1
          - description: Manages the deployment and configuration of ClusterPulse
              monitoring platform
            displayName: ClusterPulse Instance
            kind: ClusterPulse
            name: clusterpulses.charts.clusterpulse.io
            version: v1alpha1
          - description: Defines fine-grained access control policies for cluster
              monitoring
            displayName: Monitor Access Policy
            kind: MonitorAccessPolicy
            name: monitoraccesspolicies.clusterpulse.io
            version: v1alpha1
          - description: Defines a connection to a container registry for monitoring
            displayName: Registry Connection
            kind: RegistryConnection
            name: registryconnections.clusterpulse.io
            version: v1alpha1
        description: |
          ## About ClusterPulse
          ClusterPulse is a comprehensive multi-cluster monitoring and management platform designed for OpenShift environments. It provides centralized visibility, policy-based access control, and real-time monitoring across multiple clusters through a unified dashboard. For more information, visit [https://github.com/ClusterPulse/operator](https://github.com/ClusterPulse/operator).

          ## Features and Capabilities
          - **Multi-Cluster Monitoring**: Connect and monitor multiple OpenShift/Kubernetes clusters from a single interface
          - **Policy-Based Access Control**: Fine-grained access policies using MonitorAccessPolicy CRDs
          - **Real-Time Metrics**: Live cluster health, node status, and resource utilization tracking
          - **Registry Monitoring**: Monitor connected container registries
          - **OAuth Integration**: Seamless OpenShift OAuth integration for authentication
          - **Redis-Backed Storage**: Scalable data storage using Redis for metrics and state
        displayName: ClusterPulse
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - clusterpulse
        links:
        - name: ClusterPulse
          url: https://github.com/ClusterPulse/operator
        maintainers:
        - email: asmoore@redhat.com
          name: Asa Moore
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: ClusterPulse
        relatedImages:
        - docker.io/clusterpulse/clusterpulse-operator:0.1.0
        - quay.io/community-operator-pipeline-prod/clusterpulse:0.1.0
        version: 0.1.0
      entries:
      - name: clusterpulse.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: clusterpulse
    provider:
      name: ClusterPulse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:61685ecb519d17160149a5dc16ae28d6fd96264652a6f19fef861addfa4ec6f4
        - quay.io/rasen708/ncn-operator:v0.2.2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:a99a5ca46c5bd3065905588f4d0e3cdc6409d435abdc02793be88e77c1a8f0f5
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:7d2ad8e800339c15efa5d5954106e01151d34135b25e9a23e30f0027ee23f193
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:eaa77ae0f1a1f3d32e269e73eb29a73d11012d5e920af1011d1842bd904d7293
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:4d9d003842e3ffc386f4e07ce7a81cfa85e676031fb6fdc96aefa9654244974a
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:a7641fc5f4c7aed4b73d7d98ccb58ef64cf14ab16d72378f56dba3ad4b8714c6
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:7dffd170b9b7d021b1615e6d746cbdfcbd819a1306ecdbaf042bcaf2202f7daa
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:111d1f4ab8f006386102d1bec4cd6a41b446d9ae401a43c821b2f318c8ba74c8
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:d7a4f9b8e5eb9bf6d11e00898eb168e86636f0633959eeb2b36b2d6e5e505f71
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle-rhmp@sha256:1f3d005224eada1b71d04f21d07a7c4fde159725c7a44d967d996152c802bede
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle-marketplace@sha256:7864f9f1cf3ad1d551b698b2e803302d207704b448b15f6c0013cbd9e854593c
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
          createdAt: "2025-07-16T09:03:58Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <2.8.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk@sha256:99aec4a74ea4c9910b50cc035b4c0c85ed4711a4c9b456028690d0f0c7af2657
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/splunk/splunk-operator-bundle@sha256:433bec6334c4771e140a24ea8753e7bcbf358c4504c942658fd3364915ac86ac
        - splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
        version: 2.8.1
      entries:
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
          createdAt: "2025-09-19T19:24:03Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-mq-controller@sha256:702ffd931cb4baafb1eac49f9fda5588d8b02f3b8c696e477dd626b0efad9f54
        version: 1.1.1
      entries:
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ssm-controller@sha256:922fd6daa660d2250bad8a3c54733cceae845624a14651cb20e049e29bf90e1c
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        - quay.io/openshift-community-operators/community-trivy-operator@sha256:e3652e68e1e569d7f40432fd8cbb7bfdc6738a116ecc1011c71a36f7ca5663c6
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:3b538de0bbaa43d2368402ea1c0eaf767c8b9e5a28e112a309148a630dd3c052
        - registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:5b61484c6a2bbc377435093bd9abb31df980ece470325cae92f3106963b6482c
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:af8ccc7a075f11bf6f31310830704e3a27b23587094f56e518b1f264c92b0fe4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: kubernetes-nmstate-operator.4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: kubernetes-nmstate-operator.4.18.0-202508271223
        version: 4.18.0-202508271223
      - name: kubernetes-nmstate-operator.4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: kubernetes-nmstate-operator.4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: kubernetes-nmstate-operator.4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: kubernetes-nmstate-operator.4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: kubernetes-nmstate-operator.4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: kubernetes-nmstate-operator.4.18.0-202505280403
        version: 4.18.0-202505280403
      - name: kubernetes-nmstate-operator.4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: kubernetes-nmstate-operator.4.18.0-202505061007
        version: 4.18.0-202505061007
      - name: kubernetes-nmstate-operator.4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: kubernetes-nmstate-operator.4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: kubernetes-nmstate-operator.4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: kubernetes-nmstate-operator.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: kubernetes-nmstate-operator.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: kubernetes-nmstate-operator.4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: kubernetes-nmstate-operator.4.18.0-202503190933
        version: 4.18.0-202503190933
      - name: kubernetes-nmstate-operator.4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: kubernetes-nmstate-operator.4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: kubernetes-nmstate-operator.4.18.0-202502250802
        version: 4.18.0-202502250802
      - name: kubernetes-nmstate-operator.4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/cluster-observability-operator-bundle@sha256:8c27ac0dfae9e507601dc0a33ea19c8f757e744350a41f41b39c1cb8d60867b2
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        - registry.connect.redhat.com/loadcore/keysight-wap-diagnostics-bundle@sha256:4e8400a38a04c61dc20ece1ee7dac1ad8a56ac309621f3225e7389e84e20c870
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator-bundle@sha256:85e75ec797f70c5807ec1ea25208cf0bb1990daa05eca9fbc82bc38aa4a840ab
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        - quay.io/community-operator-pipeline-prod/percona-postgresql-operator@sha256:44e687b38a48f68bbbb919e6df0d3b87b75492aca7bd41e508a2aa0e80d024a9
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        - quay.io/community-operator-pipeline-prod/ack-elasticache-controller@sha256:a489c815e9d3777d775ccd65bbd0ecd5fe7c577de36c504b63cf1620531d4ccb
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:301c3b793d8068f9e50c855196da91b6c07111ef500b02a91fbb354c9fc1316b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:8edd2f89d7adda6c1a5aae7414f50e0d786ec8443c1db13834ea4924009b6419
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:227159c5a93966a43fa3a84cb69536ff0051cc261da068561241261d3837998a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:60635156d6b4e54529195af3bdd07329dcbe6239757db86e536dbe561aa77247
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:e4d49cf6a2fb8b3585ecfe3b3da3e9de090fb99de3dee2da0601b1f6b071173c
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:095e93f8dd31a0985d6fef6a6168da456630febe89ab9e0cb9480170f1036541
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:524d315efa226f97ff26b8668f9297e755c6f0c815b8f3385b317667553f0e62
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:f5f025c7f56643fab9148bb5de8ee34a84d02b78a6abb4b5592306cf54296654
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:857c1bdff46d1a1d5abb51f75a80b7944bb461687b179b6f1622103947b8c3b1
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:d101e2b9ab4d8ca7c87c86ea4e3630dcc82a9c5a78885179c8b16773c8aac587
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:ecdd5c5fc609128e0af751f87b797b71e4ab0200a9abaaca835ff9a3f293fb10
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:be099e23851cb4b6d15e9c44c4da8fb6cba74aab005597d305757d1dae8b0122
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:96119f61731cec7eb037c3a4a7c2c8bfdf9ef11fd5b08abc25e3d24ddf342b06
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:2ea267aac67763b4f160c4e6876c0226d2fa135ee00614814a59b4de1d91f4c4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:218905f7902e6ac80366856d6c5ae37cd6d167c81e9521df7c3c3d6619f6c425
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:09fe022370abfcefcf3abfe8b6debc2f201a0ac8b8dacaf1278b8a6bee2b5273
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:5b865231214649649a2a9fb06934f931503f68ed740e150be1c7c8fcbb6f1317
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:16245de1e88228c231ab2edb4e9380753eff982bb72febadd00cdaad286a8eb3
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:491e7aea221eb0e1d2992310e53d50db287cbf32f62d090bd842236b603d38bb
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:7de09e5394985cdec87f5afab65b3343b8a37c0eb7f778774b21933ae69d691c
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle-marketplace@sha256:28f388286bc4b4981182ff1ab9e927f59744dcb755392907fe65cd3b96a44519
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        - registry.connect.redhat.com/loadcore/keysight-rest-api-browser-v1-bundle@sha256:89ef7cf0dc8b27e58a63a82a701894a7cc4d0643bf0c8c8e20c9bb59d1a1b1bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-bundle@sha256:80ca5ad4634305808ac5a3e8ff656ce6ff69b3c16dccc5e03f42763fa89bc762
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/community-operator-pipeline-prod/openshift-node-upgrade-mutex-operator@sha256:5ac79c9193934440464a595434fe4241e038505d32650843afcad5f5bc2ce8a7
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        - quay.io/openshift-community-operators/mercury-operator@sha256:0c253fd002cb389db40b2438d4d1c9e953a84125dbf0f4b6f5e3a6efd87cf3a6
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:9db455e082454734d2a17811dcfec53d3d60c28df6e7fe8bc979a45b28a35061
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:8650430afd813fd749fcd80ad5f6fd0aeaab3df048e8fe34e6aa72a07ab801fc
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:ed063f507fa8e2f378d0adbde823c110484d0212d124cc72d67efa987e86444e
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.8.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:68b310ed3cfd65db893ba015ef1d5442365201c0ced006c1915e90edb99933ea
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:bcc5ed9560667c7347f3044c7fa6148e3867c17e93e83595a62794339a5a6b0d
        version: 0.8.1
      entries:
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/3scale-apicast-operator-bundle@sha256:29fe669a56a69adb537798c91b9359fa3b051da5ffa9f27789a2067a3a72330a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.7.9-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:ea2229948fea841343fcaa337bb7e2854acf5f563d3ffb0af8bda1c21ec2a91f
        - registry.redhat.io/3scale-mas/apicast-operator-bundle@sha256:d3a608b37aa089ea46e803696ba4957585b875f8a810f517e891f385ad105937
        - registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
        version: 0.7.9-mas
      entries:
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        - registry.connect.redhat.com/loadcore/keysight-wap-storage-bundle@sha256:0955dbda03bd7307364a1a458ab8d6ac5a8c9feb6150706dcede9676a6513b0d
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:591fc4b5867dff26df1952653423b33889c0f56ba3f756057694d4c0e396eeb0
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:3225c49cf30a146784533496ce478466034d4549ff9721752958bac342fdec77
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:b0fa6b0356d83626509289b33dfa669bd7d32542aba986a18d296b85ce63389a
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:fe759127885502dc66423530ff03e37973355c5566a1c27e988b59d71528dd2b
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:c0336e5079f255ab6cd38e4e2e61d0cb4c6a35a976528fb44409bfde57e70dde
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/openshift-qiskit-operator@sha256:633ce4190b02034d70b52d7e624fc9d417ef997607f9710a36e73bbc0963fe4d
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a77d39ded03271f2deb8b438abd4328a1846ee0d4886cce14d869ba61f84d600
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a6f3a3a8aa52de4441ef3249a76dcff1b6b6defba417a9fdbba37eb458adee04
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:6208f64c7fe75f39ab7511e2b480c493d2ecc8d1e76a1ac498a93a6a24e8beb5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:11d64a921a98b7ec1a2126a03e446a055a5ae1705dc2005bb8b73cbbcf7ecf99
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        - quay.io/openshift-community-operators/hedvig-operator@sha256:81d956c12057f18bf920044a4e39c27c99b2c23afdccc0ff9468f0484e1961d7
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-operator-bundle@sha256:f40e7955ac649c383134ab6e77bdd521b75b21ecd9b476436405d0674c7fc855
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-operator-bundle@sha256:d04033c2cd2d236f6319a3e35880b89c8f0c60d53cd32f6f097d07d6d27e5f2c
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-ati-updates-bundle@sha256:3d06a58b2a277a67339c22f3fe10c1daac1c8eabaac8540b01177f6f5688784c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatchlogs-controller@sha256:94145cae2b8cef1410acbb4f1c1eb571e71408341716af1742f27ddf4dc350b6
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/coherence-operator:3.5.4
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
        - quay.io/community-operator-pipeline-prod/ndb-operator@sha256:5f33c2a8181ea8e91844afd1f4042affbd9eb878266b6b2bcc6906f0829ed00b
        version: 1.6.0
      entries:
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - quay.io/openshift-community-operators/keda@sha256:50816f675c76d36dbc62846579150477018ef87df325861751f3fc02a70c8781
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/community-operator-pipeline-prod/alloydb-omni-operator@sha256:08fafb811f709780c0ccb3c22e3c6417cb14a70a9305e46a7b50457647a82102
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:ed530db9b4a29fb591ea208cd0c6b3b83b49537ff5a0a4f501b54ce667f9b303
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:31bdff225bd825d02ce0f468b1edc77541ce85b7b9111f14a583c418c984232f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:80198f65f7e3f9022a34e0342e9423807b855dcb7bda810172eebc9a27afb5c2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:f04d565eb5162fd5a9c78cd8443af6a3b184a660277537f1d6d575caa2e85441
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:773592dfb9964970f444871a8812e8b74db3f205faaa0bf148defa1a2b87b649
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:0b1771f697a700263c5b827dbde56a5c9bab610e9d7df55ef93885a4e54cc0e9
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8441322a0fce02df407573675b5b5d92d56de97c8aec72541b33d2dadc4050d5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:be93be0341649215fb001725aac0c3c5925343adbf1ad00c979b8c3d489512d6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b22190de3162d3e03481cae8904ce563912756eea7ac7e7feec0acbbdae5e6f6
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:c2f39be75173b4cc57498c5845b14c464bf9af92b49716df9661c8135ac2cf7a
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b2d58f4dc20eb7e17e09af0e8c5de55fa04f9f4f697e10c6be81a7a6a08381b2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:27efa0ab1d9c69047a924bc1dcd2b56108f5f390a6da56afb6ab7a37cb4578e2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/volume-expander-operator@sha256:b775df99f891e2cee9c2169e9c948a8d61d65b4cc90f36752b46a56e2d4be602
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/apimatic-kubernetes-operator@sha256:a5f68205291caa9bb7903180afc974fe86252ba87e671978c8b49e0206976d06
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: sonar-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        - quay.io/openshift-community-operators/sonar-operator@sha256:5e7a87f52955162f090c1d7fc69f6d0917444b553355a11c061ab60f289a17d7
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    defaultChannel: alpha
    packageName: sonar-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-documentdb-controller@sha256:8a991c76f281ec30258b6317bd623894b78a4f39d9790a1cbb32e6f1b7ff63f5
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        - quay.io/community-operator-pipeline-prod/dynatrace-operator@sha256:c475f9cc0f6caa2bb316ff42a35a40b1d4bf8ef7c8a3384d9c78eb205bda7076
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        - quay.io/community-operator-pipeline-prod/kaoto-operator@sha256:4b6cc43eba8efbe17ffec746431d6cf6b045886ff865673a08a0820abecebdcd
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:5ea2dd102074798d1ee005dc03fbbd24f4324df831f279355dc2dfc262c9d880
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:e4dec1f39f8e24043cf16703efcac2cc3b953d3f4c37b3956fce0c65ef829764
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:ed8230e90f0ab0586a743d5e760a7469345b1136971aad17ef675c1beccf1982
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:08995782514e496c24d888352e9d1e6b788dd87adb761e0c2e745be2975ffcaf
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:7b324e897fbcc881ececa529822d7b738f82cd4c3adf4cc54b66080b888f5d9c
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:e2a329fc74ed5f17e395367126c7e96197178e8f886d475b657a7910a4a14152
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:0218d68f97e89b48bbdd9c7cd750d4a666cf33af1cb8a76507c067a332a7b6bd
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:2bc56e149c14920662b549a5367f98daf9afad0df29b4746b195f6afc0c9c597
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
          createdAt: "2025-09-19T22:01:50Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
        - quay.io/community-operator-pipeline-prod/ack-eks-controller@sha256:120566a02e487aaa8823323423fccfaa377135037e2e1cb9f3ff14313cf3bfeb
        version: 1.9.2
      entries:
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        - quay.io/community-operator-pipeline-prod/pelorus-operator@sha256:f54c42a30675897a971cac278d5d8ea24edf3121570264c07cde04f11cc4e7a8
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:57af9e32c31dad769bda449ebe35e09ea135e8a7d06cddb42be4d1a09005f518
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ab3cb65b88c34fd58524a2a51cd510c611924e5860834965bad0b4418a8b6dd3
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
        - registry.redhat.io/openshift4/ose-local-storage-operator-bundle@sha256:d303fc63f5b849f073e34a7791ea92cdc3cff269cd1b593e27480ae96cac1a15
        version: 4.18.0-202509100344
      entries:
      - name: local-storage-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: local-storage-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: local-storage-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: local-storage-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: local-storage-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: local-storage-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: local-storage-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: local-storage-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: local-storage-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: local-storage-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: local-storage-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: local-storage-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: local-storage-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: local-storage-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: local-storage-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: local-storage-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
          createdAt: "2024-09-04T17:53:42Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:07cafb83ba90378c1fa60338cf951a9d7145ec245710637e90fdb508875dcc45
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:1d86075469312e84f3d16379efcd6130ec9e8b23a1343e3945f071d82beff6e3
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:dc10544e1bb24883f38373a0445944745282e5a69ed1744f428f98871a7a39c9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:31b3658dd79c18111c21e99aea52cc198f8e95b458a1a4f8d2bfc4d21e9bdacb
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:1b8008a16a0f53c09ecc5a219964194689a1aeb01acee8313aa49a9c26259429
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:a539064234088a4ee13b8a9666ca9507208fe38189ee9b8b673a38a58aa75599
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:b137e1345038c9e566767f24cce60c03ec278f5a2efc28fa08e0bd8bfa016486
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:c001dc96a1478a58f30461ab018e40bede13a78f3cdce869a70dcb8b9bba22b1
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:92a3162c705537147005a065cdddbb410e1863d80c08d8b0f27b7ff5588500cd
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:8fbfffb0bc3adfb4d72961dad381e1daba414bf4a7084f03b5208f6be4191ceb
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:035855c3d4b32a29232846d73dc398e81f941fd90a03e33241ca06f6e80b2204
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:f86e96edb0418811df328debdee93b4a51cf7f651134ee35e9c50b9dbfc29fd8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:084f1e54ef0d7f457d5a7cfc57b567ad8d85237c12f707c82f7a30a54e3652ba
        version: 1.7.17
      entries:
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:9f5479ac795eb50fa5ea597828b2d50ddaa4878faa1a07b3de0e611cb4af5fc6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/operator-bundle@sha256:bc8a47d5f8ffe3a2fe48819f08c82525677131cfb4f3c0facd82470c0f10acee
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        - registry.connect.redhat.com/loadcore/keysight-nimbusmosaic-dusim-bundle@sha256:97d5f906ef83272a8d7a2a249f38f8b9b186f8e9b64729a049a7eac0dbda8b1a
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        - registry.connect.redhat.com/loadcore/keysight-wap-appsec-bundle@sha256:e5049593a52fb26c124de7005b45919f26fa0430293f18283a20c8cde981b025
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cli-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager/cli-manager-rhel9-operator@sha256:be0fc230fef111cff36a8b846bbac9aaa98aae94ad9200416e841aa487121de4
        - registry.redhat.io/cli-manager/cli-manager-rhel9@sha256:a6b7857c1acd5173c770e6ba1d03947eb13804d2512f7f7cd6bb96e5e5cdaa17
        - registry.redhat.io/cli-manager/cli-manager-operator-bundle@sha256:74c88f1cfef4cc31b794fc5ed8dcc10d3b0e6869c228d808f1d98819e69554b6
        version: 0.1.1
      entries:
      - name: cli-manager-operator.v0.1.1
        version: 0.1.1
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:a0e4c432590e65818e7ebaab127d483d68fdc5ccae9a13820665a8aef122fde7
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:242d754d5fe1f2a07091ae93da284b8912e0f4b4c973200cf6d1241e6abb88a9
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:437a786a301ea720cdb5651dff878a9f51470bab46ba28daae4928c7074c9157
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
          createdAt: "2024-11-21T23:48:37Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVSNetwork is the Schema for the ovsnetworks API
            displayName: OVSNetwork
            kind: OVSNetwork
            name: ovsnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-operator-bundle@sha256:4c8271d0bf6d94061fe8df8110b041cb0e1a840fb6263315f552166fc5053e73
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:2e71d4125a35f6e7bd2a04cdd8cb6aae058d745eda372ae3d3fd37783778f486
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:57a57f4f3a0adb8b7234683f2968352988ec9a414824f4f444cca096825a5df5
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:45835aba95af72aa8e73b630d1d2b4c1079217531f025f4715a2eb8e8856205f
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:8b412541a797aaff1b49aa229e310957a8d7da03a3d38e6bbfd65f2ca36c2bdd
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:6eb244a301ae5395b69b1bfcd0baac37c094f4c38a1a330c0a70fb81e71f6157
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:3b87d49b200ac4fdeac97f0e97a46001e91c8f4680e2975e90bd94ec37864403
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:0f18db6a5bbc952bee8f3e4ac9e93e807d1e85cd132affe94b1f4005c1e128ce
        - registry.redhat.io/openshift4/ose-sriov-rdma-cni-rhel9@sha256:65864bc235689ff38b63d2a7b2c72874ee155241a493fa9429fbe87d5100cfcf
        version: 4.18.0-202509100344
      entries:
      - name: sriov-network-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: sriov-network-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: sriov-network-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: sriov-network-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: sriov-network-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: sriov-network-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: sriov-network-operator.v4.18.0-202505260733
        version: 4.18.0-202505260733
      - name: sriov-network-operator.v4.18.0-202505121333
        version: 4.18.0-202505121333
      - name: sriov-network-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: sriov-network-operator.v4.18.0-202504221235
        version: 4.18.0-202504221235
      - name: sriov-network-operator.v4.18.0-202504160002
        version: 4.18.0-202504160002
      - name: sriov-network-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: sriov-network-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: sriov-network-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: sriov-network-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: sriov-network-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: sriov-network-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle-marketplace@sha256:4145f7a04d7dbdda226b75af4e4cabca63a803f290b1581ea4ae0dc223eca7df
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        - registry.connect.redhat.com/questlabs/alloy-operator@sha256:fe4e83fc4a50e1aa04e6a9ba2a57acc737077b86681c1d7c597646fad337f221
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - registry.connect.redhat.com/kove/kove-operator@sha256:04bf970760c4dfe910465b0ef6a18e9c17fa70f6388131f12ae0aa0cd646f663
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vertical-pod-autoscaler.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "autoscaling.k8s.io/v1",
                "kind": "VerticalPodAutoscaler",
                "metadata": {
                  "name": "myapp-vpa"
                },
                "spec": {
                  "resourcePolicy": {
                    "containerPolicies": [
                      {
                        "containerName": "*",
                        "controlledResources": [
                          "cpu",
                          "memory"
                        ],
                        "maxAllowed": {
                          "cpu": 1,
                          "memory": "500Mi"
                        },
                        "minAllowed": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    ]
                  },
                  "targetRef": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "myapp-deployment"
                  }
                }
              },
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "minReplicas": 2,
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "safetyMarginFraction": 0.15
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
          createdAt: "2024-12-05T23:41:17Z"
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1beta2
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1beta2
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: Vertical Pod Autoscaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - verticalpodautoscaler
        - vpa
        - autoscaling
        - scaling
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-operator-bundle@sha256:f5df5d878ef0a0b494c9bafe99c35f3152d30f6c426be789fcd636dd13eff6e6
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:5da084c4dc81251a707db1bd9347c13b39a1e01660120aa2e6a47f9d1d981fe9
        version: 4.18.0-202509100344
      entries:
      - name: vertical-pod-autoscaler.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: vertical-pod-autoscaler.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: vertical-pod-autoscaler.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: vertical-pod-autoscaler.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: vertical-pod-autoscaler.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: vertical-pod-autoscaler.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: vertical-pod-autoscaler.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: vertical-pod-autoscaler.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: vertical-pod-autoscaler.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: vertical-pod-autoscaler.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: vertical-pod-autoscaler.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: vertical-pod-autoscaler.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: vertical-pod-autoscaler.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: vertical-pod-autoscaler.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: vertical-pod-autoscaler.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator-metadata@sha256:d6d0e71f5451e8255edfff87ef28ba7743fe7057efd53b25bbbd01bf9e1d0307
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/lws-operator
    name: leader-worker-set
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: leader-worker-set.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "LeaderWorkerSetOperator",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-04-14T08:12:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lws-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/lws-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LeaderWorkerSetOperator
            name: leaderworkersetoperators.operator.openshift.io
            version: v1
        description: "Leader Worker Set Operator is based on the [LWS](https://lws.sigs.k8s.io/docs/)
          open source project.\nLeader Worker Set: An API for deploying a group of
          pods as a unit of replication. \nIt aims to address common deployment patterns
          of AI/ML inference workloads, \nespecially multi-host inference workloads
          where the LLM will be sharded and run across multiple devices on multiple
          nodes.\n"
        displayName: Leader Worker Set Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - leader-worker-set
        links:
        - name: Leader Worker Set
          url: https://lws.sigs.k8s.io/docs/overview/
        - name: Operand Source Code
          url: https://github.com/openshift/kubernetes-sigs-lws
        - name: Operator Source Code
          url: https://github.com/openshift/lws-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/lws-operator
        relatedImages:
        - registry.redhat.io/leader-worker-set/lws-rhel9@sha256:affb303b1173c273231bb50fef07310b0e220d2f08bfc0aa5912d0825e3e0d4f
        - registry.redhat.io/leader-worker-set/lws-operator-bundle@sha256:f641cf3b2d6659e26eb0ae49c67a3416152a2fbc2f26f490889a2bd169b9ea06
        - registry.redhat.io/leader-worker-set/lws-rhel9-operator@sha256:c202bfa15626262ff22682b64ac57539d28dd35f5960c490f5afea75cef34309
        version: 1.0.0
      entries:
      - name: leader-worker-set.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: leader-worker-set
    provider:
      name: Red Hat
      url: https://github.com/openshift/lws-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:72a54c9457e311001b25c1b414b7001c6d7f9a34b89efb1b4198084586386f31
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:7a04dfd687dfe1646bd59a39c9e88781ffb815085bbd15f6305f6eac5f59bc49
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:d1a9246def81537e3417a6d48d1e818459583db658b54a87ca5af6911c8a3f53
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:e88676c0346c245b315f1a3bf7f85b10515bae3620c786fedda0533b39d2a400
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:66bd3eba0b68c317fe128c6713488cb6374f9647095768d3f7eed8120bf02e77
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        - quay.io/openshift-community-operators/authorino-operator@sha256:8fe34f94f87c567401cc2bbbbcc3a3e719364b80a2b17851496f86c85af5c09a
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino:v0.20.0
        - quay.io/community-operator-pipeline-prod/authorino-operator@sha256:92ad93f27c34f1406fc05a248d04709b0109adf37f2f9424f48fc6a73d9d359e
        - quay.io/kuadrant/authorino-operator:v0.16.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/logging-operator@sha256:d3ce3322f7f25c7b8bfabdd6e9fd32b833239fc39592240e0ccff3a13757fb17
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:f252d6f89a7cdcc520c3af461ca57d0dfabf736b9c38277d77477c4c880e182c
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:97d03647567ddb6f04edea39244bc0814e8f7e557251060d9b241a5fabe04dfc
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:f3497acc3216d820014cce7332e9d0683f4ca6438782ace02210192f787ca097
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:23f2f66e9f6d79d05e2be5a10c394834f093f4677ab06d8935ed33980408e8d3
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:c6d6c3e0408b851cdc9589cc0ee642d77397cdfaa40f50cf42557c5aaf9c0690
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:ff314ab983699455c32ab0ef27bb552d00825974a2d7088df2000f09afd8a2e2
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:5f10cfc1bc8571f44a99d88755adade0743106219673bbdd2e4d3a04ee714778
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:9f90cd7ab8c0ab5da6eb12f7f0dbd177df62f07c38fb759a97989899cfcbe75d
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:71ff93b137b6569f8fdefbd446cf12cd95477e4294bdea98fb5318e5d7d1cdd6
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle@sha256:2a28b64f7321df23928c8226b3eb3161cf520d38cc57d875b3cbfc828c164edc
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        - quay.io/community-operator-pipeline-prod/hazelcast-platform-operator@sha256:1209876fc9103eb2d69c34e97e4e2e0b384a9a778ae7e6a53222b8552fbf2f26
        version: 5.13.0
      entries:
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kms-controller@sha256:a2c60484a6556949318aa1d730c601b5bb05b1795b945a607209ef3434cce72f
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - k8s.gcr.io/redis:v1
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - quay.io/openshift-community-operators/ripsaw@sha256:ff8d796d5bd6c8cb0e155b3ced631122397758eb472470223e30352eeddba480
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:68cdc37990bc7da8a6bbe3a4908378c8026e5ff980bf8c44c0277f689006fc51
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:97762fcd8bfeec64fe4d3d43a6e60bc705ff2c8773354357e545adf31c8f85c5
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:4fa7a30eb3f679a495aeb96794056ffd003237686b2b60e693c8cabf6bc15bae
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6dce519730a6eb73bc18bef8372cdde65c29af3cb52eb78afc808d8582526fd3
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:9f2c01e88b29154ba7a0f70a1c69a49c86a567808cf057cb5d115159278cf2cf
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:edec37e045212f74d160724ffc7b2c8ccf4e60eb95629aae80360494d78d5433
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:465f0bef1f167b1d16ba9ae5f2d9a1133d11786b6809698a84c5c962387dbf6f
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8be098242f971aefa61a4a24d2c6630ecaa7856ece842166c38887a66bce1bda
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6afbc7ffc98918e3a557b0b3d9b2c3aa80caa47b9a73205fac9e6572c66f4df7
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:eefdd05485167d490f0011d9e5d2bb6112db1daea25940c79e1f55b3af254548
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:05e66583757da895fd23c3af9830ae1ac6a84d107107e5b6bbaece57e4d79c9a
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:02f53d3395b2b98f92a96b466c83ca15c79a75d82451e2ed8ba921375e727c7f
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8abe0994052e2a328e54c9d30c920dc331e9b816a4606234a3023b9ef65ec4ed
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:ef3f85e1227e886a131d3ddec3fb431aef6d3daed46115b288246b9ace0e69e2
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8b9b06ead5e014b920c82034ded615413dca53cf36f804005408726c9bbda66f
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:51db956f1472ae0498c57b9488448c34cf7e24afab6b9997eb3689b001d73277
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:392f9299c75a3ca4f593eaebdfe515145696f12db9f990bb004cfa58dd6bf0ef
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:df4e4dce45f87e684e7e1044a7edb2456160141afb6ba57b9af28088ffbac407
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:fecad1a330421341e6946252f95735ed177a690eab6250be097413ea29ce9473
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:9dcf24fecf3ee0d999601c276ac15e8634c275590f2c42b8a6bcb83984bf1445
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a9950264bfefc536b4e0a0adc2348f92aa41da55dd50a2324ddad0253291b67c
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:4d59e60cc75de82bd3f3a063237cef8a5598d315f6d53e803d37589a94929b0e
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:40d17d4d0632c272779da152f710fb7c2de50acc3791eb0b9fe445a13c0cb102
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:8e321dc1bf74ba331b62d21bff4d8993956ab935cf5be32f01e7f7aa6897bf22
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:16fbdaaf303ab40202d49438737f3307926594a5259081eeeb39595b7ea41f8e
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-zero-trust-workload-identity-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zero-trust-workload-identity-manager.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "ZeroTrustWorkloadIdentityManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/part-of": "zero-trust-workload-identity-manager"
                  },
                  "name": "zerotrustworkloadidentitymanager-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
          createdAt: 2025-06-16T11:15:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: zero-trust-workload-identity-manager
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/zero-trust-workload-identity-manager
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterFederatedTrustDomain
            name: clusterfederatedtrustdomains.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterSPIFFEID
            name: clusterspiffeids.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterStaticEntry
            name: clusterstaticentries.spire.spiffe.io
            version: v1alpha1
          - kind: SpiffeCSIDriver
            name: spiffecsidrivers.operator.openshift.io
            version: v1alpha1
          - kind: SpireAgent
            name: spireagents.operator.openshift.io
            version: v1alpha1
          - kind: SpireOIDCDiscoveryProvider
            name: spireoidcdiscoveryproviders.operator.openshift.io
            version: v1alpha1
          - kind: SpireServer
            name: spireservers.operator.openshift.io
            version: v1alpha1
          - kind: ZeroTrustWorkloadIdentityManager
            name: zerotrustworkloadidentitymanagers.operator.openshift.io
            version: v1alpha1
        description: The Zero Trust Workload Identity Manager is an Openshift Day-2
          Operator designed to automate the setup and management of SPIFFE/SPIRE components
          (like spire-server, spire-agent, spiffe-csi-driver, and oidc-discovery-provider)
          within OpenShift clusters. It enables zero-trust security by dynamically
          issuing and rotating workload identities, enforcing strict identity-based
          authentication across workloads. This manager abstracts complex SPIRE configurations,
          streamlining workload onboarding with secure identities and improving the
          cluster's overall security posture.
        displayName: Zero Trust Workload Identity Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zero-trust-workload-identity-manager
        - ztwim
        - spire
        - spiffe-spire
        - security
        links:
        - name: Documentation
          url: https://github.com/openshift/zero-trust-workload-identity-manager/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-agent-rhel9@sha256:0281d6ab233a685ba0f739629584aacd9254e0b13a697fe8708efa41289e746e
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-operator-bundle@sha256:1974903a3e2b692eedab2602395fdc4a6e2ffbc79ed4e633705c05ce7771e1b7
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-helper-rhel9@sha256:a39c80c6e1495ea166a858c8ed03304c6455420c077d937c99bf8732a6829790
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-controller-manager-rhel9@sha256:67f0e6670195c84515e7d78139a2951e463db11445740fe1d9a449a7f338556c
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-oidc-discovery-provider-rhel9@sha256:5ba7e94b7ea06852d32002b055df556ff2a2fdc9a2e8f793b837c4856d773782
        - registry.access.redhat.com/ubi9@sha256:861e833044a903f689ecfa404424494a7e387ab39cf7949c54843285d13a9774
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:7d170ba9b3de2e88dc0d0a6b0c13630ca3b23eed96ea5e3c725516046d7c7fc7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-csi-driver-rhel9@sha256:a16ac0a4363f7cffada39a2f17aa42f5de3faf78b3688ae5ebf2ecf3a4759cb7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-server-rhel9@sha256:70f97c613feeaadfed99d37aabd62e1cb2247825292ad9be806bc2fd065c9166
        version: 0.1.0
      entries:
      - name: zero-trust-workload-identity-manager.v0.1.0
        version: 0.1.0
      name: tech-preview-v0.1
    defaultChannel: tech-preview-v0.1
    packageName: openshift-zero-trust-workload-identity-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-operator-bundle-rhel8@sha256:0fbbb0b32adadc5746251103d25bba6bd365b2ff60e289e4694061fc92c3ce6e
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        - registry.connect.redhat.com/paloalto/pcc-operator@sha256:c7200a8c652ee544b0e0bb4d9427eab3a3b2dee6ca61ee88abb8052316382673
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/sap-hana-express-operator@sha256:0fa4f29bd7669690be87a08405bed95d3ead2293a125b28149de2b10b02d2cd6
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/community-operator-pipeline-prod/falcon-operator@sha256:3916adf5083c83838564b1ed039decbadb2efa4ec288699ef0137a66983a3cc4
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.connect.redhat.com/hitachi/hnas-csi-operator-bundle@sha256:79e6ba3db78aa55a9f3839a172843d284d91e9cdaae9cb2c8e7e705f3d03115d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:60b293c23587072b3376f83afee95b0207228c5c7b80483b025adda572c3551d
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        - quay.io/community-operator-pipeline-prod/ack-rds-controller@sha256:80b0d6e33fc94471d05e05f6d370be3a84fcb5c2044486e007d57360eb69d9d7
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-messaging-topology-operator@sha256:4c62fb6d686a20a28437905d4e1a9b93ac05853b4dbf29fcce34421a43268724
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:b10c121395f5bc46d61fdeddd35a0c8d9ae4e67a9a77d8ada90e0ea53666323e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle-marketplace@sha256:9d84d46dd1eb74afa18466c73766402b46c2885ef483600aff83065f8065d8d1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        - registry.redhat.io/jboss-eap/eap-operator-bundle@sha256:0db70367917668964e2f4f5efdcef864ccc3f065848023e52caf4abdec728d9d
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:096946dd896b6d9d937eb6529c3a23a84ccb129649c5a0389d0ab83989c28b53
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/jhipster-online-operator@sha256:5f2436297c026832e5c15241d60090da581daa5502e352ea33d77f22ef3e7a52
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        - quay.io/community-operator-pipeline-prod/argocd-operator@sha256:483854621b95dcefa03d18a970600b271726422ae3362479f73dcfca9085c92d
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/mtr/mtr-operator-bundle@sha256:190a6fa2a6a9b6b182cdf9cab18814d05c203b54ca18dc533e553fa5dcca779e
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:2808a0397495982b4ea0001ede078803a043d5c9b0285662b08044fe4c11f243
        version: 1.0.2
      deprecation:
        message: |
          The preview-1.0 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:a91cee14f47824ce49759628d06bf4e48276e67dae00b50123d3233d78531720
        version: 1.1.0
      deprecation:
        message: |
          The preview-1.1 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:4477810692429ac498b6ec1b295e6ead0f8fea25e91f6ddf17255d7e349b82e1
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:ff2bb666c2696fed365df55de78141a02e372044647b8031e6d06e7583478af4
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        version: 1.2.2
      deprecation:
        message: |
          The stable-1.2 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:8da62ba1c19c905bc1b87a6233ead475b047a766dc2acb7569149ac5cfe7f0f1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        version: 1.3.3
      deprecation:
        message: |
          The stable-1.3 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-load-core-bundle@sha256:f62a7f8eb74d0d5309fede6c9c0201d1f0c3736f93f1aeec7446089d64fce28f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/openshift-community-operators/sigstore-helm-operator@sha256:ff4e548a8a003bd23e5076362bafb2e22f77b8445c5e3585d99cd5ce26fcaf56
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:5a5e2f775760aa68c85d0f69b700d23ef2a6be7bb415bd96401d03eb6077f90c
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:79a9191f0cd5d11af004ead87ba9eb5359c67fd658c346db4c3cea3b16b23987
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:54ea11add0a1d988e0c965968f9a65e011978190f76fa2f826f1378474bac1d4
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
          createdAt: "2024-10-11T14:05:47Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8337924489a1747d4bbca6dd6ac42210731cf721b549b5830846b271621e2ac1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:6d0aeffbccaf9df83b42c530bf82707aa7c1cc16f72a5ef30b66fbd562c01ab3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8c05a303be5c536f4051ef291475b578d4d38ca0946fb4b4dd8bca376ef432aa
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:e91c2e5473f3e24c185dd25e9843f333e6d7419d9c5f4d4396669717b5969683
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:c543dbbb66fcac02345e1d3c12333780721bf049dca0f0a27d4c55b15a5d3e14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:a65780c5680a6fbbd0e3d0eed423b14a9c9b3d7d17f801c1eb6eb6b777384ca5
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fd6fb83bbeb9010f7e91f0afc91beac0bf38cb79c2de8f533a2bd07b1b238262
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:936fd68d7659683de7789fcf34d2e5873bd4bd1534c01c2d80501647f7c3c442
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1dd8f0f08ff1323d910fea8adfa13cef64c2fd2f01968e4d9d55f71cf42835d3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:365115407964d182de7e7564438c6fd089e2c0bcc0f94ec48621ca0cc2bf0fe4
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4bea07e101740afc36a1617363a79b1c474efcacd584f6e475a53b3f2395a079
        version: 2.7.2
      entries:
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:3d0409c1a0b17c51d3dbd5aa0371edd597e3c5ba194aad987fe867dd804d1923
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    - currentCSV: mtv-operator.v2.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
          createdAt: "2025-09-09T18:06:51Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:41cc528bf93000095b4eb56aaff9a6dfb124d1711a2f31f0745163ff934ac96e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:0c8ebfdd52053dbcc221a3df7d603163a821dd9e0ac6f54977e081adc6cb4504
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a299acb18821691cddba52904140cef8bf12632fbf2e752b33118a574a5ff394
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:ad3b548405e8c4c075ef7d4c0eb0d0430684066ed7eab4563a1f8378eaf3d939
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4b37b1aecbbf087257be5ba2bb559f2de005aacb4cd371192cdde1a9891739cf
        - registry.redhat.io/migration-toolkit-virtualization/mtv-vsphere-xcopy-volume-populator-rhel9@sha256:c0a4a232ff8adfb7d57e0761251ad0d79a0f02d27c4bd813f62f6b943deb6585
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f90f4eaaaa02b289704b19dd171282f3d11f370ed6cea992e55964729ad5116e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:258efbf073f78f8b5b48fdda22ef055e2315046cfecc9644da51959289c836d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d57b39674260f3bb1ae4c164022c2c7be3cc9645f1c46a83e43f40965633451a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:aeaa9ffcb0ffbb0f69b2712a3031f7e2dfc0824273b5c9c0f6349bc701a0e89e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:e438c5fccc957ad26ecfc3633561342d3b6f2a6d5a4c7839109564008de62222
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:83a523392a9c715972bba7044c7c3d3944594cc89778449e1f59fbd6a2659baf
        version: 2.9.4
      entries:
      - name: mtv-operator.v2.9.4
        version: 2.9.4
      - name: mtv-operator.v2.9.3
        version: 2.9.3
      - name: mtv-operator.v2.9.2
        version: 2.9.2
      - name: mtv-operator.v2.9.1
        version: 2.9.1
      - name: mtv-operator.v2.9.0
        version: 2.9.0
      name: release-v2.9
    defaultChannel: release-v2.9
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:0ee0bd2449f183277af845e312257b4a1ced4de865831b15e1758a397d8d29eb
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:2cf058c819b40512094f84f4d374457455370ec4994d34cee6e39b9d87841f05
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        - quay.io/openshift-community-operators/camel-karavan-operator@sha256:6dc78ecdf6e1936872626c02d41986d6553c1500002d85692c97e86fa77c129d
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:f182078b0077b6df6d47ddcab0204e963c1475486bd61b8eccb73b46eb8b2995
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c8b3505ed491c1bf6a913f161715c8ac483d8d9adf1fd29916ea4edad987cd67
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:af5a13351f960c0938c68604b772f24743c0e997701838a7784c53406471e72a
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:0d9f5293d69eebd91fffd83ae371141e640cab8f9e2d559a83714ce0f94fee08
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:d6675617c19b4ac7c1913078eaf9ffbe3d03761b570dffa002d9de11f337f19f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:2a35a0ef852cb4d76e963df50449712a06ff13391c30675aaf931e21cb561b8a
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:00f7ce0148988e93679d4b53b45ae8d2ad54f308b908940e7b28504318f718c6
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:c352f5633beef2eaaa324dda66b9b484e5ea79efbbc3b781079f2e9d10c880a9
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-bundle@sha256:47a7bafff2715fbef6130ca0b84560b6c2358275821b72ae54ad46020e02bee5
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hive-operator:1.2.4953-c248a51
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:be43412a2e4bad2cd9d15e4fbabe0292585c5bbbd65da12d519f30f17d2fe594
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:822fd8e65c241cebedd835bfebf67645199561cfdd6148d5afffb8b35e499ae3
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        - quay.io/openshift-community-operators/hive-operator@sha256:9a3cb682afc66ad662045b3d8ff214203cc00be60f66f74564820f6dcf99cd69
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        - registry.connect.redhat.com/fortanix/ccm-node-agent-dcap-operator-bundle@sha256:545724c273dfd6a1fa923d468b2d054f39ec3186d01586a8b3fe494c0f22f8dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/ndmspc-operator@sha256:b1f47ce0cd6930d902dc3538a1a87140dbe9257dd50773a518fc2350185f90cc
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        - quay.io/community-operator-pipeline-prod/klusterlet@sha256:9efef9a034c9f12505a0d81bc9697c4a3ab7e3d05d1e7aaa4180490d70ddf6a2
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:30264b9f49912cc0e6bed59001869e3168caace20b292c9dad4cc2d4947d289d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:d7cc6d2575eb252fb3e2468106e801f192da7844b0b153dc3c6a12cf7cc0a8fe
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:60e99bb91309e448c652898dac070a6d920fda164f27879d94632c88a9e61ffa
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:11edce56874ff1ee666752ec5fa521f3f6ae78c52cf1f289ceb97a86e6c083aa
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:5e4c319ce871a0dcbab7fff23325e6233c9dbcb2592fdeb1a9ba7c57ac2a63b2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:d86bb21fdeec35b1fc9789023a558750da76e036e0790dcfce97dcd137dcb6fb
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:7d94521a5cbdd34f8307653a3e1b596ec86415b2165fabdbabc8ed8bed35473d
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:3d5ad015a953708cc53ac2ea92b1b1d99c870a801b26e790d41458adc0a958e5
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:fea5b977249013181616f60b95c5e6d224cd0f248717171d946ff0e4d6eed6c0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-memorydb-controller@sha256:7f2e65d817479fcc5658cac745cef325b2bfd6256eeed8b819f1713afd59fe68
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patch-operator@sha256:437f5b7c661d3e9bc2c8ac13b0c8f3565f685b9e5ef52791312bc64d55e2e476
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:21f4648e0e71b33ef24b4921ab34b9e2b5e90eac25ade735d51331ac4f7c4cf3
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:2ed6bfdf37e36c63cde3633443553dee410201200ad351e6d91a5651f8e20ecc
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - quay.io/helmoperators/cockroachdb:v5.0.4
        - quay.io/openshift-community-operators/cockroachdb@sha256:f42337e7b85a46d83c94694638e2312e10ca16a03542399a65ba783c94a32b63
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        - quay.io/openshift-community-operators/cockroachdb@sha256:d3016b1507515fc7712f9c47fd9082baf9ccb070aaab58ed0ef6e5abdedde8ba
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/netobserv-operator@sha256:73af4040707ce79baa3115deed1e9148436b248ef3570f1175a6d31fdf9596c4
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:394d122efcf0da16bb68bfef1292395cda1bdb21909ae55f8fcc54a16814f1cf
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:62f93c5989d1d8217a3bac076addac05c0906a1a1f1eeaacb87bcf256812fdf9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:ea8c060444aa8e89dfbd90d61d6ec3900ab4d6634f6aa8067face1e6974b9065
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        - registry.connect.redhat.com/loadcore/keysight-wap-notifications-bundle@sha256:9c2ebb98476bf3374b71a838c394017e7d31aaa045bd52b3a66924b6f6b7b067
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.connect.redhat.com/cpopen/ibm-licensing-operator@sha256:a9f68d204b6acb50ac5023e5f2314ad0d06e06483f87290381d1e92476a79579
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator-certified.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-operator-bundle@sha256:4512f786fed0d9b2285b5a2827ec3b597452f90c217e27398f2f0f552377d844
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:48b37460dd7245cf661009267c3353962f0f195efd6a3007c9cc70b51b91a9e3
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:9949a581df3960a770117b6cf6b9fee1c63ea12ca09ac2d677f1cbe0a83fbd93
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - quay.io/akaris/sosreport-operator:0.0.2
        - quay.io/openshift-community-operators/sosreport-operator@sha256:f0d28a3e2ac2e1cbd0c307797b728ad89ac33375135d72054a7b765e274e4bea
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/node-maintenance-operator@sha256:80dfaab6f7a5c34cebb45c874ce134bf60e3697e14a95b69f1972dd98ecc172a
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=4.11.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:75eaf7b8b71e1a15e002f6cc5430e5e50039699d7b52c3150f5f71b1595f00c5
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-operator-bundle@sha256:41ece1ea773fd610cf8a57483a4d2846e0bb3ce81db701021a3e0dd91e13b7c8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:f7159e880ec55da49ac774db0c454800055e7cd3f37ba94fbcb629d08aac1146
        version: 4.18.0-202509100344
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-operator-bundle@sha256:4f8908ba97517826fda14cab3624ed42a6348fb96f0be4abf9cd2a7b68f7d3e9
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-wap-migration-bundle@sha256:3deff8b48ab1a6f0e5fd0f8315bc1e191da55525cb3cf0f315f2f7afdb47a6c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - registry.connect.redhat.com/noiro/aci-operator-bundle@sha256:80e25268ab5aa6ea418deae5ccc197fbffe2b9cdf4f27894f5beeb9315f6b123
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:20a4f2c56cb186fff4e8d31021d84b690512cc65c3c9629ffe623a0213ca9b0a
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        - quay.io/community-operator-pipeline-prod/mariadb-operator@sha256:3cbbff7ae75fb4f4629c7b8b58ca85ce67b23d5da558241301266d010e8b1188
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acm-controller@sha256:4d63542b39f41c458dd1eec28e44edce3a919f4dd9137ff64a9452cac1a1f954
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - docker.io/absaoss/k8gb:v0.11.4
        - quay.io/openshift-community-operators/k8gb@sha256:9743f957b00e5ad344fed0c062d616d8222990139ac4afbf14871fd9813f0ef0
        - absaoss/k8s_crd:v0.0.11
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/ucs-ci-solutions-operator@sha256:e284253263183eb2cd313f6fb009bb830b9a79a4d4e2481468821c30a9cdce03
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-operator-bundle@sha256:67723df19ab7d7871a6b639da3975c230d95dd250fa2995a9f44bdbb4b141454
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:439aeb806c3264948a15e325f7b1e71037099a085c8d971646188ace77fcf5d3
        version: 4.18.0-202509100344
      entries:
      - name: secrets-store-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: secrets-store-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: secrets-store-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: secrets-store-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: secrets-store-csi-driver-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: secrets-store-csi-driver-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: secrets-store-csi-driver-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: secrets-store-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: secrets-store-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: secrets-store-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: secrets-store-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: secrets-store-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: secrets-store-csi-driver-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: secrets-store-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: secrets-store-csi-driver-operator.v4.18.0-202502280802
        version: 4.18.0-202502280802
      - name: secrets-store-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: secrets-store-csi-driver-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-09T14:17:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:c5bd4f10120088f285a3b65e85bba825f9f55964758901f3b5827ac31516675b
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:26935de5f21a7865aba484599136df6a251c009e8fa9781e1040da3b19e7dd33
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:e05074d46e062268461d2689f7048c1a5dc4ed89a039d93e382f5ebc0726a269
        version: 4.17.13-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-multicluster-orchestrator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-multicluster-orchestrator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-multicluster-orchestrator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-multicluster-orchestrator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-multicluster-orchestrator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-multicluster-orchestrator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-multicluster-orchestrator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-multicluster-orchestrator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-multicluster-orchestrator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-multicluster-orchestrator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-multicluster-orchestrator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-multicluster-orchestrator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-multicluster-orchestrator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-multicluster-orchestrator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-11T15:26:27Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:4b78c8df4f4aa4be805a442b059b5e2822464bb029826f2f0090cf8b6b747768
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:47c754d0e2782dd8d885758e68803173c32f1fe6319ac09c786dd24e5124df25
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:957805157e7690deedaae63856c9ce71e5c299610bca401524397a40e2820537
        version: 4.18.11-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-multicluster-orchestrator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-multicluster-orchestrator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-multicluster-orchestrator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-multicluster-orchestrator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-multicluster-orchestrator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-multicluster-orchestrator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-multicluster-orchestrator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-multicluster-orchestrator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-multicluster-orchestrator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-multicluster-orchestrator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-multicluster-orchestrator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:3b683338f5849b887e1bc81ea078a9b799d272f824abd4247ff87e70ca7acbf2
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle@sha256:333a6f476d03adcbfaa4ab38802859719a7dd8857783f8487c99b008d06815c5
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1a156d67fd5beb9ffddecab724490acb7143f8141ff3b392c6f519e76c3ad374
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:645fcb18684884aef8ca4d5299f0d053d496c3a5d3f37cadd66c51d38514a680
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:5a73ce3a8cc8422aeda37fd9ef97f5aecdbf8c69fb6ff488df26342e989b5c57
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1dfe9180a9fca4f773c1282dfd0006daa0bb48dfbac87e352661e051107f9033
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - registry.connect.redhat.com/a5f1f0cc7ac3db90370970f9b/tawon@sha256:3691c0948c8eb266ada348de504622227af6516f3b3d331eb09fd258dd0609a5
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    defaultChannel: alpha
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
          createdAt: "2025-09-19T13:48:41Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dapr-kubernetes-operator@sha256:8f2087df1987437f1c5e62aa9453fea220a41012077fec67b6ba8368ec267421
        - docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
        version: 0.0.11
      entries:
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-apigatewayv2-controller@sha256:1cf12140fcd67b0f3e8c5791a91c5b705bb3af8b2e03ea4ba2d4ce6f9f7a4809
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - registry.connect.redhat.com/mongodb/enterprise-operator-bundle@sha256:23c3c8387eb9377b52327d7193f59eeea5831f4050248b3cbb54cbcd9437fc71
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-websocket-service-bundle@sha256:da100dd9c5b542ba46de7166e783e28a0f7e675f9ac896f211405e5a9e2082ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2cc940711539edd93a23355dcb06365821d140346e1df248386a77667c0f47ea
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2418015d00846dd0d7a8aca11927f1e89b4d8d525e6ae936360e3e3b3bd9e22f
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "rawDeploymentServiceConfig": "Headed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.34.0
          createdAt: "2025-09-03T09:59:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Argo
          Workflows\n* Model Mesh - ModelMesh Serving is the Controller for managing
          ModelMesh, a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks\n* Feast - Feature
          Store is the fastest path to manage existing infrastructure to productionize
          analytic data for model training and online inference.\n* Llama Stack –
          unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opendatahub-operator:2.34.0
        - quay.io/opendatahub/opendatahub-operator:v2.34.0
        version: 2.34.0
      entries:
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        - quay.io/community-operator-pipeline-prod/opendatahub-operator@sha256:d3e3d5ae02b834522481fa9046bf2ad3ea06d706c5cb06a32d32e61e229e4a6c
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:2992841660e8b1664fa2ce06da894aa828d4ea1be3eb5b27476efdd73731486a
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:bf85a9da8a100a90deb3745497bc1505852910ef23dd12c624fa17de51bd8377
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:826354c9c06a06f998e522a88d6a81239e003e1d6a2633e9b7b27b8a61d5fe38
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-operator-bundle@sha256:a59435a22c2f470b149ff721240f14144d2e56200b141e59a4f1011267a6c925
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:1346dcee82f2120bd2af22be9f041e057bdc32f9ad1c3ef717d599ba6519faeb
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:44a5c4cc3086e4eac42dc78a72b7492d13df0098017288278d9006e767a480a1
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:b9dbbce235b7a4f45f42184cf0b908e91e17094e971bc29a8710f35127a6b06f
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:026418bae8b2d425d45b2b3481f65c241d89e25f455e3fbfcf2f79536d978bca
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        - registry.connect.redhat.com/nec/webotx-operator-bundle@sha256:7e655933a7fab3ac3aa88eb5e4ca9ac3a71b98875ef1ef5bc49e87284f160ff8
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-licensing-bundle@sha256:2d3abd0263183bc92718316581675b92073f37c3786784fae24b3587a9441def
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:329d6eaeff787bcde5db0ba4f9403eee533a601b45dc2d5d349b318705452591
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:2cac5fdb55664c08928a7d924eacb18cc4c03c609261e3b6d812590e4d166375
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:b35da8aa4e6746952e5cfdbbac72950a48f7ad32a3d762a470ce62118c78643a
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:e5a0cb8af7879bf13d940df278ccdb8a5c400e5d7108f23b4abef94c964e40d6
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:91aae0500cd3cfc9a4a536241feb4fe50288e2a68f69ab5c602c02169e5f5b2e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:40c7d896355dfe67ff3616f1e1df86548bfcd0524f7e56d4d72ab4f5f1739f6a
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:3f69a124f9c31ebc0c78f249872a69b1289622bf4f5aad4c739f6e8caf9cb85b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:33a2e21a069ff24af4bf0ffc51e83ddf398df5d9735d0123120768e9dc2c27f9
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/community-operator-pipeline-prod/gitwebhook-operator@sha256:2d33e4f4f4b607ee2006481efc7ad9341049085ef988777d23e586f7406a5109
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - quay.io/bpfman/bpfman-operator:v0.5.6
        - quay.io/community-operator-pipeline-prod/bpfman-operator@sha256:6b3603c7c1917c67b5aa57202d1d2fdc0d9ac6b3a9373afa7172f322b3b3fa9b
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:74775c7851ac98ba2903435c71b9d02cc504ca8ee98b65ad7282d671392756a6
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.connect.redhat.com/cpopen/ibm-elasticsearch-operator@sha256:0bedcf95b31e28d41b124780f3b97c68a87db5626ecf1d5e81bb06b8b3fcd91a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/federatorai-operator-bundle@sha256:d0ac978d749654ddf4cc20823805ba7d33ee62e9a7ec98fdd53d9926b697c347
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/community-operator-pipeline-prod/cert-utils-operator@sha256:92a0c0b893b56330593e0a255b8df61031e92a61d172aecc58be5805d847798a
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle-marketplace@sha256:19c9d44cb9a0c7f52501fcb589750f3e85e270d066423bcadac593dd79e48ff4
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        - quay.io/openshift-community-operators/github-arc-operator@sha256:d8d97d19cea466c4c240097b79bfb7de90312d937ad4db63f9e90901686e54da
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/self-node-remediation@sha256:549585ca18d331cfbd5911fa3e79912428a0be3308588be2438ece3247746bd8
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |2

          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler,
          **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: kaliu@redhat.com
          name: KaiYiLiu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-operator-bundle@sha256:5ebbc06088e99b937327cd457324461cafdda9649b2dd2db1bcb892be8556eed
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:141b069ff9925390607315b0cf675003abbdac9555b29700e3b999fadd6c89f2
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:971174692a1aae68ac07ffa132254eb3c018a52507eb32621b6dc286fd25ca0d
        version: 0.5.0
      entries:
      - name: power-monitoring-operator.v0.5.0
        version: 0.5.0
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.2-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
          createdAt: "2025-09-11T18:04:40Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.2'
          olm.substitutesFor: cryostat-operator.v4.0.2
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:b6a6e401042daf1780ae6be19aedc551ed176d1a13bf1f0f48359552c638a156
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:8fd30d86f5468227899560071a28e92197c5790ea820e918c47559f0695cfb8a
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:d2788c8732fec62c7a27e7bd1497fea1ab3ff605e7e913bbe17d8ef97567329b
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:550de94fb20462c5cc5c52166820319544e774b8d45fc7beed7a095f5c1300aa
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:bbee66a44c2e3a5525c2fa81db72e8528bccf40583d0bd665841644473308310
        - registry.redhat.io/ubi9/nginx-124@sha256:126903359331eb9f2c87950f952442db456750190956b9301d00acfcadbc19ed
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:67e313e50c6ef100d6938dc1a734728661f328ab1f9757667c7a06de11689967
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:d9720fefa3e34d05a5536fe1a66e23b30fe0845ae2b96b9c9be4ee705be9633b
        - registry.redhat.io/cryostat/cryostat-operator-bundle@sha256:455ecde240a152987fcfe0013abee992014510de9e89c014559a5d0665a56512
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:9ba156d802c9c3eafc948afef88e0eaca6b5e73bf32ba80ea7b0651ba7ce923d
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:2a5e1d2d39a7d9d2bc74316110e1a65e9e18f3a9fb7f0a498c7457dd9e6684ee
        version: 4.0.2+4
      entries:
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cognitoidentityprovider-controller@sha256:093670fcaee6ef20e471c0df45c9ea7f40e5f0c43208c9447623324fdb39dbe0
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
          createdAt: "2024-11-27T00:00:00Z"
          description: The deployment validation operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        - registry.redhat.io/dvo/deployment-validation-operator-bundle@sha256:c04bae05ee0c24fe5bf94257b0dd2cb020f5c9b5f173ceffffdd041b6dff4ee9
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.11
        version: 0.7.11
      - name: deployment-validation-operator.0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.0.7.5
        version: 0.7.5
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CephCSI Community
      provider-url: ""
    name: cephcsi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cephcsi-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:12:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:b023d95637934b16ec8342994edd4e3198579517b68817ce6b4cc14958746b9b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:10cfe59ee2d625ed67f3804ba7e471f10fa1e12bc4f96de5e2f453e5cdefcf89
        version: 4.17.13-rhodf
      entries:
      - name: cephcsi-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: cephcsi-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: cephcsi-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: cephcsi-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: cephcsi-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      name: stable-4.17
    - currentCSV: cephcsi-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:20:52Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:903e47fa88119b5507a0d773d58318236bbce53201b5e6a18a16311f704b1120
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:562703a8b197f03b29d0564df329ecfda192a94b7818c6fd5f919329a5cc9bd7
        version: 4.18.11-rhodf
      entries:
      - name: cephcsi-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: cephcsi-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: cephcsi-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: cephcsi-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: cephcsi-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: cephcsi-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: cephcsi-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: cephcsi-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: cephcsi-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: cephcsi-operator
    provider:
      name: CephCSI Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/openshift-community-operators/pulp-operator@sha256:1926e96b185443ff0c7a2861cefc2e3aded0c0493e0d7f7d78e45447e06750d4
        - quay.io/pulp/pulp-minimal:stable
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:1e75649538f13cf3bb2e6131160f038a70689d6f92a110bdcec348e865040c4a
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    - currentCSV: pulp-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 1
                  },
                  "custom_pulp_settings": "settings",
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "ingress_type": "nodeport",
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "example-pulp",
                  "postgres_configuration_secret": "example-pulp-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "example-pulp"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.1.0
          createdAt: "2025-09-18T18:00:01Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          metadata.annotations.support: Community
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://pulpproject.org/](https://pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://pulpproject.org/pulp-oci-images/docs/admin/reference/available-images/single-process-images/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://pulpproject.org/pulp_operator/](https://pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:92a3e95cb3b1e50100933ea08cb44074fc505f6c0e7fa24d028ec46fa9e95590
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.1.0
        - quay.io/pulp/pulp-web:stable
        version: 1.1.0
      entries:
      - name: pulp-operator.v1.1.0
        version: 1.1.0
      - name: pulp-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:fbf847c41e1736fb6efab7a6acdc03acc562fa7e667306773880b277883bc4e5
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-operator-bundle@sha256:7c808f6e4ea469d88703300bdcc6bdce160634833cbd10bd110dc69d05be588c
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
        version: 4.18.0-202509100344
      entries:
      - name: clusterresourceoverride-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: clusterresourceoverride-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: clusterresourceoverride-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: clusterresourceoverride-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: clusterresourceoverride-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: clusterresourceoverride-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: clusterresourceoverride-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: clusterresourceoverride-operator.v4.18.0-202505070803
        version: 4.18.0-202505070803
      - name: clusterresourceoverride-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: clusterresourceoverride-operator.v4.18.0-202504251302
        version: 4.18.0-202504251302
      - name: clusterresourceoverride-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: clusterresourceoverride-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: clusterresourceoverride-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: clusterresourceoverride-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: clusterresourceoverride-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: clusterresourceoverride-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: clusterresourceoverride-operator.v4.18.0-202501231202
        version: 4.18.0-202501231202
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator-metadata@sha256:a2f0513cd8b96d71855464384cad7f1cd8c39f8d445be1ff1c1b91f22008d16d
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-operator-bundle@sha256:c2229c2cfad9bbbca9179cb6617bce0033f5be40731641a69b41f9a90e034c8b
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        - registry.connect.redhat.com/hazelcast/hazelcast-platform-operator-bundle@sha256:0aabd3e7f5f993c1288450678d55e8f7b36c8b30894e5d844521a474aa6b2714
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudtrail-controller@sha256:20fe3c3ba7688906de40d7fd36c00552be280595708935ff800ad5228829f781
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
kind: List
metadata:
  resourceVersion: ""
apiVersion: v1
items:
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stackwatch
      provider-url: www.kubecost.com
    name: kubecost
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost.v2.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "global": {
                    "prometheus": {
                      "enabled": true,
                      "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                    },
                    "grafana": {
                      "enabled": false,
                      "domainName": "cost-analyzer-grafana.default.svc",
                      "scheme": "http",
                      "proxy": false
                    },
                    "gcpstore": {
                      "enabled": false
                    },
                    "gmp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8085/",
                      "gmpProxy": {
                        "enabled": false,
                        "image": "gke.gcr.io/prometheus-engine/frontend:v0.4.1-gke.0",
                        "imagePullPolicy": "Always",
                        "name": "gmp-proxy",
                        "port": 8085,
                        "projectId": "YOUR_PROJECT_ID"
                      }
                    },
                    "amp": {
                      "enabled": false,
                      "prometheusServerEndpoint": "http://localhost:8005/workspaces/<workspaceId>/",
                      "remoteWriteService": "https://aps-workspaces.us-west-2.amazonaws.com/workspaces/<workspaceId>/api/v1/remote_write",
                      "sigv4": {
                        "region": "us-west-2"
                      }
                    },
                    "mimirProxy": {
                      "enabled": false,
                      "name": "mimir-proxy",
                      "image": "nginxinc/nginx-unprivileged",
                      "port": 8085,
                      "mimirEndpoint": "$mimir_endpoint",
                      "orgIdentifier": "$your_tenant_ID"
                    },
                    "notifications": {
                      "alertmanager": {
                        "enabled": false,
                        "fqdn": "http://cost-analyzer-prometheus-server.default.svc"
                      }
                    },
                    "savedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Saved Report 0",
                          "window": "today",
                          "aggregateBy": "namespace",
                          "chartDisplay": "category",
                          "idle": "separate",
                          "rate": "cumulative",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 1",
                          "window": "month",
                          "aggregateBy": "controllerKind",
                          "chartDisplay": "category",
                          "idle": "share",
                          "rate": "monthly",
                          "accumulate": false,
                          "filters": [
                            {
                              "key": "namespace",
                              "operator": "!:",
                              "value": "kubecost"
                            }
                          ]
                        },
                        {
                          "title": "Example Saved Report 2",
                          "window": "2020-11-11T00:00:00Z,2020-12-09T23:59:59Z",
                          "aggregateBy": "service",
                          "chartDisplay": "category",
                          "idle": "hide",
                          "rate": "daily",
                          "accumulate": true,
                          "filters": []
                        }
                      ]
                    },
                    "assetReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Asset Report 0",
                          "window": "today",
                          "aggregateBy": "type",
                          "accumulate": false,
                          "filters": [
                            {
                              "property": "cluster",
                              "value": "cluster-one"
                            }
                          ]
                        }
                      ]
                    },
                    "advancedReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Example Advanced Report 0",
                          "window": "7d",
                          "aggregateBy": "namespace",
                          "filters": [
                            {
                              "key": "cluster",
                              "operator": ":",
                              "value": "dev"
                            }
                          ],
                          "cloudBreakdown": "service",
                          "cloudJoin": "label:kubernetes_namespace"
                        }
                      ]
                    },
                    "cloudCostReports": {
                      "enabled": false,
                      "reports": [
                        {
                          "title": "Cloud Cost Report 0",
                          "window": "today",
                          "aggregateBy": "service",
                          "accumulate": false
                        }
                      ]
                    },
                    "podAnnotations": {},
                    "additionalLabels": {},
                    "securityContext": {
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      },
                      "fsGroup": 1001,
                      "runAsGroup": 1001,
                      "runAsUser": 1001,
                      "fsGroupChangePolicy": "OnRootMismatch"
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "privileged": false,
                      "readOnlyRootFilesystem": true,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "platforms": {
                      "openshift": {
                        "enabled": true,
                        "route": {
                          "enabled": false,
                          "annotations": {}
                        },
                        "scc": {
                          "nodeExporter": false,
                          "networkCosts": false
                        },
                        "securityContext": {
                          "runAsNonRoot": true,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          }
                        }
                      },
                      "cicd": {
                        "enabled": true,
                        "skipSanityChecks": true
                      }
                    }
                  },
                  "upgrade": {
                    "toV2": false
                  },
                  "kubecostToken": null,
                  "pricingCsv": {
                    "enabled": false,
                    "location": {
                      "provider": "AWS",
                      "region": "us-east-1",
                      "URI": "s3://kc-csv-test/pricing_schema.csv",
                      "csvAccessCredentials": "pricing-schema-access-secret"
                    }
                  },
                  "saml": {
                    "enabled": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "oidc": {
                    "enabled": false,
                    "clientID": "",
                    "clientSecret": "",
                    "existingCustomSecret": {
                      "enabled": false,
                      "name": ""
                    },
                    "skipOnlineTokenValidation": false,
                    "rbac": {
                      "enabled": false
                    }
                  },
                  "systemProxy": {
                    "enabled": false,
                    "httpProxyUrl": "",
                    "httpsProxyUrl": "",
                    "noProxy": ""
                  },
                  "kubecostFrontend": {
                    "enabled": true,
                    "deployMethod": "singlepod",
                    "haReplicas": 2,
                    "image": "gcr.io/kubecost1/frontend",
                    "imagePullPolicy": "Always",
                    "resources": {
                      "requests": {
                        "cpu": "10m",
                        "memory": "55Mi"
                      }
                    },
                    "deploymentStrategy": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 1,
                      "periodSeconds": 5,
                      "failureThreshold": 6
                    },
                    "ipv6": {
                      "enabled": true
                    },
                    "useDefaultFqdn": false
                  },
                  "kubecostMetrics": {
                    "exporter": {
                      "enabled": false,
                      "port": 9005,
                      "prometheusScrape": true,
                      "resources": {},
                      "tolerations": [],
                      "affinity": {},
                      "service": {
                        "annotations": {}
                      },
                      "serviceMonitor": {
                        "enabled": false,
                        "additionalLabels": {},
                        "metricRelabelings": [],
                        "relabelings": []
                      },
                      "priorityClassName": "",
                      "additionalLabels": {},
                      "nodeSelector": {},
                      "extraArgs": []
                    }
                  },
                  "sigV4Proxy": {
                    "image": "public.ecr.aws/aws-observability/aws-sigv4-proxy:latest",
                    "imagePullPolicy": "Always",
                    "name": "aps",
                    "port": 8005,
                    "region": "us-west-2",
                    "host": "aps-workspaces.us-west-2.amazonaws.com",
                    "extraEnv": null,
                    "resources": {}
                  },
                  "kubecostModel": {
                    "image": "gcr.io/kubecost1/cost-model",
                    "imagePullPolicy": "Always",
                    "outOfClusterPromMetricsEnabled": false,
                    "warmCache": false,
                    "etl": true,
                    "etlFileStoreEnabled": true,
                    "etlDailyStoreDurationDays": 91,
                    "etlHourlyStoreDurationHours": 49,
                    "etlWeeklyStoreDurationWeeks": 53,
                    "etlReadOnlyMode": false,
                    "plugins": {
                      "enabled": false,
                      "install": {
                        "enabled": false,
                        "fullImageName": "curlimages/curl:latest",
                        "securityContext": {
                          "allowPrivilegeEscalation": false,
                          "seccompProfile": {
                            "type": "RuntimeDefault"
                          },
                          "capabilities": {
                            "drop": [
                              "ALL"
                            ]
                          },
                          "readOnlyRootFilesystem": true,
                          "runAsNonRoot": true,
                          "runAsUser": 1001
                        }
                      },
                      "folder": "/opt/opencost/plugin",
                      "enabledPlugins": [],
                      "configSecret": "kubecost-plugin-secret"
                    },
                    "cloudCost": {
                      "labelList": {
                        "IsIncludeList": false,
                        "labels": ""
                      },
                      "topNItems": 1000
                    },
                    "allocation": null,
                    "containerStatsEnabled": true,
                    "maxQueryConcurrency": 5,
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "55Mi"
                      }
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "extraArgs": [],
                    "ingress": {
                      "enabled": false,
                      "labels": null,
                      "annotations": null,
                      "paths": [
                        "/"
                      ],
                      "pathType": "ImplementationSpecific",
                      "hosts": [
                        "cost-analyzer-model.local"
                      ],
                      "tls": []
                    },
                    "utcOffset": "+00:00",
                    "extraPorts": []
                  },
                  "etlUtils": {
                    "enabled": false,
                    "fullImageName": null,
                    "resources": {},
                    "env": {},
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {}
                  },
                  "ingress": {
                    "enabled": false,
                    "labels": null,
                    "annotations": null,
                    "paths": [
                      "/"
                    ],
                    "pathType": "ImplementationSpecific",
                    "hosts": [
                      "cost-analyzer.local"
                    ],
                    "tls": []
                  },
                  "nodeSelector": {},
                  "tolerations": [],
                  "affinity": {},
                  "topologySpreadConstraints": [],
                  "priority": {
                    "enabled": false,
                    "name": ""
                  },
                  "networkPolicy": {
                    "enabled": false,
                    "denyEgress": true,
                    "sameNamespace": true,
                    "costAnalyzer": {
                      "enabled": false,
                      "annotations": {},
                      "additionalLabels": {}
                    }
                  },
                  "extraVolumes": [],
                  "extraVolumeMounts": [],
                  "persistentVolume": {
                    "size": "32Gi",
                    "dbSize": "32.0Gi",
                    "enabled": true,
                    "labels": {},
                    "annotations": {},
                    "dbPVEnabled": false
                  },
                  "service": {
                    "type": "ClusterIP",
                    "port": 9090,
                    "targetPort": 9090,
                    "nodePort": {},
                    "labels": {},
                    "annotations": {},
                    "sessionAffinity": {
                      "enabled": false,
                      "timeoutSeconds": 10800
                    }
                  },
                  "prometheus": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccounts": {
                      "alertmanager": {
                        "create": true,
                        "name": null
                      },
                      "nodeExporter": {
                        "create": true,
                        "name": null
                      },
                      "pushgateway": {
                        "create": true,
                        "name": null
                      },
                      "server": {
                        "create": true,
                        "name": null,
                        "annotations": {}
                      }
                    },
                    "imagePullSecrets": null,
                    "extraScrapeConfigs": "- job_name: kubecost\n  honor_labels: true\n  scrape_interval: 1m\n  scrape_timeout: 60s\n  metrics_path: /metrics\n  scheme: http\n  dns_sd_configs:\n  - names:\n    - {{ template \"cost-analyzer.serviceName\" . }}\n    type: 'A'\n    port: 9003\n- job_name: kubecost-networking\n  kubernetes_sd_configs:\n    - role: pod\n  relabel_configs:\n  # Scrape only the the targets matching the following metadata\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_instance]\n      action: keep\n      regex:  kubecost\n    - source_labels: [__meta_kubernetes_pod_label_app_kubernetes_io_name]\n      action: keep\n      regex:  network-costs\n",
                    "server": {
                      "enabled": true,
                      "name": "server",
                      "sidecarContainers": null,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "image": {
                        "repository": "quay.io/prometheus/prometheus",
                        "tag": "v2.50.1",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "prefixURL": "",
                      "baseURL": "",
                      "env": [],
                      "extraFlags": [
                        "web.enable-lifecycle"
                      ],
                      "configPath": "/etc/config/prometheus.yml",
                      "global": {
                        "scrape_interval": "1m",
                        "scrape_timeout": "60s",
                        "evaluation_interval": "1m",
                        "external_labels": {
                          "cluster_id": "cluster-one"
                        }
                      },
                      "remoteWrite": {},
                      "remoteRead": {},
                      "extraArgs": {
                        "query.max-concurrency": 1,
                        "query.max-samples": 100000000
                      },
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "pathType": "Prefix",
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "32Gi",
                        "subPath": ""
                      },
                      "emptyDir": {
                        "sizeLimit": ""
                      },
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "podLabels": {},
                      "alertmanagers": [],
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "annotations": {},
                        "labels": {},
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "readinessProbeInitialDelay": 5,
                      "readinessProbeTimeout": 3,
                      "readinessProbeFailureThreshold": 3,
                      "readinessProbeSuccessThreshold": 1,
                      "livenessProbeInitialDelay": 5,
                      "livenessProbeTimeout": 3,
                      "livenessProbeFailureThreshold": 3,
                      "livenessProbeSuccessThreshold": 1,
                      "resources": {},
                      "verticalAutoscaler": {
                        "enabled": false
                      },
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP",
                        "gRPC": {
                          "enabled": false,
                          "servicePort": 10901
                        },
                        "statefulsetReplica": {
                          "enabled": false,
                          "replica": 0
                        }
                      },
                      "terminationGracePeriodSeconds": 300,
                      "retention": "97h"
                    },
                    "alertmanager": {
                      "enabled": false,
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "name": "alertmanager",
                      "image": {
                        "repository": "quay.io/prometheus/alertmanager",
                        "tag": "v0.27.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "prefixURL": "",
                      "baseURL": "http://localhost:9093",
                      "extraEnv": {},
                      "extraSecretMounts": [],
                      "configMapOverrideName": "",
                      "configFromSecret": "",
                      "configFileName": "alertmanager.yml",
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "extraLabels": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "affinity": {},
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      },
                      "podAnnotations": {},
                      "podLabels": {},
                      "replicaCount": 1,
                      "statefulSet": {
                        "enabled": false,
                        "podManagementPolicy": "OrderedReady",
                        "headless": {
                          "annotations": {},
                          "labels": {},
                          "servicePort": 80
                        }
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true,
                        "runAsGroup": 1001,
                        "fsGroup": 1001
                      },
                      "service": {
                        "annotations": {},
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 80,
                        "sessionAffinity": "None",
                        "type": "ClusterIP"
                      }
                    },
                    "alertmanagerFiles": {
                      "alertmanager.yml": {
                        "global": {},
                        "receivers": [
                          {
                            "name": "default-receiver"
                          }
                        ],
                        "route": {
                          "group_wait": "10s",
                          "group_interval": "5m",
                          "receiver": "default-receiver",
                          "repeat_interval": "3h"
                        }
                      }
                    },
                    "configmapReload": {
                      "prometheus": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {},
                        "containerSecurityContext": {}
                      },
                      "alertmanager": {
                        "enabled": false,
                        "name": "configmap-reload",
                        "image": {
                          "repository": "quay.io/prometheus-operator/prometheus-config-reloader",
                          "tag": "v0.72.0",
                          "pullPolicy": "IfNotPresent"
                        },
                        "extraArgs": {},
                        "extraVolumeDirs": [],
                        "extraConfigmapMounts": [],
                        "resources": {}
                      }
                    },
                    "nodeExporter": {
                      "enabled": false,
                      "hostNetwork": true,
                      "hostPID": true,
                      "dnsPolicy": "ClusterFirstWithHostNet",
                      "name": "node-exporter",
                      "image": {
                        "repository": "prom/node-exporter",
                        "tag": "v1.7.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "extraArgs": {},
                      "extraHostPathMounts": [],
                      "extraConfigmapMounts": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "deploymentAnnotations": {},
                      "pod": {
                        "labels": {}
                      },
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/scrape": "true"
                        },
                        "labels": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "hostPort": 9100,
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9100,
                        "type": "ClusterIP"
                      }
                    },
                    "pushgateway": {
                      "enabled": false,
                      "name": "pushgateway",
                      "image": {
                        "repository": "prom/pushgateway",
                        "tag": "v1.6.2",
                        "pullPolicy": "IfNotPresent"
                      },
                      "priorityClassName": "",
                      "extraArgs": {},
                      "ingress": {
                        "enabled": false,
                        "annotations": {},
                        "hosts": [],
                        "extraPaths": [],
                        "tls": []
                      },
                      "tolerations": [],
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "replicaCount": 1,
                      "podDisruptionBudget": {
                        "enabled": false,
                        "maxUnavailable": 1
                      },
                      "resources": {},
                      "securityContext": {
                        "runAsUser": 1001,
                        "runAsNonRoot": true
                      },
                      "service": {
                        "annotations": {
                          "prometheus.io/probe": "pushgateway"
                        },
                        "labels": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "servicePort": 9091,
                        "type": "ClusterIP"
                      },
                      "strategy": {
                        "type": "Recreate",
                        "rollingUpdate": null
                      },
                      "persistentVolume": {
                        "enabled": true,
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "existingClaim": "",
                        "mountPath": "/data",
                        "size": "2Gi",
                        "subPath": ""
                      }
                    },
                    "serverFiles": {
                      "alerting_rules.yml": {},
                      "alerts": {},
                      "recording_rules.yml": {},
                      "prometheus.yml": {
                        "rule_files": [
                          "/etc/config/recording_rules.yml",
                          "/etc/config/alerting_rules.yml",
                          "/etc/config/rules",
                          "/etc/config/alerts"
                        ],
                        "scrape_configs": [
                          {
                            "job_name": "prometheus",
                            "static_configs": [
                              {
                                "targets": [
                                  "localhost:9090"
                                ]
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes-cadvisor",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_usage_seconds_total|container_memory_working_set_bytes|container_network_receive_errors_total|container_network_transmit_errors_total|container_network_receive_packets_dropped_total|container_network_transmit_packets_dropped_total|container_memory_usage_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_periods_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_cpu_cfs_throttled_periods_total|container_cpu_cfs_periods_total|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_inodes_free|container_fs_inodes_total|container_fs_usage_bytes|container_fs_limit_bytes|container_spec_cpu_shares|container_spec_memory_limit_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|container_fs_reads_bytes_total|container_network_receive_bytes_total|container_fs_writes_bytes_total|container_fs_reads_bytes_total|cadvisor_version_info|kubecost_pv_info)",
                                "action": "keep"
                              },
                              {
                                "source_labels": [
                                  "container"
                                ],
                                "target_label": "container_name",
                                "regex": "(.+)",
                                "action": "replace"
                              },
                              {
                                "source_labels": [
                                  "pod"
                                ],
                                "target_label": "pod_name",
                                "regex": "(.+)",
                                "action": "replace"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-nodes",
                            "scheme": "https",
                            "tls_config": {
                              "ca_file": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                              "insecure_skip_verify": true
                            },
                            "bearer_token_file": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                            "kubernetes_sd_configs": [
                              {
                                "role": "node"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_node_label_(.+)"
                              },
                              {
                                "target_label": "__address__",
                                "replacement": "kubernetes.default.svc:443"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_node_name"
                                ],
                                "regex": "(.+)",
                                "target_label": "__metrics_path__",
                                "replacement": "/api/v1/nodes/$1/proxy/metrics"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(kubelet_volume_stats_used_bytes)",
                                "action": "keep"
                              }
                            ]
                          },
                          {
                            "job_name": "kubernetes-service-endpoints",
                            "kubernetes_sd_configs": [
                              {
                                "role": "endpoints"
                              }
                            ],
                            "relabel_configs": [
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scrape"
                                ],
                                "action": "keep",
                                "regex": true
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_endpoints_name"
                                ],
                                "action": "keep",
                                "regex": "(.*node-exporter|kubecost-network-costs)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_scheme"
                                ],
                                "action": "replace",
                                "target_label": "__scheme__",
                                "regex": "(https?)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_annotation_prometheus_io_path"
                                ],
                                "action": "replace",
                                "target_label": "__metrics_path__",
                                "regex": "(.+)"
                              },
                              {
                                "source_labels": [
                                  "__address__",
                                  "__meta_kubernetes_service_annotation_prometheus_io_port"
                                ],
                                "action": "replace",
                                "target_label": "__address__",
                                "regex": "([^:]+)(?::\\d+)?;(\\d+)",
                                "replacement": "$1:$2"
                              },
                              {
                                "action": "labelmap",
                                "regex": "__meta_kubernetes_service_label_(.+)"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_namespace"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_namespace"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_service_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_name"
                              },
                              {
                                "source_labels": [
                                  "__meta_kubernetes_pod_node_name"
                                ],
                                "action": "replace",
                                "target_label": "kubernetes_node"
                              }
                            ],
                            "metric_relabel_configs": [
                              {
                                "source_labels": [
                                  "__name__"
                                ],
                                "regex": "(container_cpu_allocation|container_cpu_usage_seconds_total|container_fs_limit_bytes|container_fs_writes_bytes_total|container_gpu_allocation|container_memory_allocation_bytes|container_memory_usage_bytes|container_memory_working_set_bytes|container_network_receive_bytes_total|container_network_transmit_bytes_total|DCGM_FI_DEV_GPU_UTIL|deployment_match_labels|kube_daemonset_status_desired_number_scheduled|kube_daemonset_status_number_ready|kube_deployment_spec_replicas|kube_deployment_status_replicas|kube_deployment_status_replicas_available|kube_job_status_failed|kube_namespace_annotations|kube_namespace_labels|kube_node_info|kube_node_labels|kube_node_status_allocatable|kube_node_status_allocatable_cpu_cores|kube_node_status_allocatable_memory_bytes|kube_node_status_capacity|kube_node_status_capacity_cpu_cores|kube_node_status_capacity_memory_bytes|kube_node_status_condition|kube_persistentvolume_capacity_bytes|kube_persistentvolume_status_phase|kube_persistentvolumeclaim_info|kube_persistentvolumeclaim_resource_requests_storage_bytes|kube_pod_container_info|kube_pod_container_resource_limits|kube_pod_container_resource_limits_cpu_cores|kube_pod_container_resource_limits_memory_bytes|kube_pod_container_resource_requests|kube_pod_container_resource_requests_cpu_cores|kube_pod_container_resource_requests_memory_bytes|kube_pod_container_status_restarts_total|kube_pod_container_status_running|kube_pod_container_status_terminated_reason|kube_pod_labels|kube_pod_owner|kube_pod_status_phase|kube_replicaset_owner|kube_statefulset_replicas|kube_statefulset_status_replicas|kubecost_cluster_info|kubecost_cluster_management_cost|kubecost_cluster_memory_working_set_bytes|kubecost_load_balancer_cost|kubecost_network_internet_egress_cost|kubecost_network_region_egress_cost|kubecost_network_zone_egress_cost|kubecost_node_is_spot|kubecost_pod_network_egress_bytes_total|node_cpu_hourly_cost|node_cpu_seconds_total|node_disk_reads_completed|node_disk_reads_completed_total|node_disk_writes_completed|node_disk_writes_completed_total|node_filesystem_device_error|node_gpu_count|node_gpu_hourly_cost|node_memory_Buffers_bytes|node_memory_Cached_bytes|node_memory_MemAvailable_bytes|node_memory_MemFree_bytes|node_memory_MemTotal_bytes|node_network_transmit_bytes_total|node_ram_hourly_cost|node_total_hourly_cost|pod_pvc_allocation|pv_hourly_cost|service_selector_labels|statefulSet_match_labels|kubecost_pv_info|up)",
                                "action": "keep"
                              }
                            ]
                          }
                        ]
                      },
                      "rules": {
                        "groups": [
                          {
                            "name": "CPU",
                            "rules": [
                              {
                                "expr": "sum(rate(container_cpu_usage_seconds_total{container!=\"\"}[5m]))",
                                "record": "cluster:cpu_usage:rate5m"
                              },
                              {
                                "expr": "rate(container_cpu_usage_seconds_total{container!=\"\"}[5m])",
                                "record": "cluster:cpu_usage_nosum:rate5m"
                              },
                              {
                                "expr": "avg(irate(container_cpu_usage_seconds_total{container!=\"POD\", container!=\"\"}[5m])) by (container,pod,namespace)",
                                "record": "kubecost_container_cpu_usage_irate"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"}) by (container,pod,namespace)",
                                "record": "kubecost_container_memory_working_set_bytes"
                              },
                              {
                                "expr": "sum(container_memory_working_set_bytes{container!=\"POD\",container!=\"\"})",
                                "record": "kubecost_cluster_memory_working_set_bytes"
                              }
                            ]
                          },
                          {
                            "name": "Savings",
                            "rules": [
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod))",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_cpu_allocation) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_cpu_allocation",
                                "labels": {
                                  "daemonset": "true"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind!=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod))",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "false"
                                }
                              },
                              {
                                "expr": "sum(avg(kube_pod_owner{owner_kind=\"DaemonSet\"}) by (pod) * sum(container_memory_allocation_bytes) by (pod)) / sum(kube_node_info)",
                                "record": "kubecost_savings_memory_allocation_bytes",
                                "labels": {
                                  "daemonset": "true"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    "alertRelabelConfigs": null,
                    "networkPolicy": {
                      "enabled": false
                    }
                  },
                  "networkCosts": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/kubecost-network-costs",
                      "tag": "v0.17.3"
                    },
                    "imagePullPolicy": "Always",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "prometheusScrape": false,
                    "trafficLogging": true,
                    "logLevel": "info",
                    "port": 3001,
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "20Mi"
                      }
                    },
                    "extraArgs": [],
                    "config": {
                      "destinations": {
                        "in-zone": [
                          "127.0.0.0/8",
                          "169.254.0.0/16",
                          "10.0.0.0/8",
                          "172.16.0.0/12",
                          "192.168.0.0/16"
                        ],
                        "in-region": [],
                        "cross-region": [],
                        "internet": [],
                        "direct-classification": []
                      },
                      "services": {
                        "google-cloud-services": false,
                        "amazon-web-services": false,
                        "azure-cloud-services": false
                      }
                    },
                    "tolerations": [],
                    "affinity": {},
                    "service": {
                      "annotations": {},
                      "labels": {}
                    },
                    "priorityClassName": "",
                    "podMonitor": {
                      "enabled": false,
                      "additionalLabels": {}
                    },
                    "additionalLabels": {},
                    "nodeSelector": {},
                    "annotations": {},
                    "healthCheckProbes": {},
                    "additionalSecurityContext": {}
                  },
                  "kubecostDeployment": {
                    "replicas": 1,
                    "labels": {},
                    "annotations": {}
                  },
                  "forecasting": {
                    "enabled": true,
                    "fullImageName": "gcr.io/kubecost1/kubecost-modeling:v0.1.6",
                    "resources": {
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "limits": {
                        "cpu": "1500m",
                        "memory": "1Gi"
                      }
                    },
                    "env": {
                      "GUNICORN_CMD_ARGS": "--log-level info -t 1200"
                    },
                    "priority": {
                      "enabled": false,
                      "name": ""
                    },
                    "nodeSelector": {},
                    "tolerations": {},
                    "affinity": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    }
                  },
                  "kubecostAggregator": {
                    "deployMethod": "singlepod",
                    "enabled": false,
                    "replicas": 1,
                    "stagingEmptyDirSizeLimit": "2Gi",
                    "numDBCopyPartitions": 25,
                    "env": {
                      "LOG_LEVEL": "info",
                      "DB_READ_THREADS": "1",
                      "DB_WRITE_THREADS": "1",
                      "DB_CONCURRENT_INGESTION_COUNT": "3"
                    },
                    "persistentConfigsStorage": {
                      "storageClass": "",
                      "storageRequest": "1Gi"
                    },
                    "aggregatorDbStorage": {
                      "storageClass": "",
                      "storageRequest": "128Gi"
                    },
                    "resources": {},
                    "readinessProbe": {
                      "enabled": true,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "failureThreshold": 200
                    },
                    "cloudCost": {
                      "enabled": false,
                      "resources": {},
                      "readinessProbe": {
                        "enabled": true,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 10,
                        "failureThreshold": 200
                      }
                    },
                    "jaeger": {
                      "enabled": false,
                      "image": "jaegertracing/all-in-one",
                      "imageVersion": "latest"
                    }
                  },
                  "diagnostics": {
                    "enabled": true,
                    "primary": {
                      "enabled": false,
                      "retention": "7d",
                      "readonly": false
                    },
                    "pollingInterval": "300s",
                    "keepDiagnosticHistory": false,
                    "collectHelmValues": false,
                    "deployment": {
                      "enabled": false,
                      "resources": {
                        "requests": {
                          "cpu": "10m",
                          "memory": "20Mi"
                        }
                      },
                      "env": {},
                      "labels": {},
                      "securityContext": {},
                      "containerSecurityContext": {},
                      "nodeSelector": {},
                      "tolerations": {},
                      "affinity": {}
                    }
                  },
                  "clusterController": {
                    "enabled": false,
                    "image": {
                      "repository": "gcr.io/kubecost1/cluster-controller",
                      "tag": "v0.16.0"
                    },
                    "imagePullPolicy": "Always",
                    "priorityClassName": "",
                    "tolerations": [],
                    "actionConfigs": {
                      "clusterTurndown": [],
                      "namespaceTurndown": null,
                      "clusterRightsize": null,
                      "containerRightsize": null
                    },
                    "kubescaler": {
                      "defaultResizeAll": false
                    },
                    "namespaceTurndown": {
                      "rbac": {
                        "enabled": true
                      }
                    }
                  },
                  "reporting": {
                    "logCollection": true,
                    "productAnalytics": true,
                    "errorReporting": true,
                    "valuesReporting": true
                  },
                  "serviceMonitor": {
                    "enabled": false,
                    "additionalLabels": {},
                    "metricRelabelings": [],
                    "relabelings": [],
                    "networkCosts": {
                      "enabled": false,
                      "scrapeTimeout": "10s",
                      "additionalLabels": {},
                      "metricRelabelings": [],
                      "relabelings": []
                    }
                  },
                  "prometheusRule": {
                    "enabled": false,
                    "additionalLabels": {}
                  },
                  "supportNFS": false,
                  "initChownDataImage": "busybox",
                  "initChownData": {
                    "resources": {}
                  },
                  "grafana": {
                    "rbac": {
                      "create": true
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    },
                    "replicas": 1,
                    "deploymentStrategy": "RollingUpdate",
                    "readinessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      }
                    },
                    "livenessProbe": {
                      "httpGet": {
                        "path": "/api/health",
                        "port": 3000
                      },
                      "initialDelaySeconds": 60,
                      "timeoutSeconds": 30,
                      "failureThreshold": 10
                    },
                    "image": {
                      "repository": "grafana/grafana",
                      "tag": "10.3.4",
                      "pullPolicy": "IfNotPresent"
                    },
                    "securityContext": {},
                    "priorityClassName": "",
                    "downloadDashboardsImage": {
                      "repository": "curlimages/curl",
                      "tag": "latest",
                      "pullPolicy": "IfNotPresent"
                    },
                    "podAnnotations": {},
                    "annotations": {},
                    "service": {
                      "type": "ClusterIP",
                      "port": 80,
                      "annotations": {},
                      "labels": {}
                    },
                    "ingress": {
                      "enabled": false,
                      "annotations": {},
                      "labels": {},
                      "path": "/",
                      "pathType": "Prefix",
                      "hosts": [
                        "chart-example.local"
                      ],
                      "tls": []
                    },
                    "resources": {},
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "persistence": {
                      "enabled": false
                    },
                    "adminUser": "admin",
                    "adminPassword": "strongpassword",
                    "env": {},
                    "envFromSecret": "",
                    "extraSecretMounts": [],
                    "plugins": [],
                    "dashboardProviders": {},
                    "dashboards": {},
                    "dashboardsConfigMaps": {},
                    "ldap": {
                      "existingSecret": "",
                      "config": ""
                    },
                    "smtp": {
                      "existingSecret": ""
                    },
                    "sidecar": {
                      "image": {
                        "repository": "kiwigrid/k8s-sidecar",
                        "tag": "1.26.0",
                        "pullPolicy": "IfNotPresent"
                      },
                      "resources": {},
                      "dashboards": {
                        "enabled": true,
                        "label": "grafana_dashboard",
                        "labelValue": "1",
                        "annotations": {},
                        "error_throttle_sleep": 0,
                        "folder": "/tmp/dashboards"
                      },
                      "datasources": {
                        "enabled": false,
                        "error_throttle_sleep": 0,
                        "label": "grafana_datasource"
                      }
                    },
                    "grafana.ini": {
                      "server": {
                        "serve_from_sub_path": false,
                        "root_url": "%(protocol)s://%(domain)s:%(http_port)s/grafana"
                      },
                      "paths": {
                        "data": "/var/lib/grafana/data",
                        "logs": "/var/log/grafana",
                        "plugins": "/var/lib/grafana/plugins",
                        "provisioning": "/etc/grafana/provisioning"
                      },
                      "analytics": {
                        "check_for_updates": true
                      },
                      "log": {
                        "mode": "console"
                      },
                      "grafana_net": {
                        "url": "https://grafana.net"
                      },
                      "auth.anonymous": {
                        "enabled": true,
                        "org_role": "Editor",
                        "org_name": "Main Org."
                      }
                    }
                  },
                  "serviceAccount": {
                    "create": true,
                    "annotations": {}
                  },
                  "awsstore": {
                    "useAwsStore": false,
                    "imageNameAndVersion": "gcr.io/kubecost1/awsstore:latest",
                    "createServiceAccount": false,
                    "priorityClassName": "",
                    "nodeSelector": {},
                    "annotations": {}
                  },
                  "federatedETL": {
                    "agentOnly": false,
                    "federatedCluster": false,
                    "redirectS3Backup": false,
                    "useMultiClusterDB": false
                  },
                  "kubecostAdmissionController": {
                    "enabled": false,
                    "secretName": "webhook-server-tls",
                    "caBundle": "${CA_BUNDLE}"
                  },
                  "costEventsAudit": {
                    "enabled": false
                  },
                  "extraObjects": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
          createdAt: "2024-05-19T20:46:35Z"
          description: An operator that deploys Kubecost 2.2.4 via the Helm chart.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Stackwatch
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |-
          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts, so you can provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from cloud infrastructure services for complete cost visibility—all services from AWS, GCP, and Microsoft Azure are supported.
          * Gain insights you can use to save 30-50% or more on your infrastructure spend, without exposing your private information—your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure, stay within budgeted limits, and address monitoring interruptions immediately.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubecost
        - cost
        - kubernetes
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: czoller@kubecost.com
          name: Chip Zoller
        maturity: alpha
        provider:
          name: Stackwatch
          url: www.kubecost.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - gcr.io/kubecost1/operator/kubecost-operator-controller:v2.2.4
        - quay.io/community-operator-pipeline-prod/kubecost@sha256:ff1e00f5a04e49fe9daf06db6b60c2ca14692efd098a38ecd622ef6bc5de8dd5
        version: 2.2.4
      entries:
      - name: kubecost.v2.2.4
        version: 2.2.4
      - name: kubecost.v2.1.0
        version: 2.1.0
      - name: kubecost-operator.v1.107.1
        version: 1.107.1
      - name: kubecost-operator.v1.106.2
        version: 1.106.2
      - name: kubecost-operator.v1.106.0
        version: 1.106.0
      - name: kubecost.v1.99.0
        version: 1.99.0
      - name: kubecost.v1.97.0
        version: 1.97.0
      name: alpha
    defaultChannel: alpha
    packageName: kubecost
    provider:
      name: Stackwatch
      url: www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-ui
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-ui.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Ui",
                "metadata": {
                  "name": "ui-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/opt/app-root/src/assets/config.json",
                      "name": "config",
                      "subPath": "config.json"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:58Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ui
            name: uis.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP UI Operator
        displayName: Keysight WAP UI Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Ui
          url: https://keysight-wap-ui.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-ui@sha256:876a394f5cc2f745091ad2ee9ca0d08404cfb95284222e1f3a8607e193bd986a
        - registry.connect.redhat.com/loadcore/keysight-wap-ui-bundle@sha256:7ca0365e849fc98950021f86f3acb9048303fad3dcb6eac0ee09dd4bcdaf8fe0
        version: 0.1.21
      entries:
      - name: keysight-wap-ui.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-ui
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco AppDynamics
      provider-url: https://www.appdynamics.com
    name: appdynamics-cloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-cloud-operator.v24.7.0-133
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clustermon",
                "metadata": {
                  "name": "clustermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "clusterName": "test",
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Containermon",
                "metadata": {
                  "name": "containermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ContainerMon"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbCollector",
                "metadata": {
                  "name": "dbcollector",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-db-collector:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "DbConfig",
                "metadata": {
                  "name": "dbconfig",
                  "namespace": "appdynamics"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Inframon",
                "metadata": {
                  "name": "inframon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-k8s-monitoring:latest"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "LogCollector",
                "metadata": {
                  "name": "logcollector-sample",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "image": "appdynamics/appdynamics-cloud-log-collector-agent:latest",
                  "resources": {
                    "limits": {
                      "cpu": "400m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "150Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Servermon",
                "metadata": {
                  "name": "servermon",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "name": "ServerMon"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-08-06T08:33:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clustermon is the Schema for the clustermons API
            displayName: Clustermon
            kind: Clustermon
            name: clustermons.cluster.appdynamics.com
            version: v1alpha1
          - description: Containermon is the Schema for the containermons API
            displayName: Containermon
            kind: Containermon
            name: containermons.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbCollector
            name: dbcollectors.cluster.appdynamics.com
            version: v1alpha1
          - kind: DbConfig
            name: dbconfigs.cluster.appdynamics.com
            version: v1alpha1
          - description: Inframon is the Schema for the inframons API
            displayName: Inframon
            kind: Inframon
            name: inframons.cluster.appdynamics.com
            version: v1alpha1
          - kind: LogCollector
            name: logcollectors.cluster.appdynamics.com
            version: v1alpha1
          - description: Servermon is the Schema for the servermons API
            displayName: Servermon
            kind: Servermon
            name: servermons.cluster.appdynamics.com
            version: v1alpha1
        description: The Appdynamics Cloud Operator is used to install Cluster Collector,
          Container Collector, Server Collector, Log Collector and Cluster Agent
        displayName: Cisco AppDynamics Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - appdynamics
        links:
        - name: Appdynamics Cloud Operator
          url: https://appdynamics-cloud-operator.domain
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: Cisco AppDynamics
          url: https://www.appdynamics.com
        relatedImages:
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator-bundle@sha256:3715abae1ffeea2d0bebf665b1dfaddce4686bb8a61f0697e6fd201b157798b6
        - registry.connect.redhat.com/appdynamics/appdynamics-cloud-operator@sha256:2710d63425942f60af21f060a02eb3c40f5a5f0c952ea6593b3b9769c7ca1014
        version: 24.7.0-133
      entries:
      - name: appdynamics-cloud-operator.v24.7.0-133
        version: 24.7.0-133
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-cloud-operator
    provider:
      name: Cisco AppDynamics
      url: https://www.appdynamics.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-nifi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: nifi-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
          description: Stackable Operator for Apache NiFi
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/nifi-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for NifiClusterSpec via `CustomResource`
            displayName: nificlusters.nifi.stackable.tech
            kind: NifiCluster
            name: nificlusters.nifi.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache NiFi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/nifi-operator@sha256:84bc5a81bd8a2ad8fca1ca7fa49d5d2bcc1c83d020b58ee5a7fc23fc20c6b2e8
        - registry.connect.redhat.com/stackable/stackable-apache-nifi-operator@sha256:efcce89937bf5ec062f36bdd3a5417b98d6d140c027c2cc615c87fb04f6d1602
        version: 25.7.0
      entries:
      - name: nifi-operator.v25.7.0
        version: 25.7.0
      - name: nifi-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-nifi-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: tenant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tenant-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-03-19T13:30:50Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.1
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:fb7bc294096fa78f199d909f6ff8f7fcbcd771c3dc19f3f767d77dd13a2081a9
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:a899d60b31efd5cf2358efcd48737b1c4f4a0356ab6c0abe7a5f62ea825a52cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a35b96eacb4e46114b39354f66429f8f05df09c55c98c9416b4ffa2768310ef7
        version: 1.2.1
      entries:
      - name: tenant-operator.v1.2.1
        version: 1.2.1
      name: release-1.2
    - currentCSV: tenant-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-06-11T15:25:04Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.0
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: TemplateGroupInstance is the Schema for the templategroupinstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:e88aa5b008ece406bca35ec22fdc87c842544049d613d4d81c8bc5dd2fb50812
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:10e4bdca44a1a0cdef27b415765238249f81ae6ba53f94f992e37d350bc9f800
        version: 1.3.0
      entries:
      - name: tenant-operator.v1.3.0
        version: 1.3.0
      name: release-1.3
    - currentCSV: tenant-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-08T08:14:28Z"
          description: Stakater Multi-Tenancy Operator helps to solve the complexity
            issues of doing true multi-tenancy in Openshift by providing simplified
            abstractions on top of the native primitives to provide organizations
            and platform providers the tooling, self-service capabilities and robust
            automation needed to do secure and efficient Multi-Tenancy in a single
            OpenShift cluster. OpenShift is designed to support a single tenant platform,
            hence making it difficult for cluster admins to host multi-tenancy in
            a single OpenShift cluster. If multi-tenancy is achieved by sharing a
            cluster, it can have many advantages, e.g. efficient resource utilization,
            less configuration effort and easier sharing of cluster-internal resources
            among different tenants.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.4.2
          operatorframework.io/suggested-namespace: multi-tenant-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Extensions is the Schema for the extensions API
            displayName: Extensions
            kind: Extensions
            name: extensions.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationConfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1alpha1
          - description: IntegrationConfig is the Schema for the integrationconfigs
              API
            displayName: Integration Config
            kind: IntegrationConfig
            name: integrationconfigs.tenantoperator.stakater.com
            version: v1beta1
          - description: Quota is the Schema for the quotas API
            displayName: Quota
            kind: Quota
            name: quotas.tenantoperator.stakater.com
            version: v1beta1
          - description: ResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Resource Supervisor
            kind: ResourceSupervisor
            name: resourcesupervisors.tenantoperator.stakater.com
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta2
          - kind: Tenant
            name: tenants.tenantoperator.stakater.com
            version: v1beta3
        description: |-
          OpenShift is designed to support a single tenant platform, hence making it difficult for cluster admins to host multi-tenancy in a single OpenShift cluster. If multi-tenancy is achieved by sharing a cluster, it can have many advantages, e.g. efficient resource utilization, less configuration effort and easier sharing of cluster-internal resources among different tenants. Multi Tenant Operator helps to solve the complexity issues of doing true multi-tenancy by providing simplified abstractions on top of the native primitives to provide organizations and platform providers the tooling, self-service capabilities and robust automation needed to do secure and efficient Multi-Tenancy in a single OpenShift cluster.

          ## Access Control

          Multi Tenant Operator provides several ClusterRoles that are automatically bound to the Tenants Namespaces used for managing access to the Namespaces and the resources they contain. You can also modify the default roles or create new roles to have full control and customize access control for your users and teams.

          ## Self-Service

          With Multi Tenant Operator, you can empower your users to safely provision namespaces for themselves and their teams (typically mapped to SSO groups). Team-owned namespaces and the resources inside of them count towards the team's quotas rather than the user's individual limits and are automatically shared with all team members according to the access rules you configure in tenant-operator.

          ## HashiCorp Vault Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to HashiCorp Vault where it can create vault paths and greatly ease the overhead of managing RBAC in Vault.

          ## ArgoCD Multitenancy

          Multi Tenant Operator is not only providing strong Multi-Tenancy for the OpenShift internals but also extends the tenant's permission model to ArgoCD where it can provision AppProjects and Allowed Repositories for your tenants greatly easing the overhead of managing RBAC in ArgoCD.

          ## Cost/Resource Optimization

          Multi Tenant Operator provides a mechanism for defining Resource Quotas at the tenant scope, meaning all namespaces belonging to a particular tenant share the defined quota, which is why you are able to safely enable dev teams to self-serve their namespaces whilst being confident that they can only use the resources allocated based on budget and business needs.

          ## Sandboxed Dev Namespaces

          Multi Tenant Operator can be configured to automatically provision a namespace for every member of the specific tenant, that will also be pre-loaded with any selected templates and consume the same pool of resources from the tenant's quota creating safe sandboxed dev namespaces that teams can use as a scratch namespace for rapid prototyping and development.
        displayName: Multi Tenant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tenant
        - stakater
        - SRO
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Multi Tenant Operator
          url: https://docs.stakater.com/mto/latest/index.html
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/tenant-operator-0@sha256:3e8aa02c71fc99e5eeabfce39c8fc24a9f9e8e826a58f9ff59c8b7c0e9c14e0d
        - registry.connect.redhat.com/stakater/tenant-operator@sha256:7ee14c0974d1eda849d2c199a56786b0ae777ef6569085be917449f1998b1158
        version: 1.4.2
      entries:
      - name: tenant-operator.v1.4.2
        version: 1.4.2
      name: release-1.4
    defaultChannel: release-1.4
    packageName: tenant-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Shipwright Contributors
      provider-url: https://shipwright.io
    name: shipwright-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: shipwright-operator.v0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "shipwright-build"
                },
                "spec": {
                  "targetNamespace": "shipwright-build"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
          description: Shipwright is a framework for building container images on
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/shipwright-io/operator
          support: The Shipwright Contributors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
        description: |
          Shipwright is a framework for building container images on Kubernetes.

          Read more: [https://shipwright.io](https://shipwright.io)

          ## Usage

          To deploy and manage [Shipwright Builds](https://github.com/shipwright-io/build) in your cluster,
          first make sure this operator is installed and running on your cluster.

          Next, create the following:

          ```yaml
          ---
          apiVersion: operator.shipwright.io/v1alpha1
          kind: ShipwrightBuild
          metadata:
            name: shipwright-operator
          spec:
            targetNamespace: shipwright-build
          ```

          The operator will deploy Shipwright Builds in the provided `targetNamespace`.
          When `.spec.targetNamespace` is not set, the namespace will default to `shipwright-build`.
          Refer to the [ShipwrightBuild documentation](https://github.com/shipwright-io/operator/blob/main/docs/shipwrightbuild.md) for more information about this custom resource.
        displayName: Shipwright Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://shipwright.io
        - name: Shipwright Builds
          url: https://github.com/shipwright-io/build
        maintainers:
        - email: shipwright-dev@lists.shipwright.io
          name: The Shipwright Contributors
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Shipwright Contributors
          url: https://shipwright.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/shipwright-io/operator/operator:0.13.0@sha256:12cd5f936043c96f31cf571d7a4f8b6f2e240852cd9ad1a5423dbab0c208b542
        - quay.io/community-operator-pipeline-prod/shipwright-operator@sha256:f4d3b97812c69b396c8e6d081d8fa9b412ebe1365432658288e6ab065c220fca
        version: 0.13.0
      entries:
      - name: shipwright-operator.v0.13.0
        version: 0.13.0
      - name: shipwright-operator.v0.9.0
        version: 0.9.0
      name: alpha
    defaultChannel: alpha
    packageName: shipwright-operator
    provider:
      name: The Shipwright Contributors
      url: https://shipwright.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: istio-workspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: istio-workspace-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workspace.maistra.io/v1alpha1",
                "kind": "Session",
                "metadata": {
                  "name": "session-sample"
                },
                "spec": {
                  "ref": [
                    {
                      "args": {
                        "image": "quay.io/maistra-dev/istio-workspace-test:latest"
                      },
                      "name": "ratings-v1",
                      "strategy": "prepared-image"
                    }
                  ],
                  "route": {
                    "name": "x-workspace-route",
                    "type": "header",
                    "value": "feature-x"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/maistra/istio-workspace:v0.5.3
          createdAt: "2022-11-16T13:50:35Z"
          description: Safely develop and test on any Kubernetes cluster without affecting
            others.
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/maistra/istio-workspace
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Session controls the creation of the specialized hidden routes.
            displayName: Session
            kind: Session
            name: sessions.workspace.maistra.io
            version: v1alpha1
        description: "Istio Workspace enables developers to:\n\n * run one or more
          services locally during development but make it appear like it would be
          developed in the actual cluster\n   * this way you can use your favourite
          development tools\n   * have instant feedback loop\n   * don't make your
          machine overheating trying to run the whole cluster locally\n * safely preview
          incoming changes in pull requests without affecting regular users \n * have
          confidence in testing or troubleshooting your services directly in the cluster\n\nIstio
          Workspace supports testing on multi-user environment in an unobtrusive way.\nIt
          doesn’t really matter if it is QE cluster or actual production. We give
          you confidence that your changes won’t blow up the cluster, and your users
          won’t even notice a glitch.\n\nYou will need [Kubernetes](https://k8s.io)
          or [Openshift](https://openshift.com) cluster with  [Istio](https://istio.io/)
          installed. \n\nYou can read more about our vision on [Red Hat Developer’s
          blog](https://developers.redhat.com/blog/2020/07/14/developing-and-testing-on-production-with-kubernetes-and-istio-workspace/)
          or ...\n\n## See it in action!\n\n[![https://youtu.be/XTNVadUzMCc](https://img.youtube.com/vi/XTNVadUzMCc/hqdefault.jpg)](https://youtu.be/XTNVadUzMCc)\n\nWatch
          demo: [\"How to develop on production: An introduction to Istio-Workspaces\"](https://youtu.be/XTNVadUzMCc).\n\n##
          Documentation\n\nMore details can be found on our [documentation page](https://istio-workspace-docs.netlify.com/).\n"
        displayName: Istio Workspace
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - telepresence
        links:
        - name: Istio Workspace Documentation
          url: https://istio-workspace-docs.netlify.com/
        - name: How to install CLI tool
          url: https://github.com/maistra/istio-workspace#install-in-two-easy-steps
        maintainers:
        - email: aslak@redhat.com
          name: Aslak Knutsen
        - email: bartosz@redhat.com
          name: Bartosz Majsak
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/istio-workspace-operator@sha256:7a581f048aee64701b23f1a0e94d77f2e6932ee8229fd46b1635d8df6994e174
        - quay.io/maistra/istio-workspace:v0.5.3
        version: 0.5.3
      entries:
      - name: istio-workspace-operator.v0.5.3
        version: 0.5.3
      - name: istio-workspace-operator.v0.5.2
        version: 0.5.2
      - name: istio-workspace-operator.v0.5.1
        version: 0.5.1
      - name: istio-workspace-operator.v0.5.0
        version: 0.5.0
      - name: istio-workspace-operator.v0.4.0
        version: 0.4.0
      - name: istio-workspace-operator.v0.3.0
        version: 0.3.0
      - name: istio-workspace-operator.v0.2.0
        version: 0.2.0
      - name: istio-workspace-operator.v0.1.0
        version: 0.1.0
      - name: istio-workspace-operator.v0.0.10
        version: 0.0.10
      - name: istio-workspace-operator.v0.0.9
        version: 0.0.9
      - name: istio-workspace-operator.v0.0.8
        version: 0.0.8
      - name: istio-workspace-operator.v0.0.7
        version: 0.0.7
      - name: istio-workspace-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: istio-workspace-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: assisted-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: assisted-service-operator.v0.7.71
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-x86_64-live.x86_64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-aarch64-live.aarch64.iso",
                      "version": "414.92.202310170514-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.0/rhcos-4.14.0-s390x-live.s390x.iso",
                      "version": "414.92.202309201615-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202310310037-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.15.0-ec.2/rhcos-4.15.0-ec.2-s390x-live.s390x.iso",
                      "version": "415.92.202310310037-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
          createdAt: "2023-12-15T15:26:08Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.71'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:0f4ecf9a75b6755d02a45f9b70357ac45b6b4db66f24f0e83f888269d64f4293
        - quay.io/edge-infrastructure/assisted-service@sha256:cffbdc5fd51c08f16ec2aa484cf226e3b29510919a8aa9668d766007de8f3d79
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:4e12e48d0b2b884de2a0d32682da044cff419aea53bb162915538e2179fcd7f3
        - quay.io/edge-infrastructure/assisted-image-service@sha256:027fa9c5a881848786745aafeac5e48575eefe22417abe29d9e8eb17940b8f80
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:40bafed94547a5fe96deb5492f10225501b965198227c9634869931692bda492
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:98740646fbfc0520f9d85ac5b483d1a898f22e164c5234901be51970bef7fcf3
        version: 0.7.71
      entries:
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.10
    - currentCSV: assisted-service-operator.v0.7.73
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-x86_64-live.x86_64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-aarch64-live.aarch64.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.15/rhcos-4.14.15-s390x-live.s390x.iso",
                      "version": "414.92.202402130420-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-x86_64-live.x86_64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.16.0-ec.4/rhcos-4.16.0-ec.4-x86_64-live.x86_64.iso",
                      "version": "416.94.202403071059-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-aarch64-live.aarch64.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202402130021-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.0/rhcos-4.15.0-s390x-live.s390x.iso",
                      "version": "415.92.202402130021-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
          createdAt: "2024-03-20T10:08:49Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.73'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:9eca42cc297a93293200a1c46e8ed949a61770550366044ef54e36bed9d7a116
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b33068bbe88e38c0444fb158e233e5644fb127416198d52fdc809a3f2e684fde
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:2cc14304ba27c51d61b86db05445f3f0acc68593a4d76081da8eb9d6bae41757
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:d0da07ec81877302fc376e8049dca69c1acdbf56a37aa3e320cd75f2b85a2ed0
        - quay.io/edge-infrastructure/assisted-installer@sha256:234e4680b0ff4376f2c11d6f94a75a5dd6ae16c3218d6b63d2003cf43c65e185
        - quay.io/edge-infrastructure/assisted-service@sha256:641fb65d8f0117cc6007567337ae38d5e912add454a76905617b60db99b60a64
        version: 0.7.73
      entries:
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.11
    - currentCSV: assisted-service-operator.v0.7.74
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-x86_64-live.x86_64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-aarch64-live.aarch64.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.14/4.14.34/rhcos-4.14.34-s390x-live.s390x.iso",
                      "version": "414.92.202407091253-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-x86_64-live.x86_64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-aarch64-live.aarch64.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-ppc64le-live.ppc64le.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.15",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.15/4.15.23/rhcos-4.15.23-s390x-live.s390x.iso",
                      "version": "415.92.202407091355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-x86_64-live.x86_64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-aarch64-live.aarch64.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-ppc64le-live.ppc64le.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.16",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.16/4.16.3/rhcos-4.16.3-s390x-live.s390x.iso",
                      "version": "416.94.202406251923-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-x86_64-live.x86_64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-aarch64-live.aarch64.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-ppc64le-live.ppc64le.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.17",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.17/4.17.0/rhcos-4.17.0-s390x-live.s390x.iso",
                      "version": "417.94.202408270355-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "418.94.202409162337-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.18",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.18.0-ec.2/rhcos-4.18.0-ec.2-s390x-live.s390x.iso",
                      "version": "418.94.202409162337-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service:latest
          createdAt: "2024-10-16T17:43:34Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.74'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AgentClassification is the Schema for the AgentClassifications
              API
            displayName: Agent Classification
            kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - description: Agent is the Schema for the hosts API
            displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: Infra Env
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/assisted-service-operator@sha256:0e8a33d795a2d3fa3c5fd409e4e1c40805d5c08a71481afbf9e7b8014b4179ca
        - quay.io/edge-infrastructure/assisted-service:latest
        version: 0.7.74
      entries:
      - name: assisted-service-operator.v0.7.74
        version: 0.7.74
      - name: assisted-service-operator.v0.7.73
        version: 0.7.73
      - name: assisted-service-operator.v0.7.72
        version: 0.7.72
      - name: assisted-service-operator.v0.7.71
        version: 0.7.71
      - name: assisted-service-operator.v0.7.70
        version: 0.7.70
      - name: assisted-service-operator.v0.7.69
        version: 0.7.69
      - name: assisted-service-operator.v0.7.68
        version: 0.7.68
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.13
    - currentCSV: assisted-service-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "osImages": [
                    {
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202105062123-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
          createdAt: "2021-09-23T11:11:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.0 <0.0.8'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/ocpmetal/assisted-service@sha256:64e88c5aa41c606bc2db2b0991acdbb1166559306474e13f5e2810f325616d64
        - quay.io/ocpmetal/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:31eda4f1888b180feb023c42a6ec93478b2e956f13420d55987d9abf7c7797c3
        - quay.io/ocpmetal/assisted-installer-agent@sha256:a391892766d1ce81300bf82943318e7c3be0d2945fc19065527ab9f64fc24cf8
        - quay.io/ocpmetal/assisted-installer-controller@sha256:43b65c310eee985ee0fc20ad8ee684f7557ffe704fd1a2fa53831e3dbc41e849
        - quay.io/ocpmetal/assisted-installer@sha256:db97f7a565a8d5af61f3876d6465996e90c7f79dad6ec71c5bb5cce837444974
        version: 0.0.8
      entries:
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.3
    - currentCSV: assisted-service-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.2/rhcos-4.8.2-x86_64-live.x86_64.iso",
                      "version": "48.84.202107202156-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
          createdAt: "2022-04-07T11:23:06Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.1.3'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled, for
              all other names a "Warning" event will be emitted + message logged and
              it will be ignored forever.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent".
          No other name will be accepted.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:7d5f9e02f3d6c3e2ac63f1fd33240dc696ee79563c214d21d2897ee773517e9c
        - quay.io/edge-infrastructure/assisted-image-service@sha256:add19e251cafdfc4f34264d0aec5b660f0ea2dcc13326f4a2c8d50a2e465c0a2
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:979650f2e03e457d71958e8813285fcb7185fe0035f21a6ba956b4277cb29b70
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:658787d1e6c1f733c5f03863655aff315258c0b38e1972573c3878df46e4e463
        - quay.io/edge-infrastructure/assisted-installer@sha256:461d60b21d1d3ccb15a170f937f4871088cd0f83108b2271e84bdf2883a66870
        - quay.io/edge-infrastructure/assisted-service@sha256:c3a852842b88f47c51cf1c62d3a19b102729fb355df19ced0e03e8265739c39c
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        version: 0.1.3
      entries:
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.4
    - currentCSV: assisted-service-operator.v0.3.30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-x86_64-live.x86_64.iso",
                      "version": "49.84.202110081407-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.9",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live-rootfs.aarch64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.9/4.9.0/rhcos-4.9.0-aarch64-live.aarch64.iso",
                      "version": "49.84.202110080947-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "rootFSUrl": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live-rootfs.x86_64.img",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.3/rhcos-4.10.3-x86_64-live.x86_64.iso",
                      "version": "410.84.202201251210-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
          createdAt: "2022-06-13T23:16:28Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.3.30'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer@sha256:57cf8d5fe49067a8c1345bedaf90f2b63aededdd7d615878496182c35a13892e
        - quay.io/edge-infrastructure/assisted-service@sha256:7c13ce52d82f761aec47bc5a6639d7baabe99d0f4f2e2106c4341018c606b458
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:5b061126dc3adf1fc9c8c22dfded18f4652bd62537aefca88ddcc585078b94ae
        - quay.io/edge-infrastructure/assisted-image-service@sha256:d6396f9360b877a4f57139ba5b4969f243278db709be4de3573f2fba75e50018
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:6271d639e4c0647bd28fbc85cea54f31143a876c3237376b9c670ec254a9c2fc
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:2267d2d7be38f81d530fed977b0c46f6c976369cd89c123c8e036c3b1dd988d5
        version: 0.3.30
      entries:
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.5
    - currentCSV: assisted-service-operator.v0.4.41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.9/4.9.40/rhcos-4.9.40-x86_64-live.x86_64.iso",
                      "version": "49.84.202206171736-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-x86_64-live.x86_64.iso",
                      "version": "410.84.202205191234-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.16/rhcos-4.10.16-aarch64-live.aarch64.iso",
                      "version": "410.84.202205191023-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "411.86.202207090519-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.11.0-rc.2/rhcos-4.11.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "411.86.202207090159-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
          createdAt: "2022-09-11T14:56:07Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.41'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-service@sha256:0fa869095185759c26c391f0e353f293c903ea739576f160d6668f5f20404bdb
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:4ca00219efbc0d9dcce415dadc1039e9495f3a3b00ab3bd575b41554d7934899
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:007a3d43633fe355202002737f766a7c636f537f3daeffa1f62356c0bbf969aa
        - quay.io/edge-infrastructure/assisted-image-service@sha256:104d2aae313c724fb10d3ef3bba78b2370fc0597de31335453894b679a450ab8
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:56484f99799d22f345215bb65c6b81b6ee14e07d26b625c3f283a2ddfdc37238
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:951b9d625bc300b5b72a3916cb382015527ea786f10b7acbeebf83ae854ca51a
        - quay.io/edge-infrastructure/assisted-installer@sha256:803168aeb7e01db77e3b2ff7418ceea3d93daccaceeb95f8fc02027144ebac75
        version: 0.4.41
      entries:
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.6
    - currentCSV: assisted-service-operator.v0.4.50
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "412.86.202208101039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.12.0-ec.2/rhcos-4.12.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "412.86.202208101040-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
          createdAt: "2022-11-23T19:17:47Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.4.50'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-image-service@sha256:71e8c9c76caa8fe01255fdb4a1172cda372969f7fe1399a3a54fb93e967e68cf
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:34e0341c0471b1772db11535107db06e511c8d5d9816edc3971e745928164ecd
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:c1c7281b9f06e56330cc55ac8d74185d42ee82dff105ae747db677b8eb059aea
        - quay.io/edge-infrastructure/assisted-installer@sha256:f225c86242b0bfbf58753123ef33b7c07f367a5099641d19cfda5a7121382586
        - quay.io/edge-infrastructure/assisted-service@sha256:0ca1cfe04f088cd4c7318cb91978754d6472ada29e3436d05e22335740bfff49
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:54b39c6cafefb70d258ec1d491b6a0f1dd888cd85bd442c8f40a4805cf4721ca
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:87594d78977ead074e53b39a52e905a2d2861d24d69d2f2e33b066077cd279a4
        version: 0.4.50
      entries:
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.7
    - currentCSV: assisted-service-operator.v0.6.65
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.8",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.8/4.8.14/rhcos-4.8.14-x86_64-live.x86_64.iso",
                      "version": "48.84.202109241901-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-x86_64-live.x86_64.iso",
                      "version": "411.86.202210041459-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-aarch64-live.aarch64.iso",
                      "version": "411.86.202210032347-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-s390x-live.s390x.iso",
                      "version": "411.86.202210171842-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.9/rhcos-4.11.9-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202210171844-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-x86_64-live.x86_64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-aarch64-live.aarch64.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-s390x-live.s390x.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.10/rhcos-4.12.10-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202303211731-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-x86_64-live.x86_64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-aarch64-live.aarch64.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202303281804-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.13.0-rc.2/rhcos-4.13.0-rc.2-s390x-live.s390x.iso",
                      "version": "413.92.202303281804-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
          createdAt: "2023-05-24T11:15:57Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.6.65'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:a6f99547564cc2e047df74003019817b2dbd3dad992a0c8e53aaaba4f5b02ab6
        - quay.io/edge-infrastructure/assisted-installer@sha256:f8a7e7cb6ff1a4a7daa0cb26f0fab8da17f109653af5979c9d53c35b7ca2c1bb
        - quay.io/edge-infrastructure/assisted-service@sha256:2dfaf2c3ac58b2987b0088df302c7cf717d0b6640cdb8b2327370aa110120784
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:ff8e563d97f0d0e42c690ca3e61d3625e920947264ea99c396392c440cd43201
        - quay.io/edge-infrastructure/assisted-image-service@sha256:b1ba1076fe16b2f868ed31a660b16583a957a83634c7ec3cf03574b56e02144c
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:1ad91b13c763437863480df68f5b65576e95c558d4537886432a81c0b8d3cedc
        version: 0.6.65
      entries:
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.8
    - currentCSV: assisted-service-operator.v0.7.67
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "agent-install.openshift.io/v1beta1",
                "kind": "AgentServiceConfig",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "databaseStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "filesystemStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "100Gi"
                      }
                    }
                  },
                  "imageStorage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "10Gi"
                      }
                    }
                  },
                  "mustGatherImages": [
                    {
                      "name": "cnv",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel8:v2.6.5"
                    },
                    {
                      "name": "ocs",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/ocs4/ocs-must-gather-rhel8"
                    },
                    {
                      "name": "lso",
                      "openshiftVersion": "4.8",
                      "url": "registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel8"
                    }
                  ],
                  "osImages": [
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.9",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.9/4.9.45/rhcos-4.9.45-x86_64-live.x86_64.iso",
                      "version": "49.84.202207192205-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-x86_64-live.x86_64.iso",
                      "version": "410.84.202210040010-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.10",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.10/4.10.37/rhcos-4.10.37-aarch64-live.aarch64.iso",
                      "version": "410.84.202210040011-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-x86_64-live.x86_64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-aarch64-live.aarch64.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-s390x-live.s390x.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.11",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.11/4.11.48/rhcos-4.11.48-ppc64le-live.ppc64le.iso",
                      "version": "411.86.202308081056-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-x86_64-live.x86_64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-aarch64-live.aarch64.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-s390x-live.s390x.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.12",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.12/4.12.30/rhcos-4.12.30-ppc64le-live.ppc64le.iso",
                      "version": "412.86.202308081039-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-x86_64-live.x86_64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-aarch64-live.aarch64.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-ppc64le-live.ppc64le.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.13",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/4.13/4.13.10/rhcos-4.13.10-s390x-live.s390x.iso",
                      "version": "413.92.202307260246-0"
                    },
                    {
                      "cpuArchitecture": "x86_64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/x86_64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-x86_64-live.x86_64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "arm64",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/aarch64/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-aarch64-live.aarch64.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "ppc64le",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/ppc64le/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-ppc64le-live.ppc64le.iso",
                      "version": "414.92.202305090606-0"
                    },
                    {
                      "cpuArchitecture": "s390x",
                      "openshiftVersion": "4.14",
                      "url": "https://mirror.openshift.com/pub/openshift-v4/s390x/dependencies/rhcos/pre-release/4.14.0-ec.2/rhcos-4.14.0-ec.2-s390x-live.s390x.iso",
                      "version": "414.92.202305090606-0"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
          createdAt: "2023-10-08T07:14:20Z"
          description: The Infrastructure Operator for Red Hat OpenShift is responsible
            for managing the deployment of the Assisted Service.
          olm.skipRange: '>=0.0.1 <0.7.67'
          operatorframework.io/suggested-namespace: assisted-installer
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/assisted-service
          support: https://github.com/openshift/assisted-service/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AgentClassification
            name: agentclassifications.agent-install.openshift.io
            version: v1beta1
          - kind: AgentClusterInstall
            name: agentclusterinstalls.extensions.hive.openshift.io
            version: v1beta1
          - displayName: Agent
            kind: Agent
            name: agents.agent-install.openshift.io
            version: v1beta1
          - description: AgentServiceConfig represents an Assisted Service deployment.
              Only an AgentServiceConfig with name="agent" will be reconciled. All
              other names will be rejected.
            displayName: Agent Service Config
            kind: AgentServiceConfig
            name: agentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - kind: HypershiftAgentServiceConfig
            name: hypershiftagentserviceconfigs.agent-install.openshift.io
            version: v1beta1
          - displayName: InfraEnv
            kind: InfraEnv
            name: infraenvs.agent-install.openshift.io
            version: v1beta1
          - displayName: NMStateConfig
            kind: NMStateConfig
            name: nmstateconfigs.agent-install.openshift.io
            version: v1beta1
        description: |-
          The Infrastructure Operator for Red Hat OpenShift is responsible for managing
          the deployment of the Assisted Service. Assisted Service is used to orchestrate
          baremetal OpenShift installations.

          The Assisted Service can currently install clusters with highly-available
          control planes (3 hosts and above) and can also install Single-Node
          OpenShift (SNO).  Highly available clusters are configured to use
          OpenShift's `baremetal` platform (typically used in
          [bare metal IPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_bare_metal_ipi/ipi-install-overview.html)),
          while SNO uses `none` (typically used in
          [UPI deployments](https://docs.openshift.com/container-platform/4.7/installing/installing_platform_agnostic/installing-platform-agnostic.html)).

          When creating the AgentServiceConfig CR. It is important to note that the
          controller will only ever reconcile an AgentServiceConfig named "agent",
          any other name will be ignored.
        displayName: Infrastructure Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - assisted-service
        - assisted-installer
        - OpenShift
        links:
        - name: Assisted Service
          url: https://github.com/openshift/assisted-service
        - name: Assisted Image Service
          url: https://github.com/openshift/assisted-image-service
        - name: Assisted Installer
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Controller
          url: https://github.com/openshift/assisted-installer
        - name: Assisted Installer Agent
          url: https://github.com/openshift/assisted-installer-agent
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/edge-infrastructure/assisted-installer-controller@sha256:35c118d8c07dd72fe3307424cc051d9e0426e916412171449a05e41445b9f599
        - quay.io/edge-infrastructure/assisted-installer@sha256:d85be59bc379927f46d0a88ef04172c5c6733e31e1e4a8188de2d434d5bedb66
        - quay.io/edge-infrastructure/assisted-service@sha256:fb0397a1402d42a64e3128e66a0817aa9466e67eb407cb792dda0d380bd5a189
        - quay.io/edge-infrastructure/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/openshift-community-operators/assisted-service-operator@sha256:12a02131f13e532a26572a524ae542766092ff75624113994d304824951161b4
        - quay.io/edge-infrastructure/assisted-image-service@sha256:69ff087be8e5d8458a97f49d88b8b2025ee341d362dcf5618988c011b66688e0
        - quay.io/edge-infrastructure/assisted-installer-agent@sha256:f2aeb4f7d193ea2c5702cae0f6039c02991b9f417d091468d46974408c97efe4
        version: 0.7.67
      entries:
      - name: assisted-service-operator.v0.7.67
        version: 0.7.67
      - name: assisted-service-operator.v0.7.66
        version: 0.7.66
      - name: assisted-service-operator.v0.6.66
        version: 0.6.66
      - name: assisted-service-operator.v0.6.65
        version: 0.6.65
      - name: assisted-service-operator.v0.6.64
        version: 0.6.64
      - name: assisted-service-operator.v0.6.63
        version: 0.6.63
      - name: assisted-service-operator.v0.6.62
        version: 0.6.62
      - name: assisted-service-operator.v0.6.61
        version: 0.6.61
      - name: assisted-service-operator.v0.5.61
        version: 0.5.61
      - name: assisted-service-operator.v0.5.60
        version: 0.5.60
      - name: assisted-service-operator.v0.5.59
        version: 0.5.59
      - name: assisted-service-operator.v0.5.58
        version: 0.5.58
      - name: assisted-service-operator.v0.5.57
        version: 0.5.57
      - name: assisted-service-operator.v0.5.56
        version: 0.5.56
      - name: assisted-service-operator.v0.5.55
        version: 0.5.55
      - name: assisted-service-operator.v0.5.54
        version: 0.5.54
      - name: assisted-service-operator.v0.5.53
        version: 0.5.53
      - name: assisted-service-operator.v0.5.52
        version: 0.5.52
      - name: assisted-service-operator.v0.5.51
        version: 0.5.51
      - name: assisted-service-operator.v0.5.50
        version: 0.5.50
      - name: assisted-service-operator.v0.4.50
        version: 0.4.50
      - name: assisted-service-operator.v0.4.49
        version: 0.4.49
      - name: assisted-service-operator.v0.4.48
        version: 0.4.48
      - name: assisted-service-operator.v0.4.47
        version: 0.4.47
      - name: assisted-service-operator.v0.4.46
        version: 0.4.46
      - name: assisted-service-operator.v0.4.45
        version: 0.4.45
      - name: assisted-service-operator.v0.4.44
        version: 0.4.44
      - name: assisted-service-operator.v0.4.43
        version: 0.4.43
      - name: assisted-service-operator.v0.4.42
        version: 0.4.42
      - name: assisted-service-operator.v0.4.41
        version: 0.4.41
      - name: assisted-service-operator.v0.4.40
        version: 0.4.40
      - name: assisted-service-operator.v0.4.39
        version: 0.4.39
      - name: assisted-service-operator.v0.4.38
        version: 0.4.38
      - name: assisted-service-operator.v0.3.38
        version: 0.3.38
      - name: assisted-service-operator.v0.3.37
        version: 0.3.37
      - name: assisted-service-operator.v0.3.36
        version: 0.3.36
      - name: assisted-service-operator.v0.3.35
        version: 0.3.35
      - name: assisted-service-operator.v0.3.34
        version: 0.3.34
      - name: assisted-service-operator.v0.3.33
        version: 0.3.33
      - name: assisted-service-operator.v0.3.32
        version: 0.3.32
      - name: assisted-service-operator.v0.3.31
        version: 0.3.31
      - name: assisted-service-operator.v0.3.30
        version: 0.3.30
      - name: assisted-service-operator.v0.3.29
        version: 0.3.29
      - name: assisted-service-operator.v0.3.28
        version: 0.3.28
      - name: assisted-service-operator.v0.3.27
        version: 0.3.27
      - name: assisted-service-operator.v0.3.26
        version: 0.3.26
      - name: assisted-service-operator.v0.3.25
        version: 0.3.25
      - name: assisted-service-operator.v0.2.25
        version: 0.2.25
      - name: assisted-service-operator.v0.2.24
        version: 0.2.24
      - name: assisted-service-operator.v0.2.23
        version: 0.2.23
      - name: assisted-service-operator.v0.2.22
        version: 0.2.22
      - name: assisted-service-operator.v0.2.21
        version: 0.2.21
      - name: assisted-service-operator.v0.2.20
        version: 0.2.20
      - name: assisted-service-operator.v0.2.19
        version: 0.2.19
      - name: assisted-service-operator.v0.2.18
        version: 0.2.18
      - name: assisted-service-operator.v0.2.17
        version: 0.2.17
      - name: assisted-service-operator.v0.2.16
        version: 0.2.16
      - name: assisted-service-operator.v0.2.15
        version: 0.2.15
      - name: assisted-service-operator.v0.2.14
        version: 0.2.14
      - name: assisted-service-operator.v0.2.13
        version: 0.2.13
      - name: assisted-service-operator.v0.2.12
        version: 0.2.12
      - name: assisted-service-operator.v0.2.11
        version: 0.2.11
      - name: assisted-service-operator.v0.2.10
        version: 0.2.10
      - name: assisted-service-operator.v0.2.9
        version: 0.2.9
      - name: assisted-service-operator.v0.2.8
        version: 0.2.8
      - name: assisted-service-operator.v0.2.7
        version: 0.2.7
      - name: assisted-service-operator.v0.2.6
        version: 0.2.6
      - name: assisted-service-operator.v0.2.5
        version: 0.2.5
      - name: assisted-service-operator.v0.2.4
        version: 0.2.4
      - name: assisted-service-operator.v0.2.3
        version: 0.2.3
      - name: assisted-service-operator.v0.2.2
        version: 0.2.2
      - name: assisted-service-operator.v0.2.1
        version: 0.2.1
      - name: assisted-service-operator.v0.2.0
        version: 0.2.0
      - name: assisted-service-operator.v0.1.3
        version: 0.1.3
      - name: assisted-service-operator.v0.1.2
        version: 0.1.2
      - name: assisted-service-operator.v0.1.1
        version: 0.1.1
      - name: assisted-service-operator.v0.1.0
        version: 0.1.0
      - name: assisted-service-operator.v0.0.8
        version: 0.0.8
      - name: assisted-service-operator.v0.0.7
        version: 0.0.7
      - name: assisted-service-operator.v0.0.6
        version: 0.0.6
      - name: assisted-service-operator.v0.0.5
        version: 0.0.5
      - name: assisted-service-operator.v0.0.5-rc.1
        version: 0.0.5-rc.1
      - name: assisted-service-operator.v0.0.4
        version: 0.0.4
      - name: assisted-service-operator.v0.0.3
        version: 0.0.3
      - name: assisted-service-operator.v0.0.2
        version: 0.0.2
      - name: assisted-service-operator.v0.0.1
        version: 0.0.1
      name: ocm-2.9
    defaultChannel: alpha
    packageName: assisted-service-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dpu-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-network-operator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dpu.openshift.io/v1alpha1",
                "kind": "OVNKubeConfig",
                "metadata": {
                  "name": "ovnkubeconfig-sample"
                },
                "spec": {
                  "kubeConfigFile": "tenant-cluster-1-kubeconf",
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/dpu-worker": ""
                    }
                  },
                  "poolName": "dpu"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional, Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
          createdAt: "2023-07-13T18:03:11Z"
          description: The operator is responsible for the life-cycle management of
            the ovn-kube components and the necessary host network initialization
            on DPU cards.
          olm.skipRange: '>=4.10.0-0 <4.13.0-202309181427'
          operatorframework.io/suggested-namespace: openshift-dpu-network-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/dpu-network-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVNKubeConfig is the Schema for the ovnkubeconfigs API
            displayName: OVNKube Config
            kind: OVNKubeConfig
            name: ovnkubeconfigs.dpu.openshift.io
            version: v1alpha1
        description: The operator to be responsible for the life-cycle management
          of the ovn-kube components and the necessary host network initialization
          on DPU cards.
        displayName: DPU Network Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - network
        links:
        - name: Dpu Network Operator
          url: https://dpu-network-operator.domain
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:767682dd3bd65651a8204c9fb732b9c48b99127189992b43abc6ecce027f4589
        - registry.redhat.io/openshift4/dpu-network-operator-bundle@sha256:006edc597eec839f960bfd191b75217fc8c6bfaea9d1694e7ac9fa34266f9b4f
        - registry.redhat.io/openshift4/dpu-network-rhel8-operator@sha256:d07b8ecb5d1625599878b5c1e34b0de94188b81d90c829787a2d13969f6d5971
        version: 4.13.0-202309181427
      entries:
      - name: dpu-network-operator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: stable
    defaultChannel: stable
    packageName: dpu-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-telemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-telemetry-operator.v1.5.1757650583
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infra.watch/v1beta1",
                "kind": "ServiceTelemetry",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "alerting": {
                    "alertmanager": {
                      "receivers": {
                        "snmpTraps": {
                          "alertOidLabel": "oid",
                          "community": "public",
                          "enabled": false,
                          "port": 162,
                          "retries": 5,
                          "target": "192.168.24.254",
                          "timeout": 1,
                          "trapDefaultOid": "1.3.6.1.4.1.50495.15.1.2.1",
                          "trapDefaultSeverity": "",
                          "trapOidPrefix": "1.3.6.1.4.1.50495.15"
                        }
                      },
                      "storage": {
                        "persistent": {
                          "pvcStorageRequest": "20G"
                        },
                        "strategy": "persistent"
                      }
                    },
                    "enabled": true
                  },
                  "backends": {
                    "events": {
                      "elasticsearch": {
                        "certificates": {
                          "caCertDuration": "70080h",
                          "endpointCertDuration": "70080h"
                        },
                        "enabled": false,
                        "forwarding": {
                          "hostUrl": "https://elasticsearch-es-http:9200",
                          "tlsSecretName": "elasticsearch-es-cert",
                          "tlsServerName": "",
                          "useBasicAuth": true,
                          "useTls": true,
                          "userSecretName": "elasticsearch-es-elastic-user"
                        },
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20Gi"
                          },
                          "strategy": "persistent"
                        },
                        "version": "7.16.1"
                      }
                    },
                    "metrics": {
                      "prometheus": {
                        "enabled": true,
                        "scrapeInterval": "30s",
                        "storage": {
                          "persistent": {
                            "pvcStorageRequest": "20G"
                          },
                          "retention": "24h",
                          "strategy": "persistent"
                        }
                      }
                    }
                  },
                  "clouds": [
                    {
                      "metrics": {
                        "collectors": [
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "collectd",
                            "debugEnabled": false,
                            "subscriptionAddress": "collectd/cloud1-telemetry"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 16384,
                              "verbose": false
                            },
                            "collectorType": "ceilometer",
                            "debugEnabled": false,
                            "subscriptionAddress": "anycast/ceilometer/cloud1-metering.sample"
                          },
                          {
                            "bridge": {
                              "ringBufferCount": 15000,
                              "ringBufferSize": 65535,
                              "verbose": false
                            },
                            "collectorType": "sensubility",
                            "debugEnabled": false,
                            "subscriptionAddress": "sensubility/cloud1-telemetry"
                          }
                        ]
                      },
                      "name": "cloud1"
                    }
                  ],
                  "graphing": {
                    "enabled": false,
                    "grafana": {
                      "baseImage": "registry.redhat.io/rhel8/grafana:9",
                      "dashboards": {
                        "enabled": true
                      },
                      "disableSignoutMenu": false,
                      "ingressEnabled": true
                    }
                  },
                  "highAvailability": {
                    "enabled": false
                  },
                  "observabilityStrategy": "use_redhat",
                  "transports": {
                    "qdr": {
                      "auth": "basic",
                      "certificates": {
                        "caCertDuration": "70080h",
                        "endpointCertDuration": "70080h"
                      },
                      "enabled": true,
                      "web": {
                        "enabled": false
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
          createdAt: "2025-09-12T04:17:40Z"
          description: Service Telemetry Framework. Umbrella Operator for instantiating
            the required dependencies and configuration of various components to build
            a Service Telemetry platform for telco grade monitoring.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: service-telemetry
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/service-telemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of the Service Telemetry Framework
            displayName: STF Cluster
            kind: ServiceTelemetry
            name: servicetelemetrys.infra.watch
            version: v1beta1
        description: Service Telemetry Operator for monitoring clouds
        displayName: Service Telemetry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - telecommunications
        links:
        - name: Source Code
          url: https://github.com/infrawatch/service-telemetry-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/17.1/html/service_telemetry_framework_1.5/index
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-alertmanager-rhel8:0.2.0@sha256:cfd18e83d64a55ac310937eefe83fc5b9e98660962a127cd4f8d9a382764c171
        - registry.redhat.io/cluster-observability-operator/coo-prometheus-rhel8:0.2.0@sha256:6bbb1aebb1cfdc3310e398ec7720ca1d4a9a89cecf7406412071bfaceb0d3cb3
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        - registry.redhat.io/stf/prometheus-webhook-snmp-rhel9@sha256:03199f8a3b6b22c82e01ede3adab6e98b276369d642029ae03641aac3b25581a
        - registry.redhat.io/stf/service-telemetry-rhel9-operator@sha256:822becd77fa8835e753c88b3e307f62e42e090bdedc6e1fbcbe9fd6110310846
        - registry.redhat.io/stf/service-telemetry-operator-bundle@sha256:c4ec01d0d27eaf93cf820d8da1f2ad7ef3e557c5b4ed651c5a1d07a1fc116f79
        version: 1.5.1757650583
      entries:
      - name: service-telemetry-operator.v1.5.1757650583
        version: 1.5.1757650583
      - name: service-telemetry-operator.v1.5.1743671601
        version: 1.5.1743671601
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: service-telemetry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: jaeger-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
          createdAt: 25 Aug 2025, 13:30
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.65.0-4'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-distributed-tracing
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/os-observability/konflux-jaeger
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |2-

          ## Operator deprecation warning

          This Operator is deprecated and not recommended for new deployments. This Operator will be removed from the `redhat-operators` catalog in a future release.

          If this is not the first time you are installing it, migrate to the Red Hat build of OpenTelemetry Operator and the Tempo Operator for distributed tracing collection and storage. For more information, see "Migrating" in the Red Hat build of OpenTelemetry documentation, "Installing" in the Red Hat build of OpenTelemetry documentation, and "Installing" in the distributed tracing platform (Tempo) documentation.

          ## About

          Red Hat OpenShift distributed tracing platform based on Jaeger. Jaeger is project inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/). It is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.

          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Elasticsearch, Memory.

          ### Operator features

          * **Multiple modes** - Provides `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).

          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.

          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `elasticsearch`. The operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.

          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).

          * **UI** - Optionally setup secure route to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).

          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.

          ### Support & Troubleshooting

          Red Hat OpenShift distributed tracing Jaeger is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/distributed_tracing/distributed-tracing-platform-jaeger) is recommended when installing, configuring, and managing the Operator and instances.
          Support is provided to Red Hat OpenShift entitled customers subject to the [Production Scope for Coverage](https://access.redhat.com/support/offerings/production/soc) and the [Red Hat OpenShift distributed tracing Life Cycle](https://access.redhat.com/support/policy/updates/openshift#jaeger).
        displayName: Red Hat OpenShift distributed tracing platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:d92003d6cf0b4f3dae11f849355c9179e78e768f3b7dbeaf3224a4a019ec5d5d
        - registry.redhat.io/rhosdt/jaeger-operator-bundle@sha256:f250e39033d7cb1d786e5a7ec6798c25d4c5d8c6ecbcf6828915605fc4658da5
        - registry.redhat.io/rhosdt/jaeger-rhel8-operator@sha256:51a8685e4dc70888c991996f149cd42233049d18f702ec72b066d6dad77a07c9
        - registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:1d586f9aafd1cf847a14f509692ce54956b0e754a9a04beaa7a75c480f86717c
        - registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e77efd450e757772d0753a73062753dabd08e53f21315e9b87f4cc1efa332661
        - registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:667d6545b92aa8381f65369de3b5a6781d1de5b4bcfa58ccde4721b6b0406f3d
        - registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fd8110529daa8cf3e7f9fcf71346adc5f48cb05717ebbad1c25bc61c94f31656
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:7b021ed467d35b2bcae87765dfc64f639b03bf71f7e34bbea1dfad16626847b9
        - registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:1f4a9daf152420d9ca7cf89c38ae5550b12dd274bada389c7b85ea4cbcbc4fb7
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 1.65.0-4
      entries:
      - name: jaeger-operator.v1.65.0-4
        version: 1.65.0-4
      - name: jaeger-operator.v1.65.0-3
        version: 1.65.0-3
      - name: jaeger-operator.v1.65.0-2
        version: 1.65.0-2
      - name: jaeger-operator.v1.65.0-1
        version: 1.65.0-1
      - name: jaeger-operator.v1.62.0-2
        version: 1.62.0-2
      - name: jaeger-operator.v1.62.0-1
        version: 1.62.0-1
      - name: jaeger-operator.v1.57.0-10
        version: 1.57.0-10
      - name: jaeger-operator.v1.57.0-10-0.1730817420.p
        version: 1.57.0-10+0.1730817420.p
      - name: jaeger-operator.v1.57.0-7
        version: 1.57.0-7
      - name: jaeger-operator.v1.57.0-6
        version: 1.57.0-6
      - name: jaeger-operator.v1.57.0-5
        version: 1.57.0-5
      - name: jaeger-operator.v1.53.0-4
        version: 1.53.0-4
      - name: jaeger-operator.v1.53.0-3
        version: 1.53.0-3
      - name: jaeger-operator.v1.51.0-1
        version: 1.51.0-1
      - name: jaeger-operator.v1.47.1-5
        version: 1.47.1-5
      - name: jaeger-operator.v1.42.0-5-0.1687199951.p
        version: 1.42.0-5+0.1687199951.p
      - name: jaeger-operator.v1.42.0-5
        version: 1.42.0-5
      - name: jaeger-operator.v1.34.1-5
        version: 1.34.1-5
      - name: jaeger-operator.v1.30.2
        version: 1.30.2
      name: stable
    defaultChannel: stable
    packageName: jaeger-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-druid-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: druid-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
          description: Stackable Operator for Apache Druid
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/druid-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for DruidClusterSpec via `CustomResource`
            displayName: druidclusters.druid.stackable.tech
            kind: DruidCluster
            name: druidclusters.druid.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Druid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - druid
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/druid-operator@sha256:00d4d45d5709784a9d33ac05b2ebc378762f365a69e177cdbed4cc53e9251c69
        - registry.connect.redhat.com/stackable/stackable-apache-druid-operator@sha256:f0bca5f12bf9e0cc1cd3800c2e23afaeb817ab94dcfe6aa10912be245d3ddec0
        version: 25.7.0
      entries:
      - name: druid-operator.v25.7.0
        version: 25.7.0
      - name: druid-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-druid-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-bigip-ctlr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: alpha
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: beta
    - currentCSV: f5-bigip-ctlr-operator.v1.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cis.f5.com/v1",
                "kind": "F5BigIpCtlr",
                "metadata": {
                  "name": "f5bigipctlr-sample"
                },
                "spec": {
                  "args": {
                    "agent": "as3",
                    "bigip_partition": "<BIGIP_PARTITION>",
                    "bigip_url": "<BIGIP_IP_ADDRESS>",
                    "insecure": true,
                    "log_as3_response": true,
                    "log_level": "<DEBUG_OR_INFO>",
                    "manage_routes": true,
                    "openshift_sdn_name": "<BIGIP_PARTITION/SDN_NAME>",
                    "pool_member_type": "<CLUSTER_OR_NODEPORT_OR_LOADBALANCER>",
                    "route_vserver_addr": "<IP_ADDRESS>",
                    "route_label": "<F5TYPE>"
                  },
                  "bigip_login_secret": "<BIGIP_LOGIN_K8S_SECRET>",
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "k8s-bigip-ctlr",
                    "user": "f5networks"
                  },
                  "ingressClass": {
                    "create": false,
                    "defaultController": false,
                    "ingressClassName": "f5"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "resources": {},
                  "serviceAccount": {
                    "create": true
                  },
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
          createdAt: "2025-05-28T03:01:48Z"
          description: Operator to install F5 Container Ingress Services (CIS) for
            BIG-IP.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/k8s-bigip-ctlr
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This CRD provides kind `F5BigIpCtlr` to configure and deploy
              F5 BIG-IP Controller.
            displayName: F5BigIpCtlr
            kind: F5BigIpCtlr
            name: f5bigipctlrs.cis.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 Container Ingress Services (CIS) for BIG-IP in your OpenShift Cluster. This enables to configure and deploy CIS using Helm Charts.
          ## F5 Container Ingress Services for BIG-IP
          F5 Container Ingress Services (CIS) integrates with container orchestration environments to dynamically create L4/L7 services on F5 BIG-IP systems, and load balance network traffic across the services.
          Monitoring the orchestration API server, CIS is able to modify the BIG-IP system configuration based on changes made to containerized applications.
          ## Documentation
          Refer to F5 documentation
          - CIS on OpenShift (https://clouddocs.f5.com/containers/latest/userguide/openshift/) - OpenShift Routes (https://clouddocs.f5.com/containers/latest/userguide/routes.html)
          ## Prerequisites
          Create BIG-IP login credentials for use with Operator Helm charts. A basic way be,
          ```
          oc create secret generic  <SECRET-NAME>  -n kube-system --from-literal=username=<USERNAME>  --from-literal=password=<PASSWORD>
          ```
          ### Support for Kubernetes networking.k8s.io/v1 Ingress resources
          Create ingress class resource as follows if you are using networking.k8s.io/v1 ingresses in your openshift environment
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/2.x-master/docs/config_examples/ingress/networkingV1/example-default-ingress-class.yaml
          ```
          ### Support for CIS CustomResources
          Install CRDs manually as follows if using CIS CustomResources (VirtualServer/TransportServer/IngressLink):
          ```
          export CIS_VERSION=<cis-version>
          # For example
          # export CIS_VERSION=v2.12.0
          # or
          # export CIS_VERSION=2.x-master
          #
          # the latter if using a CIS image with :latest label
          #
          oc create -f https://raw.githubusercontent.com/F5Networks/k8s-bigip-ctlr/${CIS_VERSION}/docs/config_examples/customResourceDefinitions/customresourcedefinitions.yml
          ```
        displayName: F5 Container Ingress Services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress Controller
        - BIGIP
        - F5
        - container
        - router
        - application
        - delivery
        - controller
        - waf
        - firewall
        - loadbalancer
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/
        - name: Github Repo
          url: https://github.com/F5Networks/k8s-bigip-ctlr/f5-bigip-ctlr-operator
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: beta
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/cntr-ingress-svcs@sha256:7ee53dc08596d4078da5bdb8188f3eda337d44ca8a91976667f2612bf54f9948
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator-bundle@sha256:e91f0a0d542baf43269de8dbc503307eda2f20e951792b4835228f4907758eb4
        - registry.connect.redhat.com/f5networks/k8s-bigip-ctlr-operator@sha256:9436f86d7b9909f5ec85441ed639f35ef996b4bf7b802bae903c74672f9d1959
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 1.21.0
      entries:
      - name: f5-bigip-ctlr-operator.v1.21.0
        version: 1.21.0
      name: stable
    defaultChannel: stable
    packageName: f5-bigip-ctlr-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v0.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:0.8.4
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <0.8.4
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - quay.io/openshift-community-operators/datadog-operator@sha256:b1977170e5991e1f08c49037d6f34873c8543d29d3a2c3692ebb1d66734f5b3e
        - gcr.io/datadoghq/operator:0.8.4
        version: 0.8.4
      entries:
      - name: datadog-operator.v0.8.4
        version: 0.8.4
      - name: datadog-operator.v0.8.2
        version: 0.8.2
      - name: datadog-operator.v0.8.1
        version: 0.8.1
      - name: datadog-operator.v0.8.0
        version: 0.8.0
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: gcr.io/datadoghq/operator:1.18.0
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - gcr.io/datadoghq/operator:1.18.0
        - quay.io/community-operator-pipeline-prod/datadog-operator@sha256:0a5684e5bb442c5daad6737f1560198ca380507f2da583efd0a7d66e18f2dade
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.2
        version: 1.4.2
      - name: datadog-operator.v1.4.1
        version: 1.4.1
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      - name: datadog-operator.v1.2.0
        version: 1.2.0
      - name: datadog-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: datadog-operator
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron-engine.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
          createdAt: "2025-02-24T01:37:53Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/provider-credential-controller@sha256:a5ac3999d0aefb1dd9a239f9394485f93e4252a45eb4a43a43e603aaff80b3bb
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:2dded4ad8e80f711fc18ada34ea825e844045da44f0813550a586371dac84b52
        - quay.io/stolostron/managed-serviceaccount@sha256:0e4edd9ab95c1558fba40a0c41c5cc326996e9e19e296324ea10b5a442a7ef63
        - quay.io/stolostron/addon-manager@sha256:7e347a177d8c5b80a78411a3e8dbf04e49d157a4ca67e7fd5170970c2ba62da9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/cluster-api-provider-agent@sha256:dda3e9aef611c744c4c89c2689a12d0649a173c1a64954c3037de55ec8eacec9
        - quay.io/stolostron/discovery-testserver@sha256:f757a7a11108a88de4bc0ecc5e16f74b3bd9b5110ec94bfeef2074121ec1e47f
        - quay.io/stolostron/registration@sha256:9c221a40ffa119ac154d76f578e0919b5a63137c908afb0583519a9b8a399b2b
        - quay.io/stolostron/assisted-service@sha256:58271e54eda979c43fb474c3bb255f7328e380a7af9ba833e5f7558f323bd239
        - quay.io/stolostron/backplane-operator-test@sha256:b04661ad658423c9c7557b722adb2159d988c10445de1c59f39c44583262cb42
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:65180432218d035ffc7996a7643ddb5300eb8b09e47e09cb270cb9b961aecf7a
        - quay.io/stolostron/assisted-service-el8@sha256:f50f4dece91100f45ef9e20b1e52557a55db45b17a380f32d365ee013d695d5c
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/assisted-installer-controller@sha256:3b0b43fdbbbbd5c1e7588fc449b002e1ed9ad0ea6daa774925a962bd499a4195
        - quay.io/stolostron/managedcluster-import-controller@sha256:ecfe7c1f2d24a75bfb8db2a0a2765fe5e9136f81f341882de896431570fd10e9
        - quay.io/stolostron/cluster-image-set-controller@sha256:629361c1c0ec50fe6410fc6fe6417cba89957c49ab89201c6470c0f97c8fc91d
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:b016a001a53bc37ae534b74cc832552759800746c27e16adcf0626ff9ad2e4c0
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/helloworld@sha256:4838f699ea2461f8358dcda32b3ac96ed05b55d4e1958eac2bdacae4807501d1
        - quay.io/stolostron/backplane-operator@sha256:fa1fa6ef3958fd9d50bfb38197f033a635a05b47c32e653822246e708d8598cc
        - quay.io/stolostron/image-based-install-operator@sha256:fecbff73eb97b28bedd7c17d5adbef4826f181a6838ad4cc72e7478acf9b4254
        - quay.io/stolostron/cluster-proxy@sha256:b341dd7058cdffccb00a026908031fa123fec5ba58ba9307dbb873b75d250e4d
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/multicloud-manager@sha256:0293f0852206562b891375b167cfe17843e2e17585f04c9bb6482d1fe78089d0
        - quay.io/stolostron/placement@sha256:543f09446652bff8c4149945d94d715eff7a76d82783f2bad89e712798542c73
        - quay.io/stolostron/registration-operator@sha256:52289e6cf5f159d344717b49e6226ff9d4ee33942f63cc543d1b820d7ee24fc5
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/work@sha256:6ec081826f32467ea1e1a91f6624df349a382fb8a1cbbda906141281ac2bad5c
        - quay.io/stolostron/hypershift-operator@sha256:2d1809223ce5e75b4133a39b78a269703c6b5234f7ec2a7c0072124c345fae0a
        - quay.io/stolostron/assisted-image-service@sha256:6846719f91dd954ce4b4c6e7c9654c7d79db72990270acd3c34bdd2a89af34fc
        - quay.io/stolostron/backplane-must-gather@sha256:99b558d687e1c17036b164d819cb1c9b96a814be79e3c87960ef7cdb521eee7d
        - quay.io/stolostron/hypershift-addon-operator@sha256:65f2b7a356c22413307ce7b7de38ce00c31e4c43b6ee3b4b484c0bc61dc483e9
        - quay.io/stolostron/discovery-operator@sha256:43f40b9c20b112248fcdb42d114fe4af73fdd45b05d9ff7855d827cdd5e971b4
        - quay.io/stolostron/assisted-installer-agent@sha256:cd74c8b8888c100bf567dd07cfff40097c458d18fcc29a11ed3830f697156267
        - quay.io/stolostron/cluster-proxy-addon@sha256:7da9da3e4ecd1e0df636c9f72f9f30cc7ff24b81e247ba9825cef39f18e86182
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:2781d87c1568f2e40409f5431bfd1a9a4c9ca35655537e7d4db65d9b881edd08
        - quay.io/stolostron/discovery-operator-tests@sha256:21449ecd291cf7203cc8e27a2a00ac60e74f5993708f9e7465e8efdfdca1c89f
        - quay.io/stolostron/hive@sha256:1e1f60492c6a3b43b69fe8a0640f9b81f5a1cfafc0afcd6854c3cfe362ed063a
        - quay.io/stolostron/console-mce@sha256:bfaf9180cd1e5af4d34daf4f1dcdee89951067a605f21bfe6e1ce6bda85b0334
        - quay.io/stolostron/clusterclaims-controller@sha256:b796cb2edd3d4c22b29fc6d2b45e5328a59b7ec1bdc52d5500e1a4ee74725d6c
        - quay.io/stolostron/assisted-installer@sha256:c88976ab4b35d458a39044e522d2d5ab1916d8485af07bbef373af1bcbbd6af0
        - quay.io/stolostron/cluster-curator-controller@sha256:5d05462c2c9d93b21656a436c01fbda4256757bd057e5de7014eb6f320310a1e
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        version: 0.6.3
      entries:
      - name: stolostron-engine.v0.6.3
        version: 0.6.3
      - name: stolostron-engine.v0.6.2
        version: 0.6.2
      - name: stolostron-engine.v0.6.1
        version: 0.6.1
      - name: stolostron-engine.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron-engine.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
          createdAt: "2025-06-09T15:04:58Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: stolostron-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: 'Stolostron Engine provides the foundational components that
          are necessary for the centralized management of multiple Kubernetes-based
          clusters across data centers, public clouds, and private clouds. You can
          use the engine to create Red Hat OpenShift Container Platform clusters on
          selected providers, or import existing Kubernetes-based clusters. After
          the clusters are managed, you can use the APIs that are provided by the
          engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters. '
        displayName: Stolostron Engine
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multiclusterengine
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/clusterclaims-controller@sha256:afab145b12a52f81e1f2eaa189731e3a6287b54f2f1f1ad2976ad378fa7dfa3a
        - quay.io/stolostron/ose-baremetal-cluster-api-controllers-rhel9@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/apiserver-network-proxy@sha256:dc74c7c67ca851f50b83d7a03f91d028214fd8a941f12edd4411b860dc7b763d
        - quay.io/stolostron/assisted-installer-agent@sha256:ffceb06ce038013291e092324d3c49091cd1b2ff1569abe77c29484075e28107
        - quay.io/stolostron/assisted-service@sha256:af3532d487e03a077ce23e4178abbcb41cf45726b646d7660840c036677715e8
        - quay.io/stolostron/backplane-must-gather@sha256:f4f9596eade78333628199f62f8e71e4c2b3957c20da5bad807c12a2a02560db
        - quay.io/stolostron/cluster-api-provider-agent@sha256:2b22890b948ad1d597c424d73d4fe34c7c4d265ad6e54fa0ed87a7dec920ac57
        - quay.io/stolostron/managedcluster-import-controller@sha256:4d0be693db0576eb2e9e3d409bf1479cc60784f5747ed1b25b6a2443c5e91e36
        - quay.io/stolostron/clusterlifecycle-state-metrics@sha256:1e95f55d6918263e79de1c1e2be8a96cd3d35a346dde6d3e2fb484f48e567811
        - quay.io/stolostron/ose_cluster_api_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/addon-manager@sha256:a245847f135f96c4b4fcaa673fc32013c8e407252b3259175c796e3c16111c71
        - quay.io/stolostron/hypershift-addon-operator@sha256:ed1c1580fee28310fc1b137c5c70271448e58ed29af4371515fe9c74512a2c7e
        - quay.io/stolostron/assisted-installer@sha256:17b78d73775a93b31ca872db79520ddc49d2be042fdf1599f419fbcbfd9a98ac
        - quay.io/stolostron/cluster-api-provider-azure@sha256:f6e780adc1bf3fe64303fcb8c5d080ab639e62c62d0577a1bbae5b5f5bfa393d
        - quay.io/stolostron/discovery-testserver@sha256:faef24208a65283b4157521079a7bf23bc4ebd553d70a83a78c324b5cb0e24e4
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-control-plane@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api-aws-controller@sha256:22a57cf5b74d0692025f0b0f57c5f28f8b492ea13295e9101b63f57766ffff38
        - quay.io/stolostron/discovery-operator-tests@sha256:d96fe1f4f49ecfd4a27bedd87d7070a3c7f3f82b6ee8bb59e98a262d8cedcfc0
        - quay.io/stolostron/discovery-operator@sha256:a6a37237bc06f0635a936b803d3c53f0f0caf5426c76b37951c4e157b181ecb3
        - quay.io/stolostron/helloprow-go@sha256:e055ab3773a248c22dbde9847a3289e8512ccc85c6e9015544fc4f59d14b9cad
        - quay.io/stolostron/foundation-e2e@sha256:b5502848bee71b7fb9f4b1a54f00ccbdf77a9ecbb5e3e7bfad815057f2c55e2f
        - quay.io/stolostron/image-based-install-operator@sha256:9d6796007b616de60560f55f14546295c4e48b3768d059098fce9367674a3368
        - quay.io/stolostron/aws-encryption-provider@sha256:83f11e28004fcbb31d7ec6c34ef2d2ec7dd8fc175c5e0135f6ace9894c65bfd0
        - quay.io/stolostron/console-mce@sha256:7f538a21eb22ccd57765f47692da23be9cecfd3b7b65986f1413234bdf709530
        - quay.io/stolostron/backplane-operator@sha256:9b39a5b6a502778d7884e276b276a9b44e6dbc88e7e82470e26d91ded3cb3839
        - quay.io/stolostron/cluster-image-set-controller@sha256:7f05ad13ffaa11daec066c5889b87df1b631e6eb3dc1c2913a2766c8e65218aa
        - quay.io/stolostron/backplane-operator-test@sha256:8d6514fc0db5ee79c346ddc31bb094ad87fa375519da67e32398300d70d72f6e
        - quay.io/stolostron/managed-serviceaccount-e2e@sha256:e983d2cd7f240dbbe47612055185873ffa63e5ac34083732557cdf2c20bc8a1a
        - quay.io/stolostron/registration-operator@sha256:a9fdf837e149c8659fc99deab6a284975c8eb8785549381fa082303cdb172e44
        - quay.io/stolostron/assisted-service-el8@sha256:f52c55691850a1f2f254b0e9377ba97493be337e012223e5a2b3190f03e1a349
        - quay.io/stolostron/assisted-image-service@sha256:b6df9c9fc8bbfdc58e2169b9e199e1eed188d3d847511eba01eb4920fabe3992
        - quay.io/stolostron/cluster-proxy@sha256:1528e36dfa4fa2c9063d84c0f2b9aa8d2b549a750e812d4a2e2b8c7515bfe5cb
        - quay.io/stolostron/registration@sha256:2db2c28af25007c77979cc456d929f5f077cd671cc4e0562c43efce2c42c4cd8
        - quay.io/stolostron/placement@sha256:6276e411c9f7cf7072d3d364569b6c3cd472742ab4ad48912f18cb4091d27a6f
        - quay.io/stolostron/multicloud-manager@sha256:5a95a5d1dbb17456d7ce62f1a517ae49ca780e25f3ebb74b32346de1df23b4c9
        - quay.io/stolostron/cluster-api-provider-kubevirt@sha256:44f8bb72b975456d4151a88a560fc525ad6f4da7968a38e94a86b6dd77d3c543
        - quay.io/stolostron/ip-address-manager@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/work@sha256:709c41d1a7a1eaf10b334dd0c0105a261d6db811693d5ae67eb55b9a26daed87
        - quay.io/stolostron/provider-credential-controller@sha256:3484cb0c2102f4a522efe31bc4c77910a202b42e1ad082a0a8709cc1a7698649
        - quay.io/stolostron/assisted-installer-controller@sha256:8ea9d12fa30825fe4f965a7d506af687e8de709f8d07dd75204747da2a6e09a8
        - quay.io/stolostron/ose_aws_cluster_api_controllers_rhel9@sha256:5bdd5905d5a6c6a611df6bf0aa0caf22eb768c938b52ae6a733ed5277c6e1c22
        - quay.io/stolostron/cluster-api-provider-openshift-assisted-bootstrap@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/cluster-api@sha256:7a04ac4e5bdaf29498af75b2b7c4ff19cb6223f2076c9373a3a955553243ddda
        - quay.io/stolostron/cluster-proxy-addon@sha256:c776405c4b1b623500c69a12b7d76563646b27cf23bab74dc8f1361534cfb81e
        - quay.io/stolostron/hive@sha256:86ff94d3e7c174ac4d6818183c66791edc7638965dd09e2d41172609c81f451f
        - quay.io/stolostron/cluster-curator-controller@sha256:d667bd0892812eb4e1f724679759469db335db4ab801f9d57ea8818d8cd13188
        - quay.io/stolostron/postgresql-12-centos7@sha256:94727d70e0afbf4e167e078744f3a10ac9d82edc553d57b0ecbb5443264f07e1
        - quay.io/stolostron/hypershift-addon-operator-canary-test@sha256:1d4221cfc364743b58e15bad3afe9548f4151f06b0dbf0aec7c00715ba75abba
        - quay.io/stolostron/kube-rbac-proxy-mce@sha256:1757b9ad635839087c4bef198d5e81f4abb7aee38bc74690021f9e8887366bcf
        - quay.io/stolostron/managed-serviceaccount@sha256:4bcd5c9b2044254c28215fd78bf1e3fcd919578238986831fcea8055b2764696
        - quay.io/stolostron/helloworld@sha256:a790795409f6e781f506b6a63f39f7d344f01efa30c3de87b1727dc2010d5bbf
        - quay.io/stolostron/hypershift-operator@sha256:c9a8fc943ef35126434cc767f152abd09943a4d2d194a9966495fdc0870b417c
        - quay.io/community-operator-pipeline-prod/stolostron-engine@sha256:a20986fafd0a475a5a77296b3f8a18c78edb84da5cc8246b06c805dd5f693511
        version: 0.7.1
      entries:
      - name: stolostron-engine.v0.7.1
        version: 0.7.1
      - name: stolostron-engine.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hawtio-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.14.0",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/hawtio/operator:0.5.0
          createdAt: "2022-04-08T00:00:00Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Hawtio Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - docker.io/hawtio/operator:0.5.0
        - quay.io/openshift-community-operators/hawtio-operator@sha256:38b36745547a08afd10e1620f8243e6f1ce18a3eb8883ec78cd631577177c5ee
        version: 0.5.0
      entries:
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: alpha
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: latest
    - currentCSV: hawtio-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/hawtio/operator:1.2.0
          createdAt: "2024-12-02T17:22:48Z"
          description: Hawtio eases the discovery and management of Java applications
            deployed on OpenShift.
          olm.skipRange: '>=1.0.0 <1.0.2'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Hawtio console that eases the discovery and management
              of hawtio-enabled applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Hawtio
          ==============

          The Hawtio console eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between the Hawtio console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#jvm-agent).
        displayName: Hawtio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Hawtio Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The Hawtio team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hawtio-operator@sha256:306ab7846b4c65dd9cb21d903a6cfc9b6f8f2ad409fffbdd136b30ad5989a255
        - quay.io/hawtio/operator:1.2.0
        version: 1.2.0
      entries:
      - name: hawtio-operator.v1.2.0
        version: 1.2.0
      - name: hawtio-operator.v1.1.1
        version: 1.1.1
      - name: hawtio-operator.v1.1.0
        version: 1.1.0
      - name: hawtio-operator.v1.0.1
        version: 1.0.1
      - name: hawtio-operator.v1.0.0
        version: 1.0.0
      - name: hawtio-operator.v0.5.0
        version: 0.5.0
      - name: hawtio-operator.v0.4.0
        version: 0.4.0
      name: stable-v1
    defaultChannel: stable-v1
    packageName: hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: helloworld-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: helloworld-operator.v4.18.0-202506160933
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.example.com/v1alpha1",
                "kind": "HelloWorld",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "helloworld-operator"
                  },
                  "name": "helloworld-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-05-18T20:30:40Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202506160933'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HelloWorld is the Schema for the helloworlds API
            displayName: Hello World
            kind: HelloWorld
            name: helloworlds.apps.example.com
            version: v1alpha1
        description: helloworld-operator
        displayName: helloworld-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - helloworld-operator
        links:
        - name: Helloworld Operator
          url: https://helloworld-operator.domain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/helloworld-operator-bundle@sha256:056744dd517acc3c12908e5ce06bfffe2c634d427b837e244dd7185c70915c00
        - registry.redhat.io/openshift4/helloworld-operator-rhel9@sha256:7c9a290bef3368d01e44bbecafdd99d2e18376221557938ab82c483bf189f370
        version: 4.18.0-202506160933
      entries:
      - name: helloworld-operator.v4.18.0-202506160933
        version: 4.18.0-202506160933
      - name: helloworld-operator.v4.18.0-202506090359
        version: 4.18.0-202506090359
      - name: helloworld-operator.v4.18.0-202506050817
        version: 4.18.0-202506050817
      name: stable
    defaultChannel: stable
    packageName: helloworld-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:13caed923cfbf9d5ff215c8f87fc64ecd753d183f5573c31ea89c9bc5915c6e1
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-xtradb-cluster-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle-rhmp@sha256:c47f099ae5ddd9b25fb7aa5ecfc253fdf24df2e7929318bbcf162583df91f886
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified-rhmp.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-superset-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: superset-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
          description: Stackable Operator for Apache Superset
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/superset-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SupersetClusterSpec via `CustomResource`
            displayName: supersetclusters.superset.stackable.tech
            kind: SupersetCluster
            name: supersetclusters.superset.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for DruidConnectionSpec via `CustomResource`
            displayName: druidconnections.superset.stackable.tech
            kind: DruidConnection
            name: druidconnections.superset.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Superset
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - superset
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/superset-operator@sha256:a373cfa27d482e47b489bca457fe4107a3999be1707e9617a5b3617d0cd1f408
        - registry.connect.redhat.com/stackable/stackable-apache-superset-operator@sha256:2dcb87ee6a11a84b7ac13e3818614e7e6a9692ac766d508c588c09420a0b8776
        version: 25.7.0
      entries:
      - name: superset-operator.v25.7.0
        version: 25.7.0
      - name: superset-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-superset-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Saison Technology Co., Ltd.
      provider-url: https://www.hulft.com/
    name: hulft
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hulft.v10.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hulft.saison-technology.com/v1",
                "kind": "Hulft",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hulft",
                    "app.kubernetes.io/instance": "sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hulft",
                    "app.kubernetes.io/part-of": "hulft"
                  },
                  "name": "sample",
                  "namespace": "user-apps"
                },
                "spec": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    }
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "observe": {
                      "websocket": {
                        "allowExternalAccess": false
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": false
                    }
                  }
                },
                "status": {
                  "applicationName": "hulft-apps",
                  "control": {
                    "allowExternalAccess": true,
                    "apiClientTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                    "consoleLoginSecret": "secret-hulft-sample-console-login",
                    "consoleTlsSecret": "secret-hulft-sample-control-console-tls",
                    "hostname": "",
                    "httpPort": 30080,
                    "httpsPort": 30443,
                    "replicas": {
                      "max": 1,
                      "min": 1
                    },
                    "samlSecret": ""
                  },
                  "db": {
                    "authorizationSecret": "secret-hulft-db-authorization",
                    "connectionRetryCount": 30,
                    "encryptionKeySecret": "secret-hulft-sample-db-encryption-key",
                    "hostname": "database.example.com",
                    "name": "hulft",
                    "port": 3306,
                    "tablePrefix": "hulft"
                  },
                  "licenseSecret": "secret-hulft-license",
                  "resourceIdentifier": "sample",
                  "transfer": {
                    "control": {
                      "clientTlsSecret": "secret-hulft-sample-control-console-tls",
                      "port": 30443,
                      "retryCount": 10,
                      "retryInterval": 10000,
                      "serverTlsSecret": "secret-hulft-sample-transfer-control-server-tls",
                      "timeout": 60000
                    },
                    "gracefulShutdownWait": 3600,
                    "hostname": "hulft-sample-transfer-wss-ci.apps.openshift.example.com",
                    "hulft": {
                      "port": 30000
                    },
                    "observe": {
                      "port": 31000,
                      "websocket": {
                        "allowExternalAccess": true,
                        "port": 33443,
                        "serverTlsSecret": "secret-hulft-sample-transfer-observe-wss-server-tls"
                      }
                    },
                    "replicas": {
                      "max": 5,
                      "min": 1
                    },
                    "s3": {
                      "authenticationSecret": "",
                      "defaultRegion": "ap-northeast-1"
                    },
                    "websocket": {
                      "allowExternalAccess": true,
                      "clientTlsSecret": "",
                      "port": 31443,
                      "serverTlsSecret": "secret-hulft-sample-transfer-wss-server-tls"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
          createdAt: "2025-05-30T01:26:40Z"
          description: HULFT10 for Container Platform is file integration middleware
            that provides transfer functions for the data accumulated in your IT system
            during the daily operations of your business, such as sales data, customer
            information, and various logs. File transfer from/to on-premises HULFT
            is also available.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: To use HULFT10 for Container
            Platform, you must purchase the product and have a key issued. Please
            refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS)
            for the deployment procedure.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Saison Technology Co., Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hulft is the Schema for the hulfts API
            displayName: Hulft
            kind: Hulft
            name: hulfts.hulft.saison-technology.com
            version: v1
        description: |
          HULFT10 for Container Platform is file integration middleware that provides transfer functions for the data accumulated in your IT system during the daily operations of your business, such as sales data, customer information, and various logs. File transfer from/to on-premises HULFT is also available.
          To use HULFT10 for Container Platform, you must purchase the product and have a key issued. Please refer to [Start Up Guide](https://www.hulft.com/help/ja-jp/HULFT-V10/CONP-SUGOS) for the deployment procedure.
        displayName: HULFT10 for Container Platform
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backup & Recovery
        - Storage
        - Migration
        links:
        - name: Product page
          url: https://www.hulft.com/software/hulft-10
        - name: Product manual
          url: https://www.hulft.com/help/ja-jp/HULFT-V10/index.htm
        maintainers:
        - email: salessupport_mp@hulft.com
          name: Saison Technology Co., Ltd.
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Saison Technology Co., Ltd.
          url: https://www.hulft.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/saison-technology/hulft-bundle@sha256:b7d1348b16ae564bc1092dbdb2042bc2a1ee8a3b7fdc85b7d6fe18be9bb99c19
        - registry.connect.redhat.com/saison-technology/hulft-operator@sha256:d466c27045043c663104b53c273cd7e5a98bbe2d69d82b3a5f1547fd461fb545
        version: 10.4.0
      entries:
      - name: hulft.v10.4.0
        version: 10.4.0
      - name: hulft.v10.3.0
        version: 10.3.0
      name: stable
    defaultChannel: stable
    packageName: hulft
    provider:
      name: Saison Technology Co., Ltd.
      url: https://www.hulft.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hdfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hdfs-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
          description: Stackable Operator for Apache Hadoop HDFS
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hdfs-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HdfsClusterSpec via `CustomResource`
            displayName: hdfsclusters.hdfs.stackable.tech
            kind: HdfsCluster
            name: hdfsclusters.hdfs.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hadoop HDFS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hdfs
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hdfs-operator@sha256:225c5e8029cb7a131e30dd108a4c0dd3800372d4964163a7b1043cfcae4213e8
        - quay.io/stackable/hdfs-operator@sha256:4bf788d7bec9fb2dbd8a8de0ccaddaafee7b4908cc8d47945e8f005eb14f3e0f
        version: 25.7.0
      entries:
      - name: hdfs-operator.v25.7.0
        version: 25.7.0
      - name: hdfs-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hdfs-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLABS
      provider-url: nlp.johnsnowlabs.com
    name: annotationlab
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: annotationlab.v5.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Annotationlab",
                "metadata": {
                  "name": "annotationlab"
                },
                "spec": {
                  "activelearning": {
                    "extraEnv": "",
                    "extraVolumeMounts": "",
                    "extraVolumes": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/johnsnowlabs/annotationlab",
                      "tag": "active-learning-5.7.0"
                    },
                    "lifecycleHooks": ""
                  },
                  "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"annotationlab.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: role\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"annotationlab.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: role\n              operator: NotIn\n              values:\n                - test\n        topologyKey: failure-domain.beta.kubernetes.io/zone\n",
                  "airflow": {
                    "activelearning": {
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "active-learning-5.7.0"
                      }
                    },
                    "affinity": {},
                    "airflowConfigAnnotations": {},
                    "airflowHome": "/opt/airflow",
                    "airflowLocalSettings": "{{- if semverCompare \"\u003e=2.2.0\" .Values.airflowVersion }}\n{{- if not (or .Values.webserverSecretKey .Values.webserverSecretKeySecretName) }}\nfrom airflow.www.utils import UIAlert\n\nDASHBOARD_UIALERTS = [\n  UIAlert(\n    'Usage of a dynamic webserver secret key detected. We recommend a static webserver secret key instead.'\n    ' See the \u003ca href='\n    '\"https://airflow.apache.org/docs/helm-chart/stable/production-guide.html#webserver-secret-key\"\u003e'\n    'Helm Chart Production Guide\u003c/a\u003e for more details.',\n    category=\"warning\",\n    roles=[\"Admin\"],\n    html=True,\n  )\n]\n{{- end }}\n{{- end }}",
                    "airflowPodAnnotations": {},
                    "airflowVersion": "2.5.1",
                    "allowPodLaunching": true,
                    "cleanup": {
                      "affinity": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow kubernetes cleanup-pods --namespace={{ .Release.Namespace }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "failedJobsHistoryLimit": null,
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "resources": {},
                      "schedule": "*/15 * * * *",
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "successfulJobsHistoryLimit": null,
                      "tolerations": [],
                      "topologySpreadConstraints": []
                    },
                    "config": {
                      "api": {
                        "auth_backend": "airflow.api.auth.backend.default"
                      },
                      "celery": {
                        "flower_url_prefix": "{{ .Values.ingress.flower.path }}",
                        "worker_concurrency": 16
                      },
                      "celery_kubernetes_executor": {
                        "kubernetes_queue": "kubernetes"
                      },
                      "core": {
                        "colored_console_log": "False",
                        "dags_are_paused_at_creation": "False",
                        "dags_folder": "{{ include \"airflow_dags\" . }}",
                        "executor": "{{ .Values.executor }}",
                        "load_examples": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "elasticsearch": {
                        "json_format": "True",
                        "log_id_template": "{dag_id}_{task_id}_{execution_date}_{try_number}"
                      },
                      "elasticsearch_configs": {
                        "max_retries": 3,
                        "retry_timeout": "True",
                        "timeout": 30
                      },
                      "kerberos": {
                        "ccache": "{{ .Values.kerberos.ccacheMountPath }}/{{ .Values.kerberos.ccacheFileName }}",
                        "keytab": "{{ .Values.kerberos.keytabPath }}",
                        "principal": "{{ .Values.kerberos.principal }}",
                        "reinit_frequency": "{{ .Values.kerberos.reinitFrequency }}"
                      },
                      "kubernetes": {
                        "airflow_configmap": "{{ include \"airflow_config\" . }}",
                        "airflow_local_settings_configmap": "{{ include \"airflow_config\" . }}",
                        "multi_namespace_mode": "{{ ternary \"True\" \"False\" .Values.multiNamespaceMode }}",
                        "namespace": "{{ .Release.Namespace }}",
                        "pod_template_file": "{{ include \"airflow_pod_template_file\" . }}/pod_template_file.yaml",
                        "worker_container_repository": "{{ .Values.images.airflow.repository | default .Values.defaultAirflowRepository }}",
                        "worker_container_tag": "{{ .Values.images.airflow.tag | default .Values.defaultAirflowTag }}"
                      },
                      "logging": {
                        "colored_console_log": "False",
                        "remote_logging": "{{- ternary \"True\" \"False\" .Values.elasticsearch.enabled }}"
                      },
                      "metrics": {
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "scheduler": {
                        "run_duration": 41460,
                        "standalone_dag_processor": "{{ ternary \"True\" \"False\" .Values.dagProcessor.enabled }}",
                        "statsd_host": "{{ printf \"%s-statsd\" .Release.Name }}",
                        "statsd_on": "{{ ternary \"True\" \"False\" .Values.statsd.enabled }}",
                        "statsd_port": 9125,
                        "statsd_prefix": "airflow"
                      },
                      "webserver": {
                        "authenticate": "False",
                        "base_url": "http://0.0.0.0:8080",
                        "enable_proxy_fix": "True",
                        "rbac": "False"
                      }
                    },
                    "createUserJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"users create\" \"create_user\" }} \"$@\"",
                        "--",
                        "-r",
                        "{{ .Values.webserver.defaultUser.role }}",
                        "-u",
                        "{{ .Values.webserver.defaultUser.username }}",
                        "-e",
                        "{{ .Values.webserver.defaultUser.email }}",
                        "-f",
                        "{{ .Values.webserver.defaultUser.firstName }}",
                        "-l",
                        "{{ .Values.webserver.defaultUser.lastName }}",
                        "-p",
                        "{{ .Values.webserver.defaultUser.password }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "labels": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-create-user-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "dagProcessor": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow dag-processor"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "dags": {
                      "gitSync": {
                        "branch": "v2-2-stable",
                        "containerName": "git-sync",
                        "depth": 1,
                        "enabled": false,
                        "env": [],
                        "extraVolumeMounts": [],
                        "maxFailures": 0,
                        "repo": "https://github.com/apache/airflow.git",
                        "resources": {},
                        "rev": "HEAD",
                        "securityContext": {},
                        "subPath": "tests/dags",
                        "uid": 65533,
                        "wait": 5
                      },
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "existingClaim": null,
                        "size": "1Gi",
                        "storageClassName": null,
                        "subPath": null
                      }
                    },
                    "data": {
                      "brokerUrl": null,
                      "brokerUrlSecretName": null,
                      "metadataConnection": {
                        "db": "airflow",
                        "pass": "airflow",
                        "port": 5432,
                        "protocol": "postgresql",
                        "sslmode": "disable",
                        "user": "airflow"
                      },
                      "metadataSecretName": null,
                      "resultBackendConnection": null,
                      "resultBackendSecretName": null
                    },
                    "defaultAirflowRepository": "apache/airflow",
                    "defaultAirflowTag": "2.5.1",
                    "elasticsearch": {
                      "connection": {},
                      "enabled": false,
                      "secretName": null
                    },
                    "enableBuiltInSecretEnvVars": {
                      "AIRFLOW_CONN_AIRFLOW_DB": true,
                      "AIRFLOW__CELERY__BROKER_URL": true,
                      "AIRFLOW__CELERY__CELERY_RESULT_BACKEND": true,
                      "AIRFLOW__CELERY__RESULT_BACKEND": true,
                      "AIRFLOW__CORE__FERNET_KEY": true,
                      "AIRFLOW__CORE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__DATABASE__SQL_ALCHEMY_CONN": true,
                      "AIRFLOW__ELASTICSEARCH__ELASTICSEARCH_HOST": true,
                      "AIRFLOW__ELASTICSEARCH__HOST": true,
                      "AIRFLOW__WEBSERVER__SECRET_KEY": true
                    },
                    "env": [],
                    "executor": "CeleryExecutor",
                    "extraConfigMaps": {},
                    "extraEnv": "- name: ACTIVE_LEARNING_IMAGE_PULL_POLICY\n  value: \"{{ .Values.activelearning.image.pullPolicy }}\"\n- name: ACTIVE_LEARNING_IMAGE\n  value: \"{{ .Values.activelearning.image.repository }}:{{ .Values.activelearning.image.tag }}\"\n- name: NUMBER_OF_PREANNOTATOR_DAG\n  value: {{ .Values.preannotator_queues.count | quote }}\n- name: AIRFLOW__LOGGING__LOGGING_LEVEL\n  value: INFO\n- name: ANNOTATIONLAB_FULLNAME\n  value: {{ include \"annotationlab.annotationlab.fullname\" . }}\n- name: ANNOTATIONLAB_URL\n  value: http://{{ include \"annotationlab.annotationlab.fullname\" . }}:8200\n- name: AIRFLOW_NAMESPACE\n  value: {{ .Release.Namespace }}\n- name: DATABASE_CONNECTION_STRING\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"annotationlab.annotationlab.fullname\" . }}-secret\n      key: DATABASE_CONNECTION_STRING\n",
                    "extraEnvFrom": null,
                    "extraSecrets": {},
                    "fernetKey": null,
                    "fernetKeySecretName": null,
                    "flower": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery flower\" \"flower\" }}"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.flowerUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "password": null,
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "secretName": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "flower-ui",
                            "port": "{{ .Values.ports.flowerUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "username": null
                    },
                    "fullnameOverride": "",
                    "gid": 0,
                    "global": {},
                    "image": {
                      "repository": "docker.io/johnsnowlabs/annotationlab"
                    },
                    "images": {
                      "airflow": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": "dataflows-5.7.0"
                      },
                      "flower": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "gitSync": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "k8s.gcr.io/git-sync/git-sync",
                        "tag": "v3.6.3"
                      },
                      "migrationsWaitTimeout": 60,
                      "pgbouncer": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-2021.04.28-1.14.0"
                      },
                      "pgbouncerExporter": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "apache/airflow",
                        "tag": "airflow-pgbouncer-exporter-2021.09.22-0.12.0"
                      },
                      "pod_template": {
                        "pullPolicy": "IfNotPresent",
                        "repository": null,
                        "tag": null
                      },
                      "redis": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/redis",
                        "tag": "alpine3.16"
                      },
                      "statsd": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "quay.io/prometheus/statsd-exporter",
                        "tag": "v0.22.8"
                      },
                      "useDefaultImageForMigration": false
                    },
                    "ingress": {
                      "enabled": false,
                      "flower": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      },
                      "web": {
                        "annotations": {},
                        "enabled": false,
                        "host": "",
                        "hosts": [],
                        "ingressClassName": "",
                        "path": "/",
                        "pathType": "ImplementationSpecific",
                        "precedingPaths": [],
                        "succeedingPaths": [],
                        "tls": {
                          "enabled": false,
                          "secretName": ""
                        }
                      }
                    },
                    "kerberos": {
                      "ccacheFileName": "cache",
                      "ccacheMountPath": "/var/kerberos-ccache",
                      "config": "# This is an example config showing how you can use templating and how \"example\" config\n# might look like. It works with the test kerberos server that we are using during integration\n# testing at Apache Airflow (see `scripts/ci/docker-compose/integration-kerberos.yml` but in\n# order to make it production-ready you must replace it with your own configuration that\n# Matches your kerberos deployment. Administrators of your Kerberos instance should\n# provide the right configuration.\n\n[logging]\ndefault = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_libs.log\"\nkdc = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kerberos_kdc.log\"\nadmin_server = \"FILE:{{ template \"airflow_logs_no_quote\" . }}/kadmind.log\"\n\n[libdefaults]\ndefault_realm = FOO.COM\nticket_lifetime = 10h\nrenew_lifetime = 7d\nforwardable = true\n\n[realms]\nFOO.COM = {\n  kdc = kdc-server.foo.com\n  admin_server = admin_server.foo.com\n}\n",
                      "configPath": "/etc/krb5.conf",
                      "enabled": false,
                      "keytabBase64Content": null,
                      "keytabPath": "/etc/airflow.keytab",
                      "principal": "airflow@FOO.COM",
                      "reinitFrequency": 3600
                    },
                    "kubeVersionOverride": "",
                    "labels": {},
                    "limits": [],
                    "logs": {
                      "persistence": {
                        "enabled": false,
                        "existingClaim": null,
                        "size": "100Gi",
                        "storageClassName": null
                      }
                    },
                    "migrateDatabaseJob": {
                      "affinity": {},
                      "annotations": {},
                      "applyCustomEnv": true,
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"db upgrade\" \"upgradedb\" }}"
                      ],
                      "command": null,
                      "enabled": true,
                      "extraContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "jobAnnotations": {},
                      "nodeSelector": {},
                      "resources": {},
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-migrate-database-job"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "useHelmHooks": true
                    },
                    "multiNamespaceMode": false,
                    "nameOverride": "",
                    "networkPolicies": {
                      "enabled": false
                    },
                    "nodeSelector": {},
                    "pgbouncer": {
                      "affinity": {},
                      "annotations": {},
                      "args": null,
                      "auth_file": "/etc/pgbouncer/users.txt",
                      "auth_type": "md5",
                      "ciphers": "normal",
                      "command": [
                        "pgbouncer",
                        "-u",
                        "nobody",
                        "/etc/pgbouncer/pgbouncer.ini"
                      ],
                      "configSecretName": null,
                      "enabled": false,
                      "extraIni": null,
                      "extraIniMetadata": null,
                      "extraIniResultBackend": null,
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "logConnections": 0,
                      "logDisconnections": 0,
                      "maxClientConn": 100,
                      "metadataPoolSize": 10,
                      "metricsExporterSidecar": {
                        "resources": {},
                        "sslmode": "disable"
                      },
                      "nodeSelector": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "resultBackendPoolSize": 5,
                      "revisionHistoryLimit": null,
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": null
                      },
                      "ssl": {
                        "ca": null,
                        "cert": null,
                        "key": null
                      },
                      "sslmode": "prefer",
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534,
                      "verbose": 0
                    },
                    "podTemplate": null,
                    "ports": {
                      "airflowUI": 8080,
                      "flowerUI": 5555,
                      "pgbouncer": 6543,
                      "pgbouncerScrape": 9127,
                      "redisDB": 6379,
                      "statsdIngest": 9125,
                      "statsdScrape": 9102,
                      "workerLogs": 8793
                    },
                    "postgresql": {
                      "auth": {
                        "enablePostgresUser": true,
                        "password": null,
                        "postgresPassword": "postgres",
                        "username": null
                      },
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "airflow-postgresql",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "airflow",
                      "postgresqlPassword": "airflow",
                      "postgresqlUsername": "airflow",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-airflow-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "preannotator_queues": {
                      "count": 5
                    },
                    "quotas": {},
                    "rbac": {
                      "create": true,
                      "createSCCRoleBinding": false
                    },
                    "redis": {
                      "affinity": {},
                      "enabled": true,
                      "nodeSelector": {},
                      "password": "DefaulTRediS-P0ss",
                      "passwordSecretName": null,
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "size": "1Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "resources": {},
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-redis"
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 0
                    },
                    "registry": {
                      "connection": {},
                      "secretName": "kubernetesdockerhub"
                    },
                    "registryCredentials": {
                      "pullPolicy": "IfNotPresent",
                      "registryUrl": "docker.io/annotationlab"
                    },
                    "revisionHistoryLimit": null,
                    "scheduler": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow scheduler"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-scheduler"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "secret": [],
                    "securityContext": {},
                    "statsd": {
                      "affinity": {},
                      "args": [
                        "--statsd.mapping-config=/etc/statsd-exporter/mappings.yml"
                      ],
                      "enabled": false,
                      "extraMappings": [],
                      "extraNetworkPolicies": [],
                      "nodeSelector": {},
                      "overrideMappings": [],
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "extraAnnotations": {}
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-statsd"
                      },
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "uid": 65534
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "triggerer": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow triggerer"
                      ],
                      "command": null,
                      "enabled": false,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "annotationlab-airflow-triggerer"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 60,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      }
                    },
                    "uid": 50000,
                    "volumeMounts": [],
                    "volumes": [],
                    "webserver": {
                      "affinity": {},
                      "allowPodLogReading": true,
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec airflow webserver"
                      ],
                      "command": null,
                      "defaultUser": {
                        "email": "admin@example.com",
                        "enabled": false,
                        "firstName": "admin",
                        "lastName": "user",
                        "password": "admin",
                        "role": "Admin",
                        "username": "admin"
                      },
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraNetworkPolicies": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "livenessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "networkPolicy": {
                        "ingress": {
                          "from": [],
                          "ports": [
                            {
                              "port": "{{ .Values.ports.airflowUI }}"
                            }
                          ]
                        }
                      },
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podDisruptionBudget": {
                        "config": {
                          "maxUnavailable": 1
                        },
                        "enabled": false
                      },
                      "priorityClassName": null,
                      "readinessProbe": {
                        "failureThreshold": 20,
                        "initialDelaySeconds": 15,
                        "periodSeconds": 5,
                        "scheme": "HTTP",
                        "timeoutSeconds": 30
                      },
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "loadBalancerIP": null,
                        "loadBalancerSourceRanges": [],
                        "ports": [
                          {
                            "name": "airflow-ui",
                            "port": "{{ .Values.ports.airflowUI }}"
                          }
                        ],
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-webserver"
                      },
                      "strategy": null,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "waitForMigrations": {
                        "enabled": true,
                        "env": []
                      },
                      "webserverConfig": null,
                      "webserverConfigConfigMapName": null
                    },
                    "webserverSecretKey": "f565cd64401775f93c820dc63fc3d4f6",
                    "webserverSecretKeySecretName": null,
                    "workers": {
                      "affinity": {},
                      "annotations": {},
                      "args": [
                        "bash",
                        "-c",
                        "exec \\\nairflow {{ semverCompare \"\u003e=2.0.0\" .Values.airflowVersion | ternary \"celery worker\" \"worker\" }}"
                      ],
                      "command": null,
                      "env": [],
                      "extraContainers": [],
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "hostAliases": [],
                      "keda": {
                        "advanced": {},
                        "cooldownPeriod": 30,
                        "enabled": false,
                        "maxReplicaCount": 10,
                        "minReplicaCount": 0,
                        "namespaceLabels": {},
                        "pollingInterval": 5
                      },
                      "kerberosSidecar": {
                        "enabled": false,
                        "resources": {}
                      },
                      "labels": {},
                      "livenessProbe": {
                        "command": null,
                        "enabled": true,
                        "failureThreshold": 5,
                        "initialDelaySeconds": 10,
                        "periodSeconds": 60,
                        "timeoutSeconds": 20
                      },
                      "logGroomerSidecar": {
                        "args": [
                          "bash",
                          "/clean-logs"
                        ],
                        "command": null,
                        "enabled": true,
                        "resources": {},
                        "retentionDays": 15
                      },
                      "nodeSelector": {},
                      "persistence": {
                        "annotations": {},
                        "enabled": true,
                        "fixPermissions": false,
                        "size": "5Gi",
                        "storageClassName": null
                      },
                      "podAnnotations": {},
                      "priorityClassName": null,
                      "replicas": 1,
                      "resources": {},
                      "revisionHistoryLimit": null,
                      "safeToEvict": true,
                      "securityContext": {},
                      "serviceAccount": {
                        "annotations": {},
                        "create": false,
                        "name": "airflow-worker"
                      },
                      "strategy": {
                        "rollingUpdate": {
                          "maxSurge": "100%",
                          "maxUnavailable": "50%"
                        }
                      },
                      "terminationGracePeriodSeconds": 600,
                      "tolerations": [],
                      "topologySpreadConstraints": [],
                      "updateStrategy": null,
                      "waitForMigrations": {
                        "env": []
                      }
                    }
                  },
                  "backup": {
                    "enable": false,
                    "files": true,
                    "image": {
                      "repository": null,
                      "tag": null
                    },
                    "s3_access_key": "",
                    "s3_bucket_fullpath": "",
                    "s3_secret_key": ""
                  },
                  "configuration": {
                    "FLASK_SECRET_KEY": "rpaQghA-9lmpLwB8qxW0FsW1TJyZZYuB60ppzswpTzM=",
                    "KEYCLOAK_CLIENT_ID": "annotationlab",
                    "KEYCLOAK_CLIENT_SECRET_KEY": "09a71c59-0351-4ce6-bc8f-8fd3feb9d2ff",
                    "KEYCLOAK_REALM_NAME": "master",
                    "PERSISTENT_STORAGE": "/projects"
                  },
                  "extraContainers": "",
                  "extraEnv": "",
                  "extraNetworkPolicies": "- namespaceSelector: {}\n",
                  "extraVolumeMounts": "",
                  "extraVolumes": "",
                  "github_oauth_origin": "https://github.oauth.johnsnowlabs.com",
                  "google_analytics": {
                    "api_secret": "WENfd00yeUxRRWl4N0Utbml5SWt3Zw==",
                    "measurement_id": "Ry1XODBIMDhTTFBZ"
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "docker.io/johnsnowlabs/annotationlab",
                    "tag": "5.7.0"
                  },
                  "ingress": {
                    "annotations": {},
                    "defaultBackend": false,
                    "enabled": true,
                    "hosts": [
                      {
                        "host": "annotationlab",
                        "path": "/"
                      }
                    ],
                    "ingressroute": {
                      "enabled": false
                    },
                    "servicePort": "http",
                    "tls": [],
                    "uploadLimitInMegabytes": 16
                  },
                  "installAirflow": true,
                  "installKeycloak": true,
                  "installPostgresql": true,
                  "keycloak": {
                    "affinity": "podAntiAffinity:\n  requiredDuringSchedulingIgnoredDuringExecution:\n    - labelSelector:\n        matchLabels:\n          {{- include \"keycloak.selectorLabels\" . | nindent 10 }}\n        matchExpressions:\n          - key: app.kubernetes.io/component\n            operator: NotIn\n            values:\n              - test\n      topologyKey: kubernetes.io/hostname\n  preferredDuringSchedulingIgnoredDuringExecution:\n    - weight: 100\n      podAffinityTerm:\n        labelSelector:\n          matchLabels:\n            {{- include \"keycloak.selectorLabels\" . | nindent 12 }}\n          matchExpressions:\n            - key: app.kubernetes.io/component\n              operator: NotIn\n              values:\n                - test\n        topologyKey: topology.kubernetes.io/zone\n",
                    "args": [],
                    "auththeme": {
                      "image": {
                        "repository": "docker.io/johnsnowlabs/annotationlab",
                        "tag": "auth-theme-5.7.0"
                      }
                    },
                    "autoscaling": {
                      "behavior": {
                        "scaleDown": {
                          "policies": [
                            {
                              "periodSeconds": 300,
                              "type": "Pods",
                              "value": 1
                            }
                          ],
                          "stabilizationWindowSeconds": 300
                        }
                      },
                      "enabled": false,
                      "labels": {},
                      "maxReplicas": 10,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 80,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 3
                    },
                    "cache": {
                      "stack": "default"
                    },
                    "clusterDomain": "cluster.local",
                    "command": [
                      "/opt/keycloak/bin/kc.sh",
                      "--verbose",
                      "start",
                      "--http-enabled=true",
                      "--http-port=8080",
                      "--hostname-strict=false",
                      "--hostname-strict-https=false",
                      "--spi-events-listener-jboss-logging-success-level=info",
                      "--spi-events-listener-jboss-logging-error-level=warn"
                    ],
                    "database": {
                      "enabled": true,
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak"
                    },
                    "dbchecker": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/busybox",
                        "tag": 1.32
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        },
                        "requests": {
                          "cpu": "20m",
                          "memory": "32Mi"
                        }
                      },
                      "securityContext": {}
                    },
                    "enableServiceLinks": true,
                    "extraContainers": "",
                    "extraEnv": "- name: KC_FEATURES\n  value: \"authorization,account-api,admin-api,admin-fine-grained-authz,ciba,impersonation,kerberos,preview,step-up-authentication,token-exchange,web-authn\"\n- name: KEYCLOAK_ADMIN\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: user\n- name: KEYCLOAK_ADMIN_PASSWORD\n  valueFrom:\n    secretKeyRef:\n      name: {{ include \"keycloak.fullname\" . }}-admincreds\n      key: password\n- name: JAVA_OPTS_APPEND\n  value: \u003e-\n    -XX:+UseContainerSupport\n    -XX:MaxRAMPercentage=50.0\n    -Djava.awt.headless=true\n    -Djgroups.dns.query={{ include \"keycloak.fullname\" . }}-headless\n",
                    "extraEnvFrom": "",
                    "extraInitContainers": "- name: theme-provider\n  image: {{ .Values.auththeme.image.repository }}:{{ .Values.auththeme.image.tag }}\n  imagePullPolicy: IfNotPresent\n  command:\n    - sh\n  args:\n    - -c\n    - |\n      echo \"Copying theme...\"\n      cp -R /johnsnowlabs/* /theme\n  volumeMounts:\n    - name: theme\n      mountPath: /theme\n",
                    "extraPorts": [],
                    "extraServiceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/realms/master/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "extraVolumeMounts": "- name: admincreds\n  mountPath: /secrets/admincreds\n  readOnly: true\n- name: theme\n  mountPath: /opt/keycloak/themes/johnsnowlabs\n",
                    "extraVolumes": "- name: admincreds\n  secret:\n    secretName: '{{ include \"annotationlab.keycloak.fullname\" . }}-admincreds'\n- name: theme\n  emptyDir: {}\n",
                    "fullnameOverride": "",
                    "global": {},
                    "health": {
                      "enabled": true
                    },
                    "hostAliases": [],
                    "http": {
                      "relativePath": "/auth"
                    },
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/keycloak/keycloak",
                      "tag": "20.0.3"
                    },
                    "imagePullSecrets": [
                      {
                        "name": "kubernetesdockerhub"
                      }
                    ],
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/proxy-buffer-size": "16k"
                      },
                      "console": {
                        "annotations": {},
                        "enabled": false,
                        "ingressClassName": "",
                        "rules": [
                          {
                            "host": "{{ .Release.Name }}.keycloak.example.com",
                            "paths": [
                              {
                                "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/admin",
                                "pathType": "Prefix"
                              }
                            ]
                          }
                        ],
                        "tls": []
                      },
                      "enabled": false,
                      "ingressClassName": "",
                      "labels": {},
                      "rules": [
                        {
                          "host": "keycloak.example.com",
                          "paths": [
                            {
                              "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/",
                              "pathType": "Prefix"
                            }
                          ]
                        }
                      ],
                      "servicePort": "http",
                      "tls": []
                    },
                    "lifecycleHooks": "",
                    "livenessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/live'\n  port: http\ninitialDelaySeconds: 0\ntimeoutSeconds: 5\n",
                    "metrics": {
                      "enabled": true
                    },
                    "nameOverride": "",
                    "networkPolicy": {
                      "egress": [],
                      "enabled": false,
                      "extraFrom": [],
                      "labels": {}
                    },
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podDisruptionBudget": {},
                    "podLabels": {},
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {},
                    "postgresql": {
                      "common": {
                        "exampleValue": "common-chart",
                        "global": {
                          "postgresql": {}
                        }
                      },
                      "commonAnnotations": {},
                      "enabled": true,
                      "extraEnv": [],
                      "global": {
                        "postgresql": {}
                      },
                      "image": {
                        "debug": false,
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgresql",
                        "tag": "11.20.0-debian-11-r12"
                      },
                      "ldap": {
                        "baseDN": "",
                        "bindDN": "",
                        "bind_password": null,
                        "enabled": false,
                        "port": "",
                        "prefix": "",
                        "scheme": "",
                        "search_attr": "",
                        "search_filter": "",
                        "server": "",
                        "suffix": "",
                        "tls": false,
                        "url": ""
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 30,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "master": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": [],
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "metrics": {
                        "enabled": false,
                        "extraEnvVars": {},
                        "image": {
                          "pullPolicy": "IfNotPresent",
                          "registry": "docker.io",
                          "repository": "bitnami/postgres-exporter",
                          "tag": "0.8.0-debian-10-r166"
                        },
                        "livenessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "prometheusRule": {
                          "additionalLabels": {},
                          "enabled": false,
                          "namespace": "",
                          "rules": []
                        },
                        "readinessProbe": {
                          "enabled": true,
                          "failureThreshold": 6,
                          "initialDelaySeconds": 5,
                          "periodSeconds": 10,
                          "successThreshold": 1,
                          "timeoutSeconds": 5
                        },
                        "securityContext": {},
                        "service": {
                          "annotations": {
                            "prometheus.io/port": "9187",
                            "prometheus.io/scrape": "true"
                          },
                          "loadBalancerIP": null,
                          "type": "ClusterIP"
                        },
                        "serviceMonitor": {
                          "additionalLabels": {},
                          "enabled": false
                        }
                      },
                      "nameOverride": "keycloak-postgres",
                      "networkPolicy": {
                        "allowExternal": true,
                        "enabled": false,
                        "explicitNamespacesSelector": {}
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "annotations": {},
                        "enabled": true,
                        "mountPath": "/bitnami/postgresql",
                        "size": "8Gi",
                        "subPath": ""
                      },
                      "postgresqlDataDir": "/bitnami/postgresql/data",
                      "postgresqlDatabase": "keycloak",
                      "postgresqlPassword": "keycloak",
                      "postgresqlUsername": "keycloak",
                      "psp": {
                        "create": false
                      },
                      "rbac": {
                        "create": false
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "replication": {
                        "applicationName": "my_application",
                        "enabled": false,
                        "numSynchronousReplicas": 0,
                        "password": "repl_password",
                        "slaveReplicas": 1,
                        "synchronousCommit": "off",
                        "user": "repl_user"
                      },
                      "resources": {
                        "requests": {
                          "cpu": "250m",
                          "memory": "256Mi"
                        }
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {},
                        "port": 5432,
                        "type": "ClusterIP"
                      },
                      "serviceAccount": {
                        "enabled": true,
                        "name": "annotationlab-keycloak-postgresql"
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": true
                        },
                        "enabled": true
                      },
                      "slave": {
                        "affinity": {},
                        "annotations": {},
                        "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                        "extraVolumeMounts": [],
                        "extraVolumes": [],
                        "labels": {},
                        "nodeSelector": {},
                        "podAnnotations": {},
                        "podLabels": {},
                        "priorityClassName": "",
                        "service": {},
                        "sidecars": [],
                        "tolerations": []
                      },
                      "tls": {
                        "certCAFilename": null,
                        "certFilename": "",
                        "certKeyFilename": "",
                        "certificatesSecret": "",
                        "crlFilename": null,
                        "enabled": false,
                        "preferServerCiphers": true
                      },
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      },
                      "volumePermissions": {
                        "enabled": false,
                        "image": {
                          "pullPolicy": "Always",
                          "registry": "docker.io",
                          "repository": "bitnami/minideb",
                          "tag": "buster"
                        },
                        "securityContext": {}
                      }
                    },
                    "priorityClassName": "",
                    "prometheusRule": {
                      "annotations": {},
                      "enabled": false,
                      "labels": {},
                      "namespace": "",
                      "rules": []
                    },
                    "proxy": {
                      "enabled": true,
                      "mode": "edge"
                    },
                    "rbac": {
                      "create": false,
                      "rules": []
                    },
                    "readinessProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health/ready'\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                    "replicas": 1,
                    "resources": {},
                    "restartPolicy": "Always",
                    "route": {
                      "annotations": {},
                      "enabled": false,
                      "host": "",
                      "labels": {},
                      "path": "/",
                      "tls": {
                        "enabled": true,
                        "insecureEdgeTerminationPolicy": "Redirect",
                        "termination": "edge"
                      }
                    },
                    "secrets": {
                      "admincreds": {
                        "stringData": {
                          "password": "secret",
                          "user": "admin"
                        }
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "httpNodePort": null,
                      "httpPort": 80,
                      "httpsNodePort": null,
                      "httpsPort": 8443,
                      "labels": {},
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "sessionAffinity": "",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "allowReadPods": false,
                      "annotations": {},
                      "automountServiceAccountToken": true,
                      "create": false,
                      "imagePullSecrets": [],
                      "labels": {},
                      "name": "annotationlab-keyclo"
                    },
                    "serviceHeadless": {
                      "annotations": {}
                    },
                    "serviceMonitor": {
                      "annotations": {},
                      "enabled": false,
                      "interval": "10s",
                      "labels": {},
                      "namespace": "",
                      "namespaceSelector": {},
                      "path": "{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/metrics",
                      "port": "http",
                      "scrapeTimeout": "10s"
                    },
                    "skipInitContainers": false,
                    "startupProbe": "httpGet:\n  path: '{{ tpl .Values.http.relativePath $ | trimSuffix \"/\" }}/health'\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 1\nfailureThreshold: 60\nperiodSeconds: 5\n",
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {},
                    "terminationGracePeriodSeconds": 60,
                    "test": {
                      "deletionPolicy": "before-hook-creation",
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "docker.io/seleniarm/standalone-chromium",
                        "tag": "111.0"
                      },
                      "podSecurityContext": {},
                      "securityContext": {}
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": null,
                    "updateStrategy": "RollingUpdate"
                  },
                  "licenseserver": {
                    "origin": "https://license.johnsnowlabs.com"
                  },
                  "lifecycleHooks": "# postStart:\n#   exec:\n#     command: [\"/bin/sh\", \"-c\", \"ls\"]\n",
                  "modelshub_origin": "https://modelshub.johnsnowlabs.com",
                  "myjohnsnowlabs": {
                    "client_id": "UIGgoiIL9szXW8Nnik0O7tPT4to87oVtWvYWfc6r",
                    "origin": "https://my.johnsnowlabs.com"
                  },
                  "networkPolicy": {
                    "enabled": false
                  },
                  "nodeSelector": {},
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "annotations": {},
                    "enabled": true,
                    "size": "8Gi"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "postgresql": {
                    "common": {
                      "exampleValue": "common-chart",
                      "global": {
                        "postgresql": {}
                      }
                    },
                    "commonAnnotations": {},
                    "extraEnv": [],
                    "global": {
                      "postgresql": {}
                    },
                    "image": {
                      "debug": false,
                      "pullPolicy": "IfNotPresent",
                      "registry": "docker.io",
                      "repository": "bitnami/postgresql",
                      "tag": "11.20.0-debian-11-r12"
                    },
                    "ldap": {
                      "baseDN": "",
                      "bindDN": "",
                      "bind_password": null,
                      "enabled": false,
                      "port": "",
                      "prefix": "",
                      "scheme": "",
                      "search_attr": "",
                      "search_filter": "",
                      "server": "",
                      "suffix": "",
                      "tls": false,
                      "url": ""
                    },
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "master": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": [],
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "metrics": {
                      "enabled": false,
                      "extraEnvVars": {},
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "registry": "docker.io",
                        "repository": "bitnami/postgres-exporter",
                        "tag": "0.8.0-debian-10-r166"
                      },
                      "livenessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "prometheusRule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "readinessProbe": {
                        "enabled": true,
                        "failureThreshold": 6,
                        "initialDelaySeconds": 5,
                        "periodSeconds": 10,
                        "successThreshold": 1,
                        "timeoutSeconds": 5
                      },
                      "securityContext": {},
                      "service": {
                        "annotations": {
                          "prometheus.io/port": "9187",
                          "prometheus.io/scrape": "true"
                        },
                        "loadBalancerIP": null,
                        "type": "ClusterIP"
                      },
                      "serviceMonitor": {
                        "additionalLabels": {},
                        "enabled": false
                      }
                    },
                    "networkPolicy": {
                      "allowExternal": true,
                      "enabled": false,
                      "explicitNamespacesSelector": {}
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {},
                      "enabled": true,
                      "mountPath": "/bitnami/postgresql",
                      "size": "8Gi",
                      "subPath": ""
                    },
                    "postgresqlDataDir": "/bitnami/postgresql/data",
                    "postgresqlDatabase": "annotationlab",
                    "postgresqlPassword": "annotationlab",
                    "postgresqlUsername": "annotationlab",
                    "psp": {
                      "create": false
                    },
                    "rbac": {
                      "create": false
                    },
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 6,
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "replication": {
                      "applicationName": "my_application",
                      "enabled": false,
                      "numSynchronousReplicas": 0,
                      "password": "repl_password",
                      "slaveReplicas": 1,
                      "synchronousCommit": "off",
                      "user": "repl_user"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "securityContext": {},
                    "service": {
                      "annotations": {},
                      "port": 5432,
                      "type": "ClusterIP"
                    },
                    "serviceAccount": {
                      "enabled": false
                    },
                    "shmVolume": {
                      "chmod": {
                        "enabled": true
                      },
                      "enabled": true
                    },
                    "slave": {
                      "affinity": {},
                      "annotations": {},
                      "extraInitContainers": "# - name: do-something\n#   image: busybox\n#   command: ['do', 'something']\n",
                      "extraVolumeMounts": [],
                      "extraVolumes": [],
                      "labels": {},
                      "nodeSelector": {},
                      "podAnnotations": {},
                      "podLabels": {},
                      "priorityClassName": "",
                      "service": {},
                      "sidecars": [],
                      "tolerations": []
                    },
                    "tls": {
                      "certCAFilename": null,
                      "certFilename": "",
                      "certKeyFilename": "",
                      "certificatesSecret": "",
                      "crlFilename": null,
                      "enabled": false,
                      "preferServerCiphers": true
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "volumePermissions": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "Always",
                        "registry": "docker.io",
                        "repository": "bitnami/minideb",
                        "tag": "buster"
                      },
                      "securityContext": {}
                    }
                  },
                  "preannotator_queues": {
                    "count": 5,
                    "service": {
                      "port": 5000,
                      "portName": "http",
                      "type": "ClusterIP"
                    }
                  },
                  "priorityClassName": "",
                  "readinessProbe": "httpGet:\n  path: /healthz\n  port: http\ninitialDelaySeconds: 10\ntimeoutSeconds: 1\n",
                  "registryCredentials": {
                    "create": false,
                    "password": "XXXXX",
                    "secretName": "kubernetesdockerhub",
                    "username": "XXXXX"
                  },
                  "replicas": 1,
                  "resources": {},
                  "restartPolicy": "Always",
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 65534,
                    "fsGroupChangePolicy": "OnRootMismatch",
                    "runAsUser": 65534
                  },
                  "service": {
                    "port": 8200,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": "annotationlab"
                  },
                  "sharedData": {
                    "enabled": true,
                    "images": {
                      "storageSize": "5Gi"
                    },
                    "storageSize": "100Gi"
                  },
                  "sqlalchemy_pool_size": 10,
                  "telemetry": {
                    "enabled": true
                  },
                  "terminationGracePeriodSeconds": 60,
                  "tolerations": [],
                  "updateStrategy": {
                    "type": "Recreate"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
          createdAt: "2023-12-19T10:26:54Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/annotationlab/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: johnsnowlabs-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NLP Lab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Annotationlab
            name: annotationlabs.apps.johnsnowlabs.com
            version: v1
        description: "NLP Lab is a data labeling and annotation tool that helps you\nlabel
          different types of data using a simple interface with a standardized output\nformat.\n\nIt
          contains enterprise features like Project Management and sharing, SSO\nintegration,
          OpenAPI Specification 3.0 compliant API, task management, and active\nlearning/preannotation
          using Spark NLP.\n\nFurthermore, the production deployment of the application
          is done using a helm chart.\nThe same helm chart is also used to create
          an AWS AMI for an EC2 based single server \ndeployment using packer.\n\nMake
          sure you have defined shared storage such as efs/nfs/cephfs prior to installing
          the NLP Lab Operator.\nUpdate the host property in the YAML configuration
          to define the required domain name to use instead of the \ndefault hostname
          annotationlab.\n\nNLP Lab Operator Installtion guide [NLP Lab Operator][]:\n[NLP
          Lab Operator]: https://nlp.johnsnowlabs.com/docs/en/alab/install    \n"
        displayName: NLPLab
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlplab
        links:
        - name: Annotationlab
          url: https://annotationlab.domain
        maintainers:
        - email: bhandari@johnsnowlabs.com
          name: bikash
        maturity: alpha
        provider:
          name: JohnSnowLABS
          url: nlp.johnsnowlabs.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/annotationlab@sha256:b69fde0eb341114893aba097d58455c9e316a97cb829b41c0d080987490a83d3
        - registry.johnsnowlabs.com/annotationlab/annotationlab-operator:v5.7.0
        version: 5.7.0
      entries:
      - name: annotationlab.v5.7.0
        version: 5.7.0
      - name: annotationlab.v5.4.1
        version: 5.4.1
      - name: annotationlab.v4.9.2
        version: 4.9.2
      name: alpha
    defaultChannel: alpha
    packageName: annotationlab
    provider:
      name: JohnSnowLABS
      url: nlp.johnsnowlabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/securesign/secure-sign-operator
    name: rhtas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtas-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
          createdAt: "2024-06-27T09:45:39Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:e16accb01b2ca6bf4ebd60c6ce7faab73b784828578ee74305a3a97ff9db1e26
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:ea37c68775b14e7d590b6eeec3c9b59eba8e988860fc04d40775da3957d6d040
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:1673f08529a94671f881edec6fd29b73f43c2115dc86b643f9dd79d7e36be38e
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:5eabf561c0549d81862e521ddc1f0ab91a3f2c9d99dcd83ab5a2cf648a95dd19
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:758bfd5455d0ca96f5404229a73fdcd56686564fc4c315ca3468ebe8588dd9ca
        - registry.redhat.io/openshift4/ose-tools-rhel8@sha256:486b4d2dd0d10c5ef0212714c94334e04fe8a3d36cf619881986201a50f123c7
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:0765d248fd1c4b4f8cbcbee23f35506c08c9eeb96a8b0e8ff1511319be6c0ae6
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:0ad1961d6dd1c8636f5f1bcb5ffdbea51ad29b88c6539df366305ef801c79877
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:06382df99dfff9f002365a31eee0855f8032e351bc9583e42e489ab7623fe287
        - registry.redhat.io/rhtas/client-server-cg-rhel9@sha256:f0c8f5129c0f611332f7b34dfc5f0dc39c73ed8a985a0845d1fe6ce31d2ba020
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:d4ea970447f3b4c18c309d2f0090a5d02260dd5257a0d41f87fefc4f014a9526
        - registry.redhat.io/rhtas/tuf-server-rhel9@sha256:6c26d2394b1d82995a6f3d3483d2c18befd6a4ae826af2cd25aac30dfabef1cc
        - registry.access.redhat.com/ubi9/httpd-24@sha256:7874b82335a80269dcf99e5983c2330876f5fe8bdc33dc6aa4374958a2ffaaee
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:496b29b00a9831278eb3a8316bd5fb233ad73cb9a6217b1ff4a5f72a91b87414
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e1a78636cd7cc8e7227b6b0a5ddfd93d002db0464b15e966e88337d38989ea5a
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:58140d6b7fee762b148336b342a6745b1910ab758a701b2e8077b1fca2af1180
        - registry.redhat.io/rhtas/client-server-re-rhel9@sha256:83b9f9a5ea40473c8525a16d9965b1ed1eb2d8f7c2ec1d17ae38e5f10ac1652e
        version: 1.0.2
      deprecation:
        message: channel stable-v1.0 is no longer supported.  Please switch to channel
          'stable' or channel 'stable-v1.1'.
      entries:
      - deprecation:
          message: rhtas-operator.v1.0.2 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.2
        version: 1.0.2
      - deprecation:
          message: rhtas-operator.v1.0.1 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.1
        version: 1.0.1
      - deprecation:
          message: rhtas-operator.v1.0.0 is deprecated. Uninstall and install rhtas-operator.v1.1.1
            for support.
        name: rhtas-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: rhtas-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
          createdAt: "2025-04-08T09:29:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:540556453c98db3919e5f305a99c699e767e2891a95fe7527ace578f3c67a308
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:6131053778ea04e437f3005f90d1138aa11ebc58e3a9295e2a8d8ef6713a52be
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:09b8259ec9e982a33bde84c7c901ad70c79f3e5f6dd7880b1b1e573dafcbd56c
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:571c48ecb2658ce70199c06dc483ea48a16e032a764a6830388ad845f508d981
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:e52d07e2cd8cbdc2ccd0336a4fd29cbb2e5b8a1be628350294a54579cc2636ed
        - registry.redhat.io/ubi9/httpd-24@sha256:366cc5bc3b96069fecaa0abeb9d3510c507e69a20de1d205d881da8ae8561ab4
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:dc994a95be22b0f4bab022fc362c4f44c6a7d1887a2eb0d04870d75654ec013b
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:5a752cefdaf28bfc53847185cdd5fef1ee47e3dcff8472f8a8bf7bbdc224ef57
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:75f1049431f9e92898a4217870309cbbb3b39c8362e929c0bad3b53cad4459db
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:603a2ffa54df0bd45ff42ac710dd16331560eda2be93442232b045c959773c4d
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:329a0dad7842d4ef60430b0e3f188231bea9e579427443ecd42eb62a5c282266
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1f5a30a285a16635a7234c3c1763dfb385c8bffd605fc862b782bdb5c6c61ea3
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:3b8f49c41df15022f8ffdf3a8f8605b14c14f4e10eae754a06a86b6585d158b3
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:775b261b8f258b369213246f2da5c91218678acb5e3b8bb4670b143f487767af
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:76c24a38ac89ed632d38e44049f37e4997abfa27fa8cadbb8afb42575031296f
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:796860a3e85712c60398c36983e0ff4d45325c7a4de869da2ebf1b6ba4b19825
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:7ce611aefdfedd8b2a633def482cf41f390c95b8f8c800b6163a585f117a9e2e
        version: 1.1.2
      entries:
      - name: rhtas-operator.v1.1.2
        version: 1.1.2
      - name: rhtas-operator.v1.1.1
        version: 1.1.1
      - name: rhtas-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: rhtas-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "CTlog",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "ctlog-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Fulcio",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "fulcio-sample"
                },
                "spec": {
                  "certificate": {
                    "commonName": "fulcio.hostname",
                    "organizationEmail": "jdoe@redhat.com",
                    "organizationName": "Red Hat"
                  },
                  "config": {
                    "OIDCIssuers": [
                      {
                        "ClientID": "trusted-artifact-signer",
                        "Issuer": "https://your-oidc-issuer-url",
                        "IssuerURL": "https://your-oidc-issuer-url",
                        "Type": "email"
                      }
                    ]
                  },
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "rekor-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Securesign",
                "metadata": {
                  "annotations": {
                    "rhtas.redhat.com/metrics": "true"
                  },
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "securesign-sample"
                },
                "spec": {
                  "ctlog": null,
                  "fulcio": {
                    "certificate": {
                      "commonName": "fulcio.hostname",
                      "organizationEmail": "jdoe@redhat.com",
                      "organizationName": "Red Hat"
                    },
                    "config": {
                      "OIDCIssuers": [
                        {
                          "ClientID": "trusted-artifact-signer",
                          "Issuer": "https://your-oidc-issuer-url",
                          "IssuerURL": "https://your-oidc-issuer-url",
                          "Type": "email"
                        }
                      ]
                    },
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "rekor": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    }
                  },
                  "trillian": {
                    "database": {
                      "create": true
                    }
                  },
                  "tsa": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "monitoring": {
                      "enabled": true
                    },
                    "ntpMonitoring": {
                      "enabled": true
                    },
                    "signer": {
                      "certificateChain": {
                        "intermediateCA": [
                          {
                            "commonName": "tsa.hostname",
                            "organizationEmail": "jdoe@redhat.com",
                            "organizationName": "Red Hat"
                          }
                        ],
                        "leafCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        },
                        "rootCA": {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      }
                    }
                  },
                  "tuf": {
                    "externalAccess": {
                      "enabled": true
                    },
                    "keys": [
                      {
                        "name": "rekor.pub"
                      },
                      {
                        "name": "ctfe.pub"
                      },
                      {
                        "name": "fulcio_v1.crt.pem"
                      },
                      {
                        "name": "tsa.certchain.pem"
                      }
                    ],
                    "pvc": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "retain": true,
                      "size": "100Mi"
                    },
                    "rootKeySecretRef": {
                      "name": "tuf-root-keys"
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "TimestampAuthority",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhtas-operator",
                    "app.kubernetes.io/instance": "timestampauthority-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "timestampauthority",
                    "app.kubernetes.io/part-of": "rhtas-operator"
                  },
                  "name": "timestampauthority-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "monitoring": {
                    "enabled": true
                  },
                  "ntpMonitoring": {
                    "enabled": true
                  },
                  "signer": {
                    "certificateChain": {
                      "intermediateCA": [
                        {
                          "commonName": "tsa.hostname",
                          "organizationEmail": "jdoe@redhat.com",
                          "organizationName": "Red Hat"
                        }
                      ],
                      "leafCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      },
                      "rootCA": {
                        "commonName": "tsa.hostname",
                        "organizationEmail": "jdoe@redhat.com",
                        "organizationName": "Red Hat"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Trillian",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "trillian-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "rhtas.redhat.com/v1alpha1",
                "kind": "Tuf",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "securesign-sample",
                    "app.kubernetes.io/name": "securesign-sample",
                    "app.kubernetes.io/part-of": "trusted-artifact-signer"
                  },
                  "name": "tuf-sample"
                },
                "spec": {
                  "externalAccess": {
                    "enabled": true
                  },
                  "keys": [
                    {
                      "name": "rekor.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "rekor-pub-key"
                      }
                    },
                    {
                      "name": "fulcio_v1.crt.pem",
                      "secretRef": {
                        "key": "cert",
                        "name": "fulcio-pub-key"
                      }
                    },
                    {
                      "name": "ctfe.pub",
                      "secretRef": {
                        "key": "public",
                        "name": "ctlog-pub-key"
                      }
                    },
                    {
                      "name": "tsa.certchain.pem",
                      "secretRef": {
                        "key": "certificateChain",
                        "name": "tsa-cert-chain"
                      }
                    }
                  ],
                  "pvc": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "retain": true,
                    "size": "100Mi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
          createdAt: "2025-08-17T21:52:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Artifact Signer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/securesign/secure-sign-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CTlog is the Schema for the ctlogs API
            displayName: CTlog
            kind: CTlog
            name: ctlogs.rhtas.redhat.com
            version: v1alpha1
          - description: Fulcio is the Schema for the fulcios API
            displayName: Fulcio
            kind: Fulcio
            name: fulcios.rhtas.redhat.com
            version: v1alpha1
          - description: Rekor is the Schema for the rekors API
            displayName: Rekor
            kind: Rekor
            name: rekors.rhtas.redhat.com
            version: v1alpha1
          - description: Securesign is the Schema for the securesigns API
            displayName: Securesign
            kind: Securesign
            name: securesigns.rhtas.redhat.com
            version: v1alpha1
          - description: TimestampAuthority is the Schema for the timestampauthorities
              API
            displayName: Timestamp Authority
            kind: TimestampAuthority
            name: timestampauthorities.rhtas.redhat.com
            version: v1alpha1
          - description: Trillian is the Schema for the trillians API
            displayName: Trillian
            kind: Trillian
            name: trillians.rhtas.redhat.com
            version: v1alpha1
          - description: Tuf is the Schema for the tufs API
            displayName: Tuf
            kind: Tuf
            name: tufs.rhtas.redhat.com
            version: v1alpha1
        description: Red Hat Trusted Artifact Signer enhances software supply chain
          security by simplifying cryptographic signing and verifying of software
          artifacts, such as container images, binaries and documents. Trusted Artifact
          Signer provides a production ready deployment of the Sigstore project within
          an enterprise. Enterprises adopting it can meet signing-related criteria
          for achieving Supply Chain Levels for Software Artifacts (SLSA) compliance
          and have greater confidence in the security and trustworthiness of their
          software supply chains.
        displayName: Red Hat Trusted Artifact Signer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - rhtas
        - trust
        links:
        - name: RHTAS Operator
          url: https://github.com/securesign/secure-sign-operator
        maintainers:
        - email: rcook@redhat.com
          name: Ryan Cook
        - email: tturek@redhat.com
          name: Tomas Turek
        - email: jbouska@redhat.com
          name: Jan Bouska
        - email: gpereira@redhat.com
          name: Gregory Pereira
        - email: japower@redhat.com
          name: Jason Power
        - email: somalley@redhat.com
          name: Sally O'Malley
        - email: tdalton@redhat.com
          name: Tommy Dalton
        - email: lball@redhat.com
          name: Lance Ball
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://github.com/securesign/secure-sign-operator
        relatedImages:
        - registry.redhat.io/rhtas/rekor-search-ui-rhel9@sha256:fe31830df2d99359749cb86cd1ac2f32fcc3de07aae427d2f2cd3998d31d2833
        - registry.redhat.io/rhtas/fulcio-rhel9@sha256:6270ba29adf779de5d3916d5e82f1cd9491ba89d4ae2042347b6ad6d1819dd26
        - registry.redhat.io/rhtas/trillian-redis-rhel9@sha256:fc018a45b7eb48690c46c83f52cac55fa40089e6b02686bc21265dc7c2205b8f
        - registry.redhat.io/rhtas/segment-reporting-rhel9@sha256:c22cdb9a17f85e13c00412cba83aa644e90a0faea0236029d0f1c53fe212e227
        - registry.redhat.io/rhtas/trillian-database-rhel9@sha256:640633823a54c11fe2a8b3d05e571399fda415b6f0b3adf1ba9806882a94bdd4
        - registry.redhat.io/rhtas/client-server-rhel9@sha256:1c2201d50469d70ec6e21546c7c74bd52251d420dc1dcfa5375c1cf61dd3a9fd
        - registry.redhat.io/rhtas/trillian-logserver-rhel9@sha256:12b438ea70f83ee6edb12c96b507f92510f4afba824eac34c8f50793c4199db6
        - registry.redhat.io/ubi9/httpd-24@sha256:badeee0aa3477e61de64f6a2b93e8a3673edd38fc860125f7536506f3d592132
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:030986ea26f33db3a192c67a93cd2bde352f23c68a17b062fc062955675c3c51
        - registry.redhat.io/rhtas/timestamp-authority-rhel9@sha256:531603563c69aac6905cd6b8157d30781040527404ac2313acba5a9310e0b713
        - registry.redhat.io/rhtas/rhtas-operator-bundle@sha256:7c4e739622f68cd924afcb41cc788cdadc34c725283a097e564d620f39637bac
        - registry.redhat.io/rhtas/rekor-backfill-redis-rhel9@sha256:09b4aeeb607c88c72f69e6f87cb840c82ef5752c64a58fa13551db749fac2530
        - registry.redhat.io/rhtas/rekor-server-rhel9@sha256:02fbd4772c185edf38fcee4e5c5e1c9dad82b77c9da824addaebc445b67d35ac
        - registry.redhat.io/rhtas/trillian-logsigner-rhel9@sha256:1d782a127d2f8bdab633f287cb2df2e400ae91bf5d4b913473539d8512dfaab6
        - registry.redhat.io/rhtas/tuffer-rhel9@sha256:a93df3206409de75613d398e7dcd8b30ff7cc36eb00349c435784f73f00a2335
        - registry.redhat.io/rhtas/certificate-transparency-rhel9@sha256:eb38e98dbb9828fe033a5388609132f7246fe77d2f4d258d015a94ea30752b24
        - registry.redhat.io/rhtas/rhtas-rhel9-operator@sha256:7395b169066adc88a530c421ab9534d61bd2a680c61f86b9afabf9cf8bfd4b7e
        version: 1.2.1
      entries:
      - name: rhtas-operator.v1.2.1
        version: 1.2.1
      - name: rhtas-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    defaultChannel: stable
    packageName: rhtas-operator
    provider:
      name: Red Hat
      url: https://github.com/securesign/secure-sign-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Craig Trought
      provider-url: ""
    name: prometheus
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheusoperator.0.56.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Prometheus",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "replicas": 2,
                  "serviceAccountName": "prometheus-k8s",
                  "serviceMonitorSelector": {},
                  "ruleSelector": {},
                  "podMonitorSelector": {},
                  "probeSelector": {},
                  "alerting": {
                    "alertmanagers": [
                      {
                        "namespace": "monitoring",
                        "name": "alertmanager-main",
                        "port": "web"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "ServiceMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "operated-prometheus": "true"
                    }
                  },
                  "endpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PodMonitor",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "app": "prometheus"
                    }
                  },
                  "podMetricsEndpoints": [
                    {
                      "port": "web",
                      "interval": "30s"
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "Alertmanager",
                "metadata": {
                  "name": "alertmanager-main"
                },
                "spec": {
                  "replicas": 3,
                  "alertmanagerConfigSelector": {}
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1",
                "kind": "PrometheusRule",
                "metadata": {
                  "name": "prometheus-example-rules"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "./example.rules",
                      "rules": [
                        {
                          "alert": "ExampleAlert",
                          "expr": "vector(1)"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "monitoring.coreos.com/v1alpha1",
                "kind": "AlertmanagerConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "receivers": [
                    {
                      "name": "example"
                    }
                  ],
                  "route": {
                    "receiver": "example"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/prometheus-operator/prometheus-operator:v0.56.3
          createdAt: "2022-05-24T9:40:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          repository: https://github.com/prometheus-operator/prometheus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A running Prometheus instance
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.coreos.com
            version: v1
          - description: A Prometheus Rule configures groups of sequentially evaluated
              recording and alerting rules.
            displayName: Prometheus Rule
            kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular k8s service
            displayName: Service Monitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - description: Configures prometheus to monitor a particular pod
            displayName: Pod Monitor
            kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
          - description: Configures an Alertmanager for the namespace
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.coreos.com
            version: v1
          - description: Configures subsections of Alertmanager configuration specific
              to each namespace
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.coreos.com
            version: v1alpha1
          - description: Defines how groups of ingresses and static targets should
              be monitored.
            displayName: Probe
            kind: Probe
            name: probes.monitoring.coreos.com
            version: v1
          - description: Allows recording and alerting rules to be processed across
              multiple Prometheus instances.
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.coreos.com
            version: v1
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: k8s@trought.ca
          name: Craig Trought
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Craig Trought
        relatedImages:
        - quay.io/openshift-community-operators/prometheus@sha256:1de089becfcceb3747b3a3ce310c1612861b03446906cab5e981071db10709a9
        - quay.io/prometheus-operator/prometheus-operator:v0.56.3
        version: 0.56.3
      entries:
      - name: prometheusoperator.0.56.3
        version: 0.56.3
      - name: prometheusoperator.0.47.0
        version: 0.47.0
      name: beta
    defaultChannel: beta
    packageName: prometheus
    provider:
      name: Craig Trought
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsoft
      provider-url: ""
    name: azure-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: azure-service-operator.v1.0.59040
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "APIMgmtAPI",
                "metadata": {
                  "name": "aso-apimgmtapi-sample"
                },
                "spec": {
                  "apiId": "apiId0",
                  "apiService": "apimservice-sample",
                  "location": "westus",
                  "properties": {
                    "apiVersionDescription": "API version description",
                    "description": "API description",
                    "displayName": "aso-apimgmtapi-sample",
                    "isCurrent": true,
                    "isOnline": true,
                    "path": "/api/test",
                    "protocols": [
                      "http"
                    ],
                    "subscriptionRequired": false
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ApimService",
                "metadata": {
                  "name": "apimservice-sample"
                },
                "spec": {
                  "appInsightsName": "appinsights-sample-wmm",
                  "appInsightsResourceGroup": "resourcegroup-azure-operators-wmm",
                  "location": "westus",
                  "publisherEmail": {
                    "Your email": ""
                  },
                  "publisherName": {
                    "Your name": ""
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "tier": "premium",
                  "vnetName": "virtualnetwork-sample",
                  "vnetResourceGroup": "resourcegroup-azure-operators",
                  "vnetSubnetName": "test1",
                  "vnetType": "none"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsights",
                "metadata": {
                  "name": "appinsights-sample"
                },
                "spec": {
                  "applicationType": "other",
                  "kind": "other",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AppInsightsApiKey",
                "metadata": {
                  "name": "key-sample"
                },
                "spec": {
                  "appInsights": "appinsights-sample",
                  "authSDKControlChannel": true,
                  "readTelemetry": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "writeAnnotations": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureLoadBalancer",
                "metadata": {
                  "name": "lb-sample-01"
                },
                "spec": {
                  "backendAddressPoolName": "test",
                  "backendPort": 3000,
                  "frontendPortRangeEnd": 2000,
                  "frontendPortRangeStart": 1001,
                  "inboundNatPoolName": "test",
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "lb-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-01"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-01",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-01"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureNetworkInterface",
                "metadata": {
                  "name": "nic-sample-02"
                },
                "spec": {
                  "location": "SouthCentralUS",
                  "publicIPAddressName": "nic-pip-sample-02",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnetName": "test1",
                  "vnetName": "vnet-sample-02"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "lb-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-01"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzurePublicIPAddress",
                "metadata": {
                  "name": "nic-pip-sample-02"
                },
                "spec": {
                  "idleTimeoutInMinutes": 10,
                  "location": "SouthCentralUS",
                  "publicIPAddressVersion": "IPv4",
                  "publicIPAllocationMethod": "Static",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "skuName": "Basic"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLManagedUser",
                "metadata": {
                  "name": "azuresqlmanageduser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample",
                  "managedIdentityClientId": null,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLUser",
                "metadata": {
                  "name": "sqluser-sample"
                },
                "spec": {
                  "dbName": "azuresqldatabase-sample1",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "db_owner"
                  ],
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSQLVNetRule",
                "metadata": {
                  "name": "azuresqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlAction",
                "metadata": {
                  "name": "azuresqlaction-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverName": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample1"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "valueA",
                    "tag2": "valueB"
                  },
                  "name": "azuresqldatabase-sample3"
                },
                "spec": {
                  "edition": 0,
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample777"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaselist": [
                    "azuresqldatabase-sample1",
                    "azuresqldatabase-sample2"
                  ],
                  "failovergraceperiod": 30,
                  "failoverpolicy": "Automatic",
                  "location": "eastus",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "secondaryserver": "sqlserver-samplesec",
                  "secondaryserverresourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endipaddress": "0.0.0.0",
                  "resourcegroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startipaddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample777"
                },
                "spec": {
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVMScaleSet",
                "metadata": {
                  "name": "vmss-sample-01"
                },
                "spec": {
                  "adminUserName": "azureuser7",
                  "backendAddressPoolName": "test",
                  "capacity": 3,
                  "inboundNatPoolName": "test",
                  "loadBalancerName": "lb-sample-01",
                  "location": "SouthCentralUS",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "subnetName": "test1",
                  "virtualNetworkName": "vnet-sample-01",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm1"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-02",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": {
                    "ssh key": ""
                  },
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachine",
                "metadata": {
                  "name": "hpfvm20"
                },
                "spec": {
                  "adminUserName": "azureuser",
                  "location": "SouthCentralUS",
                  "networkInterfaceName": "nic-sample-01",
                  "osType": "Linux",
                  "platformImageURN": "Canonical:UbuntuServer:16.04-LTS:latest",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sshPublicKeyData": "{ssh public key}",
                  "vmSize": "Standard_DS1_v2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "AzureVirtualMachineExtension",
                "metadata": {
                  "name": "vmext20a"
                },
                "spec": {
                  "autoUpgradeMinorVersion": true,
                  "forceUpdateTag": "Test2",
                  "location": "SouthCentralUS",
                  "protectedSettings": "{}",
                  "publisher": "Microsoft.Azure.Extensions",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "settings": "{\"commandToExecute\":\"echo 'hello, world'\"}",
                  "typeHandlerVersion": "2.1",
                  "typeName": "CustomScript",
                  "vmName": "vm20"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accesslevel": "Container",
                  "accountname": "storageaccountsample777",
                  "location": "westus",
                  "resourcegroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ConsumerGroup",
                "metadata": {
                  "name": "consumergroup-sample-1"
                },
                "spec": {
                  "eventHub": "eventhub-sample-1",
                  "namespace": "eventhubnamespace-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "name": "cosmosdb-sample-1"
                },
                "spec": {
                  "kind": "GlobalDocumentDB",
                  "location": "westus",
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "isVirtualNetworkFilterEnabled": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "CosmosDB",
                "metadata": {
                  "labels": {
                    "CosmosAccountType": "Production"
                  },
                  "name": "cosmosdb-sample-2"
                },
                "spec": {
                  "ipRules": [
                    "104.42.195.92,40.76.54.131,52.176.6.30,52.169.50.45,52.187.184.26"
                  ],
                  "kind": "GlobalDocumentDB",
                  "location": "eastus",
                  "locations": [
                    {
                      "failoverPriority": 0,
                      "isZoneRedundant": true,
                      "locationName": "eastus"
                    },
                    {
                      "failoverPriority": 1,
                      "locationName": "westus"
                    }
                  ],
                  "properties": {
                    "databaseAccountOfferType": "Standard",
                    "enableMultipleWriteLocations": true
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-1"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "Eventhub",
                "metadata": {
                  "name": "eventhub-sample-2307-10-capture"
                },
                "spec": {
                  "authorizationRule": {
                    "name": "RootManageSharedAccessKey",
                    "rights": [
                      "Listen",
                      "Manage",
                      "Send"
                    ]
                  },
                  "location": "westus",
                  "namespace": "eventhubnamespace-sample-1",
                  "properties": {
                    "captureDescription": {
                      "destination": {
                        "archiveNameFormat": "{Namespace}/{EventHub}/{PartitionId}/{Year}/{Month}/{Day}/{Hour}/{Minute}/{Second}",
                        "blobContainer": "capturecontainer",
                        "name": "EventHubArchive.AzureBlockBlob",
                        "storageAccount": {
                          "accountName": "storageaccountauv1",
                          "resourceGroup": "my-resource-group"
                        }
                      },
                      "enabled": true,
                      "intervalInSeconds": 120,
                      "sizeLimitInBytes": 524288000
                    },
                    "messageRetentionInDays": 7,
                    "partitionCount": 2
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "EventhubNamespace",
                "metadata": {
                  "name": "eventhubnamespace-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "isAutoInflateEnabled": false,
                    "kafkaEnabled": false,
                    "maximumThroughputUnits": 0
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "capacity": 1,
                    "name": "Standard",
                    "tier": "Standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123"
                },
                "spec": {
                  "accessPolicies": [
                    {
                      "objectID": "\u003cobjectID\u003e",
                      "permissions": {
                        "certificates": [
                          "list",
                          "get"
                        ],
                        "keys": [
                          "list",
                          "get"
                        ],
                        "secrets": [
                          "list",
                          "get"
                        ],
                        "storage": [
                          "list",
                          "get"
                        ]
                      },
                      "tenantID": "\u003ctenantID\u003e"
                    }
                  ],
                  "enableSoftDelete": false,
                  "location": "westus",
                  "networkPolicies": {
                    "bypass": "AzureServices",
                    "defaultAction": "Allow",
                    "ipRules": [
                      "172.16.0.0/24",
                      "172.16.1.0/24"
                    ],
                    "virtualNetworkRules": [
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet1",
                      "/subscriptions/\u003csubid\u003e/resourceGroups/rg1/providers/Microsoft.Network/virtualNetworks/test-vnet/subnets/subnet2"
                    ]
                  },
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "standard"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVault",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "keyvaultsample123-simple"
                },
                "spec": {
                  "enableSoftDelete": false,
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "KeyVaultKey",
                "metadata": {
                  "name": "keyvaultkey-sample"
                },
                "spec": {
                  "keySize": 4096,
                  "keyVault": "keyvaultsample123",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "type": "RSA"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLDatabase",
                "metadata": {
                  "name": "mysqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLFirewallRule",
                "metadata": {
                  "name": "mysqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-replica"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus2",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{SUBID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforMySQL/servers/mysqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLServer",
                "metadata": {
                  "name": "mysqlserver-sample2"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLUser",
                "metadata": {
                  "name": "mysqluser-sample"
                },
                "spec": {
                  "dbName": "mysqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "SELECT"
                  ],
                  "server": "mysqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "MySQLVNetRule",
                "metadata": {
                  "name": "mysqlvnetrule-sample"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "mysqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-vnet",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLDatabase",
                "metadata": {
                  "name": "postgresqldatabase-sample2"
                },
                "spec": {
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLFirewallRule",
                "metadata": {
                  "name": "postgresqlfirewallrule-sample2"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample2",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-replica3"
                },
                "spec": {
                  "createMode": "Replica",
                  "location": "eastus",
                  "replicaProperties": {
                    "sourceServerId": "/subscriptions/{subscription ID}/resourceGroups/resourcegroup-azure-operators/providers/Microsoft.DBforPostgreSQL/servers/postgresqlserver-sample"
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "name": "postgresqlserver-sample2"
                },
                "spec": {
                  "location": "southcentralus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 2,
                    "family": "Gen5",
                    "name": "B_Gen5_2",
                    "size": "51200",
                    "tier": "Basic"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLUser",
                "metadata": {
                  "name": "psqluser-sample"
                },
                "spec": {
                  "dbName": "postgresqldatabase-sample",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "roles": [
                    "azure_pg_admin"
                  ],
                  "server": "postgresqlserver-sample"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "PostgreSQLVNetRule",
                "metadata": {
                  "name": "postgresqlvnetrule-sample1"
                },
                "spec": {
                  "ignoreMissingServiceEndpoint": true,
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "postgresqlserver-sample",
                  "subnetName": "test1",
                  "vNetName": "virtualnetwork-sample",
                  "vNetResourceGroup": "resourcegroup-azure-operators",
                  "vNetSubscriptionID": {
                    "vnet_subscription_id": ""
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCache",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "rediscache-sample-1"
                },
                "spec": {
                  "location": "westus",
                  "properties": {
                    "enableNonSslPort": true,
                    "sku": {
                      "capacity": 1,
                      "family": "C",
                      "name": "Basic"
                    }
                  },
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheAction",
                "metadata": {
                  "name": "rediscacheaction-sample-1"
                },
                "spec": {
                  "actionName": "rollallkeys",
                  "cacheName": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "RedisCacheFirewallRule",
                "metadata": {
                  "name": "rediscachefirewallrule"
                },
                "spec": {
                  "properties": {
                    "endIP": "0.0.0.0",
                    "startIP": "0.0.0.0"
                  },
                  "redisCache": "rediscache-sample-1",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "ResourceGroup",
                "metadata": {
                  "name": "resourcegroup-azure-operators"
                },
                "spec": {
                  "location": "westus"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "StorageAccount",
                "metadata": {
                  "name": "storageaccountsample777"
                },
                "spec": {
                  "accessTier": "Hot",
                  "kind": "StorageV2",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "sku": {
                    "name": "Standard_RAGRS"
                  },
                  "supportsHttpsTrafficOnly": true
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-01"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha1",
                "kind": "VirtualNetwork",
                "metadata": {
                  "name": "vnet-sample-02"
                },
                "spec": {
                  "addressSpace": "10.0.0.0/8",
                  "location": "SouthCentralUS",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "subnets": [
                    {
                      "subnetAddressPrefix": "10.1.0.0/16",
                      "subnetName": "test1"
                    },
                    {
                      "subnetAddressPrefix": "10.2.0.0/16",
                      "subnetName": "test2"
                    }
                  ]
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "BlobContainer",
                "metadata": {
                  "name": "blobcontainer-sample"
                },
                "spec": {
                  "accessLevel": "Container",
                  "accountName": "storageaccountsample777",
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "MySQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "mysqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus2",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "8.0",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1alpha2",
                "kind": "PostgreSQLServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "postgresqlserver-sample"
                },
                "spec": {
                  "createMode": "Default",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "serverVersion": "10",
                  "sku": {
                    "capacity": 4,
                    "family": "Gen5",
                    "name": "GP_Gen5_4",
                    "size": "51200",
                    "tier": "GeneralPurpose"
                  },
                  "sslEnforcement": "Enabled"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlDatabase",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "azuresqldatabase-sample"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "sku": {
                    "name": "Basic",
                    "tier": "Basic"
                  }
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFailoverGroup",
                "metadata": {
                  "name": "azuresqlfailovergroup-sample"
                },
                "spec": {
                  "databaseList": [
                    "azuresqldatabase-sample",
                    "azuresqldatabase-sample2"
                  ],
                  "failoverGracePeriod": 30,
                  "failoverPolicy": "automatic",
                  "location": "eastus",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "secondaryServer": "sqlserver-samplesec",
                  "secondaryServerResourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-samplepri"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlFirewallRule",
                "metadata": {
                  "name": "sqlf-allowazuresvcaccess"
                },
                "spec": {
                  "endIpAddress": "0.0.0.0",
                  "resourceGroup": "resourcegroup-azure-operators",
                  "server": "sqlserver-sample-777",
                  "startIpAddress": "0.0.0.0"
                }
              },
              {
                "apiVersion": "azure.microsoft.com/v1beta1",
                "kind": "AzureSqlServer",
                "metadata": {
                  "labels": {
                    "tag1": "value1",
                    "tag2": "value2"
                  },
                  "name": "sqlserver-sample-777"
                },
                "spec": {
                  "location": "westus",
                  "resourceGroup": "resourcegroup-azure-operators"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Cloud Provider
          certified: "false"
          containerImage: mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
          createdAt: Mon 12 Jun 2023 04:24:02 UTC
          description: The Azure Service Operator helps you provision Azure resources
            and connect your applications to them from within Kubernetes.
          displayName: Azure Service Operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Azure/azure-service-operator
          support: https://github.com/Azure/azure-service-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates an API with the specified properties in the specified
              API Management service.
            displayName: APIMgmtAPI
            kind: APIMgmtAPI
            name: apimgmtapis.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an API Management instance into a specified Resource
              Group at the specified location. It also provides the option to link
              to an Application Insights instance for logging, and to place the API
              Management instance in a specified Virtual Network.
            displayName: ApimService
            kind: ApimService
            name: apimservices.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Application Insights instance into a specified
              Resource Group at the specified location.
            displayName: AppInsights
            kind: AppInsights
            name: appinsights.azure.microsoft.com
            version: v1alpha1
          - description: Creates an Api Key to access a given Application Insights
              instance.
            displayName: AppInsightsApiKey
            kind: AppInsightsApiKey
            name: appinsightsapikeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Load Balancer (LB) into a specified Resource
              Group at the specified location.
            displayName: AzureLoadBalancer
            kind: AzureLoadBalancer
            name: azureloadbalancers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Network Interface (NIC) into a specified
              Resource Group at the specified location. Users can specify underlying
              public IP address and virtual network configurations in their NIC setup.
            displayName: AzureNetworkInterface
            kind: AzureNetworkInterface
            name: azurenetworkinterfaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Public IP Address (PIP) into a specified
              Resource Group at the specified location. Users can specify IP allocation
              method, idle timeout, IP address version, and SKU.
            displayName: AzurePublicIPAddress
            kind: AzurePublicIPAddress
            name: azurepublicipaddresses.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to roll the password for the specified SQL server.
            displayName: AzureSqlAction
            kind: AzureSqlAction
            name: azuresqlactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server.
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an SQL database given the SQL server
            displayName: AzureSqlDatabase
            kind: AzureSqlDatabase
            name: azuresqldatabases.azure.microsoft.com
            version: v1beta1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a failover group on a specified Azure SQL server
              given the secondary server and the databases to failover.
            displayName: AzureSqlFailoverGroup
            kind: AzureSqlFailoverGroup
            name: azuresqlfailovergroups.azure.microsoft.com
            version: v1beta1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the SQL server
              from specific IPs.
            displayName: AzureSqlFirewallRule
            kind: AzureSqlFirewallRule
            name: azuresqlfirewallrules.azure.microsoft.com
            version: v1beta1
          - description: Binds a specified Managed Identity as a user on the SQL database
            displayName: AzureSQLManagedUser
            kind: AzureSQLManagedUser
            name: azuresqlmanagedusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure SQL server given the location and Resource
              Group.
            displayName: AzureSqlServer
            kind: AzureSqlServer
            name: azuresqlservers.azure.microsoft.com
            version: v1beta1
          - description: Creates an user on the specified Azure SQL database and stores
              the username/password as secrets.
            displayName: AzureSQLUser
            kind: AzureSQLUser
            name: azuresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a VirtualNetwork rule to allow access to the SQL
              server from specific VNets.
            displayName: AzureSQLVNetRule
            kind: AzureSQLVNetRule
            name: azuresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys Azure Virtual Machine Extensions to provide post-deployment
              configuration and automation on Azure VMs.
            displayName: AzureVirtualMachineExtension
            kind: AzureVirtualMachineExtension
            name: azurevirtualmachineextensions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine (VM) into a specified Resource
              Group at the specified location. Users can specify platform image, size,
              user name and public SSH key, etc. for the VM.
            displayName: AzureVirtualMachine
            kind: AzureVirtualMachine
            name: azurevirtualmachines.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Machine Scale Set (VMSS) into a
              specified Resource Group at the specified location. Users can specify
              platform image, size, user name and public SSH key, etc. for the VMSS.
            displayName: AzureVMScaleSet
            kind: AzureVMScaleSet
            name: azurevmscalesets.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a Blob Container into a specified Resource Group
              at the specified location. A Storage Account must first be created prior
              to creating a Blob Container.
            displayName: BlobContainer
            kind: BlobContainer
            name: blobcontainers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a consumer group given the Event Hub, Event Hubs
              namespace and Resource Group.
            displayName: ConsumerGroup
            kind: ConsumerGroup
            name: consumergroups.azure.microsoft.com
            version: v1alpha1
          - description: Provision a CosmosDB instance given the Cosmos DB instance
              type, location, and Resource Group.
            displayName: CosmosDB
            kind: CosmosDB
            name: cosmosdbs.azure.microsoft.com
            version: v1alpha1
          - kind: CosmosDBSQLDatabase
            name: cosmosdbsqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs namespace given the Resource Group
              and location. Also has the ability to configure SKU, properties, and
              network rules.
            displayName: EventhubNamespace
            kind: EventhubNamespace
            name: eventhubnamespaces.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Event Hubs instance given the Event Hubs namespace,
              Resource Group and Location.
            displayName: Eventhub
            kind: Eventhub
            name: eventhubs.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault key given the location and Resource
              Group.
            displayName: KeyVaultKey
            kind: KeyVaultKey
            name: keyvaultkeys.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Key Vault given the location and Resource
              Group.
            displayName: KeyVault
            kind: KeyVault
            name: keyvaults.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha1
          - description: Add an AAD user to an existing MySQL database.
            kind: MySQLAADUser
            name: mysqlaadusers.azure.microsoft.com
            version: v1alpha2
          - description: Deploys a database under the given Azure Database for MySQL
              server
            displayName: MySQLDatabase
            kind: MySQLDatabase
            name: mysqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for MySQL server from the specified IP range.
            displayName: MySQLFirewallRule
            kind: MySQLFirewallRule
            name: mysqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Create an AAD-based admininistrative user in a MySQL server.
            kind: MySQLServerAdministrator
            name: mysqlserveradministrators.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for MySQL server given the Location,
              Resource Group and other properties. Also helps creating read replicas
              for MySQL server.
            displayName: MySQLServer
            kind: MySQLServer
            name: mysqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Add a new user to an existing MySQL database.
            displayName: MySQLUser
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Add a new user to an existing MySQL database.
            kind: MySQLUser
            name: mysqlusers.azure.microsoft.com
            version: v1alpha2
          - description: Add virtual network rules to the MySQL server.
            displayName: MySQLVNetRule
            kind: MySQLVNetRule
            name: mysqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a database under the given Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLDatabase
            kind: PostgreSQLDatabase
            name: postgresqldatabases.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the Azure Database
              for PostgreSQL server from the specified IP range.
            displayName: PostgreSQLFirewallRule
            kind: PostgreSQLFirewallRule
            name: postgresqlfirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Database for PostgreSQL server given the
              Location and Resource Group.
            displayName: PostgreSQLServer
            kind: PostgreSQLServer
            name: postgresqlservers.azure.microsoft.com
            version: v1alpha2
          - description: Create and Manage Users for Azure Database for PostgreSQL
              server.
            displayName: PostgreSQLUser
            kind: PostgreSQLUser
            name: postgresqlusers.azure.microsoft.com
            version: v1alpha1
          - description: Create and manage VNet service endpoints and VNet rules in
              Azure Database for PostgreSQL.
            displayName: PostgreSQLVNetRule
            kind: PostgreSQLVNetRule
            name: postgresqlvnetrules.azure.microsoft.com
            version: v1alpha1
          - description: Allows you to regenerate keys and reboot the RedisCache cluster.
            displayName: RedisCacheAction
            kind: RedisCacheAction
            name: rediscacheactions.azure.microsoft.com
            version: v1alpha1
          - description: Deploys a firewall rule to allow access to the RedisCache
              from the specified IP range.
            displayName: RedisCacheFirewallRule
            kind: RedisCacheFirewallRule
            name: rediscachefirewallrules.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Cache for Redis into a specified Resource
              Group at the specified location.
            displayName: RedisCache
            kind: RedisCache
            name: rediscaches.azure.microsoft.com
            version: v1alpha1
          - description: Provisions a Resource Group given the location and a name.
            displayName: ResourceGroup
            kind: ResourceGroup
            name: resourcegroups.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Storage Account into a specified Resource
              Group at the specified location.
            displayName: StorageAccount
            kind: StorageAccount
            name: storageaccounts.azure.microsoft.com
            version: v1alpha1
          - description: Deploys an Azure Virtual Network into a specified Resource
              Group at the specified location, in the address space provided. Users
              are also able to add subnets to their virtual network through the operator.
            displayName: VirtualNetwork
            kind: VirtualNetwork
            name: virtualnetworks.azure.microsoft.com
            version: v1alpha1
        description: |
          ## Overview

          The Azure Service Operator comprises of:

          - The Custom Resource Definitions (CRDs) for each of the Azure services a Kubernetes user can provision.
          - The Kubernetes controller that watches for requests to create Custom Resources for each of these CRDs and creates them.

          The project was built using [Kubebuilder](https://book.kubebuilder.io/).

          Curious to see how it all works? Check out our [control flow diagrams](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/controlflow.md).

          **Note**: The current version of Azure Service Operator only supports watching all namespaces in the cluster.

          ## Supported Azure Services

          - [Resource Group](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/resourcegroup/resourcegroup.md)
          - [Event Hubs](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/eventhub/eventhub.md)
          - [Azure SQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/azuresql/azuresql.md)
          - [Azure Database for PostgreSQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/postgresql/postgresql.md)
          - [Azure Database for MySQL](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/mysql/mysql.md)
          - [Azure Key Vault](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/keyvault/keyvault.md)
          - [Azure Cache for Redis](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/rediscache/rediscache.md)
          - [Storage Account](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/storageaccount.md)
          - [Blob Storage](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/storage/blobcontainer.md)
          - [Virtual Network](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualnetwork/virtualnetwork.md)
          - [Application Insights](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/appinsights/appinsights.md)
          - [API Management](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/apimgmt/apimgmt.md)
          - [Cosmos DB](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/cosmosdb/cosmosdb.md)
          - [Virtual Machine](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/virtualmachine/virtualmachine.md)
          - [Virtual Machine Scale Set](https://github.com/Azure/azure-service-operator/tree/main/docs/v1/services/vmscaleset/vmscaleset.md)

          ## Instructions - Please read before installing

          Before you begin, verify you're running the Azure CLI version 2.0.53 or later. Run `az --version` to find the version. If you need to install or upgrade, see [Install Azure CLI](https://docs.microsoft.com/en-us/cli/azure/install-azure-cli).

          1. Create an Azure Service Principal. You'll need this to grant Azure Service Operator permissions to create resources in your subscription.

              First, set the following environment variables to your Azure Tenant ID and Subscription ID with your values:
              ```yaml
              AZURE_TENANT_ID=<your-tenant-id-goes-here>
              AZURE_SUBSCRIPTION_ID=<your-subscription-id-goes-here>
              ```

              You can find these values by using the [Azure CLI](https://docs.microsoft.com/en-us/cli/azure/?view=azure-cli-latest):
              ```sh
              az account show
              ```

          2. Next, we'll create a Service Principal with Contributor permissions for your subscription, so ASO can create resources in your subscription on your behalf. Note that the [ServicePrincipal](https://docs.microsoft.com/en-us/cli/azure/create-an-azure-service-principal-azure-cli) you pass to the command below needs to have access to create resources in your subscription. If you'd like to use Managed Identity for authorization instead, check out instructions [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/managedidentity.md).

              ```sh
              az ad sp create-for-rbac -n "azure-service-operator" --role contributor \
                  --scopes /subscriptions/$AZURE_SUBSCRIPTION_ID
              ```

              This should give you output like the following:
              ```sh
              "appId": "xxxxxxxxxx",
              "displayName": "azure-service-operator",
              "name": "http://azure-service-operator",
              "password": "xxxxxxxxxxx",
              "tenant": "xxxxxxxxxxxxx"
              ```

          3.  Once you have created a Service Principal, gather the following values:

              `AZURE_TENANT_ID` is the Tenant ID from Step 1.

              `AZURE_SUBSCRIPTION_ID` is the Subscription ID from Step 1.

              `AZURE_CLIENT_ID` is the appID from the Service Principal created in Step 2.

              `AZURE_CLIENT_SECRET` is the password from the service Principal we created in Step 2.

              `AZURE_CLOUD_ENV` is the Azure Environment you'd like to use, i.e. AzurePublicCloud, AzureUSGovernmentCloud, AzureChinaCloud, AzureGermanCloud.

          4.  Set those values in a `Secret` called `azureoperatorsettings` within the `operators` Namespace:
              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name: azureoperatorsettings
                namespace: operators
              stringData:
                AZURE_TENANT_ID: <your-tenant-id-goes-here>
                AZURE_SUBSCRIPTION_ID: <your-subscription-id-goes-here>
                AZURE_CLIENT_ID: <your-client-id>
                AZURE_CLIENT_SECRET: <your-client-secret>
                AZURE_CLOUD_ENV: <your-azure-cloud-environment>
              ```

          5. Now you can proceed to install the Azure Service Operator to the `operators` Namespace via the Install button on the top right of this page. After the operator is installed, you will then see the Azure Service Operator pod running in your cluster:

              ```console
              $ kubectl get pods -n operators
              NAME                                                READY   STATUS    RESTARTS   AGE
              azureoperator-controller-manager-7dd75bbd97-mk4s9   2/2     Running   0          35s
              ```

          ## About the project

          This project maintains [releases of the Azure Service Operator](https://github.com/Azure/azure-service-operator/releases).

          Please see the [FAQ](https://github.com/Azure/azure-service-operator/blob/main/docs/faq.md) for answers to commonly asked questions about the Azure Service Operator.

          Have more questions? Feel free to consult our documentation [here](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md).

          ## Contributing

          The [contribution guide](https://github.com/Azure/azure-service-operator/blob/main/CONTRIBUTING.md) covers everything you need to know about how you can contribute to Azure Service Operators. The [developer guide](https://github.com/Azure/azure-service-operator/blob/main/docs/v1/howto/contents.md#developing-azure-service-operator) will help you onboard as a developer.

          ## Support

          Azure Service Operator is an open source project that is **not** covered by the [Microsoft Azure support policy](https://support.microsoft.com/en-us/help/2941892/support-for-linux-and-open-source-technology-in-azure). Please search open issues [here](https://github.com/Azure/azure-service-operator/issues). If your issue isn't already represented, please [open a new one](https://github.com/Azure/azure-service-operator/issues/new/choose). The Azure Service Operator project maintainers will respond to the best of their abilities.

          ## Code of conduct

          This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/). For more information, see the [Code of Conduct FAQ](https://opensource.microsoft.com/codeofconduct/faq) or contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additional questions or comments.
        displayName: Azure Service Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - azure, microsoft, cloud service broker
        links:
        - name: Azure Service Operator
          url: https://github.com/Azure/azure-service-operator
        maintainers:
        - email: christian.muirhead@microsoft.com
          name: Christian Muirhead
        maturity: stable
        provider:
          name: Microsoft
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - mcr.microsoft.com/k8s/azureserviceoperator:1.0.59040
        - quay.io/openshift-community-operators/azure-service-operator@sha256:2430ebe724ead60e84f12aa485c8dae4af82af10d14400ce76356392030c8c0b
        version: 1.0.59040
      entries:
      - name: azure-service-operator.v1.0.59040
        version: 1.0.59040
      name: stable
    defaultChannel: stable
    packageName: azure-service-operator
    provider:
      name: Microsoft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/confidential-containers
    name: trustee-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustee-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
          createdAt: "2024-06-07T10:06:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trustee-operator-system
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/trustee-operator
          support: Confidential Containers Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KbsConfig is the Schema for the kbsconfigs API
            displayName: Kbs Config
            kind: KbsConfig
            name: kbsconfigs.confidentialcontainers.org
            version: v1alpha1
        description: Remote attestation operator for confidential containers
        displayName: Red Hat build of Trustee
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trustee
        - Confidential Computing
        - Attestation
        links:
        - name: Red Hat build of Trustee
          url: https://github.com/openshift/trustee-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
          url: https://github.com/confidential-containers
        relatedImages:
        - registry.redhat.io/build-of-trustee/trustee-operator-bundle@sha256:c6bc928d220fa001e2a7b5e8661a3c8a604c21385dc1bb97b8dd121ca2347916
        - registry.redhat.io/build-of-trustee/trustee-rhel9-operator@sha256:fe8e23de14ad088c79b1bd5e7aab08a1b88a86679effcbf174be81e761bd9d6d
        version: 0.4.2
      entries:
      - name: trustee-operator.v0.4.2
        version: 0.4.2
      - name: trustee-operator.v0.4.1
        version: 0.4.1
      - name: trustee-operator.v0.4.0
        version: 0.4.0
      - name: trustee-operator.v0.3.0
        version: 0.3.0
      - name: trustee-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: trustee-operator
    provider:
      name: Red Hat
      url: https://github.com/confidential-containers
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/stackgres/support?utm_source=openshift_console
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-1@sha256:a698218b7c8772fcfd1671bf0b38df30b5c113464559577c37064c6b83ec7126
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: spark-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-helm-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apachespark.apache.org/v1alpha1",
                "kind": "Spark",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "spark-operator"
                },
                "spec": {
                  "affinity": {},
                  "batchScheduler": {
                    "enable": false
                  },
                  "controllerThreads": 10,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/googlecloudplatform/spark-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "ingressUrlFormat": "",
                  "istio": {
                    "enabled": false
                  },
                  "labelSelectorFilter": "",
                  "leaderElection": {
                    "lockName": "spark-operator-lock",
                    "lockNamespace": ""
                  },
                  "logLevel": 2,
                  "metrics": {
                    "enable": true,
                    "endpoint": "/metrics",
                    "port": 10254,
                    "portName": "metrics",
                    "prefix": ""
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podMonitor": {
                    "enable": false,
                    "jobLabel": "spark-operator-podmonitor",
                    "labels": {},
                    "podMetricsEndpoint": {
                      "interval": "5s",
                      "scheme": "http"
                    }
                  },
                  "podSecurityContext": {},
                  "rbac": {
                    "create": false,
                    "createClusterRole": true,
                    "createRole": true
                  },
                  "replicaCount": 1,
                  "resourceQuotaEnforcement": {
                    "enable": false
                  },
                  "resources": {},
                  "resyncInterval": 30,
                  "securityContext": {},
                  "serviceAccounts": {
                    "spark": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "sparkoperator": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    }
                  },
                  "sparkJobNamespace": "",
                  "tolerations": [],
                  "uiService": {
                    "enable": true
                  },
                  "volumeMounts": [],
                  "volumes": [],
                  "webhook": {
                    "cleanupAnnotations": {
                      "helm.sh/hook": "pre-delete, pre-upgrade",
                      "helm.sh/hook-delete-policy": "hook-succeeded"
                    },
                    "enable": false,
                    "initAnnotations": {
                      "helm.sh/hook": "pre-install, pre-upgrade",
                      "helm.sh/hook-weight": "50"
                    },
                    "namespaceSelector": "",
                    "port": 8080,
                    "timeout": 30
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "driver": {
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "serviceAccount": "spark-operator-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.1.1"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "gcr.io/spark-operator/spark:v3.1.1",
                    "imagePullPolicy": "Always",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.1.1",
                    "type": "Scala",
                    "volumes": [
                      {
                        "hostPath": {
                          "path": "/tmp",
                          "type": "Directory"
                        },
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "pyspark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark-py:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/src/main/python/pi.py",
                  "mode": "cluster",
                  "pythonVersion": "3",
                  "restartPolicy": {
                    "onFailureRetries": 3,
                    "onFailureRetryInterval": 10,
                    "onSubmissionFailureRetries": 5,
                    "onSubmissionFailureRetryInterval": 20,
                    "type": "OnFailure"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Python",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi",
                  "namespace": "spark-operator"
                },
                "spec": {
                  "driver": {
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "serviceAccount": "spark-operator-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.1.1"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "gcr.io/spark-operator/spark:v3.1.1",
                  "imagePullPolicy": "Always",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.12-3.1.1.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.1.1",
                  "type": "Scala",
                  "volumes": [
                    {
                      "hostPath": {
                        "path": "/tmp",
                        "type": "Directory"
                      },
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/spark-helm-operator-controller:0.0.2
          createdAt: "2023-07-14T17:57:44Z"
          description: An experimental operator that installs the spark on k8s operator
            from its helm chart.
          operatorframework.io/initialization-resource: "{\n  \"apiVersion\": \"apachespark.apache.org/v1alpha1\",\n
            \ \"kind\": \"Spark\",\n  \"metadata\": {\n    \"name\": \"spark-operator\"\n
            \ },\n  \"spec\": {\n    \"affinity\": {},\n    \"batchScheduler\": {\n
            \     \"enable\": false\n    },\n    \"controllerThreads\": 10,\n    \"fullnameOverride\":
            \"\",\n    \"image\": {\n      \"pullPolicy\": \"IfNotPresent\",\n      \"repository\":
            \"ghcr.io/googlecloudplatform/spark-operator\",\n      \"tag\": \"\"\n
            \   },\n    \"imagePullSecrets\": [],\n    \"ingressUrlFormat\": \"\",\n
            \   \"istio\": {\n      \"enabled\": false\n    },\n    \"labelSelectorFilter\":
            \"\",\n    \"leaderElection\": {\n      \"lockName\": \"spark-operator-lock\",\n
            \     \"lockNamespace\": \"\"\n    },\n    \"logLevel\": 2,\n    \"metrics\":
            {\n      \"enable\": true,\n      \"endpoint\": \"/metrics\",\n      \"port\":
            10254,\n      \"portName\": \"metrics\",\n      \"prefix\": \"\"\n    },\n
            \   \"nameOverride\": \"\",\n    \"nodeSelector\": {},\n    \"podAnnotations\":
            {},\n    \"podLabels\": {},\n    \"podMonitor\": {\n      \"enable\":
            false,\n      \"jobLabel\": \"spark-operator-podmonitor\",\n      \"labels\":
            {},\n      \"podMetricsEndpoint\": {\n        \"interval\": \"5s\",\n
            \       \"scheme\": \"http\"\n      }\n    },\n    \"podSecurityContext\":
            {},\n    \"rbac\": {\n      \"create\": false,\n      \"createClusterRole\":
            true,\n      \"createRole\": true\n    },\n    \"replicaCount\": 1,\n
            \   \"resourceQuotaEnforcement\": {\n      \"enable\": false\n    },\n
            \   \"resources\": {},\n    \"resyncInterval\": 30,\n    \"securityContext\":
            {},\n    \"serviceAccounts\": {\n      \"spark\": {\n        \"annotations\":
            {},\n        \"create\": true,\n        \"name\": \"\"\n      },\n      \"sparkoperator\":
            {\n        \"annotations\": {},\n        \"create\": true,\n        \"name\":
            \"\"\n      }\n    },\n    \"sparkJobNamespace\": \"\",\n    \"tolerations\":
            [],\n    \"uiService\": {\n      \"enable\": true\n    },\n    \"volumeMounts\":
            [],\n    \"volumes\": [],\n    \"webhook\": {\n      \"cleanupAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-delete, pre-upgrade\",\n        \"helm.sh/hook-delete-policy\":
            \"hook-succeeded\"\n      },\n      \"enable\": false,\n      \"initAnnotations\":
            {\n        \"helm.sh/hook\": \"pre-install, pre-upgrade\",\n        \"helm.sh/hook-weight\":
            \"50\"\n      },\n      \"namespaceSelector\": \"\",\n      \"port\":
            8080,\n      \"timeout\": 30\n    }\n  }\n}    "
          operatorframework.io/suggested-namespace: spark-operator
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys a scheduled spark application based on the spec provided
            displayName: Scheduled Spark Application
            kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys a spark application based on the spec provided
            displayName: Spark Application
            kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - description: Deploys the spark operator based on the google cloud platform
              helm chart
            displayName: Spark Operator
            kind: Spark
            name: sparks.apachespark.apache.org
            version: v1alpha1
        description: An experimental helm based operator that wraps the spark on k8s
          helm chart from google cloud platform enabling installation and workload
          management of Apache Spark
        displayName: Spark Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - big data
        - map reduce
        - data science
        links:
        - name: Spark Helm Operator
          url: https://github.com/skattoju/spark-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - quay.io/openshift-community-operators/spark-helm-operator@sha256:9bd49276e02f2fb28bb1ff116112cd7ebe0aa12bdb9f1e46aae2501fd9e4e755
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/spark-helm-operator-controller:0.0.2
        version: 0.0.2
      entries:
      - name: spark-helm-operator.v0.0.2
        version: 0.0.2
      - name: spark-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: spark-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gatekeeper-operator-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gatekeeper-operator-product.v3.11.2-0.1725401426.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
          createdAt: 21 Mar 2024, 01:59
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.11.0
          olm.substitutesFor: gatekeeper-operator-product.v3.11.2-0.1721233953.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/gatekeeper/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/gatekeeper/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:3f7fc52c79f3e4ef7ddd6f5efc8eec1ee4836bbf2076e7dd5f68bb3b7c6e9d8f
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8-operator@sha256:c8894e4d02fb2c822ab2e02e489dd9b67c953d96bd3b96a8305219a2245dfd99
        - registry.redhat.io/gatekeeper/gatekeeper-rhel8@sha256:701b165def9e4b92d91b8eac4917721e76e634d2159b8efc69bef23e10b51b9d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 3.11.2+0.1725401426.p
      entries:
      - name: gatekeeper-operator-product.v3.11.2-0.1725401426.p
        version: 3.11.2+0.1725401426.p
      - name: gatekeeper-operator-product.v3.11.2-0.1721233953.p
        version: 3.11.2+0.1721233953.p
      - name: gatekeeper-operator-product.v3.11.2-0.1718224960.p
        version: 3.11.2+0.1718224960.p
      - name: gatekeeper-operator-product.v3.11.2
        version: 3.11.2
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.11"
    - currentCSV: gatekeeper-operator-product.v3.14.3-0.1746550072.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
          createdAt: 16 Jan 2025, 06:57
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.14.3
          olm.substitutesFor: gatekeeper-operator-product.v3.14.3-0.1744033158.p
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the Schema for the gatekeepers API
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |-
          Gatekeeper allows administrators to detect and reject non-compliant commits to an infrastructure-as-code system's source-of-truth. This strengthens compliance efforts and prevents a bad state from slowing down the organization.

          ## Support & Troubleshooting
          Gatekeeper is available and supported as part of a valid Red Hat software subscription when it is used with Red Hat Advanced Cluster Management for Kubernetes. Troubleshooting information is available in the Gatekeeper documentation.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:34d4f950bce7863f15d5f1e7d6555a3ccf97ec31dc3ee140f68442890c4f0f96
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e47f0e9da800eb0cf4a0da6af98986508a35798a41581158ee7dbc977cb651dc
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:6255f7a673e6989617176842d8f0cc897e4630b5283e812a7a2d52dbb2a1f335
        version: 3.14.3+0.1746550072.p
      entries:
      - name: gatekeeper-operator-product.v3.14.3-0.1746550072.p
        version: 3.14.3+0.1746550072.p
      - name: gatekeeper-operator-product.v3.14.3-0.1744033158.p
        version: 3.14.3+0.1744033158.p
      - name: gatekeeper-operator-product.v3.14.3-0.1742934403.p
        version: 3.14.3+0.1742934403.p
      - name: gatekeeper-operator-product.v3.14.3-0.1740676608.p
        version: 3.14.3+0.1740676608.p
      - name: gatekeeper-operator-product.v3.14.3
        version: 3.14.3
      - name: gatekeeper-operator-product.v3.14.2
        version: 3.14.2
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.14"
    - currentCSV: gatekeeper-operator-product.v3.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
          createdAt: 13 Mar 2025, 17:07
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.15.4
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: Gatekeeper allows administrators to detect and reject non-compliant
          commits to an infrastructure-as-code system's source-of-truth. This strengthens
          compliance efforts and prevents a bad state from slowing down the organization.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:29690144667b3753a7e76095e136d6fc42325270e7b9f22f97dfd411439e9f54
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:96b77663961daea3a336b97e34e9412281bfb2727854dc970a69edd029b9af85
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:8b43248aed06563e06f7a0347ff1c45fa5892e3a9adb2b5c363962a0b152eff0
        version: 3.15.4
      entries:
      - name: gatekeeper-operator-product.v3.15.4
        version: 3.15.4
      - name: gatekeeper-operator-product.v3.15.3
        version: 3.15.3
      - name: gatekeeper-operator-product.v3.15.2
        version: 3.15.2
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.15"
    - currentCSV: gatekeeper-operator-product.v3.17.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.17.3
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:e44078fc6073a79168ed52922b370773b451d69bf335aeb91d26a1462df21a9b
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:3c027ec78398df0a8fc09eb79e222a9508b4f18154d1e1b477d4f156181ae51a
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:5170255ecd45b9bb7a70abec0f6eeb0518bbc2871de759e468dd61a80261c0ad
        version: 3.17.3
      entries:
      - name: gatekeeper-operator-product.v3.17.3
        version: 3.17.3
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.17"
    - currentCSV: gatekeeper-operator-product.v3.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.18.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:0803b7219ceff7aa676488eeba58607a19c9c18a2bb8eae8b3b16cfec411cb6e
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:9a3f5b6c403fb84c9cb93c9d7cf902c80990196983607c8ca568d4b558938e17
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:49726cb2702441305e8047b24054575170fb595cfe5ebf815cdbfdaa957b02f4
        version: 3.18.1
      entries:
      - name: gatekeeper-operator-product.v3.18.1
        version: 3.18.1
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.18"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: "3.19"
    - currentCSV: gatekeeper-operator-product.v3.19.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.gatekeeper.sh/v1alpha1",
                "kind": "Gatekeeper",
                "metadata": {
                  "name": "gatekeeper"
                },
                "spec": {
                  "validatingWebhook": "Enabled"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.19.1
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/gatekeeper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gatekeeper is the schema for the gatekeepers API. Gatekeeper
              contains configurations for the Gatekeeper operator, which deploys Open
              Policy Agent Gatekeeper, a policy engine that enforces policies through
              admission controller webhooks. See https://github.com/open-policy-agent/gatekeeper.
            displayName: Gatekeeper
            kind: Gatekeeper
            name: gatekeepers.operator.gatekeeper.sh
            version: v1alpha1
        description: |
          Open Policy Agent Gatekeeper enforces CRD-based policies.

          Every organization has policies. Some are essential to meet governance and
          legal requirements. Others help ensure adherence to best practices and
          institutional conventions. Attempting to ensure compliance manually would be
          error-prone and frustrating. Automating policy enforcement ensures
          consistency, lowers development latency through immediate feedback, and
          helps with agility by allowing developers to operate independently without
          sacrificing compliance.

          Kubernetes allows decoupling policy decisions from the inner workings of the
          API Server by means of [admission controller webhooks](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/), which are executed
          whenever a resource is created, updated or deleted. Gatekeeper is a
          validating and mutating webhook that enforces CRD-based policies executed
          by [Open Policy Agent](https://github.com/open-policy-agent/opa), a policy
          engine for Cloud Native environments hosted
          by CNCF as a [graduated project](https://www.cncf.io/projects/open-policy-agent-opa/).

          In addition to the admission scenario, Gatekeeper's audit functionality
          allows administrators to see what resources are currently violating any
          given policy.

          Finally, Gatekeeper's engine is designed to be portable, allowing
          administrators to detect and reject non-compliant commits to an
          infrastructure-as-code system's source-of-truth, further strengthening
          compliance efforts and preventing bad state from slowing down the
          organization.

          Gatekeeper has the following functionality:
          * An extensible, parameterized policy library
          * Native Kubernetes CRDs for instantiating the policy library (aka "constraints")
          * Native Kubernetes CRDs for extending the policy library (aka "constraint templates")
          * Audit functionality

          For more information, see the [Gatekeeper documentation](https://open-policy-agent.github.io/gatekeeper/website/docs/).

          ## Support & Troubleshooting

          The Open Policy Agent Gatekeeper community collaborates on Slack
          [#opa-gatekeeper](https://openpolicyagent.slack.com/archives/CDTN970AX).

          Please report issues on the respective GitHub repositories for either
          the [Gatekeeper Operator](https://github.com/stolostron/gatekeeper-operator/issues) or
          [Gatekeeper](https://github.com/open-policy-agent/gatekeeper/issues) itself.
        displayName: Gatekeeper Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gatekeeper
        links:
        - name: Gatekeeper Operator
          url: https://github.com/stolostron/gatekeeper-operator
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/gatekeeper/gatekeeper-operator-bundle@sha256:68fe21212c54efd1c7a1e6f68e554722ac7cb8ffc3f5d48d608ebcc0ac8be921
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9-operator@sha256:60ee347425c6cfb36f0e1e834b33cbb3d77aa2d6c89ac513216d0b9fc55d4b52
        - registry.redhat.io/gatekeeper/gatekeeper-rhel9@sha256:583d63090eccf767cd3a1cc153a995c6102e02a49aee1579edc7645e2aca1202
        version: 3.19.1
      entries:
      - name: gatekeeper-operator-product.v3.19.1
        version: 3.19.1
      - name: gatekeeper-operator-product.v3.19.0
        version: 3.19.0
      - name: gatekeeper-operator-product.v3.18.0
        version: 3.18.0
      - name: gatekeeper-operator-product.v3.17.2
        version: 3.17.2
      - name: gatekeeper-operator-product.v3.17.1
        version: 3.17.1
      - name: gatekeeper-operator-product.v3.17.0
        version: 3.17.0
      - name: gatekeeper-operator-product.v3.15.1
        version: 3.15.1
      - name: gatekeeper-operator-product.v3.15.1-0.1727189912.p
        version: 3.15.1+0.1727189912.p
      - name: gatekeeper-operator-product.v3.15.1-0.1726639477.p
        version: 3.15.1+0.1726639477.p
      - name: gatekeeper-operator-product.v3.15.1-0.1725401534.p
        version: 3.15.1+0.1725401534.p
      - name: gatekeeper-operator-product.v3.14.1-0.1718225063.p
        version: 3.14.1+0.1718225063.p
      - name: gatekeeper-operator-product.v3.14.1
        version: 3.14.1
      - name: gatekeeper-operator-product.v3.14.1-0.1721316083.p
        version: 3.14.1+0.1721316083.p
      - name: gatekeeper-operator-product.v3.14.1-0.1725401504.p
        version: 3.14.1+0.1725401504.p
      - name: gatekeeper-operator-product.v3.14.1-0.1726638929.p
        version: 3.14.1+0.1726638929.p
      - name: gatekeeper-operator-product.v3.14.1-0.1727189868.p
        version: 3.14.1+0.1727189868.p
      - name: gatekeeper-operator-product.v3.14.0
        version: 3.14.0
      - name: gatekeeper-operator-product.v3.11.1
        version: 3.11.1
      - name: gatekeeper-operator-product.v0.2.6-0.1697738427.p
        version: 0.2.6+0.1697738427.p
      - name: gatekeeper-operator-product.v0.2.6
        version: 0.2.6
      - name: gatekeeper-operator-product.v0.2.5-0.1683051284.p
        version: 0.2.5+0.1683051284.p
      - name: gatekeeper-operator-product.v0.2.5
        version: 0.2.5
      - name: gatekeeper-operator-product.v0.2.4-0.1666670065.p
        version: 0.2.4+0.1666670065.p
      - name: gatekeeper-operator-product.v0.2.4
        version: 0.2.4
      - name: gatekeeper-operator-product.v0.2.3-0.1655383639.p
        version: 0.2.3+0.1655383639.p
      - name: gatekeeper-operator-product.v0.2.3
        version: 0.2.3
      - name: gatekeeper-operator-product.v0.2.2
        version: 0.2.2
      name: stable
    defaultChannel: stable
    packageName: gatekeeper-operator-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Veeam Kasten
      provider-url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
    name: k10-kasten-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k10-kasten-operator-rhmp.v8.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/reconcile-period": "2m",
                    "helm.sdk.operatorframework.io/rollback-force": "false"
                  },
                  "name": "k10"
                },
                "spec": {
                  "auth": {
                    "basicAuth": {
                      "enabled": false,
                      "htpasswd": "",
                      "secretName": ""
                    },
                    "tokenAuth": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "persistence": {
                      "catalog": {
                        "size": ""
                      },
                      "storageClass": ""
                    }
                  },
                  "metering": {
                    "mode": ""
                  },
                  "route": {
                    "enabled": false,
                    "host": "",
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "apik10.kasten.io/v1alpha1",
                "kind": "K10restore",
                "metadata": {
                  "name": "k10restore"
                },
                "spec": {
                  "pointInTime": null,
                  "profileName": "",
                  "skipResource": null,
                  "sourceClusterID": ""
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          console.openshift.io/plugins: '["veeam-kasten-console-plugin"]'
          containerImage: registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
          createdAt: Wed Sep 24 02:11:59 PM UTC 2025
          description: Kasten data management platform for OpenShift, The Market Leader
            for Kubernetes Backup and DR
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/k10-kasten-operator-rhmp/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: kasten-io
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: contact@kasten.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              k10restore can recover the K10 platform in case of a variety of disasters such as
              the accidental deletion of K10, failure of underlying storage that K10 uses for its catalog,
              or even the accidental destruction of the Kubernetes cluster on which K10 is deployed.

              **Before creating k10restore**: [K10 Disaster Recovery documentation](https://docs.kasten.io/latest/operating/dr.html#recovering-with-the-operator)
            kind: K10restore
            name: k10restores.apik10.kasten.io
            version: v1alpha1
          - description: |-
              [Kasten's k10](https://docs.kasten.io/) is a data lifecycle management system for all your persistence enabled container-based applications.

              **Note: The complete list of values supported by the K10 spec field in the "YAML View" are documented [here](https://docs.kasten.io/latest/install/advanced.html#complete-list-of-k10-helm-options).**
            kind: K10
            name: k10s.apik10.kasten.io
            version: v1alpha1
        description: |-
          Kasten data management platform for OpenShift, The Market Leader for Kubernetes Backup and DR.

          The supported Kubernetes versions are listed here - https://docs.kasten.io/latest/operating/support.html
          Please review release notes before an upgrade - https://docs.kasten.io/latest/releasenotes.html
        displayName: Veeam Kasten (Free)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - storage
        - data-management
        links:
        - name: Kasten Documentation
          url: https://docs.kasten.io
        - name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        maintainers:
        - email: contact@kasten.io
          name: Veeam Kasten
        maturity: stable
        provider:
          name: Veeam Kasten
          url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
        relatedImages:
        - registry.connect.redhat.com/kasten/vbrintegrationapi@sha256:1a20b0da4efb7e802516224675b37751bd56d406e57b16216388d2b5167c648d
        - registry.connect.redhat.com/kasten/catalog@sha256:11a77ac968634850c72cd10f2bde55afc632c68cdc0881aeae83c100288c6aec
        - registry.connect.redhat.com/kasten/prometheus@sha256:29d913bb06c3245cca5798772dba7d7d5dc9b76ed2e0700f5eb4d20416e24fa2
        - registry.connect.redhat.com/kasten/datamover@sha256:c9e2dc5b646aa0c465a7b2daf011462273a6f5b53fdff772d879293208e2d57c
        - registry.connect.redhat.com/kasten/executor@sha256:32c4ac1aa2ca802ff77e913991949e6841149963863ad527a67f6bab6c94a5af
        - registry.connect.redhat.com/kasten/dex@sha256:4c8522a27c0d4f0977baa074800bba96a9bd9aefc57473de283b1d95586f7517
        - registry.connect.redhat.com/kasten/jobs@sha256:c6358878e981baee4ac72abcc1269c206ee014af2c16584d1b41fe18e0f0e57f
        - registry.connect.redhat.com/kasten/kanister@sha256:1d69ff91cec8e7c131dfbe1ad3e282aab854d3412f5f6e3b808d25eb7b8c38bc
        - registry.connect.redhat.com/kasten/ocpconsoleplugin@sha256:69f6bfb9421f4264dd7d066d809311022c0db57c41319a6242532f6acfbd2145
        - registry.connect.redhat.com/kasten/state@sha256:191383c76c4ff22805c06217e9f32a2f41a1206afab5f16acec4a5dd56e4a6b4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        - registry.connect.redhat.com/kasten/metering@sha256:f809696ac1309fc54fde87e5d2030267d266386fb73a56287868bbac433756d8
        - registry.connect.redhat.com/kasten/events@sha256:7cae1c0909ea7a468e96484d99300677bdad49f34cc57c412ebaf342cb176264
        - registry.connect.redhat.com/kasten/gateway@sha256:dc2d7bdeb8abaf9454470afd53af2cd5430505351a97b6501ff65a60dd2a6db1
        - registry.connect.redhat.com/kasten/upgrade@sha256:bf2200697d38f3ee95ba2863dfbae4d982273d2fd8450c9ffbc3a85be7d6d87d
        - registry.connect.redhat.com/kasten/logging@sha256:014bfdfc3620cc2fe6540b9993b9b9e7d70c9b969714477db4de91a90bdb4a38
        - registry.connect.redhat.com/kasten/bloblifecyclemanager@sha256:7b209314b96bc8ebaa77e97426bb34176d05844cc3e8d4b462cfba5c4a8d3798
        - registry.connect.redhat.com/kasten/controllermanager@sha256:ab3b815bb12cd3c263cbfdba3939a95097450f1864431cca7380827986d704ed
        - registry.connect.redhat.com/kasten/frontend@sha256:a021964ed85133897e4956014b5c5013da3332a36a7e96a0de2e2e1b67888491
        - registry.connect.redhat.com/kasten/kasten-rhmp-bundle@sha256:c6b58e5f6c9b2a1f5c4bb770d1fa1db6eaaee8ae7d9751a48338623d0d812438
        - registry.connect.redhat.com/kasten/metric-sidecar@sha256:20c142258de56cf8c60d4b0afa5e72b9317023ebf4ab109fb84e0d57fae854a4
        - registry.connect.redhat.com/kasten/kanister-tools@sha256:7af2e37c9027f46c565901d35a3a7da4d9c7ddd2812628b67275d4acf006b80c
        - registry.connect.redhat.com/kasten/init@sha256:cc7fbbda1b6946e93dd083fb7092f38789f626c26ac68c226b61fe57064a8f11
        - registry.connect.redhat.com/kasten/crypto@sha256:86bfbbf4c4f06bab1e4b091f9f312d536813e2c11aa899690b3277ff73110a09
        - registry.connect.redhat.com/kasten/dashboardbff@sha256:e1208c60146faa8c3c7ef73b2290663840276ed8d3b3c990f495cee2967104ee
        - registry.connect.redhat.com/kasten/k10tools@sha256:59caa1920c76a1bc86e7733731602e3f0fc08026d9a16000bb748503a4d9cd9b
        - registry.connect.redhat.com/kasten/aggregatedapis@sha256:db868afd5c6d4a8d2e7c8376ed6943963c4be299fca6b26931f907430763d244
        - registry.connect.redhat.com/kasten/configmap-reload@sha256:01010f2bb580eb131d63d3271efe40eadd8269b84cdb02bdea36750d1359a43f
        - registry.connect.redhat.com/kasten/garbagecollector@sha256:6f2972bb255f5a2de33ace5ce805b15d5e04d3c259c9c8d018c1257c74562878
        - registry.connect.redhat.com/kasten/auth@sha256:f8543e67de3d5276c597c26f16053bf9bbcbd9cbdad4652214935a1e40a26817
        - registry.connect.redhat.com/kasten/restorectl@sha256:9dc3c834bd3d4543115ae1ad586ee1248fc2b637dbc2d1e2c66ecc2a459db67e
        - registry.connect.redhat.com/kasten/k10-operator@sha256:4e8ccf38e47f92d2a308587591f85fb4f15aea1e7fd5c248803764c94eaecd4a
        - registry.connect.redhat.com/kasten/repositories@sha256:924a607874e1807a877cb454e00539d64f4487e7435d4bfb524f3e569ad1f942
        version: 8.0.9
      entries:
      - name: k10-kasten-operator-rhmp.v8.0.9
        version: 8.0.9
      - name: k10-kasten-operator-rhmp.v8.0.8
        version: 8.0.8
      - name: k10-kasten-operator-rhmp.v8.0.7
        version: 8.0.7
      - name: k10-kasten-operator-rhmp.v8.0.6
        version: 8.0.6
      - name: k10-kasten-operator-rhmp.v8.0.5
        version: 8.0.5
      - name: k10-kasten-operator-rhmp.v8.0.4
        version: 8.0.4
      - name: k10-kasten-operator-rhmp.v8.0.3
        version: 8.0.3
      - name: k10-kasten-operator-rhmp.v8.0.2
        version: 8.0.2
      - name: k10-kasten-operator-rhmp.v8.0.1
        version: 8.0.1
      - name: k10-kasten-operator-rhmp.v8.0.0
        version: 8.0.0
      - name: k10-kasten-operator-rhmp.v7.5.10
        version: 7.5.10
      - name: k10-kasten-operator-rhmp.v7.5.9
        version: 7.5.9
      - name: k10-kasten-operator-rhmp.v7.5.8
        version: 7.5.8
      - name: k10-kasten-operator-rhmp.v7.5.7
        version: 7.5.7
      - name: k10-kasten-operator-rhmp.v7.5.4
        version: 7.5.4
      - name: k10-kasten-operator-rhmp.v7.5.3
        version: 7.5.3
      - name: k10-kasten-operator-rhmp.v7.5.2
        version: 7.5.2
      - name: k10-kasten-operator-rhmp.v7.5.1
        version: 7.5.1
      - name: k10-kasten-operator-rhmp.v7.5.0
        version: 7.5.0
      - name: k10-kasten-operator-rhmp.v7.0.14
        version: 7.0.14
      - name: k10-kasten-operator-rhmp.v7.0.13
        version: 7.0.13
      - name: k10-kasten-operator-rhmp.v7.0.12
        version: 7.0.12
      - name: k10-kasten-operator-rhmp.v7.0.11
        version: 7.0.11
      - name: k10-kasten-operator-rhmp.v7.0.9
        version: 7.0.9
      - name: k10-kasten-operator-rhmp.v7.0.8
        version: 7.0.8
      - name: k10-kasten-operator-rhmp.v7.0.7
        version: 7.0.7
      - name: k10-kasten-operator-rhmp.v7.0.6
        version: 7.0.6
      - name: k10-kasten-operator-rhmp.v7.0.5
        version: 7.0.5
      - name: k10-kasten-operator-rhmp.v7.0.4
        version: 7.0.4
      - name: k10-kasten-operator-rhmp.v7.0.3
        version: 7.0.3
      - name: k10-kasten-operator-rhmp.v7.0.1
        version: 7.0.1
      - name: k10-kasten-operator-rhmp.v6.5.14
        version: 6.5.14
      - name: k10-kasten-operator-rhmp.v6.5.13
        version: 6.5.13
      - name: k10-kasten-operator-rhmp.v6.5.12
        version: 6.5.12
      - name: k10-kasten-operator-rhmp.v6.5.11
        version: 6.5.11
      - name: k10-kasten-operator-rhmp.v6.5.10
        version: 6.5.10
      - name: k10-kasten-operator-rhmp.v6.5.8
        version: 6.5.8
      - name: k10-kasten-operator-rhmp.v6.5.6
        version: 6.5.6
      - name: k10-kasten-operator-rhmp.v6.5.5
        version: 6.5.5
      name: stable
    defaultChannel: stable
    packageName: k10-kasten-operator-rhmp
    provider:
      name: Veeam Kasten
      url: https://www.veeam.com/products/cloud/kubernetes-data-protection.html
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-ccs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-ccs.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccs.cpd.ibm.com/v1beta1",
                "kind": "CCS",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ccs-cr",
                    "app.kubernetes.io/managed-by": "ibm-cpd-ccs-operator",
                    "app.kubernetes.io/name": "ibm-cpd-ccs-operator"
                  },
                  "name": "ccs-cr"
                },
                "spec": {
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
          createdAt: "2025-04-18"
          description: Common Core Services
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-ccs
          olm.skipRange: '>=7.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: |-
            [
              "runtimeassemblies.runtimes.ibm.com"
            ]
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation Runtime Manager CRD. This product is not deployable
              standalone and it is part of the LICCR already accepted.  For additional
              details regarding install parameters check: http://ibm.biz/-readme.
              License By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: RuntimeAssembly
            name: runtimeassemblies.runtimes.ibm.com
            version: v1alpha1
          - description: 'Documentation This product is not deployable standalone
              and it is part of the LICCR already accepted.  For additional details
              regarding install parameters check: http://ibm.biz/-readme. License
              By installing this product you accept the license terms https://ibm.biz/BdaaLC
              or https://ibm.biz/BdaaL9.'
            kind: CCS
            name: ccs.ccs.cpd.ibm.com
            version: v1beta1
        description: |-
          # Introduction Common Core Services Operator
          Put stuff here

          # Details
          Various Details
          ## Prerequisites
          Various Prerequisites
          ## Resources Required
          Various Resources Required
          ## Installing
          Various Installation details
          # Configuration
          Various Configuration information
          ### Limitations
          Various Limitations
          ### SecurityContextConstraints Requirements
          All pods will use the restricted SCC, I think.
        displayName: Common Core Services
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commonc-core-services
        - cpd40
        links:
        - name: Ccs
          url: https://ccs.domain
        maintainers:
        - email: test@ibm.com
          name: test
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.connect.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:99e845965f7eca4b5a3361c494a385292d655e7c1906185197826c773cf9aaf6
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        version: 10.3.0
      entries:
      - name: ibm-cpd-ccs.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-ccs
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Qualcomm Technologies, Inc.
      provider-url: https://github.com/quic/aic-operator
    name: aic-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aic-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aic.quicinc.com/v1",
                "kind": "AIC",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "aic-operator",
                    "app.kubernetes.io/instance": "aic-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "aic",
                    "app.kubernetes.io/part-of": "aic-operator"
                  },
                  "name": "aic-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "false"
          containerImage: ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
          description: Automate the management of OpenShift Operator for Qualcomm
            AI 100 accelerators.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: Qualcomm Technologies, Inc.
          repository: https://github.com/quic/aic-operator
          support: Qualcomm Technologies, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AIC is the Schema for the aics API
            displayName: AIC
            kind: AIC
            name: aics.aic.quicinc.com
            version: v1
        description: The Openshift Operator enables the Qualcomm® Cloud AI series
          of hardware on OpenShift clusters by automating the configuration and installation
          of their Linux device drivers and its Kubernetes Device Plugin.
        displayName: AIC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AIC
        - AIC-Operator
        links:
        - name: AIC Operator
          url: https://github.com/quic/aic-operator
        maintainers:
        - email: quic_vmellemp@quicinc.com
          name: quic_vmellemp
        - email: quic_carls@quicinc.com
          name: quic_carls
        maturity: alpha
        provider:
          name: Qualcomm Technologies, Inc.
          url: https://github.com/quic/aic-operator
        relatedImages:
        - ghcr.io/quic/cloud_ai_openshift_operator@sha256:9cf581e9b5c9cd08819fab8f8b6313836a0a20098ad4c99b80d93dbcf5ab32bf
        - registry.connect.redhat.com/qc-qranium/aic-operator-bundle@sha256:581b21f44d6b993888681ed702ed10b9d1bb8e52dcca53a91e856febed2613a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:7a943db9bcb22f817d75341621472f4bb30abc5625c53b9a781d97d431bc4d09
        version: 0.3.1
      entries:
      - name: aic-operator.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: aic-operator
    provider:
      name: Qualcomm Technologies, Inc.
      url: https://github.com/quic/aic-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/en/products/edge/validated-patterns
    name: patterns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patterns-operator.v0.0.62
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.hybrid-cloud-patterns.io/v1alpha1",
                "kind": "Pattern",
                "metadata": {
                  "name": "pattern-sample"
                },
                "spec": {
                  "clusterGroupName": "hub",
                  "gitSpec": {
                    "targetRepo": "https://github.com/validatedpatterns/multicloud-gitops",
                    "targetRevision": "main"
                  },
                  "multiSourceConfig": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/hybridcloudpatterns/patterns-operator:0.0.62
          createdAt: "2025-09-02 16:14:40"
          description: "An operator to deploy and manage architecture patterns from
            https://validatedpatterns.io. \nThis operator collects some Analytics
            like cluster version, cloud type, etc. To disable this\nadd \"ANALYTICS:
            false\" to the environment variables in the subscription."
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/validatedpatterns/patterns-operator
          support: validatedpatterns@googlegroups.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pattern is the Schema for the patterns API
            displayName: Pattern
            kind: Pattern
            name: patterns.gitops.hybrid-cloud-patterns.io
            version: v1alpha1
        description: Deploys and manages architecture patterns from https://validatedpatterns.io
        displayName: Validated Patterns Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Validated
        - Pattern
        - Blueprint
        - Architecture
        links:
        - name: Hybrid Cloud Patterns
          url: https://validatedpatterns.io
        maintainers:
        - email: validatedpatterns@googlegroups.com
          name: patterns-team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/en/products/edge/validated-patterns
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patterns-operator@sha256:99ac398b06fa9d9a7626b99cb47fb380d0cf4a4307fa11da8e486df35966debb
        - quay.io/hybridcloudpatterns/kube-rbac-proxy@sha256:409a464692c7db3cf68893376479dca1b53802412237637725927f7e78aa0844
        - quay.io/hybridcloudpatterns/patterns-operator@sha256:15ed44938b112bcf9b3a9a566749b39ba2b9053a6a8b9b499ef247977c18496c
        version: 0.0.62
      entries:
      - name: patterns-operator.v0.0.62
        version: 0.0.62
      - name: patterns-operator.v0.0.61
        version: 0.0.61
      - name: patterns-operator.v0.0.60
        version: 0.0.60
      - name: patterns-operator.v0.0.59
        version: 0.0.59
      - name: patterns-operator.v0.0.58
        version: 0.0.58
      - name: patterns-operator.v0.0.57
        version: 0.0.57
      - name: patterns-operator.v0.0.56
        version: 0.0.56
      - name: patterns-operator.v0.0.55
        version: 0.0.55
      - name: patterns-operator.v0.0.54
        version: 0.0.54
      - name: patterns-operator.v0.0.53
        version: 0.0.53
      - name: patterns-operator.v0.0.52
        version: 0.0.52
      - name: patterns-operator.v0.0.51
        version: 0.0.51
      - name: patterns-operator.v0.0.50
        version: 0.0.50
      - name: patterns-operator.v0.0.49
        version: 0.0.49
      - name: patterns-operator.v0.0.48
        version: 0.0.48
      - name: patterns-operator.v0.0.47
        version: 0.0.47
      - name: patterns-operator.v0.0.46
        version: 0.0.46
      - name: patterns-operator.v0.0.45
        version: 0.0.45
      - name: patterns-operator.v0.0.44
        version: 0.0.44
      - name: patterns-operator.v0.0.43
        version: 0.0.43
      - name: patterns-operator.v0.0.42
        version: 0.0.42
      - name: patterns-operator.v0.0.41
        version: 0.0.41
      - name: patterns-operator.v0.0.40
        version: 0.0.40
      - name: patterns-operator.v0.0.39
        version: 0.0.39
      - name: patterns-operator.v0.0.38
        version: 0.0.38
      - name: patterns-operator.v0.0.37
        version: 0.0.37
      - name: patterns-operator.v0.0.36
        version: 0.0.36
      - name: patterns-operator.v0.0.35
        version: 0.0.35
      - name: patterns-operator.v0.0.34
        version: 0.0.34
      - name: patterns-operator.v0.0.33
        version: 0.0.33
      - name: patterns-operator.v0.0.32
        version: 0.0.32
      - name: patterns-operator.v0.0.31
        version: 0.0.31
      - name: patterns-operator.v0.0.30
        version: 0.0.30
      - name: patterns-operator.v0.0.29
        version: 0.0.29
      - name: patterns-operator.v0.0.28
        version: 0.0.28
      - name: patterns-operator.v0.0.27
        version: 0.0.27
      - name: patterns-operator.v0.0.26
        version: 0.0.26
      - name: patterns-operator.v0.0.25
        version: 0.0.25
      - name: patterns-operator.v0.0.24
        version: 0.0.24
      - name: patterns-operator.v0.0.23
        version: 0.0.23
      - name: patterns-operator.v0.0.22
        version: 0.0.22
      - name: patterns-operator.v0.0.21
        version: 0.0.21
      - name: patterns-operator.v0.0.20
        version: 0.0.20
      - name: patterns-operator.v0.0.19
        version: 0.0.19
      - name: patterns-operator.v0.0.18
        version: 0.0.18
      - name: patterns-operator.v0.0.17
        version: 0.0.17
      - name: patterns-operator.v0.0.16
        version: 0.0.16
      - name: patterns-operator.v0.0.15
        version: 0.0.15
      - name: patterns-operator.v0.0.14
        version: 0.0.14
      - name: patterns-operator.v0.0.13
        version: 0.0.13
      - name: patterns-operator.v0.0.12
        version: 0.0.12
      - name: patterns-operator.v0.0.11
        version: 0.0.11
      - name: patterns-operator.v0.0.10
        version: 0.0.10
      - name: patterns-operator.v0.0.9
        version: 0.0.9
      - name: patterns-operator.v0.0.8
        version: 0.0.8
      - name: patterns-operator.v0.0.7
        version: 0.0.7
      - name: patterns-operator.v0.0.6
        version: 0.0.6
      - name: patterns-operator.v0.0.5
        version: 0.0.5
      - name: patterns-operator.v0.0.4
        version: 0.0.4
      - name: patterns-operator.v0.0.3
        version: 0.0.3
      - name: patterns-operator.v0.0.2
        version: 0.0.2
      - name: patterns-operator.v0.0.1
        version: 0.0.1
      name: fast
    defaultChannel: fast
    packageName: patterns-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com/en/products/edge/validated-patterns
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/
    name: varnish-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: varnish-operator.v0.27.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "caching.ibm.com/v1alpha1",
                "kind": "VarnishCluster",
                "metadata": {
                  "name": "varnishcluster-sample"
                },
                "spec": {
                  "backend": {
                    "port": "web",
                    "selector": {
                      "app": "nginx"
                    }
                  },
                  "replicas": 1,
                  "service": {
                    "port": 80
                  },
                  "varnish": {
                    "args": [
                      "-p",
                      "default_ttl=3600",
                      "-p",
                      "default_grace=3600"
                    ]
                  },
                  "vcl": {
                    "configMapName": "vcl-files",
                    "entrypointFileName": "entrypoint.vcl"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: ibmcom/varnish-operator:0.27.2
          createdAt: 2021-02-15T11:15:31+0000
          description: Run and manage Varnish clusters on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/varnish-operator
          support: https://ibm.github.io/varnish-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VarnishCluster is the Schema for the varnishclusters API
            displayName: Varnish Cluster
            kind: VarnishCluster
            name: varnishclusters.caching.ibm.com
            version: v1alpha1
        description: |
          Run and manage Varnish clusters on Kubernetes

          The operator manages the whole lifecycle of the cluster: creating, deleting and keeping the cluster configuration up to date.

          The operator is responsible for building the VCL configuration using templates defined by the users and keeping the configuration up to date when relevant events occur (backend pod failure, scaling of the deployment, VCL configuration change).

          Example of a simple `VarnishCluster`:

          ```yaml
          apiVersion: caching.ibm.com/v1alpha1
          kind: VarnishCluster
          metadata:
            name: varnishcluster-sample
            namespace: varnish-ns
          spec:
            vcl:
              configMapName: vcl-files
              entrypointFileName: entrypoint.vcl
            replicas: 3
            backend:
              selector:
                app: nginx
            service:
              port: 80
          ```

          Full documentation can be found [here](https://ibm.github.io/varnish-operator/)
        displayName: Varnish Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - varnish
        - cache
        links:
        - name: Varnish Operator Docs
          url: https://ibm.github.io/varnish-operator
        - name: Varnish Operator Source Code
          url: https://github.com/IBM/varnish-operator
        maintainers:
        - email: tomash.sidei@ibm.com
          name: Tomash Sidei
        - email: cingram@us.ibm.com
          name: Craig Ingram
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/
        relatedImages:
        - ibmcom/varnish-operator:0.27.2
        - quay.io/openshift-community-operators/varnish-operator@sha256:46de72c5abad80cef3d6426be967131c51c2c15029a6a38744975578c805a44e
        version: 0.27.2
      entries:
      - name: varnish-operator.v0.27.2
        version: 0.27.2
      name: alpha
    defaultChannel: alpha
    packageName: varnish-operator
    provider:
      name: IBM
      url: https://www.ibm.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: devspaces
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devspacesoperator.v3.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "workspaceNamespaceDefault": "<username>-devspaces"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "devspaces",
                  "namespace": "openshift-devspaces"
                },
                "spec": {
                  "components": {},
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
          createdAt: "2025-09-11T22:30:49+00:00"
          description: Devfile v2 and v1 development solution, 1 instance per cluster,
            for portable, collaborative k8s workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/devspaces-images/
          skipRange: '>=3.0.0 <3.23.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Red Hat OpenShift Dev Spaces server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Red Hat OpenShift Dev Spaces instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Red Hat OpenShift Dev Spaces server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Red Hat OpenShift Dev Spaces when the CR status is set to **Available**, and you see a URL to Red Hat OpenShift Dev Spaces.

          ## Defaults
          By default, the operator deploys Red Hat OpenShift Dev Spaces with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/devspaces -n openshift-devspaces` to update Red Hat OpenShift Dev Spaces default installation options.
          See more in the [Installation guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces).

          ### Certificates
          Operator uses a default router certificate to secure Red Hat OpenShift Dev Spaces routes.
          Follow the [guide](https://docs.redhat.com/en/documentation/red_hat_openshift_dev_spaces/3.23/html/administration_guide/configuring-devspaces#importing-untrusted-tls-certificates)
          to import certificates into Red Hat OpenShift Dev Spaces.
        displayName: Red Hat OpenShift Dev Spaces
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        - devspaces
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/openshift-dev-spaces/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_openshift_dev_spaces
        - name: Operator GitHub Repo
          url: https://github.com/redhat-developer/devspaces-images/
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/devspaces/traefik-rhel9@sha256:95cf6a28028ec7a546a3d8d78377f204bb559c3d1022a73f08c646e76b85ce85
        - registry.redhat.io/devspaces/udi-rhel9@sha256:f7325ee7259a0585e8017d4442d778a37f4c8a67d81fe70c241a014001e37539
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/devspaces/devspaces-operator-bundle@sha256:a48dd94d5616183843ba939146b7e4d985210cdd902038ff80ff6df3df3d74ca
        - registry.redhat.io/devspaces-tech-preview/jetbrains-ide-rhel9@sha256:46bf7bea716df74f14d2629fb7e0e2ab692564cec348ffb8a6f4fea1739b23c2
        - registry.redhat.io/devspaces/code-rhel9@sha256:abc9b12131c8cdca0719e0404d354310f53769990b089f1a3adea31193798a8e
        - registry.redhat.io/devspaces/configbump-rhel9@sha256:cecbc3796058618da9ad0473a6f937aae5aca25776bb434894286f22baa4677c
        - registry.redhat.io/devspaces/dashboard-rhel9@sha256:c34cf5ac9bf95c1c5082afd5a2941b5e412be8ca43aef6a72e8fe5a279a43271
        - registry.redhat.io/devspaces/pluginregistry-rhel9@sha256:3ee131041761f5250b29d497043ecf7820e02e03e73ed0f929c17e2247c0d7a3
        - registry.redhat.io/devspaces-tech-preview/idea-rhel9@sha256:df24ebb239a2cc0d4543309f51a7268c7d524d428a536e0b3443c51089c3f87e
        - registry.redhat.io/devspaces/devspaces-rhel9-operator@sha256:39c95279bced68928d028ef026f82eafe53eddb3dbe57703c8007cb5a170a70a
        - registry.redhat.io/devspaces/server-rhel9@sha256:c0ec8c792f88aacf6ee31a4b6d2d847d864707a084148b48e4e402388989f868
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        version: 3.23.0
      entries:
      - name: devspacesoperator.v3.23.0
        version: 3.23.0
      - name: devspacesoperator.v3.22.0
        version: 3.22.0
      - name: devspacesoperator.v3.21.0
        version: 3.21.0
      - name: devspacesoperator.v3.20.0
        version: 3.20.0
      - name: devspacesoperator.v3.19.0
        version: 3.19.0
      - name: devspacesoperator.v3.18.1-0.1739475915.p
        version: 3.18.1+0.1739475915.p
      - name: devspacesoperator.v3.18.1
        version: 3.18.1
      - name: devspacesoperator.v3.18.0
        version: 3.18.0
      - name: devspacesoperator.v3.17.0
        version: 3.17.0
      - name: devspacesoperator.v3.16.1
        version: 3.16.1
      - name: devspacesoperator.v3.16.0
        version: 3.16.0
      - name: devspacesoperator.v3.15.0
        version: 3.15.0
      - name: devspacesoperator.v3.14.0
        version: 3.14.0
      - name: devspacesoperator.v3.13.0
        version: 3.13.0
      - name: devspacesoperator.v3.12.0-0.1712003563.p
        version: 3.12.0+0.1712003563.p
      - name: devspacesoperator.v3.12.0
        version: 3.12.0
      - name: devspacesoperator.v3.11.0
        version: 3.11.0
      - name: devspacesoperator.v3.10.1
        version: 3.10.1
      - name: devspacesoperator.v3.10.0
        version: 3.10.0
      - name: devspacesoperator.v3.9.1
        version: 3.9.1
      - name: devspacesoperator.v3.9.0
        version: 3.9.0
      - name: devspacesoperator.v3.8.0
        version: 3.8.0
      - name: devspacesoperator.v3.7.2
        version: 3.7.2
      - name: devspacesoperator.v3.7.1-0.1691407091.p
        version: 3.7.1+0.1691407091.p
      - name: devspacesoperator.v3.7.1
        version: 3.7.1
      - name: devspacesoperator.v3.7.1-0.1691548593.p
        version: 3.7.1+0.1691548593.p
      - name: devspacesoperator.v3.7.0
        version: 3.7.0
      - name: devspacesoperator.v3.6.0
        version: 3.6.0
      - name: devspacesoperator.v3.6.0-0.1685320856.p
        version: 3.6.0+0.1685320856.p
      - name: devspacesoperator.v3.6.0-0.1685322634.p
        version: 3.6.0+0.1685322634.p
      - name: devspacesoperator.v3.5.0
        version: 3.5.0
      - name: devspacesoperator.v3.4.0-0.1675914643.p
        version: 3.4.0+0.1675914643.p
      - name: devspacesoperator.v3.4.0
        version: 3.4.0
      - name: devspacesoperator.v3.3.0-0.1673528130.p
        version: 3.3.0+0.1673528130.p
      - name: devspacesoperator.v3.3.0
        version: 3.3.0
      name: stable
    defaultChannel: stable
    packageName: devspaces
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-deploy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-deploy-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.2-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39 '
          createdAt: "2025-09-02T05:43:37Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabsunsupported/central-configuration@sha256:3a35c2aada71a7f5ab8fe7d5c68441fe58b0b2c778b6ea85ac6f3f7b351f80ad
        - docker.io/xebialabsunsupported/deploy-operator@sha256:9ccca8ab999d4b6058c979bbb70568a63cf43a257ae65a787b2ff4d9bf470c39
        - docker.io/xebialabsunsupported/deploy-task-engine@sha256:15322f2263c2b9b70336177aec170e59497fa320281c2d73605918b30272b8f1
        - docker.io/xebialabsunsupported/xl-deploy@sha256:41feb1d24d70a0c78a560b338d05c96c5383d570e3ef1ff37daff0eb8c92b386
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:18452ba29f4e0f52b765e8f9359e4143ebfe1203160e6f64ce46ec9d4f095a62
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-deploy-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-deploy-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-deploy-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-deploy-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-deploy-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/os-shell",
                        "tag": "12-debian-12-r16"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665 '
          createdAt: "2024-10-16T05:39:21Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  master:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  worker:\n    replicaCount: 1\n    persistence:\n
          \     storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n      runAsUser:
          null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     runAsUser: null\n      runAsGroup: null\n    volumePermissions:\n
          \     enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n        runAsUser: null\n        runAsGroup: null\n
          \       fsGroup: null\n      containerSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \     runAsUser: null\n      runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:9716402247ed874f19574a0d5125d1035fd198f36081d9129c1d4d4ad2f714a6
        - docker.io/xebialabs/central-configuration@sha256:83359e49c66e4fab502cc75663b456eb37d81d7cb512f2c48667f4d6a324e0df
        - docker.io/xebialabs/deploy-operator@sha256:2b3413cd41f2b437e6027c8ae2cff12d1113e0ac1be69fc24840d68a05d3a665
        - docker.io/xebialabs/deploy-task-engine@sha256:8a381fd84483ea919fa7f8bee5cf19d9dec0d1d61cce73ec6b7c56fa504c6c93
        - docker.io/xebialabs/xl-deploy@sha256:77fdf6a3909367b060eba2ef698baddb77e4086d4d977f9c617d1e26a4427472
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        version: 24.1.10
      entries:
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-deploy-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-daid",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator without SCC creation and with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9 '
          createdAt: "2025-08-25T04:53:23Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind:
          DigitalaiDeploy\nmetadata:\n  name: daid-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  securityContextConstraints:\n
          \   enabled: false\n  centralConfiguration:\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  master:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  worker:\n    replicaCount:
          1\n    persistence:\n      storageClass: ''\n      size: 1Gi\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: SESSION_XLD\n
          \     haproxy.router.openshift.io/disable_cookies: \"false\"\n      haproxy.router.openshift.io/rewrite-target:
          /\n    hostname: '<mandatory-deploy-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n      securityContextConstraints:\n
          \       enabled: false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n
          \   install: true\n    persistence:\n      size: 1Gi\n      storageClass:
          ''\n    replicaCount: 1\n    podSecurityContext:\n      enabled: false\n
          \   containerSecurityContext:\n      enabled: false\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"daid-doc-digitalai-deploy-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xld-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Embedded PostgreSQL for
          DB management\n- Embedded RabbitMQ for message queue management\n\n### Security
          Configuration Details\n\n- `securityContextConstraints.enabled: false` -
          disables creation of SCCs;\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/deploy-operator@sha256:7d505a855728c761433ef7d46f0a64ffe2b4192f4819bad86c23582375099ac9
        - docker.io/xebialabs/deploy-task-engine@sha256:a97d67ea53e1b954ffb4d9514fd61426dcc0cab08922009eef02f32162026d6a
        - docker.io/xebialabs/xl-deploy@sha256:c2fe209f198fb2f42dc78a2e1544c026a6891d60682b4e47634234bbf7346540
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:397feef6fc1dffb7e2c3798bbe3bfebb86d6173c225eb61473029cc4500b0cbd
        - docker.io/xebialabs/central-configuration@sha256:581a0bc71ae3652ceb643d81927fcf8cabea92f1d301f5d354076baff32fbcde
        version: 24.1.21
      entries:
      - name: digitalai-deploy-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-deploy-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-deploy-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-deploy-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-deploy-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-deploy-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-deploy-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-deploy-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 20,
                      "probesReadinessTimeout": 20
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "classic",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "health": {
                      "enabled": true,
                      "periodScans": 10,
                      "probeFailureThreshold": 12,
                      "probesLivenessTimeout": 60,
                      "probesReadinessTimeout": 60
                    },
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470 '
          createdAt: "2025-09-26T18:18:41Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a70fceaad3b6a23e6f247b4dff41fa64349c4bbf841b66adbd0f8309c6acce74
        - docker.io/xebialabs/central-configuration@sha256:4b83d4e9a523ae3615579907c65790478a3b22315ca5afce594793424d1d33b1
        - docker.io/xebialabs/deploy-operator@sha256:41ffc26c1d99a808e537f2366d92473ecf76dbffbac699dfa04785c3016f7470
        - docker.io/xebialabs/deploy-task-engine@sha256:75fe942305bf3bb403eb243b157b88844363eb702f03227a61e76ad4321bf0ee
        - docker.io/xebialabs/xl-deploy@sha256:8536efdf530dde45b84b14191ebfdc7ec5ebb1844f84871032fe2d335b13908d
        version: 24.3.11
      entries:
      - name: digitalai-deploy-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-deploy-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-deploy-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-deploy-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-deploy-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-deploy-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-deploy-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-deploy-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "medium"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "daid-placeholders"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "centralConfiguration": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-db-host\u003e:5432/\u003cxld-database-name\u003e",
                        "username": "\u003cxld-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxld-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxld-report-db-host\u003e:5432/\u003cxld-report-database-name\u003e",
                        "username": "\u003cxld-report-username\u003e"
                      }
                    },
                    "mq": {
                      "driverClassName": "\u003cdriver-class-name\u003e",
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "master": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-master-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-master-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaimName": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaimName": ""
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-deploy-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "worker": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "diagnosticMode": {
                      "enabled": false
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "size": "\u003cdeploy-worker-storage-size\u003e",
                      "storageClass": "\u003cmandatory-deploy-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003cdeploy-worker-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "xld.digital.ai/v1alpha1",
                "kind": "DigitalaiDeploy",
                "metadata": {
                  "name": "digitalaideploy-sample"
                },
                "spec": {
                  "appContextRoot": "/",
                  "auth": {
                    "adminPassword": null
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "centralConfiguration": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/central-configuration-server/bin/run-in-operator.sh"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\necho \"Delete empty files to replace them with latest configuration\"\nfind /opt/xebialabs/central-configuration-server/centralConfiguration -maxdepth 1 -type f -empty -print -delete\n{{- if .Values.oidc.enabled }}\nif [[ ${GENERATE_XL_CONFIG,,} != \"true\" ]]; then\n  echo \"Not generating deploy-oidc.yaml as GENERATE_XL_CONFIG != 'true'\"\nelif [[ -e ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u0026\u0026 ! -f \"${APP_HOME}/centralConfiguration/deploy-oidc.yaml\" ]]; then\n  echo \"Generate configuration file deploy-oidc.yaml from environment parameters\"\n  sed -e \"s#\\${XL_DB_DRIVER}#${XL_DB_DRIVER}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${DNS_RESOLVER}#${DNS_RESOLVER}#g\" \\\n  -e \"s#\\${HOSTNAME}#${HOSTNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${JMS_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XL_CLUSTER_MODE}#${XL_CLUSTER_MODE}#g\" \\\n  -e \"s#\\${XL_DB_URL}#${XL_DB_URL}#g\" \\\n  -e \"s#\\${XL_DB_USERNAME}#${XL_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_DB_PASSWORD}#${XL_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_DB_MAX_POOL_SIZE}#${XL_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_URL}#${XL_REPORT_DB_URL}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_USERNAME}#${XL_REPORT_DB_USERNAME}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_PASSWORD}#${XL_REPORT_DB_PASSWORD}#g\" \\\n  -e \"s#\\${XL_REPORT_DB_MAX_POOL_SIZE}#${XL_REPORT_DB_MAX_POOL_SIZE}#g\" \\\n  -e \"s#\\${XL_METRICS_ENABLED}#${XL_METRICS_ENABLED}#g\" \\\n  -e \"s#\\${XLD_IN_PROCESS}#${XLD_IN_PROCESS}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_NAME}#${XLD_TASK_QUEUE_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#${XLD_TASK_QUEUE_IN_PROCESS_MAX_DISK_USAGE}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#${XLD_TASK_QUEUE_IN_PROCESS_SHUTDOWN_TIMEOUT}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#${XLD_TASK_QUEUE_DRIVER_CLASS_NAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_URL}#${XLD_TASK_QUEUE_URL}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_USERNAME}#${XLD_TASK_QUEUE_USERNAME}#g\" \\\n  -e \"s#\\${XLD_TASK_QUEUE_PASSWORD}#${XLD_TASK_QUEUE_PASSWORD}#g\" \\\n  -e \"s#\\${HOSTNAME_SUFFIX}#${HOSTNAME_SUFFIX}#g\" \\\n  -e \"s#\\${XL_LICENSE_KIND}#${XL_LICENSE_KIND}#g\" \\\n  -e \"s#\\${GENERATE_XL_CONFIG}#${GENERATE_XL_CONFIG}#g\" \\\n  -e \"s#\\${USE_IP_AS_HOSTNAME}#${USE_IP_AS_HOSTNAME}#g\" \\\n  -e \"s#\\${ENABLE_SATELLITE}#${ENABLE_SATELLITE}#g\" \\\n  -e \"s#\\${CENTRAL_CONFIG_ENCRYPT_KEY}#${CENTRAL_CONFIG_ENCRYPT_KEY}#g\" \\\n  -e \"s#\\${USE_CACHE}#${USE_CACHE}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_ID}#${OIDC_CLIENT_ID}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_SECRET}#${OIDC_CLIENT_SECRET}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}#g\" \\\n  -e \"s#\\${OIDC_ACCESS_TOKEN_SECRET_KEY}#${OIDC_ACCESS_TOKEN_SECRET_KEY}#g\" \\\n  ${APP_HOME}/central-conf/deploy-oidc.yaml.template \u003e ${APP_HOME}/centralConfiguration/deploy-oidc.yaml\nfi\n{{- end }}\n\n# copy central-conf files\ncd ${APP_HOME}/central-conf\necho \"... Copying default centralConfiguration from ${APP_HOME}/central-conf\"\nfor f in *; do\n  if [[ $f == *.template ]]; then\n    continue\n  fi\n  if [ -f ${APP_HOME}/centralConfiguration/$f ]; then\n    echo \"... Not copying $f because it already exists in the centralConfiguration directory\"\n  else\n    echo \"... Copying $f to the centralConfiguration directory\"\n    cp -R $f ${APP_HOME}/centralConfiguration/\n  fi\ndone\ncd ${APP_HOME}\n\nexec /opt/xebialabs/central-configuration-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "central-conf_deploy-oidc-yaml-template": {
                        "content": "deploy.security:\n  {{- if .Values.oidc.enabled }}\n  auth:\n    providers:\n      oidc:\n        clientId: \"${OIDC_CLIENT_ID}\"\n        clientSecret: \"${OIDC_CLIENT_SECRET}\"\n        {{- if .Values.oidc.clientAuthMethod }}\n        clientAuthMethod: {{ .Values.oidc.clientAuthMethod | quote }}\n        {{- end }}\n        {{- if .Values.oidc.clientAuthJwt.enable }}\n        clientAuthJwt:\n          jwsAlg: {{ .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n          tokenKeyId: {{ .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n          {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n          keyStore:\n            path: {{ .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\"\n            type: {{ .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.key.enable }}\n          key:\n            alias: {{ .Values.oidc.clientAuthJwt.key.alias | quote }}\n            password: \"${OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\"\n          {{- end }}\n        {{- end }}\n        issuer: {{ .Values.oidc.issuer | quote }}\n        keyRetrievalUri: {{ .Values.oidc.keyRetrievalUri | quote }}\n        accessTokenUri: {{ .Values.oidc.accessTokenUri | quote }}\n        userAuthorizationUri: {{ .Values.oidc.userAuthorizationUri | quote }}\n        logoutUri: {{ .Values.oidc.logoutUri | quote }}\n        redirectUri: {{ .Values.oidc.redirectUri | quote }}\n        postLogoutRedirectUri: {{ .Values.oidc.postLogoutRedirectUri | quote }}\n        rolesClaimName: {{ .Values.oidc.rolesClaimName | quote }}\n        userNameClaimName: {{ .Values.oidc.userNameClaimName | quote }}\n        {{- if .Values.oidc.scopes }}\n        scopes: {{ .Values.oidc.scopes }}\n        {{- end }}\n        {{- if .Values.oidc.idTokenJWSAlg }}\n        idTokenJWSAlg: {{ .Values.oidc.idTokenJWSAlg | quote }}\n        {{- end }}\n        {{- if .Values.oidc.accessToken.enable }}\n        access-token:\n          {{- if .Values.oidc.accessToken.issuer }}\n          issuer: {{ .Values.oidc.accessToken.issuer | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.audience }}\n          audience: {{ .Values.oidc.accessToken.audience | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.keyRetrievalUri }}\n          keyRetrievalUri: {{ .Values.oidc.accessToken.keyRetrievalUri | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.jwsAlg }}\n          jwsAlg: {{ .Values.oidc.accessToken.jwsAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.secretKey }}\n          secretKey: \"${OIDC_ACCESS_TOKEN_SECRET_KEY}\"\n          {{- end }}\n        {{- end }}\n        loginMethodDescription: {{ default \"External login (OpenID Connect)\" .Values.oidc.loginMethodDescription | quote }}\n        {{- if .Values.oidc.proxyHost }}\n        proxyHost: {{ .Values.oidc.proxyHost | quote }}\n        {{- end }}\n        {{- if .Values.oidc.proxyPort }}\n        proxyPort: {{ .Values.oidc.proxyPort | quote }}\n        {{- end }}\n  {{- end }}\n",
                        "mode": 432,
                        "path": "central-conf/deploy-oidc.yaml.template"
                      },
                      "central-conf_deploy-server-yaml-template": {
                        "content": "deploy.server:\n  bind-hostname: 0.0.0.0\n  bind-port: {{ .Values.master.containerPorts.deployPekko }}\n  license:\n    daysBeforeWarning: 10\n  {{- if .Values.oidc.enabled }}    \n  security:\n    auth:\n      provider: \"oidc\"\n  {{- end }}\npekko:\n  io:\n    dns:\n      resolver: \"${DNS_RESOLVER}\"\n  # loggers:\n  #  - \"org.apache.pekko.event.slf4j.Slf4jLogger\"\n  loglevel: \"INFO\"\n",
                        "mode": 432,
                        "path": "central-conf/deploy-server.yaml.template"
                      }
                    },
                    "containerPorts": {
                      "ccHttp": 8888,
                      "ccHttps": 8843
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.centralConfiguration.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "encryptKey": null,
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/central-configuration",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "migrateFromEmbedded": false,
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/tmp",
                        "/opt/xebialabs/central-configuration-server/centralConfiguration",
                        "/opt/xebialabs/central-configuration-server/conf",
                        "/opt/xebialabs/central-configuration-server/log"
                      ],
                      "enabled": false,
                      "existingClaim": "",
                      "paths": [],
                      "selector": {},
                      "single": false,
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "externalIPs": [],
                      "externalTrafficPolicy": "Cluster",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "centralConfiguration"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "nodePorts": {
                        "ccHttp": "",
                        "ccHttps": ""
                      },
                      "portEnabled": true,
                      "portNames": {
                        "ccHttp": "deploy-http-cc",
                        "ccHttps": "deploy-https-cc"
                      },
                      "ports": {
                        "ccHttp": 8888,
                        "ccHttps": 8843
                      },
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "ClusterIP"
                    },
                    "sidecars": [],
                    "ssl": {
                      "enabled": false,
                      "keystore": {
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "keystore.{{ .Values.ssl.keystoreType }}",
                            "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                          }
                        }
                      },
                      "keystoreKeypassword": "changeme",
                      "keystorePassword": "changeme",
                      "keystoreType": "pkcs12"
                    },
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 10,
                      "path": "/login",
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "centralConfiguration"
                    },
                    "terminationGracePeriodSeconds": 10,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.centralConfiguration.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.centralConfiguration.containerSecurityContext.runAsUser }}:{{ .Values.centralConfiguration.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "driverClassName": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "quorum",
                      "url": "",
                      "username": null
                    }
                  },
                  "externalCentralConfig": true,
                  "extraDeploy": [],
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-daid",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating deploy self-signed cert\"\nHOSTNAME=\"{{- include \"deploy.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-deploy -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Deploy,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-deploy -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-deploy\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/deployit-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/deployit-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-deploy-server/conf/deployit-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-deploy-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"common.names.fullname\" . }}-license"
                    }
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "master": {
                    "affinity": {},
                    "args": [
                      "--",
                      "/opt/xebialabs/xl-deploy-server/bin/run-in-operator.sh"
                    ],
                    "clusterNodeHostnameSuffix": ".{{ include \"deploy.names.master\" $ }}.{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.master.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.master.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.master.podServiceTemplate.overrideHostname .Values.master.podServiceTemplate.overrideHostnames }}\nallMasters=(\n{{- range $podNumber := untilStep 0 (int .Values.master.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $masterHostname := include \"deploy.masterHostname\" $newValues }}\n\"{{ $masterHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allMasters[$POD_NUMBER]})\necho \"Using master hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using master port: $SERVER_PORT\"\nexec /opt/xebialabs/xl-deploy-server/bin/run-in-container.sh $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      },
                      "node-conf_xl-deploy-conf-template": {
                        "content": "# The ${APP_HOME}/node-conf/xl-deploy.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-deploy.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-deploy.conf and this is specified in ${APP_HOME}/conf/xld-wrapper.conf.common\nxl {\n  server.hostname = ${HOSTNAME}${HOSTNAME_SUFFIX}\n  cluster.node {\n    id = ${XL_NODE_NAME}\n    hostname = ${XL_NODE_NAME}\n    clusterPort = 5531\n  }\n}\n",
                        "mode": 432,
                        "path": "node-conf/xl-deploy.conf.template"
                      }
                    },
                    "containerPorts": {
                      "deployClusterPekko": 25520,
                      "deployHttp": 4516,
                      "deployHttps": 4517,
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.master.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "forceUpgrade": true,
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-deploy",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/xl-deploy-server/artifacts",
                        "/opt/xebialabs/xl-deploy-server/cache",
                        "/opt/xebialabs/xl-deploy-server/centralConfiguration",
                        "/opt/xebialabs/xl-deploy-server/conf",
                        "/opt/xebialabs/xl-deploy-server/driver/jdbc",
                        "/opt/xebialabs/xl-deploy-server/driver/mq",
                        "/opt/xebialabs/xl-deploy-server/export",
                        "/opt/xebialabs/xl-deploy-server/log",
                        "/opt/xebialabs/xl-deploy-server/node-conf",
                        "/opt/xebialabs/xl-deploy-server/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/xl-deploy-server/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "master"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32180
                      },
                      "overrideHostname": "{{ include \"deploy.names.master\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.master\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32180
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "lb": {
                        "annotations": {},
                        "clusterIP": "",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployHttp": "",
                          "deployHttps": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployHttp": "deploy-http",
                          "deployHttps": "deploy-https"
                        },
                        "ports": {
                          "deployHttp": 80,
                          "deployHttps": 443
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      },
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "master"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployClusterPekko": "",
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployClusterPekko": "cluster-pekko",
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployClusterPekko": 25520,
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "path": null,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "master"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.master.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.master.containerSecurityContext.runAsUser }}:{{ .Values.master.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-daid"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-daid",
                      "name": "nginx-daid"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "loginMethodDescription": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaimName": null,
                    "scopes": null,
                    "userAuthorizationUri": null,
                    "userNameClaimName": null
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "hasReport": true,
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-deploy"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "medium",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-deploy",
                      "file": "/app/deploy_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "SESSION_XLD",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "satellite": {
                    "enabled": false
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": "changeme",
                    "keystorePassword": "changeme",
                    "keystoreType": "pkcs12"
                  },
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "usaCache": false,
                  "worker": {
                    "affinity": {},
                    "args": "- --\n- /opt/xebialabs/deploy-task-engine/bin/run-in-operator.sh",
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "configuration": {
                      "bin_run-in-operator-sh": {
                        "content": "#!/bin/bash\n\n{{- if .Values.worker.podServiceTemplate.enabled }}\nPOD_NUMBER=$(echo $POD_NAME | grep -Eo \"[0-9]+$\")\n{{- if contains .Values.worker.podServiceTemplate.serviceMode \"SingleHostname;MultiService\" }}\nexport SERVER_PORT=$(echo $SERVER_PORT+$POD_NUMBER | bc)\n{{- end }}\n{{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\nallWorkers=(\n{{- range $podNumber := untilStep 0 (int .Values.worker.replicaCount) 1 }}\n{{- $newValues := merge (dict \"podNumber\" $podNumber) $ }}\n{{- $workerHostname := include \"deploy.workerHostname\" $newValues }}\n\"{{ $workerHostname }}\"\n{{- end }}\n)\nexport OVERRIDE_HOSTNAME=$(echo ${allWorkers[$POD_NUMBER]})\necho \"Using worker hostname: $OVERRIDE_HOSTNAME\"\n{{- end }}\n{{- end }}\necho \"Using worker port: $SERVER_PORT\"\nexec /opt/xebialabs/deploy-task-engine/bin/run-in-container.sh \\\n  -name \"$POD_NAME\" \\\n  {{- if .Values.worker.podServiceTemplate.enabled }}\n  {{- if or .Values.worker.podServiceTemplate.overrideHostname .Values.worker.podServiceTemplate.overrideHostnames }}\n  -hostname \"$OVERRIDE_HOSTNAME$HOSTNAME_SUFFIX\" \\\n  {{- else }}\n  -hostname \"$POD_NAME$HOSTNAME_SUFFIX\" \\\n  {{- end }}\n  {{- end }}\n  {{- include \"deploy.workerMasters\" $ }}\n  -api \"{{ include \"deploy.masterLbUrl\" . }}\" \\\n  $@\n",
                        "mode": 493,
                        "path": "bin/run-in-operator.sh"
                      }
                    },
                    "containerPorts": {
                      "deployJmxExporter": 9100,
                      "deployPekko": 8180
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "enabled": true,
                      "readOnlyRootFilesystem": true,
                      "runAsNonRoot": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "debugMode": {
                      "enabled": false,
                      "remoteJvmParams": "{{- if .Values.worker.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=*:8001{{- end }}"
                    },
                    "defaultInitContainers": {
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "diagnosticMode": {
                      "args": [
                        "--",
                        "sleep",
                        "infinity"
                      ],
                      "command": [
                        "/opt/xebialabs/tini"
                      ],
                      "enabled": false
                    },
                    "dnsConfig": {},
                    "dnsPolicy": "",
                    "extraConfiguration": {},
                    "extraContainerPorts": [],
                    "extraEnvVars": [],
                    "extraEnvVarsCM": "",
                    "extraEnvVarsSecret": "",
                    "extraSecrets": {},
                    "extraSecretsPrependReleaseName": false,
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "hostAliases": [],
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/deploy-task-engine",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "initContainers": [],
                    "jvm": {
                      "options": {
                        "networkaddress.cache.ttl": 30
                      }
                    },
                    "jvmArgs": "",
                    "lifecycleHooks": {},
                    "livenessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "logback": {
                      "globalLoggingLevel": "info",
                      "scanEnabled": true,
                      "scanPeriod": "30 seconds"
                    },
                    "nodeAffinityPreset": {
                      "key": "",
                      "type": "",
                      "values": []
                    },
                    "nodeSelector": {},
                    "overrideName": "",
                    "pdb": {
                      "create": false,
                      "maxUnavailable": "",
                      "minAvailable": 1
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "annotations": {
                        "helm.sh/resource-policy": "keep"
                      },
                      "emptyDirPaths": [
                        "/opt/xebialabs/deploy-task-engine/cache",
                        "/opt/xebialabs/deploy-task-engine/conf",
                        "/opt/xebialabs/deploy-task-engine/driver/jdbc",
                        "/opt/xebialabs/deploy-task-engine/driver/mq",
                        "/opt/xebialabs/deploy-task-engine/export",
                        "/opt/xebialabs/deploy-task-engine/log",
                        "/opt/xebialabs/deploy-task-engine/node-conf",
                        "/opt/xebialabs/deploy-task-engine/plugins",
                        "/tmp"
                      ],
                      "enabled": true,
                      "existingClaim": "",
                      "paths": [
                        "/opt/xebialabs/deploy-task-engine/work"
                      ],
                      "selector": {},
                      "single": false,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "podAffinityPreset": "",
                    "podAnnotations": {},
                    "podAntiAffinityPreset": "soft",
                    "podLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "podManagementPolicy": "OrderedReady",
                    "podSecurityContext": {
                      "enabled": true,
                      "fsGroup": 10001,
                      "runAsUser": 10001
                    },
                    "podServiceTemplate": {
                      "annotations": {},
                      "clusterIPs": [],
                      "enabled": false,
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraPorts": [],
                      "labels": {
                        "app.kubernetes.io/component": "worker"
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}",
                      "nodePorts": {
                        "deployPekko": 32185
                      },
                      "overrideHostname": "{{ include \"deploy.names.worker\" . }}-{{ .podNumber }}",
                      "overrideHostnameSuffix": ".{{ include \"common.names.namespace\" . }}.svc.cluster.local",
                      "overrideHostnames": [],
                      "podLabels": {
                        "statefulset.kubernetes.io/pod-name": "{{ printf \"%s-%d\" (include \"deploy.names.worker\" $) .podNumber }}"
                      },
                      "portEnabled": true,
                      "portNames": {
                        "deployPekko": "deploy-pekko"
                      },
                      "ports": {
                        "deployPekko": 32185
                      },
                      "publishNotReadyAddresses": true,
                      "serviceMode": "MultiService",
                      "sessionAffinity": "None",
                      "sessionAffinityConfig": {},
                      "type": "NodePort"
                    },
                    "priorityClassName": "",
                    "readinessProbe": {
                      "enabled": true,
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "replicaCount": 3,
                    "resources": {},
                    "resourcesPreset": "micro",
                    "schedulerName": "",
                    "services": {
                      "pekko": {
                        "annotations": {},
                        "clusterIP": "None",
                        "externalIPs": [],
                        "externalTrafficPolicy": "Cluster",
                        "extraPorts": [],
                        "labels": {
                          "app.kubernetes.io/component": "worker"
                        },
                        "loadBalancerIP": "",
                        "loadBalancerSourceRanges": [],
                        "nodePorts": {
                          "deployPekko": ""
                        },
                        "portEnabled": true,
                        "portNames": {
                          "deployJmxExporter": "deploy-jmx",
                          "deployPekko": "deploy-pekko"
                        },
                        "ports": {
                          "deployPekko": 8180
                        },
                        "publishNotReadyAddresses": true,
                        "sessionAffinity": "None",
                        "sessionAffinityConfig": {},
                        "type": "ClusterIP"
                      }
                    },
                    "sidecars": [],
                    "startupProbe": {
                      "enabled": true,
                      "failureThreshold": 12,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 10,
                      "successThreshold": 1,
                      "timeoutSeconds": 20
                    },
                    "statefulsetAnnotations": {},
                    "statefulsetLabels": {
                      "app.kubernetes.io/component": "worker"
                    },
                    "terminationGracePeriodSeconds": 90,
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "updateStrategy": {
                      "type": "OnDelete"
                    },
                    "useIpAsHostname": false,
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": false,
                      "image": {
                        "digest": "",
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-deploy",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "ephemeral-storage": "2Gi",
                          "memory": "192Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "ephemeral-storage": "50Mi",
                          "memory": "128Mi"
                        }
                      },
                      "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.worker.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.worker.containerSecurityContext.runAsUser }}:{{ .Values.worker.podSecurityContext.fsGroup }}\"\ndone\n"
                    }
                  },
                  "xldInProcess": false
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "daid-min": {
                "description": "Minimal configuration for the Digital.ai Deploy operator with embedded Postgresql and RabbitMQ and disabled route setup"
              },
              "daid-placeholders": {
                "description": "Configuration for the Digital.ai Deploy operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalaideploy-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b '
          createdAt: "2025-09-26T16:30:53Z"
          description: Digital.ai Deploy enables organizations to achieve fast, secure,
            and reliable deployments to hybrid infrastructure — including containers,
            private and public clouds, middleware, and mainframe.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/deploy-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiDeploy
            name: digitalaideploys.xld.digital.ai
            version: v1alpha1
        description: "Digital.ai Deploy enables organizations to achieve fast, secure,
          and reliable deployments to hybrid infrastructure — including containers,
          private and public clouds, middleware, and mainframe.\n\n## About this Operator\n\nDigital.ai
          Deploy operator is based on helm chart that is available from [here](https://github.com/digital-ai/deploy-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Deploy containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xld.digital.ai/v1alpha1\nkind: DigitalaiDeploy\nmetadata:\n
          \ name: daid-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66C46srO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECPqEw2Wp+c6yAgMDDUB1cQB+AAUAAACQFrl6s2pnsB+GJD8vlN3Y0SItmbtfPy6n2A5qREEJWWLN9OYLu7BokScBMyFChFjIhQGwCpjMP4j+VLCgpW6GKREmYHQgKjWqWn7A+DMF9eT68ygZAD+ceIZB5buvsGM2LCYzyHJcmtujv+hpqevoTgOKKMd4U3wVV96n4B5QbkVXHYtGZWbWxk5gCHLoWhV5dAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVP+nQgVx6wurZB9hBxaIkk/6EEAPQ=='\n
          \ hooks:\n    getLicense:\n      enabled: true\n  centralConfiguration:\n
          \   replicaCount: 1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  master:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  worker:\n
          \   replicaCount: 1\n    persistence:\n      storageClass: ''\n      size:
          1Gi\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          SESSION_XLD\n      haproxy.router.openshift.io/disable_cookies: \"false\"\n
          \     haproxy.router.openshift.io/rewrite-target: /\n    hostname: '<mandatory-deploy-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"medium\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n    containerSecurityContext:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\n####
          Configuration Details\n\nThe sample configuration uses:\n\n- Autogenerated
          limited license\n- Simple administrator initial password\n- Default keystore
          setup\n- Only one replica is set for all pods\n- All Pods are missing resource
          limit setup\n- Default storage class and minimal size for persistent storage\n-
          Embedded PostgreSQL for DB management\n- Embedded RabbitMQ for message queue
          management\n\n### Security Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined spans (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Deploy, use this
          [guide](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-setup-truststore.html)\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-deploy-offline-plugin-management.html)"
        displayName: Digital.ai Deploy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai
        - digital.ai
        - deploy
        - devops
        links:
        - name: Digital.ai Deploy
          url: https://digital.ai/products/deploy/
        - name: Digital.ai Deploy Documentation
          url: https://docs.digital.ai/bundle/devops-deploy-version-v.24.1/page/deploy/operator/xl-op-before-you-begin.html
        - name: Digital.ai Deploy Helm Chart
          url: https://github.com/digital-ai/deploy-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/central-configuration@sha256:0c46275ffa3b251ccb76d2623e0e5a4cd1441bf7f94901845bda743f087bc6e4
        - docker.io/xebialabs/deploy-operator@sha256:d192ef9008d72c9792e1b7b988bdae8f1bbce45ed0e44afc48e3e268b657b38b
        - docker.io/xebialabs/deploy-task-engine@sha256:13bb779d6041836acb95c95f1eed75f4b5b2940d75687306183a9f8a2177cab3
        - docker.io/xebialabs/xl-deploy@sha256:e4212d20fad19f09cbe53796371bec7fc368cfe63316f1b7bccecf5faf6ea733
        - registry.connect.redhat.com/digital-ai/deploy-operator@sha256:a7a4d20b5acc59dbb8c99a37c19b2749f15ed0be7f841490a829796865c5a956
        version: 25.1.5
      entries:
      - name: digitalai-deploy-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-deploy-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-deploy-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-deploy-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-deploy-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-deploy-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-deploy-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: https://opdev.github.io/
    name: loki-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.example.com/v1alpha1",
                "kind": "Loki",
                "metadata": {
                  "annotations": {
                    "helm.sdk.operatorframework.io/install-disable-crds": "true"
                  },
                  "name": "loki-sample"
                },
                "spec": {
                  "global": {
                    "clusterDomain": "cluster.local",
                    "dnsNamespace": "openshift-dns",
                    "dnsService": "dns-default"
                  },
                  "loki": {
                    "auth_enabled": false,
                    "commonConfig": {
                      "replication_factor": 1
                    },
                    "storage": {
                      "type": "filesystem"
                    }
                  },
                  "monitoring": {
                    "dashboards": {
                      "enabled": false
                    },
                    "lokiCanary": {
                      "enabled": false
                    },
                    "rules": {
                      "alerting": false,
                      "enabled": false
                    },
                    "selfMonitoring": {
                      "enabled": false,
                      "grafanaAgent": {
                        "installOperator": false
                      }
                    },
                    "serviceMonitor": {
                      "enabled": false,
                      "metricsInstance": {
                        "enabled": false
                      }
                    }
                  },
                  "rbac": {
                    "sccEnabled": true
                  },
                  "singleBinary": {
                    "replicas": 1
                  },
                  "test": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/loki-helm-operator-controller:0.0.1
          createdAt: "2023-07-14T15:06:17Z"
          description: An experimental operator that installs loki from its official
            helmchart.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the loki operator helm chart
            displayName: Loki
            kind: Loki
            name: lokis.charts.example.com
            version: v1alpha1
        description: Loki operator based on the helm chart with single binary mode
          deployment optionality
        displayName: Loki Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loki
        - single-binary
        - demo
        - helm
        links:
        - name: Loki Helm Operator
          url: https://github.com/skattoju/loki-helm-operator
        maintainers:
        - email: skattoju@redhat.com
          name: Sid Kattoju
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: https://opdev.github.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/opdev/loki-helm-operator-controller:0.0.1
        - quay.io/openshift-community-operators/loki-helm-operator@sha256:fa3662ec16e6cb3f866ad46982a50d8fdefb43c68f3caa13291239013e03af13
        version: 0.0.1
      entries:
      - name: loki-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: loki-helm-operator
    provider:
      name: opdev
      url: https://opdev.github.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Infinispan
      provider-url: ""
    name: infinispan
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinispan-operator.v2.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.2.5.Final
          description: Create and manage Infinispan clusters.
          olm.skipRange: '>=2.1.x <2.2.1'
          operators.operatorframework.io/builder: operator-sdk-v1.3.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/openshift-community-operators/infinispan@sha256:3ba9d20e83825f3c041a7dfd195b0800f74fa0f5cd1e6218fbd3547fedd85d64
        - quay.io/infinispan/operator:2.2.5.Final
        version: 2.2.5
      entries:
      - name: infinispan-operator.v2.2.5
        version: 2.2.5
      - name: infinispan-operator.v2.2.4
        version: 2.2.4
      - name: infinispan-operator.v2.2.3
        version: 2.2.3
      - name: infinispan-operator.v2.2.2
        version: 2.2.2
      - name: infinispan-operator.v2.2.1
        version: 2.2.1
      - name: infinispan-operator.v2.2.0
        version: 2.2.0
      name: 2.2.x
    - currentCSV: infinispan-operator.v2.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.3.7.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:f165693aa550903d7a26b429ed4d093f60194a7a04148e65e61c52d67b527fad
        - quay.io/infinispan/operator:2.3.7.Final
        version: 2.3.8
      entries:
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.3.x
    - currentCSV: infinispan-operator.v2.4.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.17.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:5e1a1fdb5a78a34155abe77754c774c65335ea8cbfdec2a222ee22a062c2db46
        - quay.io/infinispan/operator:2.4.17.Final
        version: 2.4.17
      entries:
      - name: infinispan-operator.v2.4.17
        version: 2.4.17
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: 2.4.x
    - currentCSV: infinispan-operator.v2.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: quay.io/infinispan/operator:2.4.18.Final
          description: Create and manage Infinispan clusters.
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Infinispan is an in-memory data store and open-source project.

          ### Infinispan
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Infinispan cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Infinispan pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Infinispan Operator lets you adjust resource allocation to suit your requirements.
        displayName: Infinispan Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Blog
          url: https://blog.infinispan.org/
        - name: Documentation
          url: https://infinispan.org/documentation
        - name: Chat
          url: https://infinispan.zulipchat.com/#narrow/stream/185835-infinispan-cloud
        - name: Operator Source Code
          url: https://github.com/infinispan/infinispan-operator
        maintainers:
        - email: infinispan-dev@lists.jboss.org
          name: Infinispan Community
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Infinispan
        relatedImages:
        - quay.io/community-operator-pipeline-prod/infinispan@sha256:e43b0581415af648c0b82830377e354a951814d3eb7825bbe31b9f3e633917c5
        - quay.io/infinispan/operator:2.4.18.Final
        version: 2.4.18
      entries:
      - name: infinispan-operator.v2.4.18
        version: 2.4.18
      - name: infinispan-operator.v2.4.16
        version: 2.4.16
      - name: infinispan-operator.v2.4.15
        version: 2.4.15
      - name: infinispan-operator.v2.4.14
        version: 2.4.14
      - name: infinispan-operator.v2.4.13
        version: 2.4.13
      - name: infinispan-operator.v2.4.12
        version: 2.4.12
      - name: infinispan-operator.v2.4.11
        version: 2.4.11
      - name: infinispan-operator.v2.4.10
        version: 2.4.10
      - name: infinispan-operator.v2.4.9
        version: 2.4.9
      - name: infinispan-operator.v2.4.8
        version: 2.4.8
      - name: infinispan-operator.v2.4.7
        version: 2.4.7
      - name: infinispan-operator.v2.4.6
        version: 2.4.6
      - name: infinispan-operator.v2.4.5
        version: 2.4.5
      - name: infinispan-operator.v2.4.4
        version: 2.4.4
      - name: infinispan-operator.v2.4.3
        version: 2.4.3
      - name: infinispan-operator.v2.4.2
        version: 2.4.2
      - name: infinispan-operator.v2.4.1
        version: 2.4.1
      - name: infinispan-operator.v2.4.0
        version: 2.4.0
      - name: infinispan-operator.v2.3.8
        version: 2.3.8
      - name: infinispan-operator.v2.3.7
        version: 2.3.7
      - name: infinispan-operator.v2.3.6
        version: 2.3.6
      - name: infinispan-operator.v2.3.5
        version: 2.3.5
      - name: infinispan-operator.v2.3.4
        version: 2.3.4
      - name: infinispan-operator.v2.3.3
        version: 2.3.3
      - name: infinispan-operator.v2.3.2
        version: 2.3.2
      - name: infinispan-operator.v2.3.1
        version: 2.3.1
      - name: infinispan-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: infinispan
    provider:
      name: Infinispan
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: costmanagement-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: costmanagement-metrics-operator.4.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "true"
          containerImage: registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
          createdAt: "2025-08-18T12:10:15Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: costmanagement-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Cost Management Metrics Operator
          ## Introduction
          The `costmanagement-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Cost Management Metrics Operator (`costmanagement-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created CostManagementMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The CostManagementMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **FIPS Compliance:** Supports deployment in high-security environments when run on an OpenShift cluster with FIPS mode enabled.

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created CostManagementMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the CostManagementMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `CostManagementMetricsConfig`, creating a new `CostManagementMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `costmanagement-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: costmanagement-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the CostManagementMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the CostManagementMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon CostManagementMetricsConfig creation** (This parameter will start to appear in CostManagementMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the CostManagementMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the costmanagement-metrics-operator
          **Note** There are separate instructions for configuring the `costmanagement-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `costmanagement-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the costmanagement-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `costmanagement-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `costmanagement-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the costmanagement-metrics-operator for a restricted network
          ##### Create the CostManagementMetricsConfig
          Configure the costmanagement-metrics-operator by creating a `CostManagementMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `costmanagement-metrics-operator` -> `CostManagementMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `costmanagement-metrics-operator-data` with 10Gi of storage. To configure the costmanagement-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `costmanagement-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: costmanagement-metrics-operator
                labels:
                  app: costmanagement-metrics-operator
              spec:
                volumes:
                - name: costmanagement-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: costmanagement-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: costmanagement-metrics-operator-reports
                    mountPath: /tmp/costmanagement-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/costmanagement-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/costmanagement-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `costmanagement-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the CostManagementMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Cost Management Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Cost Management Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/costmanagement/costmanagement-metrics-operator-bundle@sha256:84d65eed99300b0ba87369cbf9a9fc17a37781f8b6f9de2819d0a159f9071a6c
        - registry.redhat.io/costmanagement/costmanagement-metrics-rhel9-operator@sha256:c5bb5291052d57390b9f4bdc660c1926827072b6dcbe09d85f5b6fc375562182
        version: 4.1.0
      entries:
      - name: costmanagement-metrics-operator.4.1.0
        version: 4.1.0
      - name: costmanagement-metrics-operator.4.0.0
        version: 4.0.0
      - name: costmanagement-metrics-operator.3.3.2
        version: 3.3.2
      - name: costmanagement-metrics-operator.3.3.1
        version: 3.3.1
      - name: costmanagement-metrics-operator.3.3.0
        version: 3.3.0
      - name: costmanagement-metrics-operator.3.2.1
        version: 3.2.1
      - name: costmanagement-metrics-operator.3.2.0
        version: 3.2.0
      - name: costmanagement-metrics-operator.3.1.1
        version: 3.1.1
      - name: costmanagement-metrics-operator.3.1.0
        version: 3.1.0
      - name: costmanagement-metrics-operator.3.0.1
        version: 3.0.1
      - name: costmanagement-metrics-operator.3.0.0
        version: 3.0.0
      - name: costmanagement-metrics-operator.2.0.0
        version: 2.0.0
      - name: costmanagement-metrics-operator.1.2.0
        version: 1.2.0
      - name: costmanagement-metrics-operator.1.1.9
        version: 1.1.9
      - name: costmanagement-metrics-operator.1.1.8
        version: 1.1.8
      - name: costmanagement-metrics-operator.1.1.7
        version: 1.1.7
      - name: costmanagement-metrics-operator.1.1.6
        version: 1.1.6
      - name: costmanagement-metrics-operator.1.1.5
        version: 1.1.5
      - name: costmanagement-metrics-operator.1.1.4
        version: 1.1.4
      - name: costmanagement-metrics-operator.1.1.3
        version: 1.1.3
      - name: costmanagement-metrics-operator.1.1.0
        version: 1.1.0
      - name: costmanagement-metrics-operator.1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: costmanagement-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.10.7-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                    "name": "ex-aao",
                    "application": "ex-aao-app"
                },
                "spec": {
                    "deploymentPlan": {
                        "image": "placeholder",
                        "size": 2,
                        "requireLogin": false,
                        "persistenceEnabled": false,
                        "journalType": "nio",
                        "messageMigration": false,
                        "jolokiaAgentEnabled": false,
                        "managementRBACEnabled": true
                    }
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                    "name": "ex-aaoaddress"
                },
                "spec": {
                    "addressName": "myAddress0",
                    "queueName": "myQueue0",
                    "routingType": "anycast"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                    "name": "ex-aaoscaledown"
                },
                "spec": {
                    "localOnly": "true"
                }
            }, {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                    "name": "ex-keycloak"
                },
                "spec": {
                    "loginModules": {
                        "propertiesLoginModules": [
                            {
                                "name": "prop-module",
                                "users": [
                                    {
                                        "name": "superman",
                                        "roles": [
                                            "root"
                                        ]
                                    }
                                ]
                            }
                        ]
                    },
                    "securityDomains": {
                        "brokerDomain": {
                            "name": "activemq",
                            "loginModules": [
                                {
                                    "name": "prop-module",
                                    "flag": "sufficient"
                                }
                            ]
                        }
                    }
                }
            }]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
          createdAt: "2021-06-18T00:02:04Z"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A stateful deployment of one or more brokers
            displayName: AMQ Broker
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: Adding and removing addresses via custom resource definitions
            displayName: AMQ Broker Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: Provides message migration on clustered broker scaledown
            displayName: AMQ Broker Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: Security configuration for the broker
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
          to deploy and manage stateful AMQ Broker broker clusters
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: rkieley@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:1a7aa54d2799d238eb5f49f7a95a78a896f6bf8d222567e9118e0e3963cc9aad
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:e140415e0361cabeab192ba6d22634283cd47749dd25e9cee4a775d3c409cd97
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        version: 7.10.7-opr-1
      entries:
      - name: amq-broker-operator.v7.10.7-opr-1
        version: 7.10.7-opr-1
      - name: amq-broker-operator.v7.10.6-opr-1
        version: 7.10.6-opr-1
      - name: amq-broker-operator.v7.10.5-opr-1
        version: 7.10.5-opr-1
      - name: amq-broker-operator.v7.10.4-opr-1
        version: 7.10.4-opr-1
      - name: amq-broker-operator.v7.10.3-opr-3
        version: 7.10.3-opr-3
      - name: amq-broker-operator.v7.10.3-opr-2
        version: 7.10.3-opr-2
      - name: amq-broker-operator.v7.10.3-opr-1
        version: 7.10.3-opr-1
      - name: amq-broker-operator.v7.10.2-opr-2-0.1676475747.p
        version: 7.10.2-opr-2+0.1676475747.p
      - name: amq-broker-operator.v7.10.2-opr-2-0.1680622941.p
        version: 7.10.2-opr-2+0.1680622941.p
      - name: amq-broker-operator.v7.10.2-opr-2
        version: 7.10.2-opr-2
      - name: amq-broker-operator.v7.10.2-opr-1
        version: 7.10.2-opr-1
      - name: amq-broker-operator.v7.10.1-opr-2
        version: 7.10.1-opr-2
      - name: amq-broker-operator.v7.10.1-opr-1
        version: 7.10.1-opr-1
      - name: amq-broker-operator.v7.10.0-opr-4
        version: 7.10.0-opr-4
      - name: amq-broker-operator.v7.10.0-opr-3
        version: 7.10.0-opr-3
      - name: amq-broker-operator.v7.10.0-opr-2
        version: 7.10.0-opr-2
      - name: amq-broker-operator.v7.10.0-opr-1
        version: 7.10.0-opr-1
      name: 7.10.x
    - currentCSV: amq-broker-operator.v7.11.8-opr-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "application": "ex-aao-app",
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisSecurity is the Schema for the activemqartemissecurities
              API
            displayName: Active MQArtemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:4e09ee3bd8ff3479fdd00e221b7b1505ae04ca6c64fd6bc9e2f1a34b8e98d137
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:97c3aa5673610875a53571449f7d5210ae8d675072d044f17b416725ed0dba88
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:5ff28e3062d3eabb9f52d8271b837fdab06cafbd8b8f732b1bc00de49ad69aad
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:35b7d96d5065f280a528626abd2c3df84e79ccea93ab6bd8b8837af66c330f30
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3219f90789690a3c5cfcc86e5f602c443e75f1fdb3a19660f43ee3996bfa8f62
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:4b7796235a0a545db83570add7a675fdd9d82b17f931eb81d0cafbedb1fbf22c
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0b2126cfb6054fdf428c1f43b69e36e93a09a49ce15350e9273c98cc08c6598b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b74d03ed852a3731467ffda95266ce49f2065972f1c37bf254f3d52b34c11991
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:61e5a8c1e46f1c18648e0406e2a57c52f911979079020cc57c7a4b06a56940bb
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:aa7af7e038a12b5202108b362c8af243954603e84034a16c0192684c86d5d28b
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:7abfdcfd7b2d3d12b9234f7b057bd199bceefa47bdf02f62236f82a13fdbf5b7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a8b2a9364fd06c8e1c29b8f8f28d3e4844bf7840a62b67137b663c558982d4a3
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:6e9e0e3fddc44fbf87591d8904176407f1ec409d578a7d108c13cdd48d85131b
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:e6f71001dccfbbe1cd7cf4fded69e181c10f46cc5c8bd3a8d3045a3f3ac41e2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ff6fefad518a6c997d4c5a6e475ba89640260167f0bc27715daf3cc30116fad1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c5f4c08e068b9721967cf7c7cbd9a9e93fb5e39b264dd13b653e99d8f3fa9e0e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:34ab1f66d8daa544e3823464383e9788e307665c875499838fbed95f3a74a4dd
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:56282c6a461e945a83bffb13d3f3d8d9621800f0fc9cd755a75d21eacd8f5fbf
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:86ff145bd68e797a0507495570063b54cbe4186fcb4acd83c900406619ef9524
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:869cda365447b0b7b53ed21d2496884c15dc9c07295a9dfa0337c2c9974fe018
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:4440b20349032bb5c0bb2a8deb89832b37767af6f0b3208b067867a85641913e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ba21779c29b431a22c824e91820705c9501fb0b7d6f7e1ae22fcd3becc8c8a2f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0d857b95084e39c011651be89482d4dcc77c2d1c3ac3c6a819f3c9d5a57a425f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:a94fa8135d3765363e442bda45df6e02a3e22bee7ea198bcb0d3584331a4320c
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:d2ae3defac36f00112c6b773e13626336ba15d6f50f905ad1c33f122bfe1a6a9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:c9eba0d1f9fba588711774f4db75221f057578407065127ed60c33eef564f499
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:83faefa67b33d93836a1be3cb438bb147242d602565dcbe59b4ed42b26072191
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:e03945756d86bd189e8c33027fa668d2065c168510085dfd06defd9f40d3e0ab
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:5d39f26f06769fe5c989ea0e0ee8f9166120e152ce2c9bc3336fc8d7cab2ff92
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b575d6eebf44dc9049445e357488efb17a690953852419b534df30e0105b9ca4
        - registry.redhat.io/amq7/amq-broker-init-openjdk-11-rhel8@sha256:d4a9176f49ffd9db6051fbda15d47a1aa32c282fb5010d0caf5b9243e80bf055
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:45a5e054546a2c1ee338518f8477f728f5f7108c9dd8b516b4c79d4a7d6f5ba7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f58678f03a250dbb296ebeb112c9150340155799ba4dad9e90bf347b9ed0ca7b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ff819e913406911735da39c51699405f747bd697116a273a00897ff176bf05ff
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:3c00b524722b500ff0922fe7a8722868557512183a8f69248f37a68c45c6a770
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:979b59325aa0f34eb05625201beba53fccbb83bd5eb80a89dcb5261ae358138f
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:34739366153b825dcb113e0ed6115ef5656085cf8d47492789ff17464042b576
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:c3ac4602b48867f3d4f7cc0efd5c9f133098e409626cfb402f80386ff460dbf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9a02decbc460ddfc501d8a4a9b41cd084776090951906b027838c0829e1fc15a
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:9465e398a55c89c121b3ae50374414190a991839fa5e1018fc02551f6aa3b9aa
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:f37f98c809c6f29a83e3d5a3ac4494e28efe9b25d33c54f533c6a08662244622
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:f2651cb097eb631fc6af90e4a8404d36516aa2205c546125bf0c37739ab525fa
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:2cb14c917246ffcbd5d861ca6f3d58cb1146a4e55715ee08160e115d3dfc1c3e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b0b3684ef1ebf8b44d3a4c46391984173792dde1bd7363ea06d43a8f95a717d0
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:ba0f5df7da08bc030ff5e0e68ecb52a931c25f6b67c6556aa8a9661c82d69d42
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:67ae74e72927557c55611d0248347b2099b117ac1aa245dac05856fe782fb59c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:137b893ad15e0cd5353bae134c8fae833257a12641b12d0f774f1565ce66f63c
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-openjdk-11-rhel8@sha256:967d840b831b310abd3a468a8652d1fa48b2ce28b342bfab53999245deb004f1
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-openj9-11-rhel8@sha256:d4010af53429502f41d722c39d69a7bb147bb06d19e3d4a69de70dbe88e3990e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:83ed89fa942ccbf4451954956db42138c83f9b015e7de07fc159ddc4bb6dc843
        - registry.redhat.io/amq7/amq-broker-init-openj9-11-rhel8@sha256:1a9f13fd565113055b2aa2022bd0709a63c1dc058a8570e0fd7eba5c7b101334
        version: 7.11.8-opr-1
      entries:
      - name: amq-broker-operator.v7.11.8-opr-1
        version: 7.11.8-opr-1
      - name: amq-broker-operator.v7.11.7-opr-2-1742200979
        version: 7.11.7-opr-2-1742200979
      - name: amq-broker-operator.v7.11.7-opr-2-1740583644
        version: 7.11.7-opr-2-1740583644
      - name: amq-broker-operator.v7.11.7-opr-2-1736956309
        version: 7.11.7-opr-2-1736956309
      - name: amq-broker-operator.v7.11.7-opr-2-1731330819
        version: 7.11.7-opr-2-1731330819
      - name: amq-broker-operator.v7.11.7-opr-2
        version: 7.11.7-opr-2
      - name: amq-broker-operator.v7.11.7-opr-1-1726157430
        version: 7.11.7-opr-1-1726157430
      - name: amq-broker-operator.v7.11.7-opr-1-1723204800
        version: 7.11.7-opr-1-1723204800
      - name: amq-broker-operator.v7.11.7-opr-1
        version: 7.11.7-opr-1
      - name: amq-broker-operator.v7.11.6-opr-2
        version: 7.11.6-opr-2
      - name: amq-broker-operator.v7.11.5-opr-1
        version: 7.11.5-opr-1
      - name: amq-broker-operator.v7.11.4-opr-1
        version: 7.11.4-opr-1
      - name: amq-broker-operator.v7.11.3-opr-1
        version: 7.11.3-opr-1
      - name: amq-broker-operator.v7.11.2-opr-1
        version: 7.11.2-opr-1
      - name: amq-broker-operator.v7.11.1-opr-2
        version: 7.11.1-opr-2
      - name: amq-broker-operator.v7.11.1-opr-1
        version: 7.11.1-opr-1
      - name: amq-broker-operator.v7.11.0-opr-3
        version: 7.11.0-opr-3
      - name: amq-broker-operator.v7.11.0-opr-2
        version: 7.11.0-opr-2
      name: 7.11.x
    - currentCSV: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 8 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 8 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:89bb769b4c33eb931cc7866eeabec9a315b913851cfa84bc6e11ae8d1e8b2cbb
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9c1ac38bf8eec67a3da57c3cc61105513776a749408cb287fe6448a17889de4e
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5e5d8b551ab8ca377e0bc939037970a2f31ee14e9f555853ebea2759d673895a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:29f07006ad62023c20d34971cb3ae900349efae5ab2918aaa8e5647d788032be
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:0dde839c2dce7cb684094bf26523c8e16677de03149a0fff468b8c3f106e1f4f
        - registry.redhat.io/amq7/amq-broker-rhel8-operator-bundle@sha256:9f9dadba5a35c8ab98ddacee22593a247a31e96a4b6e1b4f1a946c2e34006bb9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d71506472636d940b5dbd3e708d3e0448d777eeecfd38de61ba49e281d2979c6
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:442b6638133d42fa7faa41d86e47be95cdca56c2e16efcd6e7e22991c3812a36
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:eacc4fb55e8c4c30894a51944960d536508001a49fda62782c6cb2ca8767961d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:7d53744ae8c77aea09c59165486c9ed0eaa131140f6b346042293bce0bd8a745
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:cc438b4bc2b14f6bff4e5d32e7dc0fad6d2ca391c4e84b5ae925b7b40af76729
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:25c3732e945b4bc3e9aca721ce709d0217339b065d9dac9afdb6202e9fd8a799
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:90f1678c19c74454b77aed14fa6b3074a1c6f80fe884f1c6136e7e907f402aec
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ade37271426ab8f543d72f8da3e9f663d665f9427379a2539fc9323f6897cc44
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:0f443f5aa818dd85bfa93e51e37bdac3909628a283aa52a89125b527dd75d9aa
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b9f93c243ead118651083880786a81c624f390fcc3596c7ec5ebd55f34f95ea3
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:8b4c4400a02c6cacfd978147b158e2ccbe5da955578b4b970a5ee7edac865326
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:8329acdff170c5a8ae6ae16f6c37d3f7ab68d106a6e8be50c3fe131ec5ecceeb
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:5f842ebe9a6a4294a056ad10189a5f277cfd3fb4e6e0b3225e194bd5c131151d
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:b9aecae6b0ce0b241c7a5b9e6e096395940bb6a384373a4d121822cf10c26d99
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-rhel8-operator@sha256:023478b773a0c42739b86e822a4d435e01ef341b8136566ea52463f1b7e447ed
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:40ccaf2bb2e0f8132f2541b39cb835f17b38050b3ca9232f0861fd5fcaf4324a
        version: 7.12.5-opr-1+0.1757928550.p
      entries:
      - name: amq-broker-operator.v7.12.5-opr-1-0.1757928550.p
        version: 7.12.5-opr-1+0.1757928550.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1756216162.p
        version: 7.12.5-opr-1+0.1756216162.p
      - name: amq-broker-operator.v7.12.5-opr-1-0.1754272113.p
        version: 7.12.5-opr-1+0.1754272113.p
      - name: amq-broker-operator.v7.12.5-opr-1
        version: 7.12.5-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1
        version: 7.12.4-opr-1
      - name: amq-broker-operator.v7.12.4-opr-1-0.1747217191.p
        version: 7.12.4-opr-1+0.1747217191.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741881887.p
        version: 7.12.3-opr-1+0.1741881887.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739991235.p
        version: 7.12.3-opr-1+0.1739991235.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730142550.p
        version: 7.12.3-opr-1+0.1730142550.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1730813578.p
        version: 7.12.3-opr-1+0.1730813578.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1731483798.p
        version: 7.12.3-opr-1+0.1731483798.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1732654016.p
        version: 7.12.3-opr-1+0.1732654016.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1733315503.p
        version: 7.12.3-opr-1+0.1733315503.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1737643631.p
        version: 7.12.3-opr-1+0.1737643631.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1738596517.p
        version: 7.12.3-opr-1+0.1738596517.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739390944.p
        version: 7.12.3-opr-1+0.1739390944.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1739889884.p
        version: 7.12.3-opr-1+0.1739889884.p
      - name: amq-broker-operator.v7.12.3-opr-1
        version: 7.12.3-opr-1
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741884937.p
        version: 7.12.3-opr-1+0.1741884937.p
      - name: amq-broker-operator.v7.12.3-opr-1-0.1741883943.p
        version: 7.12.3-opr-1+0.1741883943.p
      - name: amq-broker-operator.v7.12.2-opr-1
        version: 7.12.2-opr-1
      - name: amq-broker-operator.v7.12.1-opr-1-0.1724054375.p
        version: 7.12.1-opr-1+0.1724054375.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1723669176.p
        version: 7.12.1-opr-1+0.1723669176.p
      - name: amq-broker-operator.v7.12.1-opr-1-0.1721247907.p
        version: 7.12.1-opr-1+0.1721247907.p
      - name: amq-broker-operator.v7.12.1-opr-1
        version: 7.12.1-opr-1
      - name: amq-broker-operator.v7.12.0-opr-2
        version: 7.12.0-opr-2
      - name: amq-broker-operator.v7.12.0-opr-1
        version: 7.12.0-opr-1
      name: 7.12.x
    defaultChannel: 7.12.x
    packageName: amq-broker-rhel8
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzograph-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v2.0.102
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
          createdAt: "2021-08-11"
          description: kubernetes operator for AnzoGraph DB
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzograph-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: AnzoGraph Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The AnzoGraph Operator provides the way to install and configure an AnzoGraph
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the AnzoGraph db Enterprise Edition License [here]( https://customercenter.cambridgesemantics.com/ ). For AnzoGraph db Free Edition questions, get help from our Anzograph User Community at Stack Overflow. When submitting a question, include the tag 'anzograph'.
        displayName: AnzoGraph Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v2.0.1/README_openshift_marketplace.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:01ca6a65a041a47efd611673a720e69464f02424d21feadc496c356065ff2f65
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-db@sha256:8a5bd26551990929fdb776d6b421a9713cb84e03b19c18f81b625582f7f41af5
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-frontend@sha256:c1443bcdc8e998fc699f61980a665f1a083350adf2d3d1d55a7cb55e31ca50b3
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph-operator@sha256:a5b212558f748ddc8156541b546290320b656506e039a18903ad9fa6e638e732
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzograph@sha256:e3bdddf811b5ba69f388d023aea1ba538edd00e9f415c33ec555321d39129a36
        version: 2.0.1
      entries:
      - name: anzograph-operator.v2.0.102
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Appian Corporation
      provider-url: https://appian.com/
    name: appian-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appian-operator.v0.184.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Application Runtime, Big Data, Integration
            & Delivery, Modernization & Migration
          createdAt: "2025-08-07T15:50:26Z"
          description: The Appian operator facilitates managing instances of Appian
            on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appian-operator
          operators.openshift.io/valid-subscription: '["Appian License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Appian Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appian
            name: appians.crd.k8s.appian.com
            version: v1beta1
        description: The Appian operator facilitates managing instances of Appian
          on Kubernetes.
        displayName: Appian operator
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - Appian
        links:
        - name: Appian
          url: https://appian.com/
        - name: Documentation
          url: https://docs.appian.com/suite/help/latest/k8s
        maintainers:
        - email: support@appian.com
          name: Appian Corporation
        maturity: planning
        provider:
          name: Appian Corporation
          url: https://appian.com/
        relatedImages:
        - 937620572175.dkr.ecr.us-east-1.amazonaws.com/appian/appian-operator@sha256:26c2ba27f9360226f34f289109cb0d10e9bf44777146ccfd5f367adf2367a400
        - registry.connect.redhat.com/appian/appian-operator-bundle@sha256:2cfb2d2e43fa207c8c06df6de8d3e90123c4cd52cdb6a4c6d618ec576b79c2eb
        version: 0.184.0
      entries:
      - name: appian-operator.v0.184.0
        version: 0.184.0
      - name: appian-operator.v0.171.0
        version: 0.171.0
      name: alpha
    defaultChannel: alpha
    packageName: appian-operator
    provider:
      name: Appian Corporation
      url: https://appian.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:c4e9e7b43800af5602dd02ada94254f707ac71b76dd277b8ac9bca4d46721b76
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:4ec16d88828e580fd9223125043ca49569324c4f06b7a65bbaa7d76923de2bad
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:7f959a4076c36d76f8f1489a92d6984d1cfd2ae5c0533a78e8b16010eb6d4d3c
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:af9215670675fe9c4fa2dac7a14d547c249f3aa0c0e72f1e5390d237e411735b
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:a3132caf9cddf6d6186b3e982a6f6232a208fc727b433ee6d200a3b72f50ca88
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:6e7e13d259f4d743524ebcd546af5c73d1a3ba90edaf4b5edd2005a3f8adcdfe
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:d89f4d647534569f401790d6e10bc42433ae791be86558b100a4027a74b92cd3
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:42728e7fd1d2302511ab08f51335ab188eef125870fa27f322df83ebb8e1f447
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "ppc64le",
                                    "s390x"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver on OpenShift.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator-bundle@sha256:05c49ac71b969e4523f21869b612cafd11be7a93b27e100427a918b136e14139
        - registry.connect.redhat.com/ibm/ibm-block-csi-operator@sha256:bd025fda98f8ef8e721c9a63f5239123dd243efc9958ffffdba0bc209d829414
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ControlPlane Ltd.
      provider-url: https://control-plane.io
    name: flux-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxInstance",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcileEvery": "1h",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "distribution": {
                    "version": "2.x",
                    "registry": "ghcr.io/fluxcd"
                  },
                  "components": [
                    "source-controller",
                    "kustomize-controller",
                    "helm-controller",
                    "notification-controller",
                    "image-reflector-controller",
                    "image-automation-controller"
                  ],
                  "cluster": {
                    "type": "openshift",
                    "size": "medium",
                    "domain": "cluster.local",
                    "networkPolicy": true,
                    "multitenant": true,
                    "tenantDefaultServiceAccount": "flux"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "FluxReport",
                "metadata": {
                  "name": "flux",
                  "namespace": "flux-system"
                },
                "spec": {
                  "distribution": {
                    "entitlement": "Issued by controlplane",
                    "managedBy": "flux-operator",
                    "status": "Installed",
                    "version": "v2.3.0"
                  },
                "components": [
                    {
                      "image": "ghcr.io/fluxcd/kustomize-controller:v1.3.0@sha256:48a032574dd45c39750ba0f1488e6f1ae36756a38f40976a6b7a588d83acefc1",
                      "name": "kustomize-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    },
                    {
                      "image": "ghcr.io/fluxcd/source-controller:v1.3.0@sha256:161da425b16b64dda4b3cec2ba0f8d7442973aba29bb446db3b340626181a0bc",
                      "name": "source-controller",
                      "ready": true,
                      "status": "Current Deployment is available. Replicas: 1"
                    }
                  ],
                "reconcilers": [
                    {
                      "apiVersion": "kustomize.toolkit.fluxcd.io/v1",
                      "kind": "Kustomization",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0
                      }
                    },
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1",
                      "kind": "GitRepository",
                      "stats": {
                        "failing": 0,
                        "running": 1,
                        "suspended": 0,
                        "totalSize": "3.7 MiB"
                      }
                    }
                  ],
                  "sync": {
                    "ready": true,
                    "id": "kustomization/flux-system",
                    "path": "clusters/production",
                    "source": "https://github.com/my-org/my-fleet.git",
                    "status": "Applied revision: refs/heads/main@sha1:a90cd1ac35de01c175f7199315d3f4cd60195911"
                  }
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSet",
                "metadata": {
                  "name": "podinfo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "30m",
                    "fluxcd.controlplane.io/reconcileTimeout": "5m"
                  }
                },
                "spec": {
                  "commonMetadata": {
                    "labels": {
                      "app.kubernetes.io/name": "podinfo"
                    }
                  },
                  "inputs": [
                    {
                      "tenant": "team1",
                      "version": "6.7.x",
                      "replicas": "2"
                    },
                    {
                      "tenant": "team2",
                      "version": "6.6.x",
                      "replicas": "3"
                    }
                  ],
                  "resources": [
                    {
                      "apiVersion": "source.toolkit.fluxcd.io/v1beta2",
                      "kind": "OCIRepository",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "10m",
                        "url": "oci://ghcr.io/stefanprodan/charts/podinfo",
                        "ref": {
                          "semver": "<< inputs.version | quote >>"
                        }
                      }
                    },
                    {
                      "apiVersion": "helm.toolkit.fluxcd.io/v2",
                      "kind": "HelmRelease",
                      "metadata": {
                        "name": "podinfo-<< inputs.tenant >>",
                        "namespace": "default"
                      },
                      "spec": {
                        "interval": "1h",
                        "releaseName": "podinfo-<< inputs.tenant >>",
                        "chartRef": {
                          "kind": "OCIRepository",
                          "name": "podinfo-<< inputs.tenant >>"
                        },
                        "values": {
                          "replicaCount": "<< inputs.replicas | int >>"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "fluxcd.controlplane.io/v1",
                "kind": "ResourceSetInputProvider",
                "metadata": {
                  "name": "demo",
                  "namespace": "default",
                  "annotations": {
                    "fluxcd.controlplane.io/reconcile": "enabled",
                    "fluxcd.controlplane.io/reconcileEvery": "10m",
                    "fluxcd.controlplane.io/reconcileTimeout": "1m"
                  }
                },
                "spec": {
                  "type": "GitHubPullRequest",
                  "url": "https://github.com/fluxcd-testing/pr-testing",
                  "filter": {
                    "limit": 10,
                    "includeBranch": "^stefanprodan-patch-.*$",
                    "labels": [
                      "enhancement"
                    ]
                  },
                  "skip": {
                    "labels": [
                        "deploy/flux-preview-pause",
                        "!test-build-push/passed"
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
          createdAt: "2025-09-30T14:21:42Z"
          description: The Flux Operator manages the lifecycle of the CNCF Flux project.
          operatorframework.io/suggested-namespace: flux-system
          repository: https://github.com/controlplaneio-fluxcd/flux-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Flux Instance
            displayName: FluxInstance
            kind: FluxInstance
            name: fluxinstances.fluxcd.controlplane.io
            version: v1
          - description: Flux Report (Autogenerated)
            displayName: FluxReport
            kind: FluxReport
            name: fluxreports.fluxcd.controlplane.io
            version: v1
          - description: Resource Set
            displayName: ResourceSet
            kind: ResourceSet
            name: resourcesets.fluxcd.controlplane.io
            version: v1
          - description: ResourceSet Input Provider
            displayName: ResourceSetInputProvider
            kind: ResourceSetInputProvider
            name: resourcesetinputproviders.fluxcd.controlplane.io
            version: v1
        description: |-
          [Flux](https://fluxcd.io) is a leading GitOps Continuous Delivery tool used to streamline and automate application deployments on Kubernetes. Flux enables teams to achieve a reliable and auditable delivery process while promoting collaboration and traceability across different environments.
          Flux is powered by the GitOps Toolkit, a set of composable APIs and specialized tools that enable a wide range of continuous delivery use-cases, from simple Kubernetes deployment pipelines to multi-tenant and multi-cluster progressive delivery rollouts.
          For more information about the CNCF-graduated Flux project, please see the [Flux architecture overview](https://fluxcd.control-plane.io/guides/flux-architecture/).
          ### Operator Capabilities
          The [Flux Operator](https://github.com/controlplaneio-fluxcd/flux-operator) provides a declarative API for the installation, configuration and upgrade of CNCF Flux and the ControlPlane [enterprise distribution](https://fluxcd.control-plane.io/).
          The operator allows the configuration of Flux multi-tenancy lockdown, network policies, persistent storage, sharding, vertical scaling, custom patches, and the synchronization of the cluster state from Git repositories, OCI artifacts and S3-compatible storage.
          The Flux Operator supervises the Flux controllers and provides a unified view of all the Flux resources that define the GitOps workflows for the target cluster. The operator generates reports, emits events, and exports Prometheus metrics to help with monitoring and troubleshooting Flux.
          ### Operator Features
          The operator extends Flux with self-service capabilities, deployment windows, and preview environments for GitHub, GitLab and Azure DevOps pull requests testing.
          The operator ResourceSet API enables platform teams to define their own application standard as a group of Flux and Kubernetes resources that can be templated, parameterized and deployed as a single unit on self-service environments. For more information, please see the [ResourceSet documentation](https://fluxcd.control-plane.io/operator/resourcesets/introduction/).
          ### OpenShift Support
          The Flux Operator should be installed in a dedicated namespace, e.g. `flux-system`.
          To deploy Flux on OpenShift clusters, create a `FluxInstance` custom resource named `flux` in the same namespace as the operator and set the `.spec.cluster.type` field to `openshift`.
          For more information on how to configure the Flux instance, please see the [Flux Operator documentation](https://fluxcd.control-plane.io/operator/flux-config/).
        displayName: Flux Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - fluxcd
        - gitops
        - continuous delivery
        links:
        - name: Source Code
          url: https://github.com/controlplaneio-fluxcd/flux-operator
        - name: Documentation
          url: https://fluxcd.control-plane.io/operator/
        - name: Enterprise Support
          url: https://fluxcd.control-plane.io/pricing/
        maintainers:
        - email: flux-enterprise@control-plane.io
          name: ControlPlane Flux Team
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: ControlPlane Ltd.
          url: https://control-plane.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/flux-operator@sha256:557ae37fa6a6942e3a9d06966df67721011ac7905e2241867c975948092f2936
        - ghcr.io/controlplaneio-fluxcd/flux-operator:v0.30.0-ubi
        version: 0.30.0
      entries:
      - name: flux-operator.v0.30.0
        version: 0.30.0
      - name: flux-operator.v0.28.0
        version: 0.28.0
      - name: flux-operator.v0.27.0
        version: 0.27.0
      - name: flux-operator.v0.26.0
        version: 0.26.0
      - name: flux-operator.v0.24.1
        version: 0.24.1
      - name: flux-operator.v0.24.0
        version: 0.24.0
      - name: flux-operator.v0.23.0
        version: 0.23.0
      - name: flux-operator.v0.22.0
        version: 0.22.0
      - name: flux-operator.v0.21.0
        version: 0.21.0
      - name: flux-operator.v0.20.0
        version: 0.20.0
      - name: flux-operator.v0.19.0
        version: 0.19.0
      - name: flux-operator.v0.18.0
        version: 0.18.0
      name: stable
    defaultChannel: stable
    packageName: flux-operator
    provider:
      name: ControlPlane Ltd.
      url: https://control-plane.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smb-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smb-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure CIFS/SMB CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.15.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/csi-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          CIFS/SMB CSI Driver Operator provides Server Message Block (SMB) CSI Driver that enables you to create and mount CIFS/SMB PersistentVolumes.
        displayName: CIFS/SMB CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - smb
        - cifs
        links:
        - name: Documentation
          url: https://github.com/openshift/csi-operator
        - name: Source Repository
          url: https://github.com/openshift/csi-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9-operator@sha256:1167114457036d6385170c11288d944a5497a19f8b8c01ec8ca6245474dce685
        - registry.redhat.io/openshift4/ose-smb-csi-driver-rhel9@sha256:ac244e06ecd2f8a5e4de7b121efb8e82e3e7d23194878cb8327ff0b2affb5f04
        - registry.redhat.io/openshift4/ose-smb-csi-driver-operator-bundle@sha256:91d04cbe0eee6c4fe0230d0d179366942e6c54afb65bb77a43a681e1a2411caf
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        version: 4.18.0-202509100344
      entries:
      - name: smb-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: smb-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: smb-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: smb-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: smb-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: smb-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: smb-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: smb-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: smb-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: smb-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: smb-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: smb-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: smb-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: smb-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: smb-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: smb-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: pg4k-pgd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pg4k-pgd.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroup",
                "metadata": {
                  "name": "region-a"
                },
                "spec": {
                  "backup": null,
                  "cnp": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "connectivity": {
                    "dns": {
                      "additional": [
                        {
                          "domain": "alternate.domain"
                        },
                        {
                          "domain": "my.domain",
                          "hostSuffix": "-dc1"
                        }
                      ]
                    },
                    "tls": {
                      "clientCert": {
                        "caCertSecret": "client-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "client-ca-issuer"
                            }
                          }
                        }
                      },
                      "mode": "verify-ca",
                      "serverCert": {
                        "caCertSecret": "server-ca-key-pair",
                        "certManager": {
                          "spec": {
                            "issuerRef": {
                              "group": "cert-manager.io",
                              "kind": "Issuer",
                              "name": "server-ca-issuer"
                            }
                          }
                        }
                      }
                    }
                  },
                  "instances": 2,
                  "pgd": {
                    "parentGroup": {
                      "create": true,
                      "name": "world"
                    }
                  },
                  "proxyInstances": 2,
                  "witness": {
                    "storage": {
                      "size": "1Gi"
                    },
                    "walStorage": {
                      "size": "1Gi"
                    }
                  },
                  "witnessInstances": 1
                }
              },
              {
                "apiVersion": "pgd.k8s.enterprisedb.io/v1beta1",
                "kind": "PGDGroupCleanup",
                "metadata": {
                  "name": "region-b-cleanup"
                },
                "spec": {
                  "executor": "region-a",
                  "target": "region-b"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
          createdAt: "2025-09-26T01:49:29Z"
          description: EDB Postgres Distributed for Kubernetes is an operator designed
            to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both
            in a single Kubernetes cluster or across multiple ones.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImageCatalog
            name: clusterimagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - kind: ImageCatalog
            name: imagecatalogs.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: Object to clean up a PGD Group
            displayName: PGD Group Cleanup
            kind: PGDGroupCleanup
            name: pgdgroupcleanups.pgd.k8s.enterprisedb.io
            version: v1beta1
          - description: A group of Postgres Clusters
            displayName: PGD Group
            kind: PGDGroup
            name: pgdgroups.pgd.k8s.enterprisedb.io
            version: v1beta1
        description: |
          EDB Postgres Distributed for Kubernetes is an operator designed to manage EDB Postgres Distributed (PGD) v5 workloads on Kubernetes both in a single Kubernetes cluster or across multiple ones.

          EDB Postgres Distributed is an extension for PostgreSQL that provides multi-master replication and data distribution with advanced conflict management, data-loss protection, and throughput up to 5X faster than native logical replication. It enables distributed PostgreSQL clusters with high availability up to five 9s.

          This operator requires access to the EDB private repositories.
        displayName: EDB Postgres Distributed for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - distributed
        - postgres
        - database
        - sql
        - cloud
        - native
        - edb
        - enterprisedb
        - cnp
        - pg4k
        - pgd
        links:
        - name: 'Documentation: EDB Postgres Distributed v5'
          url: https://www.enterprisedb.com/docs/pgd/latest/
        - name: 'Documentation: EDB Postgres Distributed for Kubernetes'
          url: https://www.enterprisedb.com/docs/postgres_distributed_for_kubernetes/latest/
        maintainers:
        - email: sales@enterprisedb.com
          name: Sales
        - email: support@enterprisedb.com
          name: Support
        maturity: alpha
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - docker.enterprisedb.com/k8s_enterprise_pgd/pg4k-pgd-openshift@sha256:5127d913011619d61fd06b27be3c709df2236a4b5c0254fa58632698d58b50fb
        - registry.connect.redhat.com/enterprisedb/pg4k-pgd@sha256:fb206b95690e9604c41715369c69125a906b866dc35cfc5c63638ce7a4337fa2
        version: 1.1.3
      entries:
      - name: pg4k-pgd.v1.1.3
        version: 1.1.3
      - name: pg4k-pgd.v1.1.2
        version: 1.1.2
      name: stable
    defaultChannel: stable
    packageName: pg4k-pgd
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nutanix
      provider-url: ""
    name: nutanixcsioperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nutanixcsioperator.v3.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crd.nutanix.com/v1alpha1",
                "kind": "NutanixCsiStorage",
                "metadata": {
                  "name": "nutanixcsistorage"
                },
                "spec": {
                  "ntnxInitConfigMap": {
                    "usePC": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
          createdAt: "2025-06-18T11:04:03Z"
          description: An Operator for deploying and managing the Nutanix CSI Driver
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <3.3.4
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi","disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nutanix/csi-operator
          support: Nutanix Support Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a new instance of Nutanix CSI Driver
            displayName: Nutanix CSI Storage
            kind: NutanixCsiStorage
            name: nutanixcsistorages.crd.nutanix.com
            version: v1alpha1
        description: "# Important\nPlease read the following instructions carefully
          before upgrading from 2.6.x to 3.3.4, for more information please refer
          to [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          \n- Please do not upgrade to the CSI 3.x operator if:\n    * You are using
          LVM volumes.\n    \n- To upgrade from the CSI 2.6.x to CSI 3.3 (interacting
          with Prism Central) operator\n    * Create a Nutanix Prism Central secret
          as explained below.\n    * Delete the csidriver object from the cluster:\n\n
          \       ```\n        oc delete csidriver csi.nutanix.com\n        ```\n\n
          \   * In the installed operators, go to Nutanix CSI Operator and change
          the subscription channel from stable to stable-3.x. \n    If you have installed
          the operator with automatic update approval, the operator will be automatically
          upgraded to CSI 3.3, and then the nutanixcsistorage resource will be upgraded.\n
          \   An update plan will be generated for manual updates. Upon approval,
          the operator will be successfully upgraded.\n    \n- Direct upgrades from
          CSI 2.6.x to CSI 3.3 interacting with Prism Element are not supported. \n
          \   The only solution is to recreate the nutanixcsistorage instance by following
          the below procedure:\n    - In the installed operators, go to Nutanix CSI
          Operator and delete the nutanixcsistorage instance.\n    - Next change the
          subscription channel from stable to stable-3.x.\n    - Verify the following
          points:\n        - Ensure a Nutanix Prism Element secret is present in the
          namespace.\n        - Ensure that all the storage classes with provisioner:
          csi.nutanix.com have a controller publish secret as explained below.\n\n
          \           ```\n            csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n            csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n            ```\n\n            If this secret
          is not present in the storage class please delete and recreate the storage
          classes with the required secrets.\n    - Create a new instance of nutanixcsistorage
          from this operator by specifying `usePC: false` in YAML spec section.\n
          \   - Caution: Moving from CSI driver interacting with Prism Central to
          CSI driver interacting with Prism Element is not supported.\n    \n- Troubleshooting:\n
          \   \n    If the upgrade was unsuccessful and you want to revert to version
          CSI 2.6.x, please delete the csidriver object as explained above, uninstall
          the operator (no need to delete the nutanixcsistorage custom resource),
          and install version CSI 2.6.x from the stable channel.\n# Overview \nThe
          Nutanix CSI Operator for Kubernetes packages, deploys, manages, and upgrades
          the Nutanix CSI Driver on Kubernetes and OpenShift for dynamic provisioning
          of persistent volumes on the Nutanix Enterprise Cloud platform.\nThe Nutanix
          Container Storage Interface (CSI) Driver for Kubernetes leverages Nutanix
          Volumes and Nutanix Files to provide scalable and persistent storage for
          stateful applications.\nWith the Nutanix CSI Provider, you can:\n*  **Provide
          persistent storage to your containers**\n    * Leverage PVC resources to
          consume dynamically Nutanix storage\n    * With Files storage classes, applications
          on multiple pods can access the same storage and have the benefit of multi-pod
          read and write access. \n## Warning\nIf you upgrade from a previous version,
          make sure to validate your existing StorageClass parameters against the
          corresponding [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)\n##
          Supported Access mode\n| Storage Class mode | ReadWriteOnce | ReadOnlyMany
          | ReadWriteMany |\n|--------------------|---------------|--------------|---------------|\n|
          Nutanix Volumes    | yes           | yes          | no            |\n| Nutanix
          Files      | yes           | yes          | yes           |\n## Supported
          orchestration platforms\nThe following table details orchestration platforms
          suitable for deployment of the Nutanix CSI driver.\n| Orchestration platform
          | Version | Architecture |\n|------------------------|---------|--------------|\n|
          Red Hat OpenShift      | 4.12    | x86          |\n| Red Hat OpenShift      |
          4.13    | x86          |\n| Red Hat OpenShift      | 4.14    | x86          |\n|
          Red Hat OpenShift      | 4.15    | x86          |\n| Red Hat OpenShift      |
          4.16    | x86          |\n| Red Hat OpenShift      | 4.17    | x86          |\n|
          Red Hat OpenShift      | 4.18    | x86          |\n| Red Hat OpenShift      |
          4.19    | x86          |\n## Configuring k8s secret and storage class\nTo
          use this driver, create the relevant storage classes and secrets and deploy
          your NutanixCsiStorage instance by following the [documentation](https://portal.nutanix.com/page/documents/details?targetId=CSI-Volume-Driver-v3_3:CSI-Volume-Driver-v3_3)
          or the example in below section:\n**1. Create a Nutanix PE secret**\n```yaml\napiVersion:
          v1\nkind: Secret\nmetadata:\n  name: ntnx-pe-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n
          \ files-key: \"fileserver01.sample.com:csi:password1\"\n```\n**2. Create
          a Nutanix PC secret**\n```yaml\napiVersion: v1\nkind: Secret\nmetadata:\n
          \ name: ntnx-pc-secret\n  namespace: openshift-cluster-csi-drivers\nstringData:\n
          \ # prism-ip:prism-port:username:password.\n  key: 1.2.3.4:9440:admin:password\n```\n**3.
          Create a NutanixCsiStorage resource to deploy your driver**\nYou can do
          it directly inside the Operator UI with the `Create instance` button or
          with the following resource\n```yaml\napiVersion: crd.nutanix.com/v1alpha1\nkind:
          NutanixCsiStorage\nmetadata:\n  name: nutanixcsistorage\n  namespace: openshift-cluster-csi-drivers\nspec:
          \n  ntnxInitConfigMap:\n    usePC : true (false for PE based installations.
          Default is true.)\n```\n**4. Create storage classes**\n- volume mode example
          for 3.x installation\n```yaml\nkind: StorageClass\napiVersion: storage.k8s.io/v1\nmetadata:\n
          \ name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n  storageContainer:
          default-container\n  storageType: NutanixVolumes\n  #description: \"description
          added to each storage object created by the driver\"\n  #hypervisorAttached:
          ENABLED\nallowVolumeExpansion: true\nreclaimPolicy: Delete\n```\n- volume
          mode example for 2.x installation\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n  name: nutanix-volume\nprovisioner: csi.nutanix.com\nparameters:\n
          \ csi.storage.k8s.io/provisioner-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/node-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/controller-expand-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/fstype: ext4\n   storageContainer:
          default-container\n   storageType: NutanixVolumes\n   #description: \"description
          added to each storage object created by the driver\"\n   #isSegmentedIscsiNetwork:
          \"false\"\n   #whitelistIPMode: ENABLED\n   #chapAuth: ENABLED\n   #isLVMVolume:
          \"false\"\n   #numLVMDisks: 4\n allowVolumeExpansion: true\n reclaimPolicy:
          Delete\n```\n- dynamic files mode example\n```yaml\nkind: StorageClass\napiVersion:
          storage.k8s.io/v1\nmetadata:\n    name: nutanix-dynfiles\nprovisioner: csi.nutanix.com\nparameters:\n
          \ dynamicProv: ENABLED\n  nfsServerName: fs\n  csi.storage.k8s.io/provisioner-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/provisioner-secret-namespace: openshift-cluster-csi-drivers\n
          \ csi.storage.k8s.io/node-publish-secret-name: ntnx-pe-secret\n  csi.storage.k8s.io/node-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-expand-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-expand-secret-namespace:
          openshift-cluster-csi-drivers\n  csi.storage.k8s.io/controller-publish-secret-name:
          ntnx-pe-secret\n  csi.storage.k8s.io/controller-publish-secret-namespace:
          openshift-cluster-csi-drivers\n  storageType: NutanixFiles\n  squashType:
          \"none\"\n  #description: \"description added to each storage object created
          by the driver\"\nallowVolumeExpansion: true\n```"
        displayName: Nutanix CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Nutanix
        - CSI
        - Volumes
        - Files
        - Storage
        links:
        - name: Documentation
          url: https://portal.nutanix.com/page/documents/list?type=software&filterKey=product&filterVal=CSI
        maintainers:
        - email: cloudnative@nutanix.com
          name: Nutanix Cloud Native
        maturity: beta
        provider:
          name: Nutanix
        relatedImages:
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator-bundle@sha256:a26586b5f2d748f56dc29fbe3871d84de422e4103ce13c93dbe477718c582487
        - registry.k8s.io/sig-storage/csi-attacher@sha256:d7325367ab72b2d469a5091d87b4fc01142d2d13d1a28b2defbbe3e6fdbc4611
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:04c55b93a032ad8075cae540d60fc740a72c31006590dbdb61ad4ab9813959d0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:10624570c0aceb03f55f1eb07147b0c537e4676869cca2e9bd4bab113f810ac4
        - registry.k8s.io/sig-storage/csi-resizer@sha256:3c116f543f0590aeff3299c8bb0683f250817d11a77d9e9071b15a0bffdabcd9
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:9e581022a4e063b9d8b4f9466a36eabe4d0ced3badc8e353f13b700074ce4553
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0f64602ea791246712b51df334bbd701a0f31df9950a4cb9c28c059f367baa9e
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:9f121929227acf5581a813dc2f952105f517647c11a39cd7ea0863e768082586
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:9af9bf28430b00a0cedeb2ec29acadce45e6afcecd8bdf31c793c624cfa75fa7
        - registry.k8s.io/sig-storage/livenessprobe@sha256:82adbebdf5d5a1f40f246aef8ddbee7f89dea190652aefe83336008e69f9a89f
        - registry.connect.redhat.com/nutanix/ntnx-csi-os@sha256:f0e191e7bf98829dbf718978fc971444436a1ecb9957ee22a5d8f9ba01a931b7
        - registry.connect.redhat.com/nutanix/nutanix-csi-operator@sha256:9cc0ffc6e9c0a82320e207dad6b6b3ddc30ab8f83fa9e099a34450d5537ca435
        version: 3.3.4
      entries:
      - name: nutanixcsioperator.v3.3.4
        version: 3.3.4
      name: stable-3.x
    defaultChannel: stable-3.x
    packageName: nutanixcsioperator
    provider:
      name: Nutanix
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Citrus Framework
      provider-url: ""
    name: yaks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yaks-operator.v0.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-instance"
                },
                "spec": {
                  "operator": {
                    "global": true
                  }
                }
              },
              {
                "apiVersion": "yaks.citrusframework.org/v1alpha1",
                "kind": "Test",
                "metadata": {
                  "labels": {
                    "app": "yaks"
                  },
                  "name": "example-test"
                },
                "spec": {
                  "source": {
                    "content": "Feature: hello world\n\n  Scenario: print slogan\n    Given YAKS does Cloud-Native BDD testing\n    Then YAKS rocks!",
                    "language": "feature",
                    "name": "example.feature"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/citrusframework/yaks:0.17.1
          createdAt: "2023-04-26T20:37:31Z"
          description: YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["instances.yaks.citrusframework.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/citrusframework/yaks
          support: Citrus Framework
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance is the Schema for the yaks instance
            displayName: Instance
            kind: Instance
            name: instances.yaks.citrusframework.org
            version: v1alpha1
          - description: Test is the Schema for the tests API
            displayName: Test
            kind: Test
            name: tests.yaks.citrusframework.org
            version: v1alpha1
        description: |
          YAKS
          ====

          YAKS is a platform to enable Cloud Native BDD testing on Kubernetes.

          ## Running a YAKS test

          With the YAKS operator installed, you can run tests by creating a `Test` custom resource on the cluster.

          Tests are defined using [Gherkin](https://cucumber.io/docs/gherkin/) syntax. YAKS provides a set of predefined steps which
          help to connect with different messaging transports (Http REST, JMS, Kafka, Knative eventing) and verify responses with
          assertions on message header and body content.

          The example below defines a simple test resource.

          ```
          apiVersion: yaks.citrusframework.org/v1alpha1
          kind: Test
          metadata:
            name: example
          spec:
            source:
              name: example.feature
              content: |-
                Feature: hello world

                  Scenario: print slogan
                    Given YAKS does Cloud-Native BDD testing
                    Then YAKS rocks!
          ```

          Refer to the [YAKS repository](https://github.com/citrusframework/yaks) for more complex examples.
        displayName: YAKS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - yaks
        - testing
        - microservices
        - knative
        - kafka
        - serverless
        - camel
        - camel-k
        links:
        - name: Project page
          url: https://citrusframework.org/
        - name: YAKS source code repository
          url: https://github.com/citrusframework/yaks
        maintainers:
        - email: cdeppisch@redhat.com
          name: Christoph Deppisch
        - email: nferraro@redhat.com
          name: Nicola Ferraro
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Citrus Framework
        relatedImages:
        - docker.io/citrusframework/yaks:0.17.1
        - quay.io/openshift-community-operators/yaks@sha256:0031de9a68502cb65923c4e17453195663f59e46b1cda37619a857add503c016
        version: 0.17.1
      entries:
      - name: yaks-operator.v0.17.1
        version: 0.17.1
      - name: yaks-operator.v0.17.0
        version: 0.17.0
      - name: yaks-operator.v0.16.0
        version: 0.16.0
      - name: yaks-operator.v0.15.1
        version: 0.15.1
      - name: yaks-operator.v0.15.0
        version: 0.15.0
      - name: yaks-operator.v0.14.3
        version: 0.14.3
      - name: yaks-operator.v0.14.2
        version: 0.14.2
      - name: yaks-operator.v0.14.1
        version: 0.14.1
      - name: yaks-operator.v0.13.0
        version: 0.13.0
      - name: yaks-operator.v0.12.2
        version: 0.12.2
      - name: yaks-operator.v0.11.0
        version: 0.11.0
      - name: yaks-operator.v0.10.0
        version: 0.10.0
      - name: yaks-operator.v0.9.0
        version: 0.9.0
      - name: yaks-operator.v0.8.0
        version: 0.8.0
      - name: yaks-operator.v0.7.0
        version: 0.7.0
      - name: yaks-operator.v0.6.0
        version: 0.6.0
      - name: yaks-operator.v0.5.0
        version: 0.5.0
      - name: yaks-operator.v0.4.0
        version: 0.4.0
      name: alpha
    defaultChannel: alpha
    packageName: yaks
    provider:
      name: Citrus Framework
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NooBaa
      provider-url: ""
    name: mcg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mcg-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:91632b89b4001e2ffa0f1aef56207eb48f0275e20c635e7f1ff7c6de57a59ec2
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:25d8929559c582a4033fba1cdcc28238eb1b2921d137f8cb9a48741b423cf9f4
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:eed8048d085eef7eefaa643a09627521fc2596ebd26280e05c79dddaa0b01afe
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        version: 4.17.13-rhodf
      entries:
      - name: mcg-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: mcg-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: mcg-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: mcg-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: mcg-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: mcg-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: mcg-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: mcg-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: mcg-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: mcg-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: mcg-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: mcg-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: mcg-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: mcg-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: mcg-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[{"kind":"NooBaa","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"noobaa","creationTimestamp":null},"spec":{"cleanupPolicy":{},"security":{"kms":{}},"loadBalancerSourceSubnets":{},"autoscaler":{},"bucketLogging":{},"bucketNotifications":{"enabled":false}},"status":{"lastKeyRotateTime":null}},{"kind":"BackingStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"NamespaceStore","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"type":""},"status":{"mode":{}}},{"kind":"BucketClass","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{},"status":{}},{"kind":"NooBaaAccount","apiVersion":"noobaa.io/v1alpha1","metadata":{"name":"default","creationTimestamp":null},"spec":{"allow_bucket_creation":false},"status":{}}]'
          capabilities: Seamless Upgrades
          categories: Storage,Big Data
          certified: "false"
          containerImage: registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
          createdAt: "2019-07-08T13:10:20.940Z"
          description: NooBaa is an object data service for hybrid and multi cloud
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: ֿ'["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/noobaa/noobaa-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A NooBaa system - Create this to start
            displayName: NooBaa
            kind: NooBaa
            name: noobaas.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos,
              PV's and more. Used in BucketClass to construct data placement policies.
            displayName: Backing Store
            kind: BackingStore
            name: backingstores.noobaa.io
            version: v1alpha1
          - description: Storage target spec such as aws-s3, s3-compatible, ibm-cos
              and more. Used in BucketClass to construct namespace policies.
            displayName: Namespace Store
            kind: NamespaceStore
            name: namespacestores.noobaa.io
            version: v1alpha1
          - description: Storage policy spec  tiering, mirroring, spreading, namespace
              policy. Combines BackingStores Or NamespaceStores. Referenced by ObjectBucketClaims.
            displayName: Bucket Class
            kind: BucketClass
            name: bucketclasses.noobaa.io
            version: v1alpha1
          - kind: NooBaaAccount
            name: noobaaaccounts.noobaa.io
            version: v1alpha1
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information.
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: "The noobaa operator creates and reconciles a NooBaa system in
          a Kubernetes/Openshift cluster.\n\nNooBaa provides an S3 object-store service
          abstraction and data placement policies to create hybrid and multi cloud
          data solutions.\n\nFor more information on using NooBaa refer to [Github](https://github.com/noobaa/noobaa-core)
          / [Website](https://www.noobaa.io) / [Articles](https://noobaa.desk.com).
          \n\n## How does it work?\n\n- The operator deploys the noobaa core pod and
          two services - Mgmt (UI/API) and S3 (object-store).\n- Both services require
          credentials which you will get from a secret that the operator creates -
          use describe noobaa to locate it.\n- The service addresses will also appear
          in the describe output - pick the one that is suitable for your client:\n
          \   - minikube - use the NodePort address.\n    - remote cluster - probably
          need one of the External addresses.\n    - connect an application on the
          same cluster - use Internal DNS (though any address should work)\n    \n-
          Feel free to email us or open github issues on any question.\n\n## Getting
          Started\n\n### Notes:\n- The following instructions are for **minikube**
          but it works on any Kubernetes/Openshift clusters.\n- This will setup noobaa
          in the **my-noobaa-operator** namespace.\n- You will need **jq**, **curl**,
          **kubectl** or **oc**, **aws-cli**.\n\n### 1. Install OLM (if you don't
          have it already):\n```\ncurl -sL https://github.com/operator-framework/operator-lifecycle-manager/releases/download/0.12.0/install.sh
          | bash -s 0.12.0\n```\n\n### 2. Install noobaa-operator:\n```\nkubectl create
          -f https://operatorhub.io/install/noobaa-operator.yaml\n```\nWait for it
          to be ready:\n```\nkubectl wait pod -n my-noobaa-operator -l noobaa-operator
          --for=condition=ready\n```\n\n### 3. Create noobaa system:\n```\ncurl -sL
          https://operatorhub.io/api/operator?packageName=noobaa-operator | \n    jq
          '.operator.customResourceDefinitions[0].yamlExample | .metadata.namespace=\"my-noobaa-operator\"'
          |\n    kubectl create -f -\n```\nWait for it to be ready:\n```\nkubectl
          wait pod -n my-noobaa-operator -l noobaa-core --for=condition=ready\nkubectl
          get noobaa -n my-noobaa-operator -w\n# NAME     PHASE   MGMT-ENDPOINTS                  S3-ENDPOINTS
          \                   IMAGE                    AGE\n# noobaa   **Ready**   [https://192.168.64.12:31121]
          \  [https://192.168.64.12:32557]   noobaa/noobaa-core:4.0   19m\n```\n\n###
          4. Get system information to your shell:\n```\nNOOBAA_SECRET=$(kubectl get
          noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.accounts.admin.secretRef.name'
          )\nNOOBAA_MGMT=$(kubectl get noobaa noobaa -n my-noobaa-operator -o json
          | jq -r '.status.services.serviceMgmt.nodePorts[0]' )\nNOOBAA_S3=$(kubectl
          get noobaa noobaa -n my-noobaa-operator -o json | jq -r '.status.services.serviceS3.nodePorts[0]'
          )\nNOOBAA_ACCESS_KEY=$(kubectl get secret $NOOBAA_SECRET -n my-noobaa-operator
          -o json | jq -r '.data.AWS_ACCESS_KEY_ID|@base64d')\nNOOBAA_SECRET_KEY=$(kubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq -r '.data.AWS_SECRET_ACCESS_KEY|@base64d')\n```\n\n###
          5. Connect to Mgmt UI:\n```\n# show email/password from the secret:\nkubectl
          get secret $NOOBAA_SECRET -n my-noobaa-operator -o json | jq '.data|map_values(@base64d)'\n\n#
          open mgmt UI login:\nopen $NOOBAA_MGMT\n```\n\n### 6. Connect to S3 with
          aws-cli:\n```\nalias s3='AWS_ACCESS_KEY_ID=$NOOBAA_ACCESS_KEY AWS_SECRET_ACCESS_KEY=$NOOBAA_SECRET_KEY
          aws --endpoint $NOOBAA_S3 --no-verify-ssl s3'\ns3 ls\ns3 sync /var/log/
          s3://first.bucket\ns3 ls s3://first.bucket\n```\n"
        displayName: NooBaa Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - noobaa
        - kubernetes
        - openshift
        - cloud
        - hybrid
        - multi
        - data
        - storage
        - s3
        - tiering
        - mirroring
        links:
        - name: Github
          url: https://github.com/noobaa/noobaa-core
        - name: Website
          url: https://www.noobaa.io
        - name: Articles
          url: https://noobaa.desk.com
        maintainers:
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: etamir@redhat.com
          name: Eran Tamir
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: NooBaa
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-operator-bundle@sha256:ed5c7c8cbbe455b6d31b1c55871585a0679a91800f53743d81d4cc36c6aac7cc
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/mcg-rhel9-operator@sha256:7b9a0973dc6b51a72a17633e5474b91e6998b0f2e1cc596267e13026671325dc
        - registry.redhat.io/odf4/odf-cosi-sidecar-rhel9@sha256:88387ba7a9e85c1bcf57966558db57c099fd301ea80a68e43afc2bca3dfe94a7
        - registry.redhat.io/rhel8/postgresql-12@sha256:af1ab437b165fa4ee9dfd6ade637d7de5fa556d0b3851fb8edc78afe0e6fa95e
        version: 4.18.11-rhodf
      entries:
      - name: mcg-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: mcg-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: mcg-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: mcg-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: mcg-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: mcg-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: mcg-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: mcg-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: mcg-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: mcg-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: mcg-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: mcg-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: mcg-operator
    provider:
      name: NooBaa
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: rhods-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: alpha
    - currentCSV: rhods-operator.2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
          createdAt: "2023-11-21T08:37:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            Data Science
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
          olm.skipRange: '>=1.33.0 <2.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "rhods",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "rhods",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Data Science"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift Data Science
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift Data Science (RHODS) is a complete platform for the entire lifecycle of your AI/ML projects. It is the flagship product of OpenShift AI.

          When using RHODS, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. RHODS can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * RHODS dashboard
          * Curated Notebook Images (incl CUDA, PyTorch, Tensorflow)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh (and Kserve) with a provided OpenVino Model Serving Runtime
          * Ability to use other runtimes for serving
          * Model Monitoring
        displayName: Red Hat OpenShift Data Science
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift Data Science
        - RHODS
        - rhods
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift Data Science
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-data-science
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhods/odh-ml-pipelines-persistenceagent-rhel8@sha256:857a8eb444375505ccb34a11cce69c1794ceabbfb00df06933d7f12cd187319b
        - registry.redhat.io/rhods/odh-kuberay-operator-controller-rhel8@sha256:bfc805aa1689a4bd6f7e617acf9eb1c2c307624dd49956e9e7442afac328151a
        - registry.redhat.io/rhods/odh-ml-pipelines-artifact-manager-rhel8@sha256:76c67d7015a2db6965219db6533e0cddb5b7cf699c88df3ef81155f9c192a3cc
        - registry.redhat.io/rhods/odh-dashboard-rhel8@sha256:b98171c5543815abf7c095fba92796ceed9ea44bd8b78db143d2815c6bcfdcf6
        - registry.redhat.io/rhods/odh-operator-bundle@sha256:b637a02d23bd8364cc8914421049eb60169c163ac70bff2f33591df1a1193002
        - registry.redhat.io/ubi8/ubi-micro@sha256:6a56010de933f172b195a1a575855d37b70a4968be8edb35157f6ca193969ad2
        - registry.redhat.io/rhel8/mariadb-103@sha256:08c90a8f133b4cdb1477a742bc44be4d8fc4e88153b8bc2a2a148cbdf19f4e7d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:eb96f4b753606528750b3c28f7f9256582f81fa030450b05afdef684634903be
        - registry.redhat.io/rhods/odh-kf-notebook-controller-rhel8@sha256:cd0ab8d70559be6074fbaf1caa1d50f8eb75e1bff1a22d6d69380f19190df0b6
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhods/odh-codeflare-operator-rhel8@sha256:2841fd3051a50a1b0d9e3caafa20a8084b5c472d49393b372828960dbf156674
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhods/odh-modelmesh-serving-controller-rhel8@sha256:d94395fb251e45aaa73d09bc6106a2947c16b90d87fc3b64bfe4f2f600be5841
        - registry.redhat.io/rhods/odh-notebook-controller-rhel8@sha256:753994fce63802f1e1121990a5c1dd722216446abee54cc1a7c50ab195134414
        - registry.redhat.io/rhods/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:c5758102609821a6fdabec0277e03c4c728f82176093d1ade0570aba5f3cd8f9
        - registry.redhat.io/rhods/odh-modelmesh-rhel8@sha256:3827221f84d3aa22440d6b364b990a4ec0666dfb5963f174f7ed518b3650d223
        - registry.redhat.io/rhods/odh-ml-pipelines-cache-rhel8@sha256:33aae32e3d348d382906f6e926281852f54006a94d7bdac188bb33debadec059
        - registry.redhat.io/rhods/odh-trustyai-service-operator-rhel8@sha256:68471555d0b2bb8150ebfae8d4f3d1dbdff9c5b961ff2305ef23ab7514cbbea1
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhods/odh-data-science-pipelines-operator-controller-rhel8@sha256:250db47cb873207cfb75c7e8a2f0de08b75b056a72db73362d2d79c08cb7b185
        - registry.redhat.io/rhods/odh-ml-pipelines-api-server-rhel8@sha256:be57f237df156474a709667fa4fb4aa142118e3459963fed56648c1cb2e38cd6
        - registry.redhat.io/rhods/odh-modelmesh-runtime-adapter-rhel8@sha256:08478a1647d0413783bcbcd966a73b444b580464de46efedb39504420f22053e
        - registry.redhat.io/rhods/odh-rhel8-operator@sha256:a4d7f020509b9720063e87e1c2f19af04e4c9b4986b199be4dc617831b971434
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhods/odh-mm-rest-proxy-rhel8@sha256:8d989f54009484d7f43fe79d7c37757418fc3c15cd4d7e80a2076677a9645bdf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhods/odh-model-controller-rhel8@sha256:fde989817cd1a6ec8eb408a7f6c73e28ba043ab2d88035586b1b7e55fb3405cb
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhods/odh-trustyai-service-rhel8@sha256:fe17d6e07ebaa088f274dd8b5e83892dacc11cb336065b06582a3c055095dba5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.4.0
      entries:
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: beta
    - currentCSV: rhods-operator.2.21.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
          createdAt: "2025-09-24T14:42:33Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.21
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:85202cb75250a1ba2691d65b52b306fd77566660dd29cdf56bf8962e8bb4939c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:a936d87646843edea2a84eb429bb44eb730bf3e04e8900fecef66721426dba89
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:7d0b8d69b6defb5f232da066573b0067ec810a8ccfb2f3cb91d472d2979ab7a7
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:6a388ee3cbe3813ac955d319f0ab65a81da321d982c9881c6c81fe135311fe1d
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:2a16c353fd13dc1f3f5c4266f1b9786a8e7a100dc05f2d554bdaef082271063f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:1b5452b87b36de2b62febc777efdc6233a168ffcc09e2067e482cc444463a57d
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:2638e3a90bbb242dd2483fefe9fc8b472f334c510d0139b9b7c875c30dab7984
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:4c192208c618b49f31cf31c9de1a33c622ce3f120204c2eec8ff7c439ea6785c
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:9c2d6ce05775ac34ee5695b3367d12bf2a9c14e318a9544d00a56653ea5f2007
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:2b4136569e10ad7d47c82bedf55c1705a833dc67c8f833656996bfe1fddbef3e
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:7da004e158006dcbdcf0c3b78c40411b82c9d56ebf54b1c30c245b6550b8ee7d
        - registry.redhat.io/ubi9/toolbox@sha256:7656f815a0d13ed37f56d11635559a87a3d00840253755c71b644b858f67d556
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fc98ed817a4e9ab84e2a66063987304f0cdbe6e5620a58a7f25cccd61b34ebf9
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:a88696cff8cd2cc2e295c9ec082b7f25f61dfc9ac12f1c208529ff7036d5af29
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:239f226328e5ac77d6dee498217d55103a6d562f4829e38abf65ba55fda0aca2
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:5c0d56f434298f6e8bf6cf7061cd1ae6df9915225cbf04b010b6b59a64775cc9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:e68492a6a75cb83be0a968b832c1478b97f14e661376328085fd5416a08cbc31
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:cb47e165935f0d9d3c88088689c37e68a69519357a20d3c52ab6f71240dab507
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:152b425725228953f43b57a4a034a6725993a5ca556af328ae5e06aa3f1acbfc
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:dc0db686b7688f98b5f8595661022cdffa5d26a078289cc7aed528f0ee1d08a4
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:52e1ba8ebf4d55c64c94c9076185c5612b537e6e5ed4944fcd6ea87ccb9cd182
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:4c57dcf4df09fe0d03578cc8b9a3ca2ea485b6af0ca0203bfb85e640058b7daf
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:1b7befcd29c9c8f783bdd3c5654b0412ea45cd88cd18eb5f8f5a411f1323c6bd
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:fe6593bc73e77d315d82b7921bb1975c19fda1e7630ef5b30b69f5901ce30178
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e9612ec99daa171c403b5ddf37788526e5375b83987e5de9b1ebe519198607b4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:003fdc6cd1821d620e1ce4050ee608e2d2a32f2b08c8fa9c147e2e14418a8d91
        - registry.redhat.io/rhel9/mariadb-105@sha256:c390b12f5c2c3ea7cd897b59eba7127ce22f474130cd8c1bbbbaadae870baca4
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:80fd53a580ce7bc4d02023d38a04912e84c2cd50679fc5ea034afba935623157
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:04859c878ffcc8c42550a98be625ac52212f9b79a6fd5b3baeafb3592926cbe9
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:cfaf4eb86ddb9c717340ab17fa4b83323e5b40977ac70265376c87329a3f6f20
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:687c8eeed55f021ecaab1307f0e88b5b16d91f72d63b3d7168d7bbee90e8947b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:c3e283c0536266195dd461042ca0036ed98cc33650d806be132db4fba45239e4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5f581957e4f8deebf6104a394272e6fbb1c54d813bf069e95875d865eac209ad
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        version: 2.21.1
      entries:
      - name: rhods-operator.2.21.1
        version: 2.21.1
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: embedded
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.16
    - currentCSV: rhods-operator.2.8.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
          createdAt: "2024-11-07T07:17:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:60507a3dddbd88560fd85204699147facd2b6371fb25e0bf6701dad02c0b4dd4
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:45498198fadfbc1457bf2a6084ae7ea515d454d86d148783aae3abdaca6e999d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:9aa5a93831992588766ac8594ae3c4dc7111d08a90fdbc8751c40484339e46d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:13b850a5a3f8776c6c8db887c666d88df72901895b5ef1be9a8bd372cc02fcb8
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c48474563d98df32e7243e33302428ea8836be5e1a2e30d20b4d796a62cbf6b
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:34d177abb56b948eaf9ce6fa62e37427ddee9fcecca48e752d09f72c69847a46
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:eeb8bdc94e7fa2b454fedf72eb75756052f8bfb9d92b116a8267a5584e2528c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:1dbf09cb1624c9f961f8ac8e371c97aa7c3fbfcad4c4d02ab4fad0eb86d2aa95
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:2f6db789ceb3f90f9c9911c6392e62d0a8e4d61be68a7f7cbd781f47faab6389
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:8ffc5da798e783c65d5be7f344219b2951d8c2233d6494a4ae2bb5c86eccd699
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:3b9e8a71ba795a2ea91f4129253ccb4be3a79c7c1c6a87082f96ccc67eabbbac
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:61a506fbf263cd72951ac8bd151127e0c74fc74f0a05f7aaecaa54fde47ee490
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:81ae2b95f2da740cb5264b9ba8b85f0f58bb4ab80d6431f66d52848ed0c21afb
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:d87e4b1cba72eb4f50b3f2d59863799929c22ca37b01a1ff4828a9ce0be31f25
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:4b1b9f05a94713292a7a2f6782fe1d3eea6f36214c36b38f402aa91a554ab18a
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:13dff890ec67fa56864c244a40fc404e9eac05f33573dd16d088b37f9340ec41
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f9c474d454d0b006e06859601ec9a64b5e5c91e9a47bfab6ad6d8cce5aa6b89d
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:df34e9d2d121a3def2dbb86f108b3fb6948bc7ba7cb44c70801dfd2e81d3fbb1
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:038ce76571217d6c624d614652e8b96ef6484562cf52ec1be88392b8ef8a3bbd
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:28ec01e514c15cae145499840597d1be1b750123bb6014de2ecc3ef0d60fa134
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:1a1074263ae26e4c8b0e3863228c35e50133169f377ece6c47b58a12e0fcf453
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        version: 2.8.5
      entries:
      - name: rhods-operator.2.8.5
        version: 2.8.5
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: eus-2.8
    - currentCSV: rhods-operator.2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
          createdAt: "2025-08-12T18:04:00Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "llamastackoperator": {
                      "managementState": "Removed"
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (including PyTorch, TensorFlow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
          * Llama Stack – unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.23
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:4847ac1be55284815670c2048c1edfc56c7227494280d3c6094aa23f1ed1a9d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:edcce7cddfa4c250108532e56541ddb95099612c180b2297e5934bc89d2bb46e
        - registry.redhat.io/rhoai/odh-llama-stack-k8s-operator-rhel9@sha256:626b9ef09ee82c8d425ddb7ed3f2ffbb3d1019f541e31604db891787c881d5c9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:ca21e218e26c46e3c63d926241846f8f307fd4a586cc4b04147da49af6018ef5
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:ccfc84028a95f1cc6a16e06ff6e9ed2ace74f74e5e9963a4fb33544c2ce005c2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:240692b86bc8a4fe0b545e6e4cd7cf4ff0ca990406bc55f0446d565fd67e37f6
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:e0c9ac73891781ef421e0a020cfd5c962f7632d38020897183cf048b1819326a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:16ff53f4f1a1e1e66826c8355c654038da83df545427a931485ec7a2d3764b16
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:84c3213e11299ec032aa0e5a4dfa9b3055c4246b04b029e888d8d9f1caf5c30c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-llama-stack-core-rhel9@sha256:43b60b1ee6f66fec38fe2ffbbe08dca8541ef162332e4bd8e422ecd24ee02646
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:467b04f4ee4118bbe0934d1ab8b80f72df8703dcdac61c2b53fdb5bd1b404c14
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:2886e40f94fde72eb5959ad564bbdd466657208c09bac731b3860b0b54f104c6
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cf904f06a9437ce5fa0bce2666a044a298d9df7a6de5344f43fecbba42678734
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:795ab1842ef675e3d8b47deb8a7485a69064428e823dfeb930a6a7e7017cde0f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:bb7c352ccc10a8c60ef8d573368d0ada59e73cd131653d11bcebdf9f564acc10
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:3243caa4c8aac75069d19225dcee361bf9c012782328a6266b7016b3026f6d56
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:fbe10282ada2fa6bf07ae97b497f46ebe735de130b4ce0182bf34301429999cb
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:0efed19630333b3d37e8c395b5e88536bda715f4b6a74cc9636649f6aff88278
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:a4649e8ad4970bc3d5e967811c9694fa82d78ee9f8f025b8b2fbd37237eb14fc
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:756b7bdaa7f6e7c4a18488e690188dd91153a33913d225bab001b0085382e514
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:e50adae13e9d773e7be8bf81b7525dac3f5e3864c9cdf4c287342bd568ee1d6b
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:da8debeb2b84a42aba0b98785ccc20e02ed520e125428fe012df3ca473cbb755
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:2d2e7803b2b1e15017ca04961f9dcd25e89c47b07b101ca54cb9e793e6027202
        - registry.redhat.io/rhel9/mariadb-105@sha256:050dd5a7a32395b73b8680570e967e55050b152727412fdd73a25d8816e62d53
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:68e74bdf518fc624cc25a59a18b2a87d0aa2a9b8aad30750ace7115d297a21ca
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:23c098f7e791441cfcfa6bf37de51763511bd8bf4d543747b7d37ed591d68886
        - registry.redhat.io/ubi9/toolbox@sha256:4213516b6fe7957a01072720efd9ecbc7d4443238e151b5fe0263bd0c36792b1
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:ff740fc3615edaa90dae976da1aa6455d64811e239eb66b422836a3aae791b41
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:80d284cac2ef6638fe028102b71a70f26041576c11a49414e77430954a3c1b99
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:a1497e64349634b2567d28f685052de118f64f6b32e7d3ea469212e2004d7eb3
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:208005da389b7d4176ad8bffc703312011a1299954df151e2abbe8f3dec01d19
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:5fb7e6c3ee1e3b170cd146ea9cd172fa953326b08163897605ca74a3644f2819
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:857560240bb45bfc2db6d7743abf11a47f929a0d3035940256a6527afd313412
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f23f8051886d0f0240937507471c818f440bca8703b8209a836d5dbe3ea2f7c8
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:b038d20def504c140e4365f2eb117a979534629332bec75a2ba750ebafdefe3d
        version: 2.23.0
      entries:
      - name: rhods-operator.2.23.0
        version: 2.23.0
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.21.0
        version: 2.21.0
      - name: rhods-operator.2.20.0
        version: 2.20.0
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.18.0
        version: 2.18.0
      - name: rhods-operator.2.17.0
        version: 2.17.0
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.15.0
        version: 2.15.0
      - name: rhods-operator.2.14.0
        version: 2.14.0
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: fast
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable
    - currentCSV: rhods-operator.2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
          createdAt: "2024-11-14T21:18:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.10.0 <2.10.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:4da77afafde99b15fd357a4ab6609644bf6330a1d9c0c03e64726ed1b9633230
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8911afbb65c8c4f10c3597624485be54b9103e87a5020b718b35ec3df2c9bf00
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:220cad7a22d40998063c6b4fb2f63e69330e64ca7118490a6680b05fc99b7733
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:267d6795bf53f9a1955d73445776a428dd042af0cd3cb8bb5b91692a1bc1e5d2
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:fdc2bc06b0dfa79ebc6451ac3793b5e8d9f165171633831dafe46ebc18c10c35
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:cba9f25ac6bb2734bb990a622b515c838b881edb688bf44ff196ebe7684ab74b
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:0ff5664c2eaadf3d3073b3d7fb711da47fb78d21f67710d8b8aad5452cb0355d
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:0af6a91553f40ea40e86d85edc204658af3e5a45e8ab94cdb6b767e545df59b9
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:932e83fe9c9b54934767a7c5af683b2ab08bd2b4f0524c191f1c766dc64d9c3d
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:9f7b1cb8623de8b0550b0ba124d3b50f5e960ed011b33a3262c6346dbe2dff95
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:a4b9553fde02d9c67a9c98f063eaf82fb1d2689f4b1269e57bf5f12c8d4adecf
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:8fed8f58732bff512289fff738d0a94be67286c41b24eaf5928bbcb30d88c88a
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:a9c6e3afdf605465766fa93334cc0f6bb5e46a5fdfd7ec64d648e2ed72b05ccf
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:4876f14ca9cff0e2b998a56f2f56d0775edf3963de610f4b810b4d94fb12d719
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:24238abd46633709feb75f2a3e41cf6c9e8528d2c0e4be2fac6f56de44e3fb9b
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:5d15175f934017c04d6a6057200ff55b2172a1f4bcbeb0f2cd9e9e545631d93f
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:8643f7ee8afacfe6ae08cbbf81da76ca97032dc3ce3da4fddf16e330e9e33a08
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:52e8deeeeaa21b17370a5d5cd644745da745b6e1cee8c6295d6df75c8d2aec04
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:12741379ab24d2286b5fae77f4383fb65713af7b3356fb43d8254ea85d4b0ee3
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:65db5e8cb2fa46d020ed4837d4479d133b52167c373e65f81cb590eec5391af2
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:d709f65e1194bc877fa4f9ec681b598467703d94139d57b58c24490d3a59fb82
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:d2514bf4b4b2d9abc19facac3f1a55405eb3ad0d9e9203e8816bfcfa08c9a6c5
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:214ca0608fcbc0fd585bf01768d28c7ba17082d1043349ec8445afd663095383
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a1d590d0e2b3fae9ecbd843a664dab083061e797f66a9f0a021b0c2c487d7e45
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:7df2f8bc2c5c0f4f347d009b62a058b5107413a2abd789f6010eb4d6c9e4f8ee
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:f4d7d3ba4a0d3322bc9f327e095e9ec56073758f6dcd8da8c7bb185cf952ead5
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:d9b55e8f25b81050380b4aab040243dc18a7432c48912c9b660c97dcc5eea008
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:807ec05e9e107afc969e718d0c2c980959e2fd462e777a5ff63e55156ea150a4
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:232ebac25084db056fdde49477b5c1e6a03a0228fea3758bcf6234b3d0eaa582
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:2007da5abf840712cd2edb50462b2d8d571a0193c3ae1550d4a279eddf2b335f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        version: 2.10.2
      entries:
      - name: rhods-operator.2.10.2
        version: 2.10.2
      - name: rhods-operator.2.10.1
        version: 2.10.1
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.10
    - currentCSV: rhods-operator.2.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
          createdAt: "2024-10-17T10:53:18Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.13.0 <2.13.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:7c895490d49beb6bd724f47ad1b17ce501650557b0889314799c4072632f2b99
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:8883bf71abf39b4a6985940c287eb6ac7dc137d491264805d48bb112b65bd008
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:a3048eaa8924ce3571c6e78014438183fae70d9e8c833a7c07774c675104e3d4
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:5c2ee4136ee0014d1ef395b5e3bcab125c846f5e258ece3631d912f431c8d208
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:7d085942e4da73c2ff4885c62f10b904dc70fd08e819e62f925eb1b4ddfa1554
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f5b17f44c5ab72d6cf658f72f24f15e06a5ac9be1790336f7228ab8e7cfc15c3
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8c19c7fa535f36bdce3b73cf3300e321909d44954cbc63a8aef6321c5f2fa693
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a744c1b386fd5e4f94e43543e829df1bfdd1b564137917372a11da06872f4bcb
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:cf3779c64bbaef1988886ed629c516e7d6ef7bdc73b9c95ad4c087b62e75750d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a37529eff3656a570d796083404e5a3f63603afecfe9e738da1a243c0cb3d7c2
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:6b2c0d1dd40e2c276684e17818b14375bc778e6c88df0bdc3cc24bfc9294804f
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:c2bd493e9345eccd124bbe0440e45cce8cf57588827d8eb9533f8bbadaa3ebbe
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:28b5fff602ad6fb36d462e1905f453509ec22fc251a1e0cf64fedba86ff420d7
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:5d6354e65106460ab9285b42124faf288298387a8b16317565bd591f009c39a6
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:a55f73d72ad2b07f01c49949cdb00969bfe4789375b3d7cdf9814b969fcc808d
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:8faab013568ee023968a5ec5b3a79977d7ddae3deafff6328e6fd56a27198e24
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:74e6cfa0bc481588cb8dc61c206ad567fbd2b28f43c2a86bee4a8e0eff55447c
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:cd6b535d95eb1b44fe261a5336270fd898b966f431c14e173944484381c26d25
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:9ccdd7fe8223057a2234a3a2eb3b492e7b69e6ce714cc26a805fe2dd82935195
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:71abac4fb3d613edef7f3ba117a73bbeea61f341d6182e8310360a76576c789e
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:030c53690fe25e15bc9907f51c3f04fab39b4d085ff2098d26316ba32cc8b09f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:bf25fce160dd7b8cce11e68a501006e4fa81a7858645038cdefdbba03df7ecdb
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:2566b73e98b783871fb500bb1a8cc4e3d58a513c1bb18abfb525da0818e6138b
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:850901511bde065cf73e4c18d787dc10b6e91ca6139fc3ba68da28c41dd28236
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:fd4ce524acc1769475079c7231f5ea79ac29551c513855ef34e5694d2c9fbbd3
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:363a35c332fc80b068157951d5e50e6b7209947649cc64803ffc02726e7581ab
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:77bca19180f30ef08f4d1598ac6c365e190bb7909ac4a1fb35d9ba9ba8d6ff7a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e1c19bbcd4c010ef2bd4f760eb24b2c96cdb82b4e025361a433e6a241eb23cf5
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:59e9cc9cd57ba48ca7e87ae1e797a902f826bb890cb198ccf549f76ea581ba6e
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:688389195aebf3f43e74e40d591d2fd5f9ae41994d8ac2db661f292f6eda012a
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:fb630d6f5817512733b8ea05f7147121928917a24ac8d98239f43d920880c359
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        version: 2.13.1
      entries:
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.13
    - currentCSV: rhods-operator.2.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
          createdAt: "2025-09-24T13:27:25Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Removed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.16
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:ee01e89f98feb185f6cd59c564e590a13e4d8d9ea760cca8de51426eb71b83a1
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:c8456d98e90b6505957ab3686e9fd2f156e29f123c5558e581c206daf1e7d93a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:d7d4fa406e0fcf0507894a7676532b27f45be742467e603a86f98ea5d2615df8
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:633984b538f027c93c5886ea1045dcbe81eeda74acf80001d5fc5f765bdbe0be
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:97e2bd9b587f08e135a9aeb9b3e0dc6eafa1a9bdacbb5ecb681ce9bd5aa37fc9
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:de5f91180ead7d73a1825fe8b032fde9b8b01392569c9789f5ce1b4b9c08a98f
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:22557a6fa52d2f311750a9ba253860f423ba697d26efa02ef8524a8258d2a909
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:11a301728573adf2b64ea72d0cb2d83ea5d4dbebea759f346e99f18c3d368c6e
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:4a7599f8a866eb169c9a62885906adbf6df0417c0be15857df1eef20cd9b1be2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:a0f72ffefb2b74b488dd949493f5d295a39bb9c97f578bf219d0138601f65468
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:efd039012559786deb3c521a130886c265e88d635e08baace2e573a9df270134
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:f37e4048f3a152798286793f3abfc6ed814453fcbe2667255a7e78eee483c5a3
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:4a212ec634225c14beac09be24ddf336e562f2aa9a13555fb1196f366ddae23c
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:4c5ff3496b2a2a739939d94ee9dafc02b682100785d228dc2fde480fb597b7a5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:83f193823c6e8f1c4772c46db78e29d3a186f4d8b41eceeba405ec519c1bcbff
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:6cf74044ae8d5308a2dfe03fa5d81086c89302db7f4cdbe2f4174a1c48b77869
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:7b70af8847b0806d8b43c399e2b3109f016fd864f5e9d30c44e2baca5d1359dc
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:f738aab1eab25854c93e1d8d4d98100a8ae7bb45a6b83f0326774e4220b1183b
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:65482864055021272a18b5b26792ab00cfa5fc9cc005d8d3a884cc82991506f3
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:5dcdcc2424602a69451f16d31dbfa1d43cb72c095ba561eb9076f0cd1e8182ed
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:2eedd1e2465d2e78fafb39e234dbb413a82713f187c64d70a80340ec94807d30
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:27b113d96453c2054d4c965a963d2badba6daac235eddaf23234c76e87bcb069
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:23d307a36b69e0df04f72a7d3b35e28d8417a8bbe23dba31e8e977569785c078
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:5192f1dbd2a9ab92ae390c4ae506efbed0970545b6122e95b014728ac937e777
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:cebc8815e03b772343b15d0a7dce8fad6fcc71dd437d871db5a3691472350803
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:2ba2b2c4db8bb334c50f4dfb54059f060361186900a44c06eed00b7a3c43977e
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:04f305c44413ae7dcb7017e53570ee49a509701792c5f50efadd64f47395730b
        version: 2.16.3
      entries:
      - name: rhods-operator.2.16.3
        version: 2.16.3
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.16
    - currentCSV: rhods-operator.2.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
          createdAt: "2025-09-24T13:27:42Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-envoy-proxy-image: registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.19
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel8@sha256:49c02285fcef6949892cecca20eeee49d03a9bb4d17bb9ceff7c7a0dbbb8ab40
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:b30d60cd458133430d4c92bf84911e03cecd02f60e88a58d1c6c003543cf833a
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:a6e02e1876fd5382ba4921dd65650b2d747666feb96c839f34e99d2972dfc323
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:baefbe4e7dd5aa74c71cb81da4a7a2b888a02d3eb1398ec4419bd8003b77f0c2
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel8@sha256:876ca7ce7b0b975688fb4a77d6cd1a6f87cf9eb40c2a9a5b84d8cffdc80eeebd
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:c1c97b6beb759d3f6f2dd53c4258e4ccbb3bf35cf11873bd446b135defaa69d6
        - registry.redhat.io/rhoai/odh-model-registry-rhel8@sha256:520ed21b80e5487b3a0a13b7e5c7d791d4dd28e0b6799d57abb65d53fca5a03f
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel8@sha256:6ad35cf5c4c62ed5ad4e192afe8cebead9796cb24a964401c76ff8656c47742f
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:4e8f6c6ca0ed516b8310af0332637683b5bfc8f2946eb935c647698dffbbcc79
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:6f01ac7cc62ab961b9c8a630a558338a61f6d7902301c4a1ad337495ac7ce85a
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:c605e5fdb3d66302781146347df6da5a91dee97e4f8f5d9ac33f105b8cdee383
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:8ce44de8c683f198bf24ba36cd17e89708153d11f5b42c0a27e77f8fdb233551
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:8a6732ac3ca60f13dcb1fd8f186d2fc5dc601b9b5981a58f039af9e237f20044
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:af4721d61dfb61c75488a1bd6a5f62b23e20d7b7586a9cf51ba062ece2837715
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:de0e4b7fb822101198817e83cd7e9f4f7406e75ec91d07a833a27a417cffe66b
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:c8c23c07038f59db848dea0d4247b62840dbe58e1c581dc0722fb9d7f9be3b75
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:f49483df1d342d8e39f9a8085920aad79313a475df02a2292a136ecb784402b5
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel8@sha256:6166f525948a018c41d735014d97fdfa52024f7be31e18976d85d1a0d88901a0
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:d47514cbceb3cc96e9d33b7db6fc1df60893e5c4118bf6086cdd27f04c472773
        - registry.redhat.io/rhel8/mariadb-103@sha256:f0ee0d27bb784e289f7d88cc8ee0e085ca70e88a5d126562105542f259a1ac01
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel8@sha256:d24e277f468b0a48f38801e8e0bcfeb8264c3f7c128073ddaa17934b8572cc3c
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel8@sha256:b24c25dfdcefb4ec26e7f311c00798df108f7685def2b1762b7624fdfb160ebf
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:6275a19ef06114b33d79833a94ef10cc8203368eee98e5d415e73d20cae1db85
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel8@sha256:87685c5cab105d876ff0d14b0e24d8dc2a1ff5b55bebe36028c5f3376b2a5271
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel8@sha256:bbef4839f24facce111d21aca43898dcf4e6cf57adb16200e3f1590bd5f89eae
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:b20ee16b776871d7c49b23f2bc91726812998561129f5dbd14eddbfe57964c93
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:b6039c3c33a5dc345ceb8e757437bb13023d2bf85cc2b4c10b9ab3887e78e0c0
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel8@sha256:bf6231bfdbe52b597d86e03e4730ec105ebe7a639fcd529268b7c0423f2a890c
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:53ac36baa374159b9065c718a9ede821bbb61d9ebe9502b2243e0a9f7aca0d16
        - registry.redhat.io/rhoai/odh-training-operator-rhel8@sha256:4313cae03c041d461bc276abc23d706b582b5d16feab316796ca43478b9dc705
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel8@sha256:71cd15f0b1acba3a48ebedff037bf8a846b426c18363a4de8896a376b6b7b3d5
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        version: 2.19.3
      entries:
      - name: rhods-operator.2.19.3
        version: 2.19.3
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.19
    - currentCSV: rhods-operator.2.22.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Removed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed",
                      "registriesNamespace": "rhoai-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
          createdAt: "2025-09-25T11:28:22Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          feast-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Managed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "feastoperator": {
                    "managementState": "Removed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "OpenshiftDefaultIngress"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Managed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "modelregistry": {
                    "managementState": "Removed",
                    "registriesNamespace": "rhoai-model-registries"
                  },
                  "ray": {
                    "managementState": "Managed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  },
                  "trainingoperator": {
                    "managementState": "Managed"
                  },
                  "trustyai": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: |-
            ["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",  "workbenches.components.platform.opendatahub.io",
            "monitorings.services.platform.opendatahub.io","modelcontrollers.components.platform.opendatahub.io", "feastoperators.components.platform.opendatahub.io",
            "llamastackoperators.components.platform.opendatahub.io"]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          ose-cli-etcd-image: registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-etcd-image: registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
          ose-oauth-proxy-dw-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
          ubi9-image: registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.

          ### Components
          * Dashboard
          * Curated Workbench Images (including CUDA, PyTorch, TensorFlow, code-server, TrustyAI)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines (including Elyra notebook interface)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue, Training Operator)
          * XAI explanations of predictive models (TrustyAI)
          * Index and manage models, versions, and artifacts metadata (Model Registry)
          * Feast - Feature Store is the fastest path to manage existing infrastructure to productionize analytic data for model training and online inference.
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - ODH
        - opendatahub
        - Red Hat OpenShift AI
        - RHOAI
        - OAI
        - ML
        - Machine Learning
        - Data Science
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        - trustyai
        - modelregistry
        - RHOAI
        - ODH
        - OAI
        - AI
        - ML
        - Machine Learning
        - Data Science
        - Feast
        - featurestore
        - llamastack
        links:
        - name: Red Hat OpenShift AI
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2.22
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat Openshift AI
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel9@sha256:b2c2ea9d35edbcd16e1ae559c88a453129b877fc049ab7a7cd6c2ebb8f0bcb59
        - registry.redhat.io/rhoai/odh-rhel9-operator@sha256:9292ff588ed93a002187b070b644c16de2696c0490c1c2c4d29a56e217963436
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-v2-rhel9@sha256:74d4e0174655cf6f6273857325a1e714a1531c5a06ad38bf139bee9c7d8f5f19
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel9@sha256:bd6d1599b4ff9eb890be9808014715b28ca19247222d65f40df024db4dd3525a
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-argoexec-rhel9@sha256:2f3f305c97e50057781553416fb4026cde306ead0fbe0213627b9b5bbcc8a0c3
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel9@sha256:cb5be3dd7098f4c8d66f9e85ad04656460df6102652deddc6cccb884a12be733
        - registry.redhat.io/rhelai1/instructlab-nvidia-rhel9@sha256:3e6eb035c69b204746a44b3a58b2751c20050cfb6af2ba7989ba327809f87c0b
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel9@sha256:6234e15f001122ab222481321153fb0eba7f4d86bf3dfd64edf3f6cecbe7d62f
        - registry.redhat.io/openshift4/ose-cli@sha256:4cfb4219f46c8cc25a5e567fd4cb8babe9a3778b0b86a1e354a3403994ef3677
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel9@sha256:d59ed40de694bc0e0564bab159a5b586cffa0721b6d761fb2d81dd42e6225d1c
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhoai/odh-model-registry-rhel9@sha256:5d79f0bcca0a213b53aee6d88834e503d343821596b857b13d3a6362de8ebeab
        - registry.redhat.io/rhoai/odh-training-operator-rhel9@sha256:b2ca229bb4c4dbf04b96766b8af8d9f55f2edef59ec4110de0d20f8c857e7d27
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel9@sha256:f9965177de24f3165ec4c001c14cf1e2c2c01ffe847839af837c442e3c96beae
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-v2-rhel9@sha256:32c9511846077ef6035eb19be7653f6a78e0f207eb420ab93b18f742ab5779b6
        - registry.redhat.io/openshift4/ose-etcd@sha256:d3275cd886d13865937d225d8138db7f6b7bf59ac1a94d9fbe61e35286bee6ff
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-dashboard-rhel9@sha256:2f1a0a0803b2f5118b167e9d9c2e86660add5f79ee6f47044f63b0e8abe0b75f
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:ab57c24b4ec39c376bce0e718deed8f0e629be231acfe3ba8f430215672e0efd
        - registry.redhat.io/ubi9@sha256:770cf07083e1c85ae69c25181a205b7cdef63c11b794c89b3b487d4670b4c328
        - registry.redhat.io/rhoai/odh-modelmesh-rhel9@sha256:1709fa3c79aad4ba7eb9be8299949396092c8e20210124e0c0936385bc04e839
        - registry.redhat.io/rhoai/odh-model-controller-rhel9@sha256:e6fbe6723b4801f6854303b3fd7d4a4a36727bdd28dc28002b7de4929739ccc0
        - registry.redhat.io/rhoai/odh-ml-pipelines-driver-rhel9@sha256:ccba67a6a5aa3403e0018f29dcdfef483220dea8065d8b572e80c653ca992add
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bd49cfc8452b3d96467cc222db9487e120abc6cc5ba81349c6b3703706f36a08
        - registry.redhat.io/rhoai/odh-feature-server-rhel9@sha256:7fd540586a1ca111370100a473923b3dc6cec4255e9e4f3654bfbabd9158a2bb
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:e87a96c8a61042e83b8db0194adf3fd84c3a311b1f3793345a35bfd0e7ea56d9
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel9@sha256:5914f71e37c815ce2567826896e051303bdef52fae1acf8d02e672fc24c7eb9d
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel9@sha256:2fe9473cd1d6c445d2ca7cc357ee4bb009aa1838f9cb69ea418296b2a044816b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        - registry.redhat.io/rhoai/odh-ml-pipelines-launcher-rhel9@sha256:a44d32a18b1e906c1411cafc82102881573fcd04e040a4e2ba6bd1bd617b70fa
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e3b90530c6e38f6940d7f9408e9fa2c81eb06a857ec2b0bab5d006b1e9a63a4f
        - registry.redhat.io/rhoai/odh-data-science-pipelines-argo-workflowcontroller-rhel9@sha256:2065b96bd6f0a830eab66e36217a927448cfc5a7dabf2a5e7d907aa6c40657a4
        - registry.redhat.io/rhoai/odh-mlmd-grpc-server-rhel9@sha256:463e9be9c4409ba1e5a42713bd65844a023e7a3e9032d4e862edbaa5fe35300a
        - registry.redhat.io/rhoai/odh-ml-pipelines-runtime-generic-rhel9@sha256:2b3aace0d80474470e71c2c8291b25738e096109229fe6d5d3e06e9885ee58b6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel9@sha256:27415b1c763cf89b79391f0b039936c3e078a078bda4420f13bf10ef9be31d5e
        - registry.redhat.io/ubi9/toolbox@sha256:ea4ce8cb8f59cf1b0709f1606d7cab074c3806d4016964eedc0555410cb5dfeb
        - registry.redhat.io/rhoai/odh-feast-operator-rhel9@sha256:d7c4ac937add511460a36375d9b6b3c475a3a9a7cc96c00df2fbad6219aee0e9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1ea6a01bf3e63cdcf125c6064cbd4a4a270deaf0f157b3eabb78f60556840366
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-v2-rhel9@sha256:9104cbc2d595933f0a2acb7313e41ed69905e76a9fae6bd9820e7717b56a632b
        - registry.redhat.io/openshift4/ose-cli@sha256:bc35a9fc663baf0d6493cc57e89e77a240a36c43cf38fb78d8e61d3b87cf5cc5
        - registry.redhat.io/rhel9/mariadb-105@sha256:5e77b33ce04c1eafc2e8840586ad6a5b51148dea8a46dda151071e28858331f0
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel9@sha256:abe1f9dd11c7313fdd55b678be511a5cdc3546654a68e2b53d42c1013fec8d00
        - registry.redhat.io/rhoai/odh-model-registry-operator-rhel9@sha256:13f3679ab289eea9de54fe129a7c83f6c8ec5d26b61b430637a7c4cfcd31a74a
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel9@sha256:7b3d7787ba95156de9c284ad78f2d63a0d3bcb525f4d8d11ebf0203d4c5c1fe0
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        version: 2.22.2
      entries:
      - name: rhods-operator.2.22.2
        version: 2.22.2
      - name: rhods-operator.2.22.1
        version: 2.22.1
      - name: rhods-operator.2.22.0
        version: 2.22.0
      - name: rhods-operator.2.19.2
        version: 2.19.2
      - name: rhods-operator.2.19.1
        version: 2.19.1
      - name: rhods-operator.2.19.0
        version: 2.19.0
      - name: rhods-operator.2.16.2
        version: 2.16.2
      - name: rhods-operator.2.16.1
        version: 2.16.1
      - name: rhods-operator.2.16.0
        version: 2.16.0
      - name: rhods-operator.2.13.1
        version: 2.13.1
      - name: rhods-operator.2.13.0
        version: 2.13.0
      - name: rhods-operator.2.12.0
        version: 2.12.0
      - name: rhods-operator.2.11.0
        version: 2.11.0
      - name: rhods-operator.2.11.0-0.1727935135.p
        version: 2.11.0+0.1727935135.p
      - name: rhods-operator.2.10.0
        version: 2.10.0
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.22
    - currentCSV: rhods-operator.2.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default-dsci",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "redhat-ods-applications",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "redhat-ods-monitoring"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "rhods-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "rhods-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          anaconda-cronjob-image: registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          containerImage: registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
          createdAt: "2024-08-01T06:23:03Z"
          description: Operator for deployment and management of Red Hat OpenShift
            AI
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          odh-etcd-image: registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
          odh-ml-pipeline-mariadb-image: registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
          odh-ml-pipeline-moveresult-image: registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
          olm.skipRange: '>=2.8.0 <2.8.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "datasciencecluster.opendatahub.io/v1",
              "kind": "DataScienceCluster",
              "metadata": {
                "name": "default-dsc",
                "labels": {
                  "app.kubernetes.io/name": "datasciencecluster",
                  "app.kubernetes.io/instance": "default-dsc",
                  "app.kubernetes.io/part-of": "rhods-operator",
                  "app.kubernetes.io/managed-by": "kustomize",
                  "app.kubernetes.io/created-by": "rhods-operator"
                }
              },
              "spec": {
                "components": {
                  "codeflare": {
                    "managementState": "Removed"
                  },
                  "dashboard": {
                    "managementState": "Managed"
                  },
                  "datasciencepipelines": {
                    "managementState": "Managed"
                  },
                  "kserve": {
                    "managementState": "Managed",
                    "serving": {
                      "ingressGateway": {
                        "certificate": {
                          "type": "SelfSigned"
                        }
                      },
                      "managementState": "Managed",
                      "name": "knative-serving"
                    }
                  },
                  "kueue": {
                    "managementState": "Removed"
                  },
                  "modelmeshserving": {
                    "managementState": "Managed"
                  },
                  "ray": {
                    "managementState": "Removed"
                  },
                  "workbenches": {
                    "managementState": "Managed"
                  }
                }
              }
            }
          operatorframework.io/suggested-namespace: redhat-ods-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift AI"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          ose-cli-image: registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
          ose-oauth-proxy-dashboard-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-dsp-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
          ose-oauth-proxy-image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
          prometheus-alertmanager-image: registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
          prometheus-base-image: registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
          prometheus-config-reloader-image: registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
          prometheus-operator-image: registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
          repository: https://github.com/red-hat-data-services/rhods-operator
          support: Red Hat OpenShift AI
          ubi-minimal-image: registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: |
          Red Hat OpenShift AI is a complete platform for the entire lifecycle of your AI/ML projects.

          When using Red Hat OpenShift AI, your users will find all the tools they would expect from a modern AI/ML platform in an interface that is intuitive, requires no local install, and is backed by the power of your OpenShift cluster.

          Your Data Scientists will feel right at home with quick and simple access to the Notebook interface they are used to. They can leverage the default Notebook Images (Including PyTorch, tensorflow, and CUDA), or add custom ones. Your MLOps engineers will be able to leverage Data Science Pipelines to easily parallelize and/or schedule the required workloads. They can then quickly serve, monitor, and update the created AI/ML models. They can do that by either using the provided out-of-the-box OpenVino Server Model Runtime or by adding their own custom serving runtime instead. These activities are tied together with the concept of Data Science Projects, simplifying both organization and collaboration.

          But beyond the individual features, one of the key aspects of this platform is its flexibility. Not only can you augment it with your own Customer Workbench Image and Custom Model Serving Runtime Images, but you will also have a consistent experience across any infrastructure footprint. Be it in the public cloud, private cloud, on-premises, and even in disconnected clusters. Red Hat OpenShift AI can be installed on any supported OpenShift. It can scale out or in depending on the size of your team and its computing requirements.

          Finally, thanks to the operator-driven deployment and updates, the administrative load of the platform is very light, leaving everyone more time to focus on the work that makes a difference.


          ### Components
          * Dashboard
          * Curated Workbench Images (incl CUDA, PyTorch, Tensorflow, code-server)
          * Ability to add Custom Images
          * Ability to leverage accelerators (such as NVIDIA GPU)
          * Data Science Pipelines. (including Elyra notebook interface, and based on standard OpenShift Pipelines)
          * Model Serving using ModelMesh and Kserve.
          * Ability to use other runtimes for serving
          * Model Monitoring
          * Distributed workloads (KubeRay, CodeFlare, Kueue)
        displayName: Red Hat OpenShift AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        - Red Hat OpenShift AI
        - notebooks
        - serving
        - training
        - kserve
        - distributed-workloads
        links:
        - name: Red Hat OpenShift AI
          url: https://www.redhat.com/en/technologies/cloud-computing/openshift/openshift-ai
        maintainers:
        - email: managed-open-data-hub@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhoai/odh-modelmesh-runtime-adapter-rhel8@sha256:850b78a201effcafe33705dcd2670f1c470d8b46ba6e307032403627b55a7cfe
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:bd543e039c7c2b0857a7e0ce34d582b8954a8ad3a3f13a2eaaff2904cdae7c53
        - registry.redhat.io/rhoai/odh-kf-notebook-controller-rhel8@sha256:080377d42d782bf525b4072384d0cf8409b48019b3748ca7c855f340bcb94251
        - registry.redhat.io/rhoai/odh-ml-pipelines-scheduledworkflow-rhel8@sha256:5f39b11f0cd6db2d7d6e51ee666c030e2643aee92a090b49656c1e7dfd8a5430
        - registry.redhat.io/rhoai/odh-mm-rest-proxy-rhel8@sha256:b6ecb3858c201498024e3bfc4a93e630699be4b1fdc3a65fd2d70e4b375eb9c2
        - registry.redhat.io/rhoai/odh-notebook-controller-rhel8@sha256:ab8792fcd16b1aa0065444ba54335cb3e747379740c3ec2d56beb46ff80cb044
        - registry.redhat.io/rhoai/odh-ml-pipelines-artifact-manager-rhel8@sha256:889f046b255b13d99801108588d3abf577611b828657b5240a18ea21f0e3cbbe
        - registry.redhat.io/rhoai/odh-trustyai-service-operator-rhel8@sha256:987c768c0688431a53e48a80591f2a41e074994169f9e691b28b89528d8e736f
        - registry.redhat.io/rhoai/odh-modelmesh-rhel8@sha256:0b8665e8978deb73be2cc65830f08a74a5a48f3e3ee7b2b817f7bbacba257d22
        - registry.redhat.io/rhoai/odh-operator-bundle@sha256:cb711fcb32b5ca0f34c4e3c67624bc4e035de4247617d988a9e841fd58c1c911
        - registry.redhat.io/rhoai/odh-kueue-controller-rhel8@sha256:91525fac5fadece75fbdafc40b94445fb92c88cfc8c2f06b596f8c6377e59c9e
        - registry.redhat.io/rhoai/odh-ml-pipelines-api-server-rhel8@sha256:90833e85e0de3d7de870074c8b04319cd11b913b9652af6a5727aae41a91458c
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager@sha256:169b788b3f8eb8909ecc1d698ab0a6bf103b49043c7f5fd569789b179928def6
        - registry.redhat.io/ubi8/ubi-micro@sha256:396baed3d689157d96aa7d8988fdfea7eb36684c8335eb391cf1952573e689c1
        - registry.redhat.io/rhoai/odh-trustyai-service-rhel8@sha256:eb4da3a645b53acfd6ce8758c043ae19b4126e62bcca09165114a141c2c8d014
        - registry.redhat.io/openshift4/ose-cli@sha256:75bf9b911b6481dcf29f7942240d1555adaa607eec7fc61bedb7f624f87c36d4
        - registry.redhat.io/rhoai/odh-kuberay-operator-controller-rhel8@sha256:06a6d5cbbe80d49bca9e4db9929571a69e35ae7951fcae84dcd6460f7ddf622c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:5d2d4d4dbec470f8ffb679915e2a8ae25ad754cd9193fa966deee1ecb7b3ee00
        - registry.redhat.io/rhoai/odh-modelmesh-serving-controller-rhel8@sha256:68e1331f13dfaf6641a1e7d54f4cf29c0ef42663abc98b6dc69167892d2d8adb
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4bef31eb993feb6f1096b51b4876c65a6fb1f4401fee97fa4f4542b6b7c9bc46
        - registry.redhat.io/rhoai/odh-dashboard-rhel8@sha256:9799003449bfb61ae8f59efd2a611b0da3bccadd9b20cc1519f6c9279d085337
        - registry.redhat.io/rhoai/odh-codeflare-operator-rhel8@sha256:1a6e29214e06a2665baadbed598ba758d6141bfa522984eb05f05a80558a403c
        - registry.redhat.io/rhoai/odh-rhel8-operator@sha256:b3563e201bf2720f60e73c0b04cb70566ea52eda7d09a28753f3a1487993c2e8
        - registry.redhat.io/openshift4/ose-prometheus-operator@sha256:4aeb07f66dc39889e3d0e66eee7c4ea2e39ce4025852d1d1eaf662af8bd8eeb6
        - registry.redhat.io/rhel7/etcd@sha256:d3495b263b103681f1b09a558be43c21989bfc269eb90f84c2609042cebdc626
        - registry.redhat.io/openshift4/ose-prometheus@sha256:c432e571eb737d55323b5f350b5f714803bf525144cfee3f20782660d9bf10ad
        - registry.redhat.io/rhoai/odh-model-controller-rhel8@sha256:df2edcaf4402b11ce24439263956e3b0cc43efbac2fdf152afacb1a398dcc59a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ab112105ac37352a2a4916a39d6736f5db6ab4c29bad4467de8d613e80e9bb33
        - registry.redhat.io/rhoai/odh-data-science-pipelines-operator-controller-rhel8@sha256:1c49888ba2e9449c04d5588040c099290fce4793c2ba8874c1037316e550d6e1
        - registry.redhat.io/rhoai/odh-ml-pipelines-cache-rhel8@sha256:92d9e4b53d580df6cc1f14a849b477749a1306941a2e8f9d1b6160f4911c3cff
        - registry.redhat.io/rhel8/mariadb-103@sha256:3d30992e60774f887c4e7959c81b0c41b0d82d042250b3b56f05ab67fd4cdee1
        - registry.redhat.io/rhoai/odh-ml-pipelines-persistenceagent-rhel8@sha256:9c90873fbd5af8fac8efc3c7c27fba742a06fad2d47f336319dd05d2b116867b
        - registry.redhat.io/openshift4/ose-cli@sha256:25fef269ac6e7491cb8340119a9b473acbeb53bc6970ad029fdaae59c3d0ca61
        version: 2.8.4
      entries:
      - name: rhods-operator.2.8.4
        version: 2.8.4
      - name: rhods-operator.2.8.3
        version: 2.8.3
      - name: rhods-operator.2.8.2
        version: 2.8.2
      - name: rhods-operator.2.8.1
        version: 2.8.1
      - name: rhods-operator.2.8.0
        version: 2.8.0
      - name: rhods-operator.2.7.0
        version: 2.7.0
      - name: rhods-operator.2.6.0
        version: 2.6.0
      - name: rhods-operator.2.5.0
        version: 2.5.0
      - name: rhods-operator.2.4.0
        version: 2.4.0
      - name: rhods-operator.1.33.0
        version: 1.33.0
      - name: rhods-operator.1.32.0
        version: 1.32.0
      - name: rhods-operator.1.31.0
        version: 1.31.0
      - name: rhods-operator.1.30.0
        version: 1.30.0
      - name: rhods-operator.1.29.0
        version: 1.29.0
      - name: rhods-operator.1.28.1
        version: 1.28.1
      - name: rhods-operator.1.28.0
        version: 1.28.0
      - name: rhods-operator.1.27.0
        version: 1.27.0
      - name: rhods-operator.1.26.0
        version: 1.26.0
      - name: rhods-operator.1.25.0
        version: 1.25.0
      - name: rhods-operator.1.24.0
        version: 1.24.0
      - name: rhods-operator.1.23.0
        version: 1.23.0
      - name: rhods-operator.1.22.1-4
        version: 1.22.1-4
      - name: rhods-operator.1.22.1-2
        version: 1.22.1-2
      - name: rhods-operator.1.22.0-2
        version: 1.22.0-2
      - name: rhods-operator.1.21.0-22
        version: 1.21.0-22
      - name: rhods-operator.1.20.1-8
        version: 1.20.1-8
      name: stable-2.8
    defaultChannel: stable
    packageName: rhods-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubemq-operator-marketplace-rhmp/support?utm_source=openshift_console
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-uni:v2.2.12
        - registry.redhat.io/openshift4/ose-grafana:v4.3
        - registry.redhat.io/openshift4/ose-prometheus:v4.3
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:26af4e950a33755114298386a70229672a09ce45b6203025856f460a55c4c8a0
        - registry.marketplace.redhat.com/rhm/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace-rhmp
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AMD
      provider-url: ""
    name: amd-t2-january
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-t2-january.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfect2.amd.com/v1",
                "kind": "SriovT2Card",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "amd-t2-january",
                    "app.kubernetes.io/instance": "sriovt2card-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sriovt2card",
                    "app.kubernetes.io/part-of": "amd-t2-january"
                  },
                  "name": "sriovt2card",
                  "namespace": "amd-t2"
                },
                "spec": {
                  "acceleratorSelector": {
                    "pciAddress": "0000:01:00.0"
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "7c-c2-55-85-ba-fe.amd.com"
                  },
                  "physicalFunction": {
                    "pfDriver": "qdma-pf",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-07-14T09:46:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovT2Card is the Schema for the sriovt2cards API
            displayName: Sriov T2 Card
            kind: SriovT2Card
            name: sriovt2cards.sriovfect2.amd.com
            version: v1
        description: The AMD-T2 Operator for accelerator enables the seamless integration
          and management of AMD T2 accelerator within Openshift/Kubernetes envirnoment.
        displayName: AMD-T2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - SRIOV
        - Xilinx
        - FPGA
        - Telco
        - T2
        - MultiArch
        - AMD
        - TelcoAccelerator
        links:
        - name: Amd T2 January
          url: https://amd-t2-january.domain
        maintainers:
        - email: dhirajs.lakhane@amd.com
          name: DhirajLakhane
        maturity: alpha
        provider:
          name: AMD
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:928e64203edad8f1bba23593c7be04f0f8410c6e4feb98d9e9c2d00a8ff59048
        - quay.io/amdt2operator/amd-t2-january@sha256:318f56e2670582c217b8973f69f0675e7fff7dc19d07875294f2ff2fc5403902
        - registry.connect.redhat.com/amd/amd-t2-january-bundle@sha256:7c87354aa07e783a2cbf4ab29877a8ba56ea4269bc1f41bfbe08c558b3666841
        version: 1.0.1
      entries:
      - name: amd-t2-january.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-t2-january
    provider:
      name: AMD
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lenovo (Shanghai) Computer Technology Co., Ltd.
      provider-url: ""
    name: lenovo-locd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lenovo-locd-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "locd.lenovo.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample",
                  "namespace": "locd-operator-system"
                },
                "spec": {
                  "affinity": {},
                  "argocd": {
                    "password": "xxxxxxxxxxx",
                    "server": "argocd.svc.cluster.local",
                    "user": "admin"
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "core": {
                    "image": "registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609",
                    "pullPolicy": "Always"
                  },
                  "fullnameOverride": "",
                  "global": {
                    "managementNamespace": "loc-devops",
                    "serviceEndpoint": "loc-devops.domain",
                    "serviceaccount": {
                      "annotations": {
                        "managedBy": "LOC DevOps"
                      },
                      "create": true,
                      "name": "loc-devops-admin"
                    }
                  },
                  "gui": {
                    "image": "registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44",
                    "pullPolicy": "Always"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "nginx",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "loc-devops.domain",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "Prefix",
                            "servicePort": 8080
                          },
                          {
                            "path": "/api",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          },
                          {
                            "path": "/swagger",
                            "pathType": "Prefix",
                            "servicePort": 8000
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": {
                      "core": 8000,
                      "gui": 8080
                    },
                    "type": "ClusterIP"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Cluster
            name: clusters.locd.lenovo.com
            version: v1alpha1
        description: This is operator used to manage components of Lenovo Open Cloud
          DevOps
        displayName: lenovo-locd-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Application management
        links:
        - name: LOCD Operator
          url: https://locd-operator.domain
        maintainers:
        - email: xujt5@lenovo.com
          name: jintaoxu
        maturity: alpha
        provider:
          name: Lenovo (Shanghai) Computer Technology Co., Ltd.
        relatedImages:
        - registry.connect.redhat.com/loc-devops/gui@sha256:65142a36095188e879467038936b7b6b6ee39dcea6d42a437fa7066fb2bc0f44
        - registry.connect.redhat.com/loc-devops/operator-bundle@sha256:b9f7a678ceec2d367d55499a48a15a9dd729bc34b0e707989be23c97fc3c5f4e
        - registry.connect.redhat.com/loc-devops/operator@sha256:4c17a3688ae3098e4b5c2c0530f49b1da8dedcc6fc1d3cd848513a1f189812a7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.connect.redhat.com/loc-devops/core@sha256:f771217b3f7849e1b63cc6abb93b8b0fc9636fd4a201e31807a15f773138c609
        version: 0.1.0
      entries:
      - name: lenovo-locd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: lenovo-locd-operator
    provider:
      name: Lenovo (Shanghai) Computer Technology Co., Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: edb-hcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edb-hcp-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edbpgai.edb.com/v1alpha1",
                "kind": "HybridControlPlane",
                "metadata": {
                  "name": "placeholder"
                },
                "spec": {
                  "beaconAgent": {
                    "applianceProviderEnabled": true,
                    "beaconDBConnected": true,
                    "beaconServerAddress": "upm-beacon-server-ingestion.upm-beacon.svc.cluster.local:8080",
                    "inCluster": true,
                    "location": "devspatcher",
                    "logLevel": "debug",
                    "plaintext": true,
                    "providers": [
                      "appliance"
                    ],
                    "provisioning": {
                      "imageDiscovery": true,
                      "imagesetDiscoveryAuthenticationType": "token",
                      "imagesetDiscoveryContainerRegistryURL": "placeholder",
                      "loadBalancersEnabled": true,
                      "nodePortDomain": "nodeportdomain.com",
                      "openshift": true,
                      "provider": "aws"
                    },
                    "tlsInsecure": false,
                    "transparentDataEncryptionMethods": [
                      "passphrase",
                      "hashicorp_vault"
                    ]
                  },
                  "beaconServer": {
                    "logLevel": "debug"
                  },
                  "componentsParameters": {
                    "edb-migration-portal": {
                      "enable_pdb": "false"
                    },
                    "transporter-data-operator": {
                      "fips_enabled": "false"
                    },
                    "transporter-dp-agent": {
                      "rw_service_url": "placeholder"
                    },
                    "transporter-rw-service": {
                      "domain_name": "placeholder"
                    },
                    "upm-api-pnxt-metric": {
                      "log_level": "info",
                      "tsdb_host": "thanos-query.monitoring.svc.cluster.local",
                      "tsdb_port": "9090"
                    },
                    "upm-api-pnxt-pd": {
                      "log_level": "info"
                    },
                    "upm-beaco-ff-base": {
                      "aes_key_rotation_interval": "2160h",
                      "usage_generator_interval": "24h"
                    },
                    "upm-beaco-usages": {
                      "usage_generator_interval": "24h"
                    },
                    "upm-beacon": {
                      "beacon_db_hibernation": "\"off\"",
                      "beacon_location_id": "devspatcher",
                      "beacon_server_replicas": "1",
                      "external_metrics_replicas": "1",
                      "image_discovery_credentials_secret_name": "placeholder",
                      "server_host": "placeholder"
                    },
                    "upm-istio": {
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288"
                    },
                    "upm-istio-gateway": {
                      "cookie_aeskey": "rzkutHl8NJNztPMEJYykZouHslNiA7xmIXH+58ISUVo=",
                      "enable_server_session": "true",
                      "ingress_grpc_tls_node_port": "30290",
                      "ingress_http_node_port": "32542",
                      "ingress_https_node_port": "30288",
                      "openshift_console_domain_name": "placeholder"
                    },
                    "upm-prometheus": {
                      "cpu": "200m",
                      "memory": "6Gi",
                      "prometheus_retention_size": "15GB",
                      "prometheus_storage_size": "20Gi"
                    },
                    "upm-thanos": {
                      "memcached_cpu": "500m",
                      "memcached_memory": "2Gi",
                      "query_frontend_cpu": "200m",
                      "query_frontend_memory": "128Mi"
                    },
                    "upm-trust-manager": {
                      "cert_manager_namespace": "cert-manager"
                    }
                  },
                  "flavour": "placeholder",
                  "globalParameters": {
                    "default_ingress_domain": "placeholder",
                    "internal_backup_folder": "a7462dbc7106",
                    "portal_certificate_secret": "beaco-istio-gw-cert",
                    "portal_domain_name": "placeholder",
                    "portal_port": "443",
                    "session_duration_seconds": "86400",
                    "storage_class": "placeholder"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "placeholder",
                      "namespace": "default"
                    }
                  ],
                  "imageRegistry": "placeholder",
                  "pgai": {
                    "portal": {
                      "authentication": {
                        "clientSecret": "edb-postgres-ai-secret",
                        "idpConnectors": [],
                        "staticPasswords": [
                          {
                            "email": "owner@mycompany.com",
                            "hash": "$2y$10/JLTqiuGblH.RemIutWxkztN5m",
                            "userID": "c5998173-a605-449a-a9a5-4a9c33e26df7",
                            "username": "Owner MyCompany"
                          },
                          {
                            "email": "admin@mycompany.com",
                            "hash": "$2y$10/yYwtXRrOagYO2jNtwVzGQPPEfDvX0QllV8NXbku",
                            "userID": "b7e2c2e2-2e7a-4c7b-9e2a-7e6b8c1f2a3d",
                            "username": "Admin MyCompany"
                          }
                        ]
                      }
                    }
                  },
                  "source": {
                    "useLocalKustomizations": true
                  },
                  "version": "placeholder"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database,AI/Machine Learning
          containerImage: docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
          createdAt: "2025-09-24T08:25:29Z"
          description: Unify transactional, analytical, and AI workloads and bring
            cloud agility to your hybrid data infrastructure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/EnterpriseDB/edb-hcp-operator
          support: techsupport@enterprisedb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybridControlPlane is the Schema for the hybridcontrolplanes
              API
            displayName: Hybrid Control Plane
            kind: HybridControlPlane
            name: hybridcontrolplanes.edbpgai.edb.com
            version: v1alpha1
        description: |
          EDB Postgres AI is a sovereign data and AI platform that brings analytics and AI closer to core operational data and enables cloud agility in private environments.
          Unify your structured and unstructured data in one multi-model data platform with centralized management and observability to:
          - Accelerate enterprise-grade transactional databases
          - Enable rapid analytics and powerful lakehouse integrations
          - Automate AI pipelines and semantic search for faster GenAI development
          Using EDB’s unmatched Postgres expertise, EDB Postgres AI automates database operations, ensuring expert-level execution across deployments – boosting operational efficiency by up to 30% and accelerating innovation.
        displayName: EDB Postgres AI
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Postgres
        - Database
        - DBaaS
        - EDB Postgres AI
        links:
        - name: EDB Postgres AI Software Deployment
          url: https://www.enterprisedb.com/products/software-deployment
        maintainers:
        - email: techsupport@enterprisedb.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/edb-hcp-operator@sha256:87833c7edfa714b433fb066533bd415049b4ae96588ea8457a85a62326801d5c
        - docker.enterprisedb.com/pgai-platform/edb-hcp-operator/manager@sha256:289cb29b7bafdbb865a61c5f07bfe17fe031652a0810977c6b0620858b049ae9
        version: 1.3.0
      entries:
      - name: edb-hcp-operator.v1.3.0
        version: 1.3.0
      - name: edb-hcp-operator.v1.2.3
        version: 1.2.3
      - name: edb-hcp-operator.v1.2.2
        version: 1.2.2
      - name: edb-hcp-operator.v1.2.1
        version: 1.2.1
      - name: edb-hcp-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: edb-hcp-operator
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: quay.io/deployment-validation-operator/dv-operator:0.7.12
          createdAt: "2025-09-11T00:00:00Z"
          description: The deployment validation operator
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: https://quay.io/deployment-validation-operator/dv-operator:0.7.12
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/deployment-validation-operator@sha256:2803298c7cf3eeaffe9da3153142058d94d65ff1dae90efd8819a4cd23e0518f
        - quay.io/deployment-validation-operator/dv-operator:0.7.12
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.v0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.v0.7.7
        version: 0.7.7
      - name: deployment-validation-operator.v0.7.6
        version: 0.7.6
      - name: deployment-validation-operator.v0.7.5
        version: 0.7.5
      - name: deployment-validation-operator.v0.7.4
        version: 0.7.4
      - name: deployment-validation-operator.v0.7.3
        version: 0.7.3
      - name: deployment-validation-operator.v0.7.2
        version: 0.7.2
      - name: deployment-validation-operator.v0.7.1
        version: 0.7.1
      - name: deployment-validation-operator.v0.7.0
        version: 0.7.0
      - name: deployment-validation-operator.v0.6.0
        version: 0.6.0
      - name: deployment-validation-operator.v0.5.0
        version: 0.5.0
      - name: deployment-validation-operator.v0.4.0
        version: 0.4.0
      - name: deployment-validation-operator.v0.3.0
        version: 0.3.0
      - name: deployment-validation-operator.v0.2.2
        version: 0.2.2
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator:0.0.11
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - docker.io/eginnovations/universal-agent-operator:0.0.11
        - quay.io/community-operator-pipeline-prod/eginnovations-operator@sha256:19b12c75f7a2dfab027881ed9f0b09d25e3baa936d21f7be3ec1e0177ed9fdf2
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      - name: eginnovations-operator.v0.0.10
        version: 0.0.10
      - name: eginnovations-operator.v0.0.9
        version: 0.0.9
      - name: eginnovations-operator.v0.0.8
        version: 0.0.8
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.5.0
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.5.0
        - quay.io/kuadrant/limitador:v1.2.0
        - quay.io/openshift-community-operators/limitador-operator@sha256:7aa0a4fe55e5e099e2e6e4c2cd44b9bfeba69d8cfade10500ccba3092fb60dd2
        version: 0.5.0
      entries:
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: alpha
    - currentCSV: limitador-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/limitador-operator:v0.11.0
          createdAt: "2024-10-02T10:46:31Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: The Limitador operator installs and maintains limitador instances
        displayName: Limitador
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        links:
        - name: Limitador Operator
          url: https://github.com/Kuadrant/limitador-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - quay.io/kuadrant/limitador-operator:v0.11.0
        - quay.io/kuadrant/limitador:v1.6.0
        - quay.io/community-operator-pipeline-prod/limitador-operator@sha256:450f9bf43b57b83e9975a1865981af22e28c725193573354d1180f406887664a
        version: 0.11.0
      entries:
      - name: limitador-operator.v0.11.0
        version: 0.11.0
      - name: limitador-operator.v0.10.0
        version: 0.10.0
      - name: limitador-operator.v0.6.0
        version: 0.6.0
      - name: limitador-operator.v0.5.0
        version: 0.5.0
      - name: limitador-operator.v0.4.0
        version: 0.4.0
      - name: limitador-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/fluxcd/source-controller:v1.3.0
          description: Flux is a Continuous Delivery solution for Kubernetes.
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta2
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta2
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta3
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        provider:
          name: The Flux project
        relatedImages:
        - ghcr.io/fluxcd/notification-controller:v1.3.0
        - ghcr.io/fluxcd/source-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/flux@sha256:cd94037ff3055ab6ce40b0670f900248a5fd233cc67a2a4f92c1559cf9d2f3ad
        - ghcr.io/fluxcd/helm-controller:v1.0.1
        - ghcr.io/fluxcd/image-automation-controller:v0.38.0
        - ghcr.io/fluxcd/image-reflector-controller:v0.32.0
        - ghcr.io/fluxcd/kustomize-controller:v1.3.0
        version: 2.3.0
      entries:
      - name: flux.v2.3.0
        version: 2.3.0
      - name: flux.v2.2.2
        version: 2.2.2
      - name: flux.v2.2.0
        version: 2.2.0
      - name: flux.v2.0.1
        version: 2.0.1
      - name: flux.v0.41.2
        version: 0.41.2
      - name: flux.v0.40.0
        version: 0.40.0
      - name: flux.v0.39.0
        version: 0.39.0
      - name: flux.v0.38.3
        version: 0.38.3
      - name: flux.v0.37.0
        version: 0.37.0
      - name: flux.v0.36.0
        version: 0.36.0
      - name: flux.v0.35.0
        version: 0.35.0
      - name: flux.v0.31.4
        version: 0.31.4
      - name: flux.v0.31.3
        version: 0.31.3
      - name: flux.v0.30.2
        version: 0.30.2
      - name: flux.v0.28.5
        version: 0.28.5
      - name: flux.v0.28.4
        version: 0.28.4
      - name: flux.v0.28.3
        version: 0.28.3
      - name: flux.v0.27.4
        version: 0.27.4
      - name: flux.v0.27.3
        version: 0.27.3
      - name: flux.v0.27.2
        version: 0.27.2
      - name: flux.v0.27.0
        version: 0.27.0
      - name: flux.v0.26.1
        version: 0.26.1
      - name: flux.v0.26.0
        version: 0.26.0
      - name: flux.v0.25.3
        version: 0.25.3
      - name: flux.v0.25.2
        version: 0.25.2
      - name: flux.v0.24.1
        version: 0.24.1
      - name: flux.v0.24.0
        version: 0.24.0
      - name: flux.v0.23.0
        version: 0.23.0
      - name: flux.v0.22.1
        version: 0.22.1
      - name: flux.v0.22.0
        version: 0.22.0
      - name: flux.v0.21.1
        version: 0.21.1
      - name: flux.v0.21.0
        version: 0.21.0
      - name: flux.v0.20.1
        version: 0.20.1
      - name: flux.v0.20.0
        version: 0.20.0
      - name: flux.v0.19.1
        version: 0.19.1
      - name: flux.v0.18.3
        version: 0.18.3
      - name: flux.v0.18.1
        version: 0.18.1
      - name: flux.v0.17.2
        version: 0.17.2
      - name: flux.v0.17.1
        version: 0.17.1
      - name: flux.v0.17.0
        version: 0.17.0
      - name: flux.v0.16.2
        version: 0.16.2
      - name: flux.v0.16.1
        version: 0.16.1
      - name: flux.v0.16.0
        version: 0.16.0
      - name: flux.v0.15.3
        version: 0.15.3
      - name: flux.v0.15.2
        version: 0.15.2
      - name: flux.v0.14.2
        version: 0.14.2
      - name: flux.v0.14.1
        version: 0.14.1
      - name: flux.v0.14.0
        version: 0.14.0
      - name: flux.v0.13.4
        version: 0.13.4
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-framework
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-framework.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Framework",
                "metadata": {
                  "name": "framework-sample"
                },
                "spec": {
                  "kcos-eula": {
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "service": {
                      "nodePort": 0,
                      "type": "ClusterIP"
                    }
                  },
                  "kcos-logging-framework": {
                    "image": {
                      "remoteRepository": "docker.io",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    },
                    "ingress": {
                      "annotations": {
                        "nginx.ingress.kubernetes.io/auth-method": "POST",
                        "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                        "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate",
                        "nginx.ingress.kubernetes.io/cors-allow-credentials": "true",
                        "nginx.ingress.kubernetes.io/cors-allow-methods": "PUT, GET, POST, OPTIONS, PATCH",
                        "nginx.ingress.kubernetes.io/cors-allow-origin": "https://admin.example.com,https://localhost:4200",
                        "nginx.ingress.kubernetes.io/enable-cors": "true",
                        "nginx.ingress.kubernetes.io/proxy-body-size": "0",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                      },
                      "enabled": true
                    }
                  },
                  "kcos-postgresql": {
                    "jobCopySecret": {
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "postgresql": {
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "initdbScripts": {
                        "init_script.sh": "#!/bin/sh\nexport PGUSER=postgres\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'db'\" | grep -q 1 || psql -c \"CREATE DATABASE db\"\npsql -c \"SELECT 1 FROM pg_database WHERE datname = 'kcoslicensingdb'\" | grep -q 1 || psql -c \"CREATE DATABASE kcoslicensingdb\"\n"
                      },
                      "metrics": {
                        "image": {
                          "registry": "docker.io"
                        }
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                      "securityContext": {
                        "enabled": false
                      },
                      "shmVolume": {
                        "chmod": {
                          "enabled": false
                        }
                      },
                      "volumePermissions": {
                        "image": {
                          "registry": "docker.io"
                        },
                        "securityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    }
                  },
                  "kcos-system-diagnostics": {
                    "environment": {
                      "ELASTICSEARCH_ENDPOINT": "elasticsearch.keysight-wap.svc.cluster.local.",
                      "IGNORE_DATA_PREFIX": "wireless_stats"
                    },
                    "image": {
                      "initContainer": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                      "repository": "prod.harbor.keysight.digital/loadcore"
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-coredns-fixup": false,
                      "kcos-eula": true,
                      "kcos-local-storage": false,
                      "kcos-logging-framework": true,
                      "kcos-logging-rotate": false,
                      "kcos-naas": false,
                      "kcos-postgresql": true,
                      "kcos-shell": false,
                      "kcos-system-diagnostics": true,
                      "kcos-vital-service": false,
                      "nfs-server-provisioner": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:27:28Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Framework
            name: frameworks.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Framework Operator
        displayName: Keysight KCOS Framework Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Framework
          url: https://keysight-kcos-framework.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-framework-bundle@sha256:53d056ac37177263d79098cda775deff86a327c65106ea5a40109b08a856cf84
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-framework@sha256:531f278def85e33d453479fce74127e12e3ed62a251872eab421c6561bedd0c5
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        version: 0.1.21
      entries:
      - name: keysight-kcos-framework.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-framework
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: secret-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
          description: Stackable Secret Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/secret-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret classes managed by the operator.
            displayName: Secret classes managed by the operator.
            kind: SecretClass
            name: secretclasses.secrets.stackable.tech
            version: v1alpha1
          - description: Trust stores managed by the operator.
            displayName: Trust stores managed by the operator.
            kind: TrustStore
            name: truststores.secrets.stackable.tech
            version: v1alpha1
        description: |-
          This is a Kubernetes operator to provide secret functionality to other Stackable operators. The Stackable Secret Operator
          is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or Spark, all
          working together seamlessly. Based on Kubernetes, it runs everywhere - on prem or in the cloud.

          You can install the operator using [stackablectl or helm](https://docs.stackable.tech/home/stable/secret-operator/installation.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secret
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/secret-operator@sha256:ac78409c6d7a2510f7381a6cd75e4538a31ec9e99612190e9f123978bce74b5a
        - quay.io/stackable/secret-operator@sha256:c4a10ffad1f336b7287758b57d01cb3a07572c33d389573776187181738663cd
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: secret-operator.v25.7.0
        version: 25.7.0
      - name: secret-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-secret-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: New England Research Cloud
      provider-url: https://nerc.mghpcc.org/
    name: keycloak-permissions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-permissions-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.nerc.mghpcc.org/v1",
                "kind": "KeycloakAuthorization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keycloak-permissions-operator",
                    "app.kubernetes.io/instance": "keycloakauthorization-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "keycloakauthorization",
                    "app.kubernetes.io/part-of": "keycloak-permissions-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "cleanupUndefinedAuthorizations": false,
                  "description": "",
                  "keycloak": {
                    "adminSecret": {
                      "name": "keycloak-initial-admin",
                      "passwordKey": "password",
                      "usernameKey": "username"
                    }
                  },
                  "managementState": "Managed",
                  "realm": null,
                  "validateCerts": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          com.redhat.openshift.versions: v4.11
          containerImage: quay.io/nerc-images/keycloak-permissions-operator:1.4.1
          createdAt: "2024-05-07T21:09:02Z"
          description: Managing Keycloak resources, scopes, policies, and permissions
            for fine-grained resource permissions
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/nerc-images/keycloak-permissions-operator
          support: New England Research Cloud
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Keycloak resources, scopes, policies, and permissions
              for fine-grained resource permissions.
            displayName: Keycloak Authorization
            kind: KeycloakAuthorization
            name: keycloakauthorizations.keycloak.nerc.mghpcc.org
            version: v1
        description: "An OpenShift Operator for managing Keycloak resources, scopes,
          policies,\nand permissions for fine-grained resource permissions.\nAccording
          to the [Keycloak Authorization Services documentation](https://www.keycloak.org/docs/23.0.7/authorization_services/):\n\n-
          A resource can be a web page, a RESTFul resource, a file in your file system,
          or other entity you wish to authorize access to.\n- Scopes usually represent
          the actions that can be performed on a resource, but they are not limited
          to that. You can also use scopes to represent one or more attributes within
          a resource.\n- Policies define the conditions that must be satisfied to
          access or perform operations on something (resource or scope), but they
          are not tied to what they are protecting. They are generic and can be reused
          to build permissions or even more complex policies.\n- Permissions are coupled
          with the resource they are protecting. Here you specify what you want to
          protect (resource or scope) and the policies that must be satisfied to grant
          or deny permission.\n\n## Getting started with the Keycloak Permissions
          Operator\n\nStart by defining the name of your KeycloakAuthorization and
          the same namespace where Keycloak is deployed.\n\n```yaml\napiVersion: keycloak.nerc.mghpcc.org/v1\nkind:
          KeycloakAuthorization\nmetadata:\n  name: example\n  namespace: keycloak\n```\n\n##
          KeycloakAuthorization spec\n\nNext, define the spec. \n\n- `validateCerts:
          false` If your Keycloak does not have valid certificates, if you are deploying
          to OpenShift Local for example. You don't need this line if Keycloak has
          valid certificates.\n- `cleanupUndefinedAuthorizations: true` if you would
          like to remove any undefined authorizations that may have been added manually.
          This is useful for keeping in sync with GitOps. If you do not want the operator
          removing any resources, policies, or permissions, then remove this line.\n\n```yaml\nspec:\n
          \ validateCerts: false\n  cleanupUndefinedAuthorizations: true\n```\n\n##
          Keycloak admin credentials\n\nSpecify the details to connect to keycloak,
          including the `baseUrl`, and the Secret name containing the Keycloak admin
          usernameKey and passwordKey.\n\n```yaml\n  keycloak:\n    baseUrl: https://keycloak-service:8443\n
          \   adminSecret:\n      name: keycloak-initial-admin\n      usernameKey:
          username\n      passwordKey: password\n```\n\n## Realm\n\nSpecify the realm
          id you wish to add authorization configuration to. If you want to add authorization
          configuration to multiple realms, then create a KeycloakAuthorization resource
          for each realm.\n\n```yaml\n  realm:\n    id: NERC\n```\n\n## Clients\n\nFor
          each client in the realm you wish to configure, add authorizationSettings.\n\n```yaml\n
          \   clients:\n      - id: ai4cloudops\n        authorizationSettings:\n```\n\n##
          Authorization Scopes\n\nList the scopes you wish to define for the client.\n\n```yaml\n
          \         scopes:\n            - id: prod\n              name: prod\n              displayName:
          Production\n            - id: test\n              name: test\n              displayName:
          Test\n```\n\n## Authorization Resources\n\nList the resources and corresponding
          scopes you wish to define for the client.\n\n```yaml\n          resources:\n
          \           - name: cluster\n              displayName: cluster\n              scopes:\n
          \               - id: prod\n                  name: prod\n                  displayName:
          Production\n                - id: test\n                  name: test\n                  displayName:
          Test\n```\n\n## Authorization Policies\n\nDefine the policies you wish to
          use in your permissions. You can choose from all the different types of
          permissions, including: aggregated, client, client-scope, group, regex,
          role, time, and user.\n\n```yaml\n          policies:\n            - id:
          client-example\n              name: client-example\n              logic:
          POSITIVE\n              decisionStrategy: UNANIMOUS\n              type:\n
          \               client:\n                  clients:\n                    -
          example\n            - id: group-example\n              name: group-example\n
          \             logic: POSITIVE\n              decisionStrategy: UNANIMOUS\n
          \             type:\n                group:\n                  groups:\n
          \                   - id: example\n```\n\n## Authorization Permissions\n\nDefine
          the permissions, connecting the resources, scopes, and policies together.\n\n```yaml\n
          \         permissions:\n            - name: group-example-cluster-test\n
          \             client: test\n              policy: group-example\n              resource:
          cluster\n            - name: client-example-cluster-test\n              client:
          example\n              policy: client-test\n              resource: cluster\n
          \             scope: test\n```\n\n## Other details\n\nResources, scopes,
          policies, and permissions will all be applied to Keycloak if you have configured
          the admin credentials to Keycloak successfully.\nWhen the KeycloakAuthorization
          is deleted, then the configured resources, policies, and permissions will
          be deleted as well. "
        displayName: Keycloak Permissions Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - fine-grained resource permissions
        - keycloak
        - resources
        - scopes
        - policies
        - permissions
        links:
        - name: Keycloak Permissions Operator
          url: https://keycloak-permissions-operator.domain
        maintainers:
        - email: computate@computate.org
          name: Christopher Tate
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: New England Research Cloud
          url: https://nerc.mghpcc.org/
        relatedImages:
        - quay.io/nerc-images/keycloak-permissions-operator:1.4.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/keycloak-permissions-operator@sha256:4de19e52e84cdba0963bdd852a63912f097280800d65add8d474c0953da478a0
        version: 1.4.1
      entries:
      - name: keycloak-permissions-operator.v1.4.1
        version: 1.4.1
      - name: keycloak-permissions-operator.v1.3.0
        version: 1.3.0
      - name: keycloak-permissions-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: keycloak-permissions-operator
    provider:
      name: New England Research Cloud
      url: https://nerc.mghpcc.org/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Airlock
      provider-url: ""
    name: airlock-microgateway
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airlock-microgateway.v4.5.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-06-13T17:33:31Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.5/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.5/
        - name: Airlock Microgateway Labs
          url: https://play.instruqt.com/airlock/invite/hyi9fy4b4jzc?icp_referrer=openshift
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:806366a0b3f78844fe14775fc290fbfbde8d2a701d34a6478b019a0e7140b699
        - registry.connect.redhat.com/airlock/microgateway@sha256:80d49a299a97bed33c28941f93312e86b949ce945e843c93da6bca49865c81d0
        version: 4.5.4
      entries:
      - name: airlock-microgateway.v4.5.4
        version: 4.5.4
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.5"
    - currentCSV: airlock-microgateway.v4.6.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControl","metadata":{"name":"access-control-example"},"spec":{"policies":[{"requestConditions":{"path":{"matcher":{"exact":"/secure"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}},{"oidc":{"claim":{"name":"roles","value":{"matcher":{"contains":"\"admin\""}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"oidc-example"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"token-exchange-example"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"identity-propagation-example"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"jwt-example"}}}}},{"authorization":{}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurity","metadata":{"name":"content-security-example"},"spec":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1"},"spec":{"value":{"name":"ext-authz","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}}}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"lifetime":"16h"},"prefix":"session_prefix_example"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SidecarGateway","metadata":{"name":"sidecar-gateway-example"},"spec":{"podSelector":{"matchLabels":{"app":"example-app"}},"sessionHandlingRef":{"name":"session-handling-example"},"applications":[{"containerPort":8443,"routes":[{"pathPrefix":"/metrics","unsecured":{}},{"pathPrefix":"/","secured":{"accessControlRef":{"name":"access-control-example"},"contentSecurityRef":{"name":"content-security-example"}}}],"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}},"restrictions":{"http":{"headersLength":"80Ki"}},"protocol":{"auto":{}},"tls":{"enable":true,"secretRef":{"name":"example-tls-secret"},"protocol":{"minimum":"TLSv1_3"},"clientCertificate":{"ignored":{}},"xfcc":"AlwaysForwardOnly"}},"upstream":{"tls":{"enable":true,"protocol":{"minimum":"TLSv1_2"}}},"envoyHTTPFilterRefs":{"prepend":[{"name":"envoy-http-filter-example-1"}]},"telemetryRef":{"name":"telemetry-example"}}],"envoyClusterRefs":[{"name":"envoy-cluster-example-1"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-23T06:27:21Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: AccessControl
            name: accesscontrols.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurity
            name: contentsecurities.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: SidecarGateway
            name: sidecargateways.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056)
          * Gateway API
          * (Optional) Airlock Microgateway CNI (Required for [data plane mode sidecar](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000137)).

          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.3.0/standard-install.yaml
          ```

          ### (Optional) Install Airlock Microgateway CNI
          ***Note:*** *The Airlock Microgateway CNI is not certified for RedHat OpenShift.*

          Refer to [our documentation](https://docs.airlock.com/microgateway/4.6/?topic=MGW-00000140) for installation instructions

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD_MONITOR\_CREATE | "false" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.6/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:8fa300ced5b69acbad824c037d2b344fd1926bcfb39e264e9f0f055ceb9bd68d
        - registry.connect.redhat.com/airlock/microgateway@sha256:087d4d726f8080111d2b18cd0f9ab290f4560333fef308e553991484b6210c61
        version: 4.6.4
      entries:
      - name: airlock-microgateway.v4.6.4
        version: 4.6.4
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.6"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: "4.7"
    - currentCSV: airlock-microgateway.v4.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"APIProtection","metadata":{"name":"api-protection-example"},"spec":{"policies":[{"name":"My
            REST API exception","requestConditions":{"path":{"matcher":{"prefix":"/rest/no-validation"}}},"noValidation":{}},{"name":"My
            REST API","requestConditions":{"path":{"matcher":{"prefix":"/rest"}}},"openAPIRef":{"name":"open-api-example"}},{"name":"My
            GraphQL API","requestConditions":{"path":{"matcher":{"prefix":"/graphql"}}},"graphQLRef":{"name":"graphql-example"}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"AccessControlPolicy","metadata":{"name":"secured"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"policies":[{"requestConditions":{"invert":true,"remoteIP":{"cidrRanges":["192.168.0.0/16"]}},"authorization":{"deny":{}}},{"requestConditions":{"path":{"matcher":{"regex":"^/test/"}}},"authorization":{"requireAny":[{"oidc":{"claim":{"name":"roles","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"group","value":{"matcher":{"exact":"superusers"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-1"}}}},"tokenExchange":{"onFailure":"Block","actions":[{"tokenExchangeRef":{"name":"test-token-exchange"}}]},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-metadata"}},{"identityPropagationRef":{"name":"test-header-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"exact":"/public"}}},"authorization":{}},{"requestConditions":{"path":{"matcher":{"regex":"^/api/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name"}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-defaults"}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-header-oidc-idtoken"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-accesstoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"regex":"^/admin/"}}},"authorization":{"requireAll":[{"oidc":{"claim":{"name":"name","value":{"matcher":{"exact":"admin"}}}}},{"oidc":{"claim":{"name":"email","value":{"matcher":{"suffix":"@company.com"}}}}}],"authentication":{"oidc":{"oidcRelyingPartyRef":{"name":"test-2"},"introspection":{"strategy":{"random":{"probability":"0.1%"}}}}}},"identityPropagation":{"actions":[{"identityPropagationRef":{"name":"test-bearertoken-metadata"}},{"identityPropagationRef":{"name":"test-bearertoken-oidc-idtoken"}}],"onFailure":"Pass"}},{"requestConditions":{"path":{"matcher":{"prefix":"/api"}}},"authorization":{"requireAll":[{"jwt":{"claim":{"name":"sub","value":{"matcher":{"prefix":"/group/a/"}}}}}],"authentication":{"jwt":{"jwtRef":{"name":"test-jwt"}}}}},{"authorization":{"deny":{}}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"CSRFProtection","metadata":{"name":"csrf-protection-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"exceptions":[{"requestConditions":{"path":{"matcher":{"regex":"^/member/","ignoreCase":true}},"invert":false,"remoteIP":{"cidrRanges":["192.168.1.0/24","10.0.0.0/16"],"invert":false},"method":["DELETE"]}}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"ContentSecurityPolicy","metadata":{"name":"content-security-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"secured":{"parserRef":{"name":"parser-example"},"limitsRef":{"name":"limits-example"},"headerRewritesRef":{"name":"header-rewrites-example"},"filter":{"denyRulesRef":{"name":"deny-rules-example"}},"apiProtectionRef":{"name":"api-protection-example"},"csrfProtectionRef":{"name":"csrf-protection-example"}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"DenyRules","metadata":{"name":"deny-rules-example"},"spec":{"request":{"builtIn":{"settings":{"level":"Strict","threatHandlingMode":"Block"},"overrides":[{"conditions":{"ruleKeys":["XSS"],"types":["Parameter"]},"settings":{"level":"Standard"}},{"conditions":{"types":["GraphQL"]},"settings":{"level":"Standard"}}],"exceptions":[{"blockedData":{"parameter":{"name":{"matcher":{"exact":"search"}},"source":"Query"}},"requestConditions":{"path":{"matcher":{"prefix":"/member/"}},"method":["GET"]},"ruleKeys":["SQL"]},{"blockedData":{"graphQL":{"field":{"matcher":{"exact":"films"}},"argument":{"matcher":{"exact":"search"}},"value":{"matcher":{"exact":"abc"}}}}}]},"custom":{"rules":[{"ruleKey":"CM_REFERRER_BLOCK","blockData":{"header":{"name":{"matcher":{"exact":"referer"}},"value":{"matcher":{"regex":".*bad.tv"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyCluster","metadata":{"name":"envoy-cluster-example-1","namespace":"default"},"spec":{"value":{"name":"envoycluster/default/envoy-cluster-example-1","connect_timeout":"1s","type":"LOGICAL_DNS","load_assignment":{"cluster_name":"ext-authz","endpoints":[{"lb_endpoints":[{"endpoint":{"address":{"socket_address":{"address":"opa.service.internal","port_value":8099}}}}]}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyExtensionPolicy","metadata":{"name":"envoy-extension-policy-example"},"spec":{"targetRefs":[{"group":"gateway.networking.k8s.io","kind":"HTTPRoute","name":"backend"}],"envoyHTTPFilterRefs":{"prepend":[{"name":"geoip-blocker"}],"append":[{"name":"lua-transform-request"}]},"envoyClusterRefs":[{"name":"geoip-server"}]}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"EnvoyHTTPFilter","metadata":{"name":"envoy-http-filter-example-1"},"spec":{"value":{"name":"envoy.filters.http.ext_authz","typed_config":{"@type":"type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz","http_service":{"path_prefix":"/auth/login","server_uri":{"uri":"http://opa.service.internal:8099","cluster":"ext-authz","timeout":"0.25s"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GatewayParameters","metadata":{"name":"gateway-parameters-example"},"spec":{"logging":{"level":"debug"},"kubernetes":{"service":{"type":"LoadBalancer","annotations":{"service.beta.kubernetes.io/aws-load-balancer-type":"external","service.beta.kubernetes.io/aws-load-balancer-nlb-target-type":"ip","service.beta.kubernetes.io/aws-load-balancer-scheme":"internet-facing"}},"deployment":{"replicas":3,"strategy":{"type":"RollingUpdate","rollingUpdate":{"maxUnavailable":1}},"placement":{"nodeSelector":{"node.kubernetes.io/instance-type":"m3.medium"},"topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"}]},"volumes":[{"name":"vol-1","emptyDir":{}}],"engineContainer":{"resources":{"limits":{"cpu":"500m","memory":"128Mi"},"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"vol-1","mountPath":"/tmp/data"}]}}},"defaults":{"sessionHandlingRef":{"name":"session-handling-example"},"telemetryRef":{"name":"telemetry-example"},"downstream":{"remoteIP":{"xff":{"numTrustedHops":1}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"GraphQL","metadata":{"name":"graphql-example"},"spec":{"settings":{"threatHandlingMode":"Block","allowMutations":true,"allowIntrospection":false,"schema":{"source":{"configMapRef":{"name":"graphql-schema"}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"HeaderRewrites","metadata":{"name":"header-rewrites-example"},"spec":{"request":{"allow":{"matchingHeaders":{"builtIn":{"standardHeaders":true,"tracingHeaders":true},"custom":[{"name":"Allow
            X-CSRF-Token header","headers":[{"name":{"matcher":{"exact":"X-CSRF-TOKEN"}}}]}]}},"remove":{"builtIn":{"alternativeForwardedHeaders":true},"custom":[{"name":"Remove
            X-Forwarded-Host","headers":[{"name":{"matcher":{"exact":"X-Forwarded-Host"}}}]}]},"add":{"custom":[{"name":"Add
            headers with TLS information of the downstream connection to member requests","headers":[{"name":"X-TLS-DOWNSTREAM-PEER-CERT","value":"%DOWNSTREAM_PEER_CERT%"}],"mode":"OverwriteOrAdd","requestConditions":{"path":{"matcher":{"regex":"^/member/"}}}}]}},"response":{"allow":{"allHeaders":{}},"remove":{"builtIn":{"auth":{"basic":true}}},"add":{"custom":[{"name":"Add
            application version as header","headers":[{"name":"X-APP-VERSION","value":"3.7.4"}],"mode":"OverwriteOrAdd"}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"IdentityPropagation","metadata":{"name":"identity-propagation-example"},"spec":{"header":{"name":"X-USER-NAME","value":{"source":{"oidc":{"idToken":{"claim":"name"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWKS","metadata":{"name":"jwks-example"},"spec":{"provider":{"remote":{"uri":"https://example.com/.well-known/jwks.json","timeouts":{"connect":"1s"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"JWT","metadata":{"name":"jwt-example"},"spec":{"extractionSources":[{"bearerToken":{}},{"cookie":{"name":"creds"}}],"requireAny":[{"jwksRef":{"name":"example-jwks"},"issuer":"https://token.example.com","audiences":["bookstore_web.apps.example.com"],"subject":{"matcher":{"prefix":"spiffe://spiffe.example.com/"}}}],"expirationRequired":true,"maxLifetime":"24h","clockSkewTolerance":"10s"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Limits","metadata":{"name":"limits-example"},"spec":{"settings":{"threatHandlingMode":"Block"},"request":{"limited":{"general":{"bodySize":"10Mi","pathLength":"1Ki"},"parameter":{"bodySize":"10Ki","count":128,"nameLength":128,"valueLength":"1Ki"},"json":{"bodySize":"10Ki","nestingDepth":100,"keyCount":250,"elementCount":150000,"keyLength":128,"valueLength":"8Ki"},"multipart":{"bodySize":"10Mi"},"graphQL":{"querySize":512,"nestingDepth":5,"valueLength":128},"exceptions":[{"length":{"json":{"jsonPath":"$.logo"}},"requestConditions":{"path":{"matcher":{"prefix":"/project/"}}}},{"length":{"parameter":{"name":{"matcher":{"exact":"txtbdy"}},"source":"Post"}}},{"length":{"graphQL":{"field":{"matcher":{"exact":"search"}},"argument":{"matcher":{"exact":"text"}}}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCProvider","metadata":{"name":"oidc-provider-example"},"spec":{"static":{"issuer":"https://iam.airlock.com/","endpoints":{"authorization":{"uri":"https://iam.airlock.com/authorize"},"token":{"uri":"https://iam.airlock.com/token"},"introspection":{"uri":"https://iam.airlock.com/introspection"}},"tokenValidation":{"idToken":{"signatureVerification":{"jwksRef":{"name":"jwks-example"}}}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OIDCRelyingParty","metadata":{"name":"oidc-rp-example"},"spec":{"oidcProviderRef":{"name":"example-provider"},"clientID":"my-id","scopes":["openid","profile","email","roles"],"credentials":{"clientSecret":{"method":"BasicAuth","secretRef":{"name":"client-password"}}},"pathMapping":{"logoutPath":{"matcher":{"exact":"/logout"}},"redirectPath":{"matcher":{"exact":"/callback"}}},"redirectURI":"%REQ(:x-forwarded-proto)%://%REQ(:authority)%/callback","flowTimeout":"7m"}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"OpenAPI","metadata":{"name":"open-api-example"},"spec":{"settings":{"schema":{"source":{"configMapRef":{"name":"openapi3-spec"}}},"threatHandlingMode":"Block","logging":{"maxFailedSubvalidations":10},"validation":{"authentication":{"oAuth2":{"allowedParameters":{"builtIn":{"standardParameters":true},"custom":["param1","param2"]}},"oidc":{"allowedParameters":{"builtIn":{"standardParameters":false},"custom":["param3","param4"]}}}}},"response":{"unsecured":{}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Parser","metadata":{"name":"parser-example"},"spec":{"request":{"defaultContentType":"application/json","parsers":{"json":{"enable":true,"mediaTypePattern":".*json.*"},"form":{"enable":true,"mediaTypePattern":".*form-urlencoded.*"},"multipart":{"enable":true,"mediaTypePattern":".*multipart.*"}},"custom":{"rules":[{"requestConditions":{"method":["POST"],"path":{"matcher":{"exact":"/company/contact.php"}}},"action":{"parse":{"json":{}}}},{"requestConditions":{"path":{"matcher":{"exact":"/support/upload.php"}}},"action":{"skip":{}}}]}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"RedisProvider","metadata":{"name":"redis-provider-example"},"spec":{"mode":{"standalone":{"host":"redis.example.com","port":1111}},"tls":{"certificateVerification":{"custom":{"trustedCA":{"certificates":[{"secretRef":{"name":"ca"}}]}}},"clientCertificate":{"secretRef":{"name":"client-certificate"}}},"timeouts":{"connect":"10s"},"auth":{"username":"user1","password":{"secretRef":{"name":"redis-password"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"SessionHandling","metadata":{"name":"session-handling-example"},"spec":{"mode":"Enforce","persistence":{"redisProviderRef":{"name":"redis-provider-example"}},"defaultTimeouts":{"idle":"30s","lifetime":"16h"},"prefix":"session_prefix_example","cookie":{"name":"SESS","attributes":{"domain":"example.org","path":"/","sameSite":"Strict","secure":true}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"Telemetry","metadata":{"name":"telemetry-example"},"spec":{"logging":{"accessLog":{"format":{"json":{"@timestamp":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","ecs":{"version":"8.5"},"log":{"logger":"access","level":"info"},"event":{"kind":"event","category":["web"],"type":"%EVENT_TYPE%","module":"envoy","dataset":"envoy.access","outcome":"success","start":"%START_TIME(%Y-%m-%dT%T.%3f%z)%","end":"%END_TIME(%Y-%m-%dT%T.%3f%z)%","duration":"%DURATION_IN_NANOSECONDS%"},"airlock":{"access_control":"%ACCESS_CONTROL%","actions":{"block":"%BLOCK_ACTION%","header_rewrites":"%HEADER_REWRITES%","log_only":"%LOG_ONLY_ACTIONS%"},"http":{"request":{"accept_language":"%REQ(ACCEPT-LANGUAGE):100%","correlation_id":"%DYNAMIC_METADATA(com.airlock.microgateway.telemetry:correlation_id)%"},"response":{"redirect_url":"%RESP(LOCATION):1000%"},"session":{"id":"%SESSION_ID%"}},"log_correlation":"%LOG_CORRELATION%","summary":{"action":"%SUMMARY_ACTION%","details":"%RESPONSE_CODE_DETAILS%","flags":"%RESPONSE_FLAGS%"},"upstream":{"destination":{"ip":"%UPSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%UPSTREAM_REMOTE_PORT%"},"http":{"version":"%UPSTREAM_HTTP_VERSION%"}}},"destination":{"ip":"%DOWNSTREAM_LOCAL_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_LOCAL_PORT%"},"http":{"request":{"body":{"bytes":"%BYTES_RECEIVED%"},"bytes":"%REQUEST_HEADERS_AND_BODY_BYTES%","id":"%STREAM_ID%","method":"%REQ(:METHOD):100%","mime_type":"%REQ_MIME_TYPE:500%","referrer":"%REQ(REFERER):1000%"},"response":{"body":{"bytes":"%BYTES_SENT%"},"bytes":"%RESPONSE_HEADERS_AND_BODY_BYTES%","mime_type":"%RESP_MIME_TYPE:500%","status_code":"%RESPONSE_CODE%"},"version":"%HTTP_VERSION%"},"network":{"forwarded_ip":"%DOWNSTREAM_REMOTE_ADDRESS_WITHOUT_PORT%"},"observer":{"product":"Airlock
            Microgateway","type":"waap","vendor":"Ergon Informatik AG","version":"%ENVIRONMENT(ENGINE_VERSION)%"},"source":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"},"url":{"domain":"%HTTP_HOST:500%","path":"%REQ_WITHOUT_QUERY(:PATH):1000%","query":"%REQ_QUERY(:PATH):1000%"},"user_agent":{"original":"%REQ(USER-AGENT):500%"},"custom":{"downstream":{"tls":{"protocol":"%DOWNSTREAM_TLS_VERSION%","cipher":"%DOWNSTREAM_TLS_CIPHER%","session_id":"%DOWNSTREAM_TLS_SESSION_ID%"},"direct_remote_address":{"ip":"%DOWNSTREAM_DIRECT_REMOTE_ADDRESS_WITHOUT_PORT%","port":"%DOWNSTREAM_DIRECT_REMOTE_PORT%"}},"http":{"request":{"cookie":"%REQ(cookie):500%"}}}}}}},"correlation":{"idSource":{"header":{"name":"X-Request-Id"}}}}},{"apiVersion":"microgateway.airlock.com/v1alpha1","kind":"TokenExchange","metadata":{"name":"token-exchange-example"},"spec":{"tokenEndpoint":{"endpoint":{"uri":"https://iam.airlock.com/token","tls":{"protocol":{"minimum":"TLSv1_3"}},"timeouts":{"connect":"2s"}},"credentials":{"clientPassword":{"method":"BasicAuth","clientID":"my-id","clientSecret":{"secretRef":{"name":"client-password"}}}}},"requestedTokenType":"urn:ietf:params:oauth:token-type:jwt","audiences":["https://api.airlock.com/realms/a","my-audience"],"resources":["https://resource.airlock.com/"],"scopes":["db"]}}]'
          capabilities: Seamless Upgrades
          categories: Security
          createdAt: "2025-09-03T13:02:23Z"
          description: Airlock Microgateway is a Kubernetes native WAAP (Web Application
            and API Protection) solution to protect microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: airlock-microgateway-system
          operators.openshift.io/valid-subscription: '["Airlock Microgateway Community
            License", "Airlock Microgateway Premium License"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/internal-objects: '["envoyconfigurations.microgateway.airlock.com"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/airlock/microgateway
          support: Airlock
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessControlPolicy
            name: accesscontrolpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: APIProtection
            name: apiprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: ContentSecurityPolicy
            name: contentsecuritypolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: CSRFProtection
            name: csrfprotections.microgateway.airlock.com
            version: v1alpha1
          - kind: DenyRules
            name: denyrules.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyCluster
            name: envoyclusters.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyConfiguration
            name: envoyconfigurations.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyExtensionPolicy
            name: envoyextensionpolicies.microgateway.airlock.com
            version: v1alpha1
          - kind: EnvoyHTTPFilter
            name: envoyhttpfilters.microgateway.airlock.com
            version: v1alpha1
          - kind: GatewayParameters
            name: gatewayparameters.microgateway.airlock.com
            version: v1alpha1
          - kind: GraphQL
            name: graphqls.microgateway.airlock.com
            version: v1alpha1
          - kind: HeaderRewrites
            name: headerrewrites.microgateway.airlock.com
            version: v1alpha1
          - kind: IdentityPropagation
            name: identitypropagations.microgateway.airlock.com
            version: v1alpha1
          - kind: JWKS
            name: jwks.microgateway.airlock.com
            version: v1alpha1
          - kind: JWT
            name: jwts.microgateway.airlock.com
            version: v1alpha1
          - kind: Limits
            name: limits.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCProvider
            name: oidcproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: OIDCRelyingParty
            name: oidcrelyingparties.microgateway.airlock.com
            version: v1alpha1
          - kind: OpenAPI
            name: openapis.microgateway.airlock.com
            version: v1alpha1
          - kind: Parser
            name: parsers.microgateway.airlock.com
            version: v1alpha1
          - kind: RedisProvider
            name: redisproviders.microgateway.airlock.com
            version: v1alpha1
          - kind: SessionHandling
            name: sessionhandlings.microgateway.airlock.com
            version: v1alpha1
          - kind: Telemetry
            name: telemetries.microgateway.airlock.com
            version: v1alpha1
          - kind: TokenExchange
            name: tokenexchanges.microgateway.airlock.com
            version: v1alpha1
        description: |-
          Airlock Microgateway is a Kubernetes native WAAP (Web Application and API Protection) solution to protect microservices.

          Modern application security is embedded in the development workflow and follows DevSecOps paradigms. Airlock Microgateway is the perfect fit for these requirements.
          It is a lightweight WAAP solution (formerly known as WAF), optimized for Kubernetes environments.
          Airlock Microgateway protects your applications and microservices with the tried-and-tested Airlock security features against attacks, while also providing a high degree of scalability.

          For a list of all features, view the [comparison of the community and premium edition](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056).

          The instructions below provide a quick start guide.

          ***Note:*** Sidecar-based data plane mode is *not* supported by this OLM Bundle.
          If you want to use the Sidecar-based data plane mode, use Helm instead. Please note that the Airlock Microgateway CNI and the Helm charts are not certified for RedHat OpenShift.
          Refer to https://github.com/airlock/microgateway for installation instruction with Helm.

          ## Prerequisites

          * [Airlock Microgateway License](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056)
          * [Gateway API](https://gateway-api.sigs.k8s.io/)

          In order to use Airlock Microgateway you need a license. You may either request a community license free of charge or purchase a premium license.

          ### Obtain Airlock Microgateway License
          1. Either request a community or premium license
              * Community license: [airlock.com/microgateway-community](https://airlock.com/en/microgateway-community)
              * Premium license: [airlock.com/microgateway-premium](https://airlock.com/en/microgateway-premium)
          2. Check your inbox and save the license file microgateway-license.txt locally.

          *See [Community vs. Premium editions in detail](https://docs.airlock.com/microgateway/4.7/?topic=MGW-00000056) to choose the right license type.*

          ### Install Gateway API standard channel
          ***Note:*** *This step is only required for OpenShift versions below 4.19, as Gateway API is included by default in OpenShift 4.19+.*

          ```console
          kubectl apply -f https://github.com/kubernetes-sigs/gateway-api/releases/download/v1.2.1/standard-install.yaml
          ```

          ## Configuring Airlock Microgateway Operator deployed by OLM
          ***Note:*** *It is possible to configure how OLM deploys an Operator via the [Subscription Config](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md).*

          The following environment variables are supported:
          | Name        | Default     | Example     | Description |
          | ----------- | ----------- | ----------- | ----------- |
          | LOG\_LEVEL   | "info"      | "debug"     | Operator application log level. Allowed Values: “error”, “warn”, “info”, “debug” |
          | GATEWAY\_API\_ENABLED |  "true"  |  "true"    | Whether to enable the Kubernetes Gateway API related controllers. Requires that the gateway.networking.k8s.io/v1 resources are installed on the cluster. See [https://gateway-api.sigs.k8s.io/guides/#install-standard-channel](https://gateway-api.sigs.k8s.io/guides/#install-standard-channel). |
          | GATEWAY\_API\_CONTROLLER\_NAME |  "microgateway.airlock.com/gatewayclass-controller" | "true" | Controller name referred in the GatewayClasses managed by this operator. The value must be a path prefixed by the domain microgateway.airlock.com. |
          | GATEWAY\_API\_POD\_MONITOR\_CREATE | "true" | "true" | Whether to create a PodMonitor resource for monitoring the Gateway API resources. |
          | GATEWAY\_API\_POD\_MONITOR\_ADDITIONAL\_LABELS | "" | "release=my-microgateway,env=dev" | Labels to add to the PodMonitor. Allowed Values: Comma-separated list of key=value pairs|
        displayName: Airlock Microgateway
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Gateway API
        - WAF
        - Web Application Firewall
        - WAAP
        - Web Application and API protection
        - OWASP
        - Airlock
        - Microgateway
        - Security
        - Filtering
        - DevSecOps
        - shift left
        - control plane
        - Operator
        links:
        - name: Airlock Microgateway Documentation
          url: https://docs.airlock.com/microgateway/4.7/
        - name: Airlock Microgateway Labs
          url: https://airlock.instruqt.com/pages/airlock-microgateway-labs
        - name: Airlock Microgateway Forum
          url: https://forum.airlock.com/
        maintainers:
        - email: support@airlock.com
          name: Airlock
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Airlock
        relatedImages:
        - quay.io/airlock/microgateway-operator@sha256:d4ce35184922953c209a9ebb234ca3370dbfc5054de411b5503c26ac2730b7ae
        - registry.connect.redhat.com/airlock/microgateway@sha256:c782627a1eb9565b5d1f13f3cd08cb01e010a616b099d216e98238c1c7f8c1c7
        version: 4.7.1
      entries:
      - name: airlock-microgateway.v4.7.1
        version: 4.7.1
      - name: airlock-microgateway.v4.7.0
        version: 4.7.0
      - name: airlock-microgateway.v4.6.3
        version: 4.6.3
      - name: airlock-microgateway.v4.6.2
        version: 4.6.2
      - name: airlock-microgateway.v4.6.1
        version: 4.6.1
      - name: airlock-microgateway.v4.6.0
        version: 4.6.0
      - name: airlock-microgateway.v4.5.3
        version: 4.5.3
      - name: airlock-microgateway.v4.5.2
        version: 4.5.2
      - name: airlock-microgateway.v4.5.1
        version: 4.5.1
      - name: airlock-microgateway.v4.5.0
        version: 4.5.0
      name: stable
    defaultChannel: "4.7"
    packageName: airlock-microgateway
    provider:
      name: Airlock
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JohnSnowLabs
      provider-url: ""
    name: nlp-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nlp-server.v0.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.johnsnowlabs.com/v1",
                "kind": "Nlpserver",
                "metadata": {
                  "name": "nlpserver-sample"
                },
                "spec": {
                  "ExtraEnv": {
                    "johnsnowlabs_client_id": "sI4MKSmLHOX2Pg7XhM3McJS2oyKG5PHcp0BlANEW",
                    "nb_uid": "1000",
                    "nb_user": "johnsnowlabs",
                    "pyspark_version": "3.0.2",
                    "spark_nlp_version": "3.4.4"
                  },
                  "affinity": {},
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "johnsnowlabs/nlp-server",
                    "tag": "0.8.0rc1"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {
                      "kubernetes.io/ingress.class": "nginx",
                      "nginx.ingress.kubernetes.io/ssl-redirect": "true"
                    },
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": [
                      {
                        "hosts": [
                          "nlp-server.jsl.aiplatform.eks.test.johnsnowlabs.com"
                        ],
                        "secretName": "annotationlab-tls-secret"
                      }
                    ]
                  },
                  "livenessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "readinessProbe": "httpGet:\n  path: /\n  port: http\ninitialDelaySeconds: 15\ntimeoutSeconds: 3\n",
                  "replicaCount": 1,
                  "resources": {},
                  "securityContext": {},
                  "service": {
                    "port": 5000,
                    "portName": "http",
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": "nlp-server-sa"
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nlpserver
            name: nlpservers.apps.johnsnowlabs.com
            version: v1
        description: NLPServer from JohnSnow labs
        displayName: nlp-server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nlp-server
        links:
        - name: Nlp Server
          url: https://johnsnowlabs.com
        maturity: alpha
        provider:
          name: JohnSnowLabs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/openshift-community-operators/nlp-server@sha256:6f80f44822301252540c6b808a20cea327f0eedeadc7643d31fc743de5c5e696
        - registry.johnsnowlabs.com/nlp-server/nlp-server-operator:0.8.6
        version: 0.8.6
      entries:
      - name: nlp-server.v0.8.6
        version: 0.8.6
      name: alpha
    defaultChannel: alpha
    packageName: nlp-server
    provider:
      name: JohnSnowLabs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator-rh.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
          createdAt: "2025-07-23T09:04:08Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.4
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:ea906dcf8d45956251df22e269fe0b8fa8d49734e96a61b10db7b25d626b2e6a
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:c5b3b3907e8135d3eedba40e27ba18a232ea4f30854d652059fcb14f82a8bf15
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:0f4683d18404b4a9412e42ed417f6d3ac0024b3c57aa72c9162ff72d6ea8d394
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:047b6c378e2b37339dd498c9025060f20812014bde4d0c98183f1a1201c21e66
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:2be8bc79667018386235823e322ae10788e4e7e889c10e63566ae27a65dcabf1
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:fdc0c228ca38db30f89812e41c09ec96d9276ed844b0801013375bbffd6c1a16
        version: 1.4.2
      entries:
      - name: multicluster-global-hub-operator-rh.v1.4.2
        version: 1.4.2
      - name: multicluster-global-hub-operator-rh.v1.4.1
        version: 1.4.1
      - name: multicluster-global-hub-operator-rh.v1.4.0
        version: 1.4.0
      name: release-1.4
    - currentCSV: multicluster-global-hub-operator-rh.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
          createdAt: "2025-08-18T08:30:32Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: release-1.5
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-operator-bundle@sha256:ea606dcfa4cbe4df7b27c10d3fc45f464845518c4896845079ca9194600ceb6f
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-agent-rhel9@sha256:ab3ed25a4448a7d545648ae4cea52b493918cc4d3b0c5adc56bb2510c0838c80
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-grafana-rhel9@sha256:eae390ec42c3582f7896a8346df8782b287d64440794983ae93755490639e21e
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-manager-rhel9@sha256:921163d139bca9e47bf95df43d7b408ff6fe2bf9b0902fb6af9f4cfc9d4598cc
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-postgres-exporter-rhel9@sha256:c95052569312d805b6750477c8bc57f7c72e1efd26143f3899ea0eddc2aed37b
        - registry.redhat.io/multicluster-globalhub/multicluster-globalhub-rhel9-operator@sha256:8af7e57b8c6285ad232084e7776fef32d026e080eb0bf14b91986ac28d00bb6a
        - registry.redhat.io/rhel9/postgresql-16@sha256:2161182c88022f38d6d9b8b467c0d59156edcededc908d383b80caa4eedf57f9
        version: 1.5.1
      entries:
      - name: multicluster-global-hub-operator-rh.v1.5.1
        version: 1.5.1
      - name: multicluster-global-hub-operator-rh.v1.5.0
        version: 1.5.0
      name: release-1.5
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator-rh
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-ingress
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-ingress.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Ingress",
                "metadata": {
                  "name": "ingress-sample"
                },
                "spec": {
                  "admissionWebhooks": {
                    "patch": {
                      "image": {
                        "repository": "docker.io/jettech/kube-webhook-certgen"
                      }
                    }
                  },
                  "configMapMerger": {
                    "registry": "prod.harbor.keysight.digital/loadcore",
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "environment": "openshift",
                  "ingress-nginx": {
                    "controller": {
                      "config": {
                        "global-auth-signin": "https://$host/eula/static/",
                        "global-auth-url": "http://kcos-framework-v1-kcos-eula.kcos-framework.svc.cluster.local/v1/check"
                      },
                      "daemonset": {
                        "useHostPort": false
                      },
                      "hostNetwork": false,
                      "image": {
                        "registry": "prod.harbor.keysight.digital/loadcore"
                      },
                      "ingressClassResource": {
                        "default": true
                      },
                      "nodeSelector": "",
                      "service": {
                        "nodePorts": {
                          "http": 80,
                          "https": 443
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "defaultBackend": {
                      "image": {
                        "repository": "k8s.gcr.io/defaultbackend-amd64"
                      }
                    }
                  },
                  "jobDelWebhook": {
                    "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:28:18Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ingress
            name: ingresses.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Ingress Operator
        displayName: Keysight KCOS Ingress Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Ingress
          url: https://keysight-kcos-ingress.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-ingress@sha256:c6f6f8b44276efd8f5ffaa099b393e7e97d9cfa627952c0d049b98d92dc3ab59
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - registry.connect.redhat.com/loadcore/keysight-kcos-ingress-bundle@sha256:f04b8f15b069c06eac5d9a97f81a2f069704237eea998d35727b89854c793746
        version: 0.1.21
      entries:
      - name: keysight-kcos-ingress.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-ingress
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-msg-broker
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-msg-broker.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Messagebroker",
                "metadata": {
                  "name": "messagebroker-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "chiselServer": {
                    "enabled": false,
                    "port": 8080
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": false
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "natsCore": {
                    "reloader": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-server-config-reloader",
                      "tag": "0.8.3-3747-30"
                    },
                    "server": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats",
                      "tag": "2.9.0-3747-21"
                    }
                  },
                  "natsHttpProxy": {
                    "enabled": false,
                    "port": 8080
                  },
                  "natsStreaming": {
                    "image": {
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "nats-streaming",
                      "tag": "0.23.2-3747-31"
                    }
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "tagSuffix": "-ocp"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:31:32Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Messagebroker
            name: messagebrokers.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Message Broker Operator
        displayName: Keysight WAP Message Broker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Msg Broker
          url: https://keysight-wap-msg-broker.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-msg-broker@sha256:4701b6e456b099697bf1e25aeb4bc8e32b1892bb0956e36d84f9f80666b3a1fc
        - registry.connect.redhat.com/loadcore/keysight-wap-msg-broker-bundle@sha256:a7dfca98b4e09c20317cbb486e7893de8d88faf5db2ee26549ba81332ec526cb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-msg-broker.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-msg-broker
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-opa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: opa-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
          description: Stackable Operator for the Open Policy Agent
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/opa-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for OpaClusterSpec via `CustomResource`
            displayName: opaclusters.opa.stackable.tech
            kind: OpaCluster
            name: opaclusters.opa.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for the Open Policy Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opa
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/opa-operator@sha256:88622835662d645198ed7543302cf5166821294284441ef5ddeb5c45c89e5281
        - registry.connect.redhat.com/stackable/stackable-opa-operator@sha256:434d1d26c0267c04b8ebec0d2ee450235ab973dd0183eb84682aa0485e480a18
        version: 25.7.0
      entries:
      - name: opa-operator.v25.7.0
        version: 25.7.0
      - name: opa-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-opa-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle@sha256:a90a748aa70ee5f1b8fd375a7877db1292792ba46f7b49f629533c82a25c43a0
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-zen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-zen-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.3 <6.1.3'
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: represents zen instance
            kind: ZenService
            name: zenservices.zen.cpd.ibm.com
            version: v1
          - description: represents zen extension instance
            kind: ZenExtension
            name: zenextensions.zen.cpd.ibm.com
            version: v1
        description: |
          The Zen operator is a component in IBM® Cloud Pak foundational services. It serves a framework for user access management, defining user roles and for powering persona driven user experiences for Cloud Paks and Service offerings
        displayName: IBM Zen Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - IBM
        - Access management
        - Platform UI
        - Zen framework
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:c8baecad9c97ddbbb3f9f02c90abbf483d46974b7bc9605622a302cbeba1de74
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:aaee5240471872a8f1a77d4440a696d45041c5ce06849d516daccb90f5341b6d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:1df782d0d923544767696d719bd0ac10802b8e06508a104af5e023a519f32dee
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:c4ff5dae4560e5a943c7218df3a5aa9ca1a9ee2d1f7dfac73db05892725c34a9
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:82afac013ef9b6b9d0eacb4575c979b65abd6f2821971182b90c71bd488ef366
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:10ede3c4e3c7e0cbc26e7b6e797cad35e39face1a1b9d4a4e2440ab0feeeb814
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:a8d2e42757c6ccbfe4372465b1827c633aad012c5194bf8b08afcd5691caa6f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:5f48081ffa453ec740c2562c24603d78b481b5e236d6bcc73e96203cabe153f2
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:45ca5370b726e31c55f9d7d249374189cb48dc05ab89d675c35225d53b90e7c9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:f06a1f50cd18d6f48762fdd0394307b44c0573d0cc91f53724c5b3c8c5e2b8e4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e9173d81ef2a75839c0e1d677b72bb00b8e6d7ef0511afab103f08a884a03738
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.connect.redhat.com/cpopen/ibm-zen-operator@sha256:5a9120d5977e934ca3801eab44ed48ecd4a9fee10c0b4980594d33cb02e64855
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:0522fc5482404fa9be179c7225a5692c8d71d755fda0b2d3928e452adcf030fb
        version: 6.1.3
      entries:
      - name: ibm-zen-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: ibm-zen-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - quay.io/community-operator-pipeline-prod/dell-csm-operator@sha256:6e2e813b7ca426a793da717d42f3620edbc3eaa84be710600f2a5262924378b4
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        version: 1.10.0
      entries:
      - name: dell-csm-operator.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.6.5-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:0739d5035590cece6d2a573aa1e799b0b0cab8b548a41cc2488cf465486cf4a5
        - turbonomic/kubeturbo-operator:8.6.5
        version: 8.6.5-beta.1
      entries:
      - name: kubeturbo-operator.v8.6.5-beta.1
        version: 8.6.5-beta.1
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: beta
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    - currentCSV: kubeturbo-operator.v8.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.4
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:3028579f4909be6b8dcd970e778bc88d0576c76e55b98c886738b8f601cdf4c1
        - turbonomic/kubeturbo-operator:8.4
        version: 8.4.0
      entries:
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.4.0
    - currentCSV: kubeturbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.5
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:e50343879b4c3c4002c5e7ff67f4664df441cf5f239f59fa98b7fddd75516d41
        - turbonomic/kubeturbo-operator:8.5
        version: 8.5.0
      entries:
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.5.0
    - currentCSV: kubeturbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/kubeturbo-operator:8.6
          createdAt: "2019-05-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/kubeturbo@sha256:f7c7a998ce5b4590c80ae301c6e1bd81bb577a54059a8b375c5468aad1f2bf6f
        - turbonomic/kubeturbo-operator:8.6
        version: 8.6.0
      entries:
      - name: kubeturbo-operator.v8.6.0
        version: 8.6.0
      - name: kubeturbo-operator.v8.5.0
        version: 8.5.0
      - name: kubeturbo-operator.v8.4.0
        version: 8.4.0
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: v8.6.0
    defaultChannel: stable
    packageName: kubeturbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: openshift
      provider-url: ""
    name: pf-status-relay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pf-status-relay-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pfstatusrelay.openshift.io/v1alpha1",
                "kind": "PFLACPMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "pf-status-relay-operator"
                  },
                  "name": "pflacpmonitor-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.19.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PFLACPMonitor is the Schema for the pflacpmonitors API
            displayName: PFLACPMonitor
            kind: PFLACPMonitor
            name: pflacpmonitors.pfstatusrelay.openshift.io
            version: v1alpha1
        description: It deploys the pf-status-relay application.
        displayName: pf-status-relay-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pf-status-relay
        - sriov
        - lacp
        links:
        - name: Pf Status Relay Operator
          url: https://pf-status-relay-operator.domain
        maintainers:
        - email: marguerr@redhat.com
          name: Marcelo Guerrero
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: openshift
        relatedImages:
        - registry.redhat.io/openshift4/pf-status-relay-rhel9@sha256:8cdc0714d02144f5b32e26c97f7678d0377e0c3826c475a784e76a42f781d3e3
        - registry.redhat.io/openshift4/pf-status-relay-operator-bundle@sha256:fb9635a614b6676fbe6a532d07b3206f12ee525b7947378785e79951d43940b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/pf-status-relay-rhel9-operator@sha256:e6719af6fdc99c652238f3c31a014782811eb46da91e65f93c083a7a001e2496
        version: 4.18.0-202509100344
      entries:
      - name: pf-status-relay-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: pf-status-relay-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: pf-status-relay-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: pf-status-relay-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: pf-status-relay-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: pf-status-relay-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: pf-status-relay-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: pf-status-relay-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: pf-status-relay-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: pf-status-relay-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: pf-status-relay-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: pf-status-relay-operator
    provider:
      name: openshift
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: businessautomation-operator.7.13.5-16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
          createdAt: "2025-07-30 19:54:27"
          description: Deploys and manages Red Hat Process Automation Manager and
            Red Hat Decision Manager environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an RHPAM/RHDM environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages Red Hat Process Automation Manager and Red Hat Decision Manager environments.

          * **Red Hat Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.

          [See more](https://www.redhat.com/en/products/process-automation).
        displayName: Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/#category-deploying-red-hat-process-automation-manager-on-openshift
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:228c6ab5b83e215a98f6eb76250feb2ae3096fd21398c44e5b2701257c9c3493
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:3d7bdbe38895e50549e97573bd40da47b11691932164db7dafff0155a051aab2
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:cb37360843e9b062c59755593d9d412e1bf2c1dc1d6902e8a65a2825812c0989
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:32912270eba0f2464da88f315bde8bf1ef5bbded814db156811d7a415a458f86
        - registry.redhat.io/rhpam-7/rhpam-process-migration-rhel8@sha256:bd5c6586b4859dec23ad9530ee98bab33497ab251d544f863fc0af01cbd45101
        - registry.redhat.io/openshift4/ose-cli@sha256:c3eb8807caf73e30012ae53c6d4f5114471290a5596929c27acef5a699a5b813
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:4edac282119898ab1395030eecaee584990064e88fc8e65fb999f19ea2ef3f8b
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:3fbef82d1c5da8da1606d1a2d5dcfd20000da574737eb32b772d4f0123b179cb
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-rhel8@sha256:1abbaba9d4b2ecaec38205bcccf5ef4c7b6388ed9725b4f4c6dc8dc664588eb1
        - registry.redhat.io/rhpam-7/rhpam-kieserver-rhel8@sha256:2526d12da63b67ae9ed9294c914e9406d833c090ffa777623303ea0be22755d8
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:9d7efdd97f163770d8a84b1ce5236bd4f31508154dced1f11dc6efe75a7b5089
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:3a901b54a9d35d38b5aca275df2160e3190a874e139fcb606b9969b38284b3dc
        - registry.redhat.io/rhpam-7/rhpam-rhel8-operator@sha256:3b4f420def9e5d2c768d5209264bd5f9af6d6a37b22ef9a988a167ff7d9959cf
        - registry.redhat.io/rhpam-7/rhpam-businesscentral-monitoring-rhel8@sha256:8d4481a11269565f1690ad007f152e24326bae4b2099cb26ddf7b21be67031e8
        - registry.redhat.io/rhpam-7/rhpam-operator-bundle@sha256:46e7d899d64ea8cbfc35221d05535db1023c41c1d3a7c580cb03e7e7d769144b
        - registry.redhat.io/rhpam-7/rhpam-dashbuilder-rhel8@sha256:46396dc8060f2f9cc6f87ad905ef55be5b8e5216e5162bdf54df98c9d70e29ef
        - registry.redhat.io/rhpam-7/rhpam-controller-rhel8@sha256:989dc6dc64a2d5b6723fb5c000b0411fbdc6f78c0056666d831ffeabeab4b612
        - registry.redhat.io/rhpam-7/rhpam-smartrouter-rhel8@sha256:d1e5461db8983a13f98259152dc3cfa2aa677c29f799104190e5cefa83cf97b5
        version: 7.13.5-16
      entries:
      - name: businessautomation-operator.7.13.5-16
        version: 7.13.5-16
      - name: businessautomation-operator.7.13.5-15
        version: 7.13.5-15
      - name: businessautomation-operator.7.13.5-14
        version: 7.13.5-14
      - name: businessautomation-operator.7.13.5-13
        version: 7.13.5-13
      - name: businessautomation-operator.7.13.5-12
        version: 7.13.5-12
      - name: businessautomation-operator.7.13.5-11
        version: 7.13.5-11
      - name: businessautomation-operator.7.13.5-10
        version: 7.13.5-10
      - name: businessautomation-operator.7.13.5-9
        version: 7.13.5-9
      - name: businessautomation-operator.7.13.5-8
        version: 7.13.5-8
      - name: businessautomation-operator.7.13.5-7
        version: 7.13.5-7
      - name: businessautomation-operator.7.13.5-6
        version: 7.13.5-6
      - name: businessautomation-operator.7.13.5-5
        version: 7.13.5-5
      - name: businessautomation-operator.7.13.5-4
        version: 7.13.5-4
      - name: businessautomation-operator.7.13.5-3
        version: 7.13.5-3
      - name: businessautomation-operator.7.13.5-2
        version: 7.13.5-2
      - name: businessautomation-operator.7.13.5-1
        version: 7.13.5-1
      - name: businessautomation-operator.7.13.4-4
        version: 7.13.4-4
      - name: businessautomation-operator.7.13.4-3
        version: 7.13.4-3
      - name: businessautomation-operator.7.13.4-2
        version: 7.13.4-2
      - name: businessautomation-operator.7.13.4-1
        version: 7.13.4-1
      - name: businessautomation-operator.7.13.3-5
        version: 7.13.3-5
      - name: businessautomation-operator.7.13.3-4
        version: 7.13.3-4
      - name: businessautomation-operator.7.13.3-3
        version: 7.13.3-3
      - name: businessautomation-operator.7.13.3-2
        version: 7.13.3-2
      - name: businessautomation-operator.7.13.3-1
        version: 7.13.3-1
      - name: businessautomation-operator.7.13.2-3
        version: 7.13.2-3
      - name: businessautomation-operator.7.13.2-2
        version: 7.13.2-2
      - name: businessautomation-operator.7.13.2-1
        version: 7.13.2-1
      - name: businessautomation-operator.7.13.1-2
        version: 7.13.1-2
      - name: businessautomation-operator.7.13.1-1
        version: 7.13.1-1
      - name: businessautomation-operator.7.13.0-1
        version: 7.13.0-1
      - name: businessautomation-operator.7.12.1-4
        version: 7.12.1-4
      - name: businessautomation-operator.7.12.1-3
        version: 7.12.1-3
      - name: businessautomation-operator.7.12.1-2
        version: 7.12.1-2
      - name: businessautomation-operator.7.12.1-1
        version: 7.12.1-1
      - name: businessautomation-operator.7.12.0-2
        version: 7.12.0-2
      - name: businessautomation-operator.7.12.0-1
        version: 7.12.0-1
      - name: businessautomation-operator.7.11.1-3
        version: 7.11.1-3
      - name: businessautomation-operator.7.11.1-2
        version: 7.11.1-2
      - name: businessautomation-operator.7.11.1-1
        version: 7.11.1-1
      - name: businessautomation-operator.7.11.0-2
        version: 7.11.0-2
      - name: businessautomation-operator.7.11.0-1
        version: 7.11.0-1
      - name: businessautomation-operator.7.10.1-2
        version: 7.10.1-2
      - name: businessautomation-operator.7.10.1-1
        version: 7.10.1-1
      - name: businessautomation-operator.7.10.0-2
        version: 7.10.0-2
      - name: businessautomation-operator.7.10.0-1
        version: 7.10.0-1
      - name: businessautomation-operator.7.9.1-1
        version: 7.9.1-1
      - name: businessautomation-operator.7.9.0
        version: 7.9.0
      - name: businessautomation-operator.7.9.0-2
        version: 7.9.0-2
      - name: businessautomation-operator.7.8.1
        version: 7.8.1
      - name: businessautomation-operator.7.8.0
        version: 7.8.0
      - name: businessautomation-operator.1.4.1
        version: 1.4.1
      - name: businessautomation-operator.1.4.0
        version: 1.4.0
      - name: businessautomation-operator.1.3.0
        version: 1.3.0
      - name: businessautomation-operator.1.2.1
        version: 1.2.1
      - name: businessautomation-operator.1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: businessautomation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-activemqartemis: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com/
    name: amq-broker-rhel9
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          description: AMQ Broker Operator for RHEL 9 (Multiarch) provides the ability
            to deploy and manage stateful AMQ Broker clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
          operators.openshift.io/valid-subscription: '["Red Hat Runtimes", "Red Hat
            Integration", "Red Hat Process Automation", "Red Hat AMQ"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-messaging/activemq-artemis-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: |
          **Red Hat AMQ Broker** is a pure-Java multiprotocol message broker. It’s built on an efficient, asynchronous core with a fast journal for message persistence.

          You can provision multiple instances and update the broker configuration on the fly using the associated Custom Resource Definitions.
          In addition, you can use the following CRDs to configure your broker environment:

          - **Address CRD**: define addresses and queues for the brokers in your deployment
          - **Security CRD**: define the security configuration for the brokers in your deployment, including:
              - users and roles
              - role based access control
              - console access control

          ## Selecting a channel
          You can specify the channel used to track and receive updates for the Operator by selecting either to track the minor or patch level versions.

          ## Supported Features
          - **Console** - Allows you view the broker topology, clients and messages. Red Hat Fuse users can integrate the broker console into the Fuse console for OpenShift.
          - **Persistence** - ephemeral and stateful deployments.
          - **Advanced queueing** - Last value queues, message groups, topic hierarchies, and large message support.
          - **Multiprotocol** - AMQP 1.0, MQTT, STOMP, OpenWire, and HornetQ Core.
          - **Address Management** - Creates and manages addresses and queues.
          - **User Management** - Creates and manages broker users.
          - **Monitoring** - Support for monitoring using Prometheus and provided Grafana dashboards.
          - **Scaledown** - Ensure consistency of messages using message migration.

          ## Storage
          An efficient data storage infrastructure is essential to the optimal performance of AMQ Broker.  AMQ Broker supports:
            - **Ephemeral** (Recommended for development only!)
            - **Persistent**
              - For example, block or NFS storage.
          Note: Your cloud provider might provide options that are not suitable for AMQ Broker, for example Amazon Elastic File System.  If in doubt, contact Red Hat Support

          ## Documentation
          Documentation for the current release is available from the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq/2021.q4/#category-amq-broker).
        displayName: Red Hat Integration - AMQ Broker for RHEL 9 (Multiarch)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - activemqartemis
        - operator
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:40920687402fd747ca02d384e63692a5e1db625723a2bf3056a607c92ee8eee1
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:755d0f7421efba76f52818bb3a8667cd7d2dcff2c7054dac1a4637b74ecac51d
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:b2402179b1d0ff9c950defe5d8808195ef9a4c3e13e3c469cae31ab51d5f5014
        - registry.redhat.io/amq7/amq-broker-rhel9@sha256:4304a5414911fe55ce4c0133723be234ff288463f4916bd18e743d6359489138
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:7f9b921d89b6d4d91b365659afff127d99f1455b94b48a10a6a7f4c38dd8b700
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:ff80e0a7accda6825bf834d9aec27f48205dd6329991890dd017b4d21d0da5c8
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:d42d713da0ce6806fdc6492b6342586783e6865a82a8647d3c4288439b1751ee
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:dc13871dfed52402c40fd192900fc3cb5a49da4fa08e4ec4d0e23a879e83db25
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:47fd4ce2533496828aba37bd1f9715e2164d5c90bd0fc6b25e7e0786d723bf01
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:bb6fbd68475a7852b4d99eea6c4ab313f9267da7963162f0d75375d7063409e7
        - registry.redhat.io/amq7/amq-broker-rhel9-operator-bundle@sha256:896a615487c0ffd5335b869f0dc0c0ec1c9200dbcafa52f2493003fecbc030bc
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:16253895bc113f395dc6ab5cd511901e43b6ee676369ec80d598a8741c3c9bf4
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:9cc48eecf1442ae04b8543fa5d4381a13bc2831390850828834d387006d1342b
        - registry.redhat.io/amq7/amq-broker-init-rhel9@sha256:e2cb996e0bf131e71fb8e80c128af18ab2b9330c9fca7a5437842796158e2870
        - registry.redhat.io/amq7/amq-broker-rhel9-operator@sha256:f34672152017a94ed51056435e218341e22787788ceed52439c1ff066480c715
        - registry.redhat.io/amq7/amq-broker-init-rhel8@sha256:ffffa9875f0379e9373f89f05eb06e5a193273bb04bc3aa5f85b044357b79098
        version: 7.13.1-opr-1+0.1758027574.p
      entries:
      - name: amq-broker-operator.v7.13.1-opr-1
        version: 7.13.1-opr-1
      - name: amq-broker-operator.v7.13.1-opr-1-0.1754564924.p
        version: 7.13.1-opr-1+0.1754564924.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756374447.p
        version: 7.13.1-opr-1+0.1756374447.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1756381365.p
        version: 7.13.1-opr-1+0.1756381365.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758024989.p
        version: 7.13.1-opr-1+0.1758024989.p
      - name: amq-broker-operator.v7.13.1-opr-1-0.1758027574.p
        version: 7.13.1-opr-1+0.1758027574.p
      - name: amq-broker-operator.v7.13.0-opr-1
        version: 7.13.0-opr-1
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747315188.p
        version: 7.13.0-opr-1+0.1747315188.p
      - name: amq-broker-operator.v7.13.0-opr-1-0.1747658456.p
        version: 7.13.0-opr-1+0.1747658456.p
      name: 7.13.x
    defaultChannel: 7.13.x
    packageName: amq-broker-rhel9
    provider:
      name: Red Hat
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Alvearie
      provider-url: https://alvearie.github.io
    name: alvearie-imaging-ingestion
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: imaging-ingestion-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventBridge",
                "metadata": {
                  "name": "events"
                },
                "spec": {
                  "dicomEventDrivenIngestionName": "core",
                  "eventBridge": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "events",
                  "natsTokenSecret": "nats-events-secure-bound-token",
                  "natsUrl": "jetstream.imaging-ingestion.svc.cluster.local:4222",
                  "role": "hub"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomEventDrivenIngestion",
                "metadata": {
                  "name": "core"
                },
                "spec": {
                  "databaseConfigName": "db-config",
                  "databaseSecretName": "db-secret",
                  "eventProcessor": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "imagePullPolicy": "Always"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomInstanceBinding",
                "metadata": {
                  "name": "stow"
                },
                "spec": {
                  "bindingConfigName": "instance-binding-config",
                  "bindingSecretName": "instance-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "instanceBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomStudyBinding",
                "metadata": {
                  "name": "fhir"
                },
                "spec": {
                  "bindingConfigName": "study-binding-config",
                  "bindingSecretName": "study-binding-secret",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "studyBinding": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DicomwebIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "imagePullPolicy": "Always",
                  "providerName": "provider",
                  "stowService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  },
                  "wadoService": {
                    "concurrency": 0,
                    "maxReplicas": 3,
                    "minReplicas": 0
                  }
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseIngestionService",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DICOM-INGEST",
                  "bucketConfigName": "imaging-ingestion",
                  "bucketSecretName": "imaging-ingestion",
                  "dicomEventDrivenIngestionName": "core",
                  "dimseService": {},
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "providerName": "provider"
                }
              },
              {
                "apiVersion": "imaging-ingestion.alvearie.org/v1alpha1",
                "kind": "DimseProxy",
                "metadata": {
                  "name": "ingestion"
                },
                "spec": {
                  "applicationEntityTitle": "DCM4CHEE",
                  "imagePullPolicy": "Always",
                  "natsSecure": true,
                  "natsSubjectRoot": "DIMSE",
                  "natsTokenSecret": "ingestion-nats-secure-bound-token",
                  "natsUrl": "nats-secure.imaging-ingestion.svc.cluster.local:4222",
                  "proxy": {},
                  "targetDimseHost": "arc.dcm4chee.svc.cluster.local",
                  "targetDimsePort": 11112
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: alvearie/imaging-ingestion-operator:0.0.3
          createdAt: "2021-04-08T15:29:55Z"
          description: The Alvearie Imaging Ingestion provides a collection of components
            for extending enterprise medical imaging informatics systems to Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Alvearie/imaging-ingestion
          support: alvearie.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DicomEventBridge is the Schema for the dicomeventbridges
              API
            displayName: Dicom Event Bridge
            kind: DicomEventBridge
            name: dicomeventbridges.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Event driven manifest of all DICOM data across all associated
              S3 buckets
            displayName: Dicom Event Driven Ingestion
            kind: DicomEventDrivenIngestion
            name: dicomeventdriveningestions.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM instance data (STOW-RS)
            displayName: Dicom Instance Binding
            kind: DicomInstanceBinding
            name: dicominstancebindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Fan-out notification of DICOM studies (FHIRv4 ImagingStudy)
            displayName: Dicom Study Binding
            kind: DicomStudyBinding
            name: dicomstudybindings.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides DICOMweb WADO-RS and STOW-RS access to a S3 bucket
            displayName: Dicomweb Ingestion Service
            kind: DicomwebIngestionService
            name: dicomwebingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a proxied DIMSE Application Entity (AE) in the cluster
              for C-STORE operations to a storage space
            displayName: Dimse Ingestion Service
            kind: DimseIngestionService
            name: dimseingestionservices.imaging-ingestion.alvearie.org
            version: v1alpha1
          - description: Provides a bidirectional proxied DIMSE Application Entity
              (AE) in the cluster
            displayName: Dimse Proxy
            kind: DimseProxy
            name: dimseproxies.imaging-ingestion.alvearie.org
            version: v1alpha1
        description: |
          The Alvearie Imaging Ingestion provides a collection of components for extending enterprise medical imaging informatics systems to Kubernetes.

          ## About this Operator
          This operator enables standards-based medical imaging application interfaces such as DICOM Message Service Element (DIMSE) and DICOMweb.
          The API's provided allow for extending enterprise imaging systems to the cloud with custom, fit for purpose deployments.
          This is possible at speed, low cost, and little or zero integration effort.or extending enterprise medical imaging informatics systems to a modern cloud.


          Some potential usages:

          - Perform lightweight secure bridging of DICOM from the enterprise imaging zone to a cloud availability zone.
          - Utilize S3 buckets for segregating DICOM data to be used for clinical or scientific purposes.
          - Fan-out DICOM data to a number of different DICOM subscribers
          - Generate FHIRv4 *ImagingStudy* resources from ingested content and publish to a FHIRv4 based patient logitudinal record.
          - Distribute imaging study arrival and update notifications to imaging study subscribers


          ## Prerequisites

          - [Knative](https://operatorhub.io/operator/knative-operator)
          - ProgresSQL such as [Crunchy](https://operatorhub.io/operator/postgresql)
          - S3 storage such as [MinIO](https://operatorhub.io/operator/minio-operator)
          - [NATS operator](https://github.com/nats-io/nats-operator/) (Optional:  only required for DIMSE Proxy component)



          ## Documentation

          Visit the [official github repository](https://github.com/Alvearie/imaging-ingestion) for documentation and examples.
        displayName: Alvearie Imaging Ingestion Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - dicom
        - medical imaging
        - fhir
        - dimse-c
        - dimse
        - imaging study
        links:
        - name: Alvearie Imaging Ingestion Operator
          url: https://github.com/Alvearie/imaging-ingestion
        maintainers:
        - email: rduggan@ca.ibm.com
          name: Richard Duggan
        - email: jjacob@us.ibm.com
          name: Jeesmon Jacob
        maturity: alpha
        provider:
          name: Alvearie
          url: https://alvearie.github.io
        relatedImages:
        - alvearie/imaging-ingestion-operator:0.0.3
        - quay.io/openshift-community-operators/alvearie-imaging-ingestion@sha256:f65cc3cb83b1c75385970ce70c6a84b15b143e682ad0d1784148f1643df0991e
        version: 0.0.3
      entries:
      - name: imaging-ingestion-operator.v0.0.3
        version: 0.0.3
      - name: imaging-ingestion-operator.v0.0.2
        version: 0.0.2
      - name: imaging-ingestion-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alvearie-imaging-ingestion
    provider:
      name: Alvearie
      url: https://alvearie.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: project-quay
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.10.0-nightly.20231014
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
          createdAt: 2023-10-14 06:26 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.10.0-nightly.20231014'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.0-nightly.20231014
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:add5d14653f57ac1e3254c2d1c70fc9232959f76ee6286fe5cb572696a5a4fd3
        - quay.io/projectquay/quay-operator:3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      entries:
      - name: quay-operator.v3.10.0-nightly.20231014
        version: 3.10.0-nightly.20231014
      - name: quay-operator.v3.10.0-nightly.20231013
        version: 3.10.0-nightly.20231013
      - name: quay-operator.v3.10.0-nightly.20231012
        version: 3.10.0-nightly.20231012
      - name: quay-operator.v3.10.0-nightly.20231011
        version: 3.10.0-nightly.20231011
      - name: quay-operator.v3.10.0-nightly.20231010
        version: 3.10.0-nightly.20231010
      - name: quay-operator.v3.10.0-nightly.20231009
        version: 3.10.0-nightly.20231009
      - name: quay-operator.v3.10.0-nightly.20231008
        version: 3.10.0-nightly.20231008
      - name: quay-operator.v3.10.0-nightly.20231007
        version: 3.10.0-nightly.20231007
      - name: quay-operator.v3.10.0-nightly.20231006
        version: 3.10.0-nightly.20231006
      - name: quay-operator.v3.10.0-nightly.20231005
        version: 3.10.0-nightly.20231005
      - name: quay-operator.v3.10.0-nightly.20231002
        version: 3.10.0-nightly.20231002
      - name: quay-operator.v3.10.0-nightly.20230930
        version: 3.10.0-nightly.20230930
      - name: quay-operator.v3.10.0-nightly.20230929
        version: 3.10.0-nightly.20230929
      - name: quay-operator.v3.10.0-nightly.20230928
        version: 3.10.0-nightly.20230928
      - name: quay-operator.v3.10.0-nightly.20230926
        version: 3.10.0-nightly.20230926
      - name: quay-operator.v3.10.0-nightly.20230925
        version: 3.10.0-nightly.20230925
      - name: quay-operator.v3.10.0-nightly.20230923
        version: 3.10.0-nightly.20230923
      - name: quay-operator.v3.10.0-nightly.20230922
        version: 3.10.0-nightly.20230922
      - name: quay-operator.v3.10.0-nightly.20230921
        version: 3.10.0-nightly.20230921
      - name: quay-operator.v3.10.0-nightly.20230920
        version: 3.10.0-nightly.20230920
      - name: quay-operator.v3.10.0-nightly.20230919
        version: 3.10.0-nightly.20230919
      - name: quay-operator.v3.10.0-nightly.20230918
        version: 3.10.0-nightly.20230918
      - name: quay-operator.v3.10.0-nightly.20230917
        version: 3.10.0-nightly.20230917
      - name: quay-operator.v3.10.0-nightly.20230916
        version: 3.10.0-nightly.20230916
      - name: quay-operator.v3.10.0-nightly.20230915
        version: 3.10.0-nightly.20230915
      - name: quay-operator.v3.10.0-nightly.20230914
        version: 3.10.0-nightly.20230914
      - name: quay-operator.v3.10.0-nightly.20230913
        version: 3.10.0-nightly.20230913
      name: candidate-3.10
    - currentCSV: quay-operator.v3.9.0-nightly.20230812
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
          createdAt: 2023-08-12 06:06 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.9.0-nightly.20230812'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.0-nightly.20230812
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:7fc4d97b5ee63085672fc9a562ebdad3d4c00d1b555e06679d1dd9ac945f1bb9
        - quay.io/projectquay/quay-operator:3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      entries:
      - name: quay-operator.v3.9.0-nightly.20230812
        version: 3.9.0-nightly.20230812
      - name: quay-operator.v3.9.0-nightly.20230724
        version: 3.9.0-nightly.20230724
      - name: quay-operator.v3.9.0-nightly.20230723
        version: 3.9.0-nightly.20230723
      - name: quay-operator.v3.9.0-nightly.20230722
        version: 3.9.0-nightly.20230722
      - name: quay-operator.v3.9.0-nightly.20230721
        version: 3.9.0-nightly.20230721
      - name: quay-operator.v3.9.0-nightly.20230720
        version: 3.9.0-nightly.20230720
      - name: quay-operator.v3.9.0-nightly.20230717
        version: 3.9.0-nightly.20230717
      - name: quay-operator.v3.9.0-nightly.20230716
        version: 3.9.0-nightly.20230716
      - name: quay-operator.v3.9.0-nightly.20230715
        version: 3.9.0-nightly.20230715
      - name: quay-operator.v3.9.0-nightly.20230714
        version: 3.9.0-nightly.20230714
      - name: quay-operator.v3.9.0-nightly.20230712
        version: 3.9.0-nightly.20230712
      - name: quay-operator.v3.9.0-nightly.20230711
        version: 3.9.0-nightly.20230711
      name: candidate-3.9
    - currentCSV: quay-operator.v3.7.0-rc.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.0-rc.4
          createdAt: 2022-05-17 14:05 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.0-rc.4'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.0-rc.4
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:05e12cc4caaebb7f185f651f8ca876c442c02d79794e561b743beba0d189fa04
        - quay.io/projectquay/quay-operator:3.7.0-rc.4
        version: 3.7.0-rc.4
      entries:
      - name: quay-operator.v3.7.0-rc.4
        version: 3.7.0-rc.4
      - name: quay-operator.v3.7.0-rc.3
        version: 3.7.0-rc.3
      - name: quay-operator.v3.7.0-rc.2
        version: 3.7.0-rc.2
      - name: quay-operator.v3.7.0-rc.1
        version: 3.7.0-rc.1
      name: preview-3.7
    - currentCSV: quay-operator.v3.8.0-rc.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.0-rc.7
          createdAt: 2022-11-29 15:11 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.0-rc.7'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.0-rc.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:84796005b6e71efa5d783c55b1f72bcc3762944bf32740cb75498bf8b908bffb
        - quay.io/projectquay/quay-operator:3.8.0-rc.7
        version: 3.8.0-rc.7
      entries:
      - name: quay-operator.v3.8.0-rc.7
        version: 3.8.0-rc.7
      - name: quay-operator.v3.8.0-rc.6
        version: 3.8.0-rc.6
      - name: quay-operator.v3.8.0-rc.5
        version: 3.8.0-rc.5
      - name: quay-operator.v3.8.0-rc.4
        version: 3.8.0-rc.4
      - name: quay-operator.v3.8.0-rc.3
        version: 3.8.0-rc.3
      - name: quay-operator.v3.8.0-rc.2
        version: 3.8.0-rc.2
      - name: quay-operator.v3.8.0-rc.1
        version: 3.8.0-rc.1
      name: preview-3.8
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.10.15
          createdAt: "2025-09-25T14:06:05Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:424e138a17e2af91603e834ff65d529400d6793771ddaab92d3988f4e6d2f2e3
        - quay.io/projectquay/quay-operator:3.10.15
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.11.13
          createdAt: "2025-09-02T06:02:15Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.11.13
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:9d505bf561650eb9785811ba50c920480d85679eac6a611a1c5794544e175934
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.12.12
          createdAt: "2025-09-23T09:59:30Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3abd8c3f97c9bba127b6e8c7efcc078c90c9d7cc35d74cfcd9bc0d303f62d2ed
        - quay.io/projectquay/quay-operator:3.12.12
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.13.8
          createdAt: "2025-08-28T03:28:40Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:21c17448932f053f1d4845111288a317786b190154bae9da7947c315ba12f796
        - quay.io/projectquay/quay-operator:3.13.8
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.14.5
          createdAt: "2025-09-22T14:19:49Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:023b9cc1e4ec2542b8d68b636b21a2e8c4eb143424581c62d16598fddb47da6a
        - quay.io/projectquay/quay-operator:3.14.5
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.15.2
          createdAt: "2025-09-22T13:44:32Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/projectquay/quay-operator:3.15.2
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:3843fd48c57be102576db93e8483a30f8efe8cd46b5e09c704ac0bc5f18553a3
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"quay.redhat.com/v1\",\n    \"kind\":
            \"QuayRegistry\",\n    \"metadata\": {\n      \"name\": \"example-registry\"\n
            \   },\n    \"spec\": {\n      \"components\": [\n        {\"kind\": \"clair\",
            \"managed\": true},\n        {\"kind\": \"postgres\", \"managed\": true},\n
            \       {\"kind\": \"objectstorage\", \"managed\": true},\n        {\"kind\":
            \"redis\", \"managed\": true},\n        {\"kind\": \"horizontalpodautoscaler\",
            \"managed\": true},\n        {\"kind\": \"route\", \"managed\": true},\n
            \       {\"kind\": \"mirror\", \"managed\": true},\n        {\"kind\":
            \"monitoring\", \"managed\": true},\n        {\"kind\": \"tls\", \"managed\":
            true}\n      ]\n    }\n  }, \n  {\n    \"apiVersion\": \"redhatcop.redhat.io/v1alpha1\",\n
            \   \"kind\": \"QuayEcosystem\",\n    \"metadata\": {\n      \"name\":
            \"example-quayecosystem\"\n    },\n    \"spec\": {}\n  }\n]"
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:v3.6.2
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.5.x <3.6.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.6.2
          repository: https://github.com/quay/quay-operator
          support: Quay Team
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Project Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        - quay
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:55bb4133abc945ce7b75e6c01c6beaf63c19c129fc8fbfb48f1b53b8e28aae97
        - quay.io/projectquay/quay-operator:v3.6.2
        version: 3.6.2
      entries:
      - name: quay-operator.v3.6.2
        version: 3.6.2
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.7.11
          createdAt: 2023-02-22 10:02 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.3.x <3.7.11'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.7.11
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:43b959fcdcf107b70e4224d80b63d1c123560c3af055109bbcc7d68476546584
        - quay.io/projectquay/quay-operator:3.7.11
        version: 3.7.11
      entries:
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      - name: quay-operator.v3.7.9
        version: 3.7.9
      - name: quay-operator.v3.7.8
        version: 3.7.8
      - name: quay-operator.v3.7.7
        version: 3.7.7
      - name: quay-operator.v3.7.6
        version: 3.7.6
      - name: quay-operator.v3.7.5
        version: 3.7.5
      - name: quay-operator.v3.7.4
        version: 3.7.4
      - name: quay-operator.v3.7.3
        version: 3.7.3
      - name: quay-operator.v3.7.2
        version: 3.7.2
      - name: quay-operator.v3.7.1
        version: 3.7.1
      - name: quay-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.8.13
          createdAt: 2023-10-17 16:13 UTC
          description: Opinionated deployment of Quay on Kubernetes.
          olm.skipRange: '>=3.6.x <3.8.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.8.13
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/project-quay@sha256:2845843f32b895dee0242e31380b44625e3f8381caee7ecb3258648311c302f7
        - quay.io/projectquay/quay-operator:3.8.13
        version: 3.8.13
      entries:
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/projectquay/quay-operator:3.9.16
          createdAt: "2025-09-25T14:31:58Z"
          description: Opinionated deployment of Quay on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: Opinionated deployment of Quay on Kubernetes.
        displayName: Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: quay-sig@googlegroups.com
          name: Project Quay Contributors
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/project-quay@sha256:c7cd3ffa88eae188e94abcc0af56d328aff43463b98a3bc3c0b574f29373944b
        - quay.io/projectquay/quay-operator:3.9.16
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: project-quay
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator3.v3.0.0-tp.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
          createdAt: "2024-12-12T10:28:38Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: IstioRevisionTag references a Istio or IstioRevision object
              and serves as an alias for sidecar injection.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-rhel9-operator@sha256:f5dfe495ba92ecb880485b66cfafbf6451f8142b8e0d5fd7f29db94171c51336
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:c21481d6dae58b82dac93b1b47b9d9f53a30617fa9d930a0b5d075620cffa74b
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-proxyv2-rhel9@sha256:cbd6bc47341ab8618609f89d063552e17232541c727429855c01deab73e4d06c
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:e3978aecaafe2678355e21a7303f8620f53d8b774f6592f10527c15bae957224
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:ea9372fabe06cd8e65ea7b4567ce3c8fbddd9d673dc34a12743fd95c548059d1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-must-gather-rhel9@sha256:2bf3704ba8d5945973ccf6bd13418457e929836352acf0ea6a47976ffa246616
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-operator-bundle@sha256:f596dc9323da89f5e45e71192f0fa698b7bb29b8205eaa0fc1659d68fa7401f4
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-pilot-rhel9@sha256:376f64eab0ebeb06035983a60d1e7647c5d41354f31a49cb34f4deb7b6708e3e
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:d0ab36246c556249b6d67346eb0c72ebd22483a6502caa0fc75112a8d327b053
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-cni-rhel9@sha256:f16cd8a29ddcbb3de77cbfcf7e1da78e42cc0c1c502132d477a01c52689cdf23
        version: 3.0.0-tp.2
      entries:
      - name: servicemeshoperator3.v3.0.0-tp.2
        version: 3.0.0-tp.2
      - name: servicemeshoperator3.v3.0.0-tp.1
        version: 3.0.0-tp.1
      name: candidates
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable
    - currentCSV: servicemeshoperator3.v3.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.6"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.6"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
          createdAt: "2025-09-03T07:53:34Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:d716f3558fbdf8514f225de708135ba87fdcac4c718ef101f394daf546514afb
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:8e223a537f487bc3cef10769adc6fb619085dafe3ea797cfd52c20e3066f694f
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        version: 3.0.5
      entries:
      - name: servicemeshoperator3.v3.0.5
        version: 3.0.5
      - name: servicemeshoperator3.v3.0.4
        version: 3.0.4
      - name: servicemeshoperator3.v3.0.3
        version: 3.0.3
      - name: servicemeshoperator3.v3.0.2
        version: 3.0.2
      - name: servicemeshoperator3.v3.0.1
        version: 3.0.1
      - name: servicemeshoperator3.v3.0.0
        version: 3.0.0
      name: stable-3.0
    - currentCSV: servicemeshoperator3.v3.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.4"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.4"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
          createdAt: "2025-09-10T06:22:42Z"
          description: The OpenShift Service Mesh Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh. OpenShift
            Service Mesh is based on the open source Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          Red Hat OpenShift Service Mesh is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### More Information

          * [Documentation](https://docs.redhat.com/en/documentation/red_hat_openshift_service_mesh/latest/html/about/ossm-about-openshift-service-mesh)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 3
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:4434ef90ad92a9031b0fb1a9dae0b283cfb40fc095f7b6d3682a85e61ef2c5b4
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d518f3d1539f45e1253c5c9fa22062802804601d4998cd50344e476a3cc388fe
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:7ea9b82e192402566e69063a4787351be9f1ef50719bfd1a8f5d5940362b3f70
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:d757e700a124d7583b9456ec9e08e5e7e249c1de202bc4e973102922b9f69191
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:8054eebd5226912aa60f000938c2efb69db235ddb7297708450aec1c9f9c7e0b
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2ea112ab90b8540f11e9949d77c3e7e3b3ef57ac3bf23f6cf1e883a88430e1f9
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:b757b81d34711cb8abfcb0e8896c023d39d4778f237dd80b637ba4bf613df893
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:e44a257de52e1778350f983f851465096caa3c00b21be80b8b2505f413525c91
        - registry.redhat.io/openshift-service-mesh-dev-preview-beta/istio-ztunnel-rhel9@sha256:45321183af38ef306276a879505ed0bc4cf58d6bda3bcc915cad778d489ac62f
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:366266f10e658c4cea86bddf6ee847e1908ea4dd780cb5f7fb0e466bac8995f1
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:26747627ae22bbdffcf9de58077454fc0c890cda83659d7834b15dea2b5aaaf2
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:6881fb9f5a4b447be4c468b8585fff86171bd3d82671a48559f3ceabcb9dad24
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:33feac0c9c888b375e207e86b3f9a1a289126ca46b64632a7f693e2191cfbbfd
        - registry.redhat.io/openshift-service-mesh/istio-sail-operator-bundle@sha256:d4ce6357b07ea13906b157c38766a5ac100bfe5dec5dab7730bfe8289445e1f6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:f592b91a6f28559abdf6546ff6c3ae70a44d8f811592746e9eb02cdb2e5dec9d
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:2b60b0f33affda8b5b2cc30a4b09b2198b053fb1a914d386d35fcf03c09234e6
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:7fa655f5efb1175ff1e1c138371fc1233e5d4313c5feb07194428d0d1fdd33a3
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:d6693c31d6137c4219812cf69097d04654b296e0ab059dc48596a09b9cd124b1
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:2ca96716812d339775d3dfd396c863834dd3810720e98bfcf87bc67c5fbd29b5
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel9@sha256:6bb44c84363a4113c5bce2ae0d6528b345fddd12bc559bbfaacf6222060ddcb0
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:17db00a219de14e5901b5a5eca64d3c6684ca74c1dba3ae3b03787c655fe9c80
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:ecc6a22a471f4f4b836648f3a72ea7f57a4ea960ebcbc7dbe3fe729500902d0b
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:36557fa4817c3d0bac499dc65a4a9d6673500681641943d2f8cec5bfec4355be
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:54cada48e5c9824f255f82daa2ef5bea236919e521d3ea49885f2883ced2b7bc
        - registry.redhat.io/openshift-service-mesh/istio-rhel9-operator@sha256:3c031fc57f78e32f203f8ae0b21215cd323ba522d8fb57d6b8f956d95b757e9f
        - registry.redhat.io/openshift-service-mesh-tech-preview/istio-ztunnel-rhel9@sha256:f7ab868455b5e887e2d89b7078678531290365f9166cae99ce56119f31fa2ba4
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:0d6930075b8906bee0ab1aeac82f6e2431c16163cf87f7f0c68c10b2c40552a1
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:07477329f2043adb27a609966f90cda6afcc67484f9cd52207eb3dd3f21a1dcd
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:87b967785d7cc222f9df9cb49f0373a9819bf67910ce523dc3b8345849e881dd
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel9@sha256:14c5a52faf20267baa43d9a72ee6416f56fbc41dd640adc2aba3c4043802a0e9
        - registry.redhat.io/openshift-service-mesh/istio-pilot-rhel9@sha256:028e10651db0d1ddb769a27c9483c6d41be6ac597f253afd9d599f395d9c82d8
        - registry.redhat.io/openshift-service-mesh/istio-proxyv2-rhel9@sha256:a020acb4a74040df03077924e10465454092f6c34a902b391fba55563c08f826
        version: 3.1.2
      entries:
      - name: servicemeshoperator3.v3.1.2
        version: 3.1.2
      - name: servicemeshoperator3.v3.1.1
        version: 3.1.1
      - name: servicemeshoperator3.v3.1.0
        version: 3.1.0
      name: stable-3.1
    defaultChannel: stable
    packageName: servicemeshoperator3
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.33
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle-rhmp@sha256:81476f897726c85293e64e6188ed314e3e1e637c568c6b0ff8f475a876bee29c
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        version: 0.0.33
      entries:
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: spyre-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spyre-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "icr.io/ibmaiu_internal",
                      "version": "0.1.1"
                    },
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.2"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "icr.io/ibmaiu_internal",
                    "version": "0.1.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
          createdAt: "2025-07-31T10:32:48Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu_internal/spyre-operator@sha256:0f657c98668a8358c37b3fc2ed3f31ec78b7ef7f9a9678e7cd35d383d22ce2da
        - icr.io/ibmaiu_internal/x86_64/release/dd2/e2e_stable@sha256:bcb685a542492160e2b34c6a7f8b05f202a01be6476827e047bba53d6792cbe2
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:4ae986761a69d25697b9b6a3a99de88d8587beba136c3de29b90624e2ea6738f
        version: 0.1.2
      entries:
      - name: spyre-operator.v0.1.2
        version: 0.1.2
      name: stable-v0.1
    - currentCSV: spyre-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreClusterPolicy",
                "metadata": {
                  "name": "spyreclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb",
                      "spyreFilter": ".",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "latest"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "spyre-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "spyre-device-plugin-init",
                      "repository": "quay.io/ibm-aiu",
                      "version": "0.2.0"
                    },
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "spyre-exporter",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "spyre-webhook-validator",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.0"
                  },
                  "scheduler": {
                    "image": "spyre-scheduler",
                    "repository": "quay.io/ibm-aiu",
                    "version": "0.2.1"
                  }
                }
              },
              {
                "apiVersion": "spyre.ibm.com/v1alpha1",
                "kind": "SpyreNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
          createdAt: "2025-09-02T13:25:34Z"
          description: Operator for IBM Spyre devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM Spyre
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["spyrenodestates.spyre.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpyreClusterPolicy is the Schema for the Spyre API
            displayName: Spyre Cluster Policy
            kind: SpyreClusterPolicy
            name: spyreclusterpolicies.spyre.ibm.com
            version: v1alpha1
          - description: SpyreNodeState is the Schema for the SpyreNodeState API
            displayName: Spyre Node State
            kind: SpyreNodeState
            name: spyrenodestates.spyre.ibm.com
            version: v1alpha1
        description: Operator for IBM Spyre devices
        displayName: IBM Spyre Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: IBM Spyre Operator
          url: https://spyre-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: Satyam.Gupta.3@ibm.com
          name: Satyam Gupta
        - email: Rehan.Khan7@ibm.com
          name: Rehan Khan
        - email: Nishan.Acharya@ibm.com
          name: Nishan Acharya
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/release/dd2/spyredriver-image@sha256:0d098b4be6eeee55fb19b922ad623a3138c0495543e72b01f61378b0b0c17bfb
        - quay.io/ibm-aiu/spyre-operator@sha256:d0cdd449f1d8f0dd9281c4d9a809333db6c4ed201bb14187b6c70ae686a30e99
        - registry.connect.redhat.com/aiu/spyre-operator-bundle@sha256:adbfebc73bc5a9cd12d17e9e1fbdf87c2176aae7a7217bcefcec3068e51bc904
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.2.1
      entries:
      - name: spyre-operator.v0.2.1
        version: 0.2.1
      name: stable-v0.2
    defaultChannel: stable-v0.2
    packageName: spyre-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
          createdAt: "2025-09-19T19:15:31Z"
          description: AWS Route53 controller is a service controller for managing
            Route53 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HealthCheck represents the state of an AWS route53 HealthCheck
              resource.
            displayName: HealthCheck
            kind: HealthCheck
            name: healthchecks.route53.services.k8s.aws
            version: v1alpha1
          - description: HostedZone represents the state of an AWS route53 HostedZone
              resource.
            displayName: HostedZone
            kind: HostedZone
            name: hostedzones.route53.services.k8s.aws
            version: v1alpha1
          - description: RecordSet represents the state of an AWS route53 RecordSet
              resource.
            displayName: RecordSet
            kind: RecordSet
            name: recordsets.route53.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Route53 resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53**

          Amazon Route 53 is a highly available and scalable Domain Name System (DNS) web service. You can use Route 53 to perform three main functions in any combination: domain registration, DNS routing, and health checking.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53 Developer Resources
          url: https://docs.aws.amazon.com/route53/index.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53-controller@sha256:227a306cc90998489bfb681cb29b416238b404bc2db28a09b8949d4cca2d87a4
        version: 1.1.1
      entries:
      - name: ack-route53-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53-controller.v1.1.0
        version: 1.1.0
      - name: ack-route53-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53-controller.v0.1.0
        version: 0.1.0
      - name: ack-route53-controller.v0.0.28
        version: 0.0.28
      - name: ack-route53-controller.v0.0.27
        version: 0.0.27
      - name: ack-route53-controller.v0.0.26
        version: 0.0.26
      - name: ack-route53-controller.v0.0.25
        version: 0.0.25
      - name: ack-route53-controller.v0.0.24
        version: 0.0.24
      - name: ack-route53-controller.v0.0.23
        version: 0.0.23
      - name: ack-route53-controller.v0.0.22
        version: 0.0.22
      - name: ack-route53-controller.v0.0.20
        version: 0.0.20
      - name: ack-route53-controller.v0.0.19
        version: 0.0.19
      - name: ack-route53-controller.v0.0.18
        version: 0.0.18
      - name: ack-route53-controller.v0.0.17
        version: 0.0.17
      - name: ack-route53-controller.v0.0.16
        version: 0.0.16
      - name: ack-route53-controller.v0.0.15
        version: 0.0.15
      - name: ack-route53-controller.v0.0.14
        version: 0.0.14
      - name: ack-route53-controller.v0.0.13
        version: 0.0.13
      - name: ack-route53-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53-controller.v0.0.6
        version: 0.0.6
      - name: ack-route53-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53-controller.v0.0.3
        version: 0.0.3
      - name: ack-route53-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: rhcl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhcl-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "authpolicy-sample"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        },
                        "credentials": {
                          "authorizationHeader": {
                            "prefix": "APIKEY"
                          }
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "protocol": "HTTP"
                  },
                  "providerRefs": [
                    {
                      "name": "provider-ref"
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "ratelimitpolicy-sample"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "limit": 50,
                          "window": "1m"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
          createdAt: "2025-04-28T13:10:44Z"
          description: Red Hat Connectivity Link enables you to secure, protect, and
            connect your APIs and applications in multicluster, multicloud, and hybrid
            cloud environments
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1
        description: Red Hat Connectivity Link enables you to secure, protect, and
          connect your APIs and applications in multicluster, multicloud, and hybrid
          cloud environments
        displayName: Red Hat Connectivity Link
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: GitHub
          url: https://github.com/kuadrant/kuadrant-operator
        - name: Documentation
          url: https://github.com/Kuadrant/kuadrant-operator/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/wasm-shim-rhel9@sha256:845bb8af57f3d219aa09b9c0bb20fa945306f3b413e34e196760c8cdf624a532
        - registry.redhat.io/rhcl-1/rhcl-console-plugin-rhel9@sha256:ae92d638d993d782986e761d5aa3c7c5da2b8af0aabc1decb332a4f1487dacd3
        - registry.redhat.io/rhcl-1/rhcl-operator-bundle@sha256:e83707d0b396c38cf55046d3159f1f427188404aa11dc47d7b1c1f25605182de
        - registry.redhat.io/rhcl-1/rhcl-rhel9-operator@sha256:92f14fd8de74bc72c407947cf5fa4739a7b455e5d351baeb64521e4c07f02a9e
        version: 1.1.0
      entries:
      - name: rhcl-operator.v1.1.0
        version: 1.1.0
      - name: rhcl-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: rhcl-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.3-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lvms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lvms-operator.v4.17.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:64a623d5d536c790dee04c1604ec2f4429f8b8f1973b4232b543e8cbe23d6590
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:4a0c845c5d1597277a5fa346abbac1b91ab2a24fce932f2a328a84dcf8be37e2
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:7c9b4327d8677f48c48888230fc38eec8c20c7f646dde568dbd96b5e4c8f7967
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:8797a3be94ce6f4e4da15e4efefe41ea5b6ffb4cdc6d698638c80976bb22491c
        version: 4.17.6
      entries:
      - name: lvms-operator.v4.17.6
        version: 4.17.6
      - name: lvms-operator.v4.17.5
        version: 4.17.5
      - name: lvms-operator.v4.17.4
        version: 4.17.4
      - name: lvms-operator.v4.17.3
        version: 4.17.3
      - name: lvms-operator.v4.17.2
        version: 4.17.2
      - name: lvms-operator.v4.17.1
        version: 4.17.1
      - name: lvms-operator.v4.17.0
        version: 4.17.0
      name: stable-4.17
    - currentCSV: lvms-operator.v4.18.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "my-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "default": true,
                        "fstype": "xfs",
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
          description: Logical volume manager storage provides dynamically provisioned
            local storage for container workloads
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
                "apiVersion": "lvm.topolvm.io/v1alpha1",
                "kind": "LVMCluster",
                "metadata": {
                  "name": "test-lvmcluster"
                },
                "spec": {
                  "storage": {
                    "deviceClasses": [
                      {
                        "name": "vg1",
                        "thinPoolConfig": {
                          "name": "thin-pool-1",
                          "overprovisionRatio": 10,
                          "sizePercent": 90
                        }
                      }
                    ]
                  }
                }
              }
          operatorframework.io/suggested-namespace: openshift-storage
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-storage",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["logicalvolumes.topolvm.io",
            "lvmvolumegroups.lvm.topolvm.io", "lvmvolumegroupnodestatuses.lvm.topolvm.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lvm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LogicalVolume
            name: logicalvolumes.topolvm.io
            version: v1
          - description: LVMCluster is the Schema for the lvmclusters API
            displayName: LVMCluster
            kind: LVMCluster
            name: lvmclusters.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroupNodeStatus
            name: lvmvolumegroupnodestatuses.lvm.topolvm.io
            version: v1alpha1
          - kind: LVMVolumeGroup
            name: lvmvolumegroups.lvm.topolvm.io
            version: v1alpha1
        description: Logical volume manager storage provides dynamically provisioned
          local storage.
        displayName: LVM Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - local storage
        - operator
        - LVM
        links:
        - name: Source Repository
          url: https://github.com/openshift/lvm-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/lvms4/lvms-must-gather-rhel9@sha256:5d0534e08790e91d5d355eac2e5c187a217c1ec6eb59464ea9a771490ac040e9
        - registry.redhat.io/lvms4/lvms-operator-bundle@sha256:6e809f8393e9c3004a9b8d80eb6ea708c0ab1e124083c481b48c01a359684588
        - registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:80f5e48600d9b42add118f1325ced01f49d544a9a4824a7da4e3ba805d64371f
        - registry.redhat.io/lvms4/topolvm-rhel9@sha256:b3b1c4836a82a7ec6c9b02e4a86c352f4ae8c45a3fe027ae7200869afe3ab028
        version: 4.18.3
      entries:
      - name: lvms-operator.v4.18.3
        version: 4.18.3
      - name: lvms-operator.v4.18.2
        version: 4.18.2
      - name: lvms-operator.v4.18.1
        version: 4.18.1
      - name: lvms-operator.v4.18.0
        version: 4.18.0
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: lvms-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/instana-agent-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-rhmp@sha256:1502ccda3c067d74689d0341008f05362bc23904c8e35b20e020fe0dca82aef7
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:ab8b35f0b45c68f69210d1b17c5dd609799f6bb0bc9f12d0d4e5302c910607a1
        version: 2.1.36
      entries:
      - name: instana-agent-operator.v2.1.36
        version: 2.1.36
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator-rhmp.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator-rhmp
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: https://www.nec.com
    name: sanstoragecsi-operator-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sanstoragecsi-operator-bundle.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "SANStorageCSI",
                "metadata": {
                  "name": "sanstoragecsi"
                },
                "spec": {
                  "driverName": "san.csi.nec.com",
                  "images": {
                    "csiDriver": "",
                    "externalAttacher": "registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff",
                    "externalProvisioner": "registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8",
                    "livenessProbe": "registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c",
                    "nodeRegistrar": "registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864"
                  },
                  "parameters": {
                    "controller": {
                      "logLevel": 5
                    },
                    "node": {
                      "livenessProbePort": 9808,
                      "logLevel": 5,
                      "maxVolumesPerNode": 1024
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          createdAt: "2022-06-15T00:00:00Z"
          description: An operator for managing the NEC Storage M Series CSI Driver
          operatorframework.io/suggested-namespace: nec-san-csi-operator-system
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SANStorageCSI is the Schema for the sanstoragecsis API
            displayName: NEC Storage M Series CSI Driver
            kind: SANStorageCSI
            name: sanstoragecsis.csi.nec.com
            version: v1
        description: |
          This operator deploys the NEC Storage M Series CSI Driver on OpenShift Container Platform.

          NEC Storage M Series CSI Driver provides persistent storage for stateful applications using NEC Storage M Series.
          Please refer to the documentation of the NEC Storage M Series CSI Driver for detail information about supported NEC Storage Platforms and supported CSI features.
        displayName: NEC Storage M Series CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - M Series
        - SAN Storage
        - CSI Driver
        links:
        - name: NEC Storage M Series
          url: https://www.nec.com/en/global/prod/storage/product/san/index.html
        maturity: stable
        provider:
          name: NEC
          url: https://www.nec.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:dc7f0992be6c2853dc0e798afda368e36005a94ad76b21fa2f09790621318864
        - registry.connect.redhat.com/nec/sanstoragecsi-operator-bundle@sha256:1957586c13e3bf80bdc688f519be440ec1dcc4ffe63343f526018dcb93e21cea
        - registry.connect.redhat.com/nec/sanstoragecsi-operator@sha256:6656c2b5036bdc013712ecb640bbbebe0e3e90cf625c73697dbb49e3e201e2ac
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:bd8262cc4b6c405e175c39623323dd98a74bbf70e6aa90851a72612056c184ff
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:85f1cc8f21568e7885e492b29660f29e09796fb23ebdf5fd526aa2e99d6b2fa8
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:1a0d15012ba70955af754b03a7488b7c6ace813646ec7c3ce845f80597050c3c
        version: 1.1.0
      entries:
      - name: sanstoragecsi-operator-bundle.v1.1.0
        version: 1.1.0
      - name: sanstoragecsi-operator-bundle.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: sanstoragecsi-operator-bundle
    provider:
      name: NEC
      url: https://www.nec.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/gitlab/gitlab-runner-operator-bundle@sha256:68b22c412a44d5fa09b45e204d7c8e7b6ca6afa4ed6879de608dbcb0b1929676
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
          createdAt: "2021-11-15"
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/openshift-community-operators/pcc-operator@sha256:e99a466ddb52ad44c435ccf952754d685744fb5b7b7e4cf77dd542e530e4f968
        - quay.io/prismacloud/pcc-operator@sha256:b8fcfbd6c51286c874e00db1bd35523386cec406fa4050ef44c0a887730cf9b8
        version: 0.2.0
      entries:
      - name: pcc-operator.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: prometheus-exporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometheus-exporter-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.3scale.net/v1alpha1",
                "kind": "PrometheusExporter",
                "metadata": {
                  "name": "example-memcached"
                },
                "spec": {
                  "dbHost": "your-memcached-host",
                  "dbPort": 11211,
                  "grafanaDashboard": {
                    "label": {
                      "key": "autodiscovery",
                      "value": "enabled"
                    }
                  },
                  "type": "memcached"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/3scale/prometheus-exporter-operator:v0.7.0
          createdAt: "2023-09-07 00:00:00"
          description: Operator to setup 3rd party prometheus exporters, with a collection
            of grafana dashboards
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/3scale-ops/prometheus-exporter-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a prometheus exporter to monitor a memcached instance
            displayName: PrometheusExporter
            kind: PrometheusExporter
            name: prometheusexporters.monitoring.3scale.net
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK to centralize the setup of 3rd party prometheus exporters on **Kubernetes/OpenShift**, with a collection of grafana dashboards.

          By just providing a few parameters like *dbHost* or *dbPort* (operator manage the container image, port, argument, command, volumes... and also prometheus `ServiceMonitor` and `GrafanaDashboard` k8s objects), you can setup different prometheus exporters to monitor:
          * The **internals from different databases**
          * **HTTP/TCP endpoints** (availability, latency, SSL/TLS certificate expiration...)
          * Any available **cloudwatch metric from any AWS Service**
          * Sendgrid email statistics (delivered, bounces, errors, spam...)

          Current prometheus exporters types supported, managed by same prometheus-exporter-operator:
          * memcached
          * redis
          * mysql
          * postgresql
          * sphinx
          * es (elasticsearch)
          * cloudwatch
          * probe (blackbox)
          * sendgrid

          The operator manages the lifecycle of the following objects:
          * Deployment (one per CR)
          * Service (one per CR)
          * ServiceMonitor (optional, one per CR)
          * GrafanaDashboard (optional, one per Namespace)

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale-ops/prometheus-exporter-operator#documentation).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale-ops/prometheus-exporter-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:
          * Raising any issues you find using Prometheus Exporter Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale-ops/prometheus-exporter-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving [documentation](https://github.com/3scale-ops/prometheus-exporter-operator)
          * Talking about Prometheus Exporter Operator
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale-ops/prometheus-exporter-operator/issues).

          ### License
          Prometheus Exporter Operator is licensed under the [Apache 2.0 license](https://github.com/3scale-ops/prometheus-exporter-operator/blob/main/LICENSE)
        displayName: Prometheus Exporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - metrics
        - observability
        - prometheus
        - exporter
        - grafana
        - mysql
        - postgresql
        - memcached
        - redis
        - sphinx
        - elasticsearch
        - cloudwatch
        - blackbox
        - sendgrid
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/prometheus-exporter-operator
        maintainers:
        - email: 3scale-operations@redhat.com
          name: 3scale Ops
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/3scale/prometheus-exporter-operator:v0.7.0
        - quay.io/openshift-community-operators/prometheus-exporter-operator@sha256:e62276f4c7e583d1fb24997fa970662f1a8d510de0baea159ba796b48cff709b
        version: 0.7.0
      entries:
      - name: prometheus-exporter-operator.v0.7.0
        version: 0.7.0
      - name: prometheus-exporter-operator.v0.3.4
        version: 0.3.4
      name: alpha
    defaultChannel: alpha
    packageName: prometheus-exporter-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-dynamodb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-dynamodb-controller.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "GlobalTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "dynamodb.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
          createdAt: "2025-09-19T19:19:47Z"
          description: AWS DynamoDB controller is a service controller for managing
            DynamoDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup represents the state of an AWS dynamodb Backup resource.
            displayName: Backup
            kind: Backup
            name: backups.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: GlobalTable represents the state of an AWS dynamodb GlobalTable
              resource.
            displayName: GlobalTable
            kind: GlobalTable
            name: globaltables.dynamodb.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS dynamodb Table resource.
            displayName: Table
            kind: Table
            name: tables.dynamodb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DynamoDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DynamoDB**

          Amazon DynamoDB is a fully managed NoSQL database service that provides fast and predictable performance with seamless scalability. DynamoDB lets you offload the administrative burdens of operating and scaling a distributed database so that you don't have to worry about hardware provisioning, setup and configuration, replication, software patching, or cluster scaling. DynamoDB also offers encryption at rest, which eliminates the operational burden and complexity involved in protecting sensitive data. For more information, see [DynamoDB Encryption at Rest](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/EncryptionAtRest.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DynamoDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dynamodb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DynamoDB Developer Resources
          url: https://aws.amazon.com/dynamodb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: dynamodb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/dynamodb-controller:1.5.2
        - quay.io/community-operator-pipeline-prod/ack-dynamodb-controller@sha256:547bdd3d1037dbc3b189f4c2f8e1d49379e1d547cd9063391d6c745e1a103b80
        version: 1.5.2
      entries:
      - name: ack-dynamodb-controller.v1.5.2
        version: 1.5.2
      - name: ack-dynamodb-controller.v1.5.1
        version: 1.5.1
      - name: ack-dynamodb-controller.v1.5.0
        version: 1.5.0
      - name: ack-dynamodb-controller.v1.4.10
        version: 1.4.10
      - name: ack-dynamodb-controller.v1.4.9
        version: 1.4.9
      - name: ack-dynamodb-controller.v1.4.7
        version: 1.4.7
      - name: ack-dynamodb-controller.v1.4.6
        version: 1.4.6
      - name: ack-dynamodb-controller.v1.4.5
        version: 1.4.5
      - name: ack-dynamodb-controller.v1.4.4
        version: 1.4.4
      - name: ack-dynamodb-controller.v1.4.3
        version: 1.4.3
      - name: ack-dynamodb-controller.v1.4.2
        version: 1.4.2
      - name: ack-dynamodb-controller.v1.4.1
        version: 1.4.1
      - name: ack-dynamodb-controller.v1.3.0
        version: 1.3.0
      - name: ack-dynamodb-controller.v1.2.23
        version: 1.2.23
      - name: ack-dynamodb-controller.v1.2.22
        version: 1.2.22
      - name: ack-dynamodb-controller.v1.2.21
        version: 1.2.21
      - name: ack-dynamodb-controller.v1.2.20
        version: 1.2.20
      - name: ack-dynamodb-controller.v1.2.15
        version: 1.2.15
      - name: ack-dynamodb-controller.v1.2.14
        version: 1.2.14
      - name: ack-dynamodb-controller.v1.2.13
        version: 1.2.13
      - name: ack-dynamodb-controller.v1.2.12
        version: 1.2.12
      - name: ack-dynamodb-controller.v1.2.11
        version: 1.2.11
      - name: ack-dynamodb-controller.v1.2.9
        version: 1.2.9
      - name: ack-dynamodb-controller.v1.2.8
        version: 1.2.8
      - name: ack-dynamodb-controller.v1.2.6
        version: 1.2.6
      - name: ack-dynamodb-controller.v1.2.5
        version: 1.2.5
      - name: ack-dynamodb-controller.v1.2.4
        version: 1.2.4
      - name: ack-dynamodb-controller.v1.2.2
        version: 1.2.2
      - name: ack-dynamodb-controller.v1.2.1
        version: 1.2.1
      - name: ack-dynamodb-controller.v1.2.0
        version: 1.2.0
      - name: ack-dynamodb-controller.v1.1.1
        version: 1.1.1
      - name: ack-dynamodb-controller.v1.1.0
        version: 1.1.0
      - name: ack-dynamodb-controller.v1.0.3
        version: 1.0.3
      - name: ack-dynamodb-controller.v1.0.2
        version: 1.0.2
      - name: ack-dynamodb-controller.v1.0.1
        version: 1.0.1
      - name: ack-dynamodb-controller.v1.0.0
        version: 1.0.0
      - name: ack-dynamodb-controller.v0.1.10
        version: 0.1.10
      - name: ack-dynamodb-controller.v0.1.9
        version: 0.1.9
      - name: ack-dynamodb-controller.v0.1.8
        version: 0.1.8
      - name: ack-dynamodb-controller.v0.1.7
        version: 0.1.7
      - name: ack-dynamodb-controller.v0.1.6
        version: 0.1.6
      - name: ack-dynamodb-controller.v0.1.5
        version: 0.1.5
      - name: ack-dynamodb-controller.v0.1.4
        version: 0.1.4
      - name: ack-dynamodb-controller.v0.1.3
        version: 0.1.3
      - name: ack-dynamodb-controller.v0.1.1
        version: 0.1.1
      - name: ack-dynamodb-controller.v0.1.0
        version: 0.1.0
      - name: ack-dynamodb-controller.v0.0.20
        version: 0.0.20
      - name: ack-dynamodb-controller.v0.0.19
        version: 0.0.19
      - name: ack-dynamodb-controller.v0.0.18
        version: 0.0.18
      - name: ack-dynamodb-controller.v0.0.17
        version: 0.0.17
      - name: ack-dynamodb-controller.v0.0.16
        version: 0.0.16
      - name: ack-dynamodb-controller.v0.0.15
        version: 0.0.15
      - name: ack-dynamodb-controller.v0.0.14
        version: 0.0.14
      - name: ack-dynamodb-controller.v0.0.13
        version: 0.0.13
      - name: ack-dynamodb-controller.v0.0.12
        version: 0.0.12
      - name: ack-dynamodb-controller.v0.0.11
        version: 0.0.11
      - name: ack-dynamodb-controller.v0.0.10
        version: 0.0.10
      - name: ack-dynamodb-controller.v0.0.9
        version: 0.0.9
      - name: ack-dynamodb-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-dynamodb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com
    name: aws-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0
    - currentCSV: aws-load-balancer-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
          createdAt: 2022-06-28T17:56:41
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:95c45fae0ca9e9bee0fa2c13652634e726d8133e4e3009b363fcae6814b3461d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:d7bc364512178c36671d8a4b5a76cf7cb10f8e56997106187b0fe1f032670ece
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:50b9402635dd4b312a86bed05dcdbda8c00120d3789ec2e9b527045100b3bdb4
        version: 0.0.1
      entries:
      - name: aws-load-balancer-operator.v0.0.1
        version: 0.0.1
      name: stable-v0.1
    - currentCSV: aws-load-balancer-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
          createdAt: 2022-12-09T15:32:32
          olm.skipRange: <0.2.0
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat << EOF| oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
            EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note:__ For UPI based installs, additional documentation can be found
          [here](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets)
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:ab38b37c14f7f0897e09a18eca4a232a6c102b76e9283e401baed832852290b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:cfd210a2cb705156883fda126d052e05eb84345cc27ba04688067f8ed9a802ea
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:3939323bc858e2a71b9ef0f858c5020ffd501e664a3652bd0be5f17176eb63f5
        version: 0.2.0
      entries:
      - name: aws-load-balancer-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1
    - currentCSV: aws-load-balancer-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
          createdAt: 2024-01-24T09:13:21
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify management of aws-load-balancer-controller

          ### Prerequisites for installation
          The operator requires AWS credentials in the installation namespace before it can be installed.

          1. Create the operator namespace with the following command:
          ```bash
          oc create namespace aws-load-balancer-operator
          ```
          2. Create the `CredentialsRequest` in the `openshift-cloud-credential-operator`
             namespace with the following command:

          ```
          cat <<EOF | oc create -f -
          apiVersion: cloudcredential.openshift.io/v1
          kind: CredentialsRequest
          metadata:
            name: aws-load-balancer-operator
            namespace: openshift-cloud-credential-operator
          spec:
            providerSpec:
              apiVersion: cloudcredential.openshift.io/v1
              kind: AWSProviderSpec
              statementEntries:
                - action:
                    - ec2:DescribeSubnets
                  effect: Allow
                  resource: "*"
                - action:
                    - ec2:CreateTags
                    - ec2:DeleteTags
                  effect: Allow
                  resource: arn:aws:ec2:*:*:subnet/*
                - action:
                    - ec2:DescribeVpcs
                  effect: Allow
                  resource: "*"
            secretRef:
              name: aws-load-balancer-operator
              namespace: aws-load-balancer-operator
            serviceAccountNames:
              - aws-load-balancer-operator-controller-manager
          EOF
          ```
          3. Ensure the credentials have been correctly provisioned
          ```bash
          oc get secret -n aws-load-balancer-operator aws-load-balancer-operator
          ```

          After this the operator can be installated through the console or through the command line.

          __Note about UPI:__ For UPI based installs, additional documentation can be found in
          [VPC and subnets](https://github.com/openshift/aws-load-balancer-operator/blob/release-1.0/docs/prerequisites.md#vpc-and-subnets)

          __Note about STS:__ For clusters using STS, use the [STS documentation](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#for-sts-clusters) instead.
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:73db97d20ee9415e9a8941cc6a36ed81d901d760d9949e95e10c90866ef8bf64
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:59ff4966f4f653e49d42d97bfed6351d3730128db74f666ee4ecacff11790baf
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:6e15df39a3d25755a79888f739e1af8b052c3fba57bf8f6b3434ef58fd4288cf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9f9f8f69ab98c1f25cc301572339e0cc5c5b76f2e29ed26e10ac91ad20e0090c
        version: 1.0.1
      entries:
      - name: aws-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: aws-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: aws-load-balancer-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
          createdAt: 2024-01-23T11:21:00
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:16e9ffed36107527a37713ac5bd34a7bc20f042269a81077429fb5884914c4d0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:2e0b9332a44d8d9c23e19c7accab0813a651f39210257820db508cac28876595
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:01f2ca529d2486f113bcefc9fedce6a6fd07bcb48aaf534394b5b04c353f8853
        version: 1.1.1
      entries:
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: aws-load-balancer-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.k8s.aws/v1alpha1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "IngressClassParams",
                "metadata": {
                  "name": "awesome-class"
                },
                "spec": {
                  "group": {
                    "name": "my-group"
                  }
                }
              },
              {
                "apiVersion": "elbv2.k8s.aws/v1beta1",
                "kind": "TargetGroupBinding",
                "metadata": {
                  "name": "my-tgb"
                },
                "spec": {
                  "serviceRef": {
                    "name": "awesome-service",
                    "port": 80
                  },
                  "targetGroupARN": "\u003carn-to-targetGroup\u003e"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              },
              {
                "apiVersion": "networking.olm.openshift.io/v1alpha1",
                "kind": "AWSLoadBalancerController",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "subnetTagging": "Auto"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
          createdAt: 2024-12-20T15:52:44
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: aws-load-balancer-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/aws-load-balancer-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API.
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1
          - description: AWSLoadBalancerController is the Schema for the awsloadbalancercontrollers
              API
            displayName: AWSLoad Balancer Controller
            kind: AWSLoadBalancerController
            name: awsloadbalancercontrollers.networking.olm.openshift.io
            version: v1alpha1
          - kind: IngressClassParams
            name: ingressclassparams.elbv2.k8s.aws
            version: v1beta1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1alpha1
          - kind: TargetGroupBinding
            name: targetgroupbindings.elbv2.k8s.aws
            version: v1beta1
        description: |-
          Operator to simplify the management of aws-load-balancer-controller.

          ### Prerequisites for installation
          In certain scenarios, the operator requires extra steps to be executed before it can be installed.

          #### Instructions for STS cluster (including ROSA in STS mode)

          [Bootstrapping AWS Load Balancer Operator on Security Token Service cluster](https://docs.openshift.com/container-platform/latest/networking/aws_load_balancer_operator/installing-albo-sts-cluster.html#nw-bootstra-albo-on-sts-cluster_albo-sts-cluster).

          #### Instructions for UPI based installation

          Additional documentation can be found in [VPC and Subnets](https://github.com/openshift/aws-load-balancer-operator/blob/main/docs/prerequisites.md#vpc-and-subnets).
        displayName: AWS Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aws
        - load-balancer
        - ALB
        - ingress
        links:
        - name: Aws Load Balancer Operator
          url: https://aws-load-balancer-operator.domain
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat Inc.
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/albo/aws-load-balancer-operator-bundle@sha256:e371c45e259eaf54d79b1bfc6c47b1093d87965a8d6076205d4860047146ed43
        - registry.redhat.io/albo/aws-load-balancer-rhel8-operator@sha256:53f14e88f05641545b675ff4edbcff6909d752659c02a1babf7491734f347153
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a2324acaea757bae3b01b1aec59f49f4dd79bd1868e69d3418d57c210a6dfd9
        - registry.redhat.io/albo/aws-load-balancer-controller-rhel8@sha256:f0b47b8200f85b557497a4f03d7b186cf415c4137e9cc0361633b67891cf2a24
        version: 1.2.0
      entries:
      - name: aws-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: aws-load-balancer-operator.v1.1.1
        version: 1.1.1
      - name: aws-load-balancer-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.2
    defaultChannel: stable-v1
    packageName: aws-load-balancer-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com/in-en/products/cloud-pak-for-data
    name: ibm-cpd-datastage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-datastage-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ds.cpd.ibm.com/v1",
                "kind": "DataStage",
                "metadata": {
                  "name": "datastage"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "fileStorageClass": "managed-nfs-storage",
                  "scaleConfig": "small"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudpakId: 49a42b864bb94569bef0188ead948f11
          cloudpakName: IBM DataStage Enterprise Plus Cartridge for IBM Cloud Pak
            for Data
          containerImage: registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
          description: The IBM DataStage Operator is used to deploy IBM DataStage
            services
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-datastage-operator
          olm.skipRange: '>=1.0.0 <7.3.0'
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.s390x: supported
          operatorframework.io/os.linux: supported
          operatorframework.io/suggested-namespace: ibm-common-services
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation: [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License: [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic),
              [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)'
            kind: DataStage
            name: datastages.ds.cpd.ibm.com
            version: v1
          - description: Documentation [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)
              License [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)
            kind: PXRuntime
            name: pxruntimes.ds.cpd.ibm.com
            version: v1
        description: |-
          # Introduction
          Industry-leading batch and real-time data integration service for data engineers to build trusted pipelines (ETL/ELT) for AI and operational use cases across on-premises and hybrid cloud environments.
          With DataStage, you can design and run data flows that move and transform data anywhere, at any scale.
          No matter how complex your data landscape, DataStage can streamline your data movement costs and increase productivity. DataStage offers:
          <ul><li>A best-in-breed parallel processing engine that enables you to process your data where it resides</li><li>Automated job design</li><li>Simple integration with cloud data lakes, real-time data sources, relational databases, big data, and NoSQL data stores</li></ul>
          The DataStage service uses Cloud Pak for Data platform connections and integration points, with services like Data Virtualization, to simplify the process of connecting to and accessing your data.

          With DataStage, Data Engineers can use the simple user interface to build no-code/low-code data pipelines. The interface offers hundreds of functions and connectors that reduce development time and inconsistencies across pipelines. The interface also makes it easy to collaborate with your peers and control access to specific analytics projects.
          The service also provides automatic workload balancing to provide high performance pipelines that make efficient use of available compute resources.

          ## Details
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=SSQNUZ_4.0/svc-welcome/dsent.html)

          ## Documentation

          [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=services-datastage)

          ## Installing
          To install the operator follow the the installation and configuration instructions within the [IBM Knowledge Center](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=datastage-installing).

          ## Prerequisites

          - OpenShift Version >= 4.6
          - IBM Cloud Pak for Data >= 4.0.2
          - IBM Cloud Pak for Data Common Core Services (CCS) >= 2.0.2 (The DataStage service automatically installs the CCS if they are not already installed.)

          ### Resources Required

          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=requirements-hardware)

          ## Configuration

          To scale DataStage services, you can specify spec.scaleConfig in the DataStage custom resource to small, medium, or large.

          ## License
          [IBM DataStage Enterprise Cartridge for Cloud Pak for Data](https://ibm.biz/Bdahic)
          [IBM DataStage Enterprise Plus Cartridge for Cloud Pak for Data](https://ibm.biz/BdahsW)

          ## Limitations
          - Only the `x86-64` platform is supported by this operator.
          - See [Limitations and known issues in DataStage](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-datastage)

          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: true
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - KILL
          - MKNOD
          - SETUID
          - SETGID
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: IBM DataStage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DataStage
        - Analytics
        - Cloud Pak
        - Data
        - Cloud Pak for Data
        - IBM
        maintainers:
        - email: cp4d@ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
          url: https://www.ibm.com/in-en/products/cloud-pak-for-data
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:8b72cafe9de355634bc3218cc1dc701ffb376a29aecb5cd12a5719d43a1304bd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:001bf35df08cd397397c3fdd1ddbd7953cdf4bde06965427ae640e5608e60ddf
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:6c569c4687546a03a5f4e127acc13b219ce3efa057fbdfd55e96868e549c70cb
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:3f8133fcd904befa579b99d3a828fe737faf13ae14e842ed0408eeeee72de5fb
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.connect.redhat.com/cpopen/ds-operator@sha256:6ec2387e45d9557cea2df749831e136120945794773ed0d48e0b04f0b3d9e115
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:5f5faf7d15f46d79eea873c12ff631ed59bc55bc57f1f7aee9fd0a8f35fdd3cf
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:3eecc54f26cd87870c201b5ade1d037d2733ccfda40fc1aedd9a2164e582275d
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:1b39e1d5d7c13451252250ceeb29b91c1e7ef3739fc17baf9958e1bf07c4c2fd
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:7015b410b947338dcafde98b68b95cb4199c08511ae46e733691e8fa835ad988
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:4604a076c837600cd4b3e29b31e22470915fe5b6b1191ae9401407f8620f2472
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:046eb99c8f1f37dd4cffa1c6c9a93aea50bb5d10a549aa12c3c90134d03c6d1c
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:99fc418fb1a09ee6c45e90dfba47f73ea8e1da1b19f9f570d2a749b68118b7a3
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:10492010b28f57f8badbc7cadbe15d1447269052b387d878c6d9ecbd2f62923b
        version: 7.3.0
      entries:
      - name: ibm-cpd-datastage-operator.v7.3.0
        version: 7.3.0
      name: v7.3
    defaultChannel: v7.3
    packageName: ibm-cpd-datastage-operator
    provider:
      name: IBM
      url: https://www.ibm.com/in-en/products/cloud-pak-for-data
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cfm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cfm-operator.v2.9.0-b96
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "stateManagement": {
                    "clusterProvider": {
                      "class": "org.apache.nifi.kubernetes.state.provider.KubernetesConfigMapStateProvider",
                      "id": "kubernetes-provider"
                    }
                  },
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.8.0-b94-nifi_1.25.0.2.3.13.0-36"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.8.0-b94"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
          createdAt: "2024-12-06T19:03:24Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:890d81188143ccf9ebe2c2fe5e039cdff3a2df53f7512555673f0468147e991e
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:6493d23e20fca8daf31823403fca3dd9b42da983464fbe788a6863e141e8b66b
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:ca70ebce43d451d040aedcec036413f986e620234fd9a56d45983805396db0b5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:0d700b7eff096dfc362868345826c3a6460fa43a700273b1422e95015b7727ee
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:914dcde9ee97894fa6f5deec285e53bf81d196e071a997cabb40fedd52deccc3
        version: 2.9.0-b96
      entries:
      - name: cfm-operator.v2.9.0-b96
        version: 2.9.0-b96
      name: alpha
    - currentCSV: cfm-operator.v2.10.0-b135
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "Nifi",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifi-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifi",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifi-sample"
                },
                "spec": {
                  "hostName": "nifi.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifi-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              },
              {
                "apiVersion": "cfm.cloudera.com/v1alpha1",
                "kind": "NifiRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cfm-operator",
                    "app.kubernetes.io/instance": "nifiregistry-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nifiregistry",
                    "app.kubernetes.io/part-of": "cfm-operator"
                  },
                  "name": "nifiregistry-sample"
                },
                "spec": {
                  "hostName": "registry.apps-crc.testing",
                  "image": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s",
                    "tag": "2.9.0-b96-nifi_1.27.0.2.3.14.0-14"
                  },
                  "replicas": 1,
                  "tiniImage": {
                    "repository": "container.repository.cloudera.com/cloudera/cfm-tini",
                    "tag": "2.9.0-b96"
                  },
                  "uiConnection": {
                    "routeConfig": {
                      "tls": {
                        "termination": "passthrough"
                      }
                    },
                    "type": "Route"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning,Big Data
          com.redhat.openshift.versions: v4.10
          containerImage: container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
          createdAt: "2025-05-05T14:39:51Z"
          description: Cloudera Flow Management - Kubernetes Operator for Apache NiFideploys
            and operates Apache NiFi
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: A valid Cloudera license is required.
            Contact a Cloudera account team to learn more.
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NifiRegistry is the Schema for the nifiregistries API.
            displayName: Nifi Registry
            kind: NifiRegistry
            name: nifiregistries.cfm.cloudera.com
            version: v1alpha1
          - description: Nifi is the Schema for the nifis API.
            displayName: Nifi
            kind: Nifi
            name: nifis.cfm.cloudera.com
            version: v1alpha1
        description: |-
          Cloudera Flow Management - Kubernetes Operator for Apache NiFi brings enterprise-grade NiFi and NiFi Registry deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of NiFi and NiFi Registry for enterprise deployments.
          To learn more about Cloudera Flow Management - Kubernetes Operator for Apache NiFi read this
          [deep dive blog post](https://community.cloudera.com/t5/Community-Articles/Cloudera-Flow-Management-Operator-A-technical-deep-dive/ta-p/390473).

          Cloudera Flow Management - Kubernetes Operator for Apache NiFi consists of and ships multiple components including Apache NiFi, Apache NiFi Registry,
          and a custom sidecar image for logging aggregation.

          ## Installation prerequisites

          Installing Cloudera Flow Management - Kubernetes Operator for Apache NiFi requires the a valid license as well as Cloudera credentials
          (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly.
          The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) from which installation artifacts are pulled.

          ### Obtaining a license
          Contact [Cloudera sales](https://cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets
          Before installation, create two `Secrets`. One containing your Cloudera credentials and one containing your license. Cloudera credentials are provided to you
          together with your license.

          1. Create a `Secret` containing your Cloudera credentials.
              `oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***]
               --docker-server=container.repository.cloudera.com --namespace [***OPERATOR INSTALL NAMESPACE***]`

          2. Create a `Secret` containing your license.
              `oc create secret generic cfm-operator-license --from-file=license.txt=[***PATH TO LICENSE FILE***] --namespace [***OPERATOR INSTALL NAMESPACE***]`

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces in which you
            plan to deploy NiFi clusters or NiFi Registry deployments. Cloudera recommends that you create the `Secret` in all required namespaces before installation
            if you know which namespaces you will be using to deploy NiFi clusters or NiFi Registry deployments.

          ## Learn more
          To learn more, visit the [Cloudera Flow Management - Kubernetes Operator for Apache NiFi documentation library](https://docs.cloudera.com/cfm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Flow Management - Kubernetes Operator for Apache NiFi, visit the following resources:
          * [Apache NiFi](https://nifi.apache.org/)
          * [Apache NiFi Registry](https://nifi.apache.org/registry.html)
        displayName: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nifi
        - nifiregistry
        - cloudera
        - cfm
        links:
        - name: Cloudera Flow Management - Kubernetes Operator for Apache NiFi
          url: https://docs.cloudera.com/cfm-operator
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86af06fb3b6c17b22be47118b0a8c797fa7fdd3c00dbe99a2d2a593ca8fc3d63
        - container.repository.cloudera.com/cloudera/cfm-nifi-k8s@sha256:8607c81906422d65784c6dca74a37af3b7c87012bca65ce28a995366014aec86
        - container.repository.cloudera.com/cloudera/cfm-nifiregistry-k8s@sha256:52e00f102579b972500eb187e64a7662a6a46caa883df743e454b1d5111b34e5
        - container.repository.cloudera.com/cloudera/cfm-operator@sha256:0312b6dadd5904ea9507bb857ab89d7266bf244c93b86855e0528e72813b7a85
        - container.repository.cloudera.com/cloudera/cfm-tini@sha256:d42ad5f05a9bb283fed6dbbd820855714a31662f5b851e691f13703153e90d13
        - registry.connect.redhat.com/cloudera/cfm-operator-bundle@sha256:4aec6aa2d2b9a24167bb6bc92db7c2c323847b7a2f3ad7f998c96bf1d2903366
        version: 2.10.0-b135
      entries:
      - name: cfm-operator.v2.10.0-b135
        version: 2.10.0-b135
      name: stable
    defaultChannel: stable
    packageName: cfm-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp
      provider-url: https://www.netapp.com/
    name: netapp-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netapp-spark-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.netapp.com/v1alpha1",
                "kind": "SparkOperator",
                "metadata": {
                  "name": "sparkoperator-sample"
                },
                "spec": {
                  "spark-operator": {
                    "affinity": {},
                    "controllerThreads": 20,
                    "disableExecutorReporting": false,
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "public.ecr.aws/f4k1p1n4/spark-operator",
                      "tag": "108810d-ubi9-rel20240916"
                    },
                    "imagePullSecrets": [],
                    "istio": {
                      "enabled": true
                    },
                    "jobAffinity": {},
                    "podLabels": {},
                    "pprof": {
                      "enable": true,
                      "port": 6060,
                      "portName": "pprof"
                    },
                    "rbac": {
                      "createClusterRole": true,
                      "createRole": true
                    },
                    "resources": {
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    },
                    "securityContext": {},
                    "serviceAccounts": {
                      "spark": {
                        "create": true
                      }
                    },
                    "sparkJobNamespace": "",
                    "tolerations": [],
                    "uiService": {
                      "enable": false
                    },
                    "webhook": {
                      "enable": true,
                      "jobTTLSecondsAfterFinished": 60,
                      "port": 9443
                    },
                    "workqueue": {
                      "maxDelay": {
                        "enabled": false,
                        "value": "6h"
                      },
                      "tokenBucketSize": 500,
                      "tokenRefillRate": 50
                    }
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "ScheduledSparkApplication",
                "metadata": {
                  "name": "spark-pi-scheduled"
                },
                "spec": {
                  "concurrencyPolicy": "Allow",
                  "schedule": "@every 5m",
                  "template": {
                    "arguments": [
                      "1000"
                    ],
                    "driver": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "coreLimit": "1200m",
                      "cores": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "serviceAccount": "sparkoperator-sample-spark",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "executor": {
                      "annotations": {
                        "openshift.io/required-scc": "sparkoperator-sample-spark"
                      },
                      "cores": 1,
                      "instances": 1,
                      "labels": {
                        "version": "3.5.0"
                      },
                      "memory": "512m",
                      "volumeMounts": [
                        {
                          "mountPath": "/tmp",
                          "name": "test-volume"
                        }
                      ]
                    },
                    "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                    "imagePullPolicy": "IfNotPresent",
                    "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                    "mainClass": "org.apache.spark.examples.SparkPi",
                    "mode": "cluster",
                    "restartPolicy": {
                      "type": "Never"
                    },
                    "sparkVersion": "3.5.0",
                    "type": "Scala",
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "test-volume"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sparkoperator.k8s.io/v1beta2",
                "kind": "SparkApplication",
                "metadata": {
                  "name": "spark-pi"
                },
                "spec": {
                  "arguments": [
                    "1000"
                  ],
                  "driver": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "coreLimit": "1200m",
                    "cores": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "serviceAccount": "sparkoperator-sample-spark",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "executor": {
                    "annotations": {
                      "openshift.io/required-scc": "sparkoperator-sample-spark"
                    },
                    "cores": 1,
                    "instances": 1,
                    "labels": {
                      "version": "3.5.0"
                    },
                    "memory": "512m",
                    "volumeMounts": [
                      {
                        "mountPath": "/tmp",
                        "name": "test-volume"
                      }
                    ]
                  },
                  "image": "public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce",
                  "imagePullPolicy": "IfNotPresent",
                  "mainApplicationFile": "local:///opt/spark/examples/jars/spark-examples_2.13-3.5.0.jar",
                  "mainClass": "org.apache.spark.examples.SparkPi",
                  "mode": "cluster",
                  "restartPolicy": {
                    "type": "Never"
                  },
                  "sparkVersion": "3.5.0",
                  "type": "Scala",
                  "volumes": [
                    {
                      "emptyDir": {},
                      "name": "test-volume"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Integration & Delivery, AI/Machine Learning, Streaming
            & Messaging
          com.redhat.openshift.versions: v4.14
          containerImage: public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
          createdAt: "2024-10-21T13:47:47Z"
          description: Red Hat and NetApp have partnered to offer enterprise-grade
            supported Apache Spark on OpenShift.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ScheduledSparkApplication
            name: scheduledsparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkApplication
            name: sparkapplications.sparkoperator.k8s.io
            version: v1beta2
          - kind: SparkOperator
            name: sparkoperators.charts.netapp.com
            version: v1alpha1
        description: |-
          # NetApp Supported Apache Spark on OpenShift

          Red Hat and NetApp have partnered to offer enterprise-grade supported Apache Spark on OpenShift.
          Use this certified operator to launch Apache Spark applications on OpenShift.

          ## Documentation

          Please find the official documentation at https://www.instaclustr.com/support/documentation/apache-spark/apache-spark-on-red-hat-openshift/getting-started/

          ## Support

          With questions about Apache Spark, contact NetApp's Instaclustr support team on the [web](https://support.instaclustr.com/) or by email at support@instaclustr.com

          With questions about Operators in the OpenShift OperatorHub, contact Red Hat's support team on the [web](https://www.redhat.com/en/services/support) or by email at customerservice@redhat.com

          With questions about pricing, email ng-netapp-apache-spark-support-redhat-openshift@netapp.com
        displayName: NetApp Supported Apache Spark on OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Spark
        - Big Data
        - Apache Spark
        - AI
        - ML
        - Data Engineering
        links:
        - name: NetApp Supported Apache Spark on OpenShift
          url: https://www.instaclustr.com/support-solutions
        maintainers:
        - email: thorsteinn.sigurdsson@netapp.com
          name: Thorsteinn Sigurdsson
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: NetApp
          url: https://www.netapp.com/
        relatedImages:
        - public.ecr.aws/f4k1p1n4/spark@sha256:96f2b1185df5a9328b566df3023684291b452c8fe23dfa214d234420ab8a08ce
        - registry.connect.redhat.com/netapp/netapp-supported-apache-spark-on-openshift-operator@sha256:4f15a0fc375156fafdf2d504f9980a8eb4451808b5554357877345cacf8f3594
        - public.ecr.aws/f4k1p1n4/netapp-spark-operator-controller@sha256:9fe1d0793f11b1c525a9a92cbcde6f4bc1bf3c887617a8496bed3a5a0008a67e
        version: 1.1.3
      entries:
      - name: netapp-spark-operator.v1.1.3
        version: 1.1.3
      name: alpha
    defaultChannel: alpha
    packageName: netapp-spark-operator
    provider:
      name: NetApp
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-key-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-key-manager-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "KeyManager",
                "metadata": {
                  "name": "my-kms",
                  "namespace": "ns-1"
                },
                "spec": {
                  "apiKeySecret": {
                    "name": "my-api-key"
                  },
                  "configuration": {
                    "name": "my-kms-server-config"
                  },
                  "hsmSecret": {
                    "name": "my-kms-hsm"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "registry-creds"
                    }
                  ],
                  "replicas": 2,
                  "volumeClaimTemplate": {
                    "metadata": {
                      "name": "key-manager-volume"
                    },
                    "spec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "100Mi"
                        }
                      },
                      "storageClassName": "standard"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security, Storage
          createdAt: "2025-08-27T14:50:59Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeyManager
            name: keymanagers.aistor.min.io
            version: v1alpha1
        description: |-
          Highly available, powerful and operationally simple, MinIO's Enterprise Key Management Server is optimized for large storage infrastructures where billions of cryptographic keys are required.

          MinIO's AIStor KMS establishes its foundational trust using the concept of an hardware security module (but given KES is software, this is only a concept). That module assumes a pivotal role in sealing and unsealing the KMS root encryption key. The module responsibility extends to safeguarding the integrity of KMS by allowing the unsealing of its encrypted on-disk state and facilitating communication among nodes within a KMS cluster.

          It solves the challenges associated with billions of cryptographic keys and hundreds of thousands of cryptographic operations per node per second - which are commonplace in larger deployments.

          <b>High Availability and Fault Tolerance</b>
          In the dynamic landscape of large-scale systems, network or node outages are inevitable. Taking down a cluster for maintenance is rarely feasible. MinIO's AIStor KMS ensures uninterrupted availability, even when faced with such disruptions, mitigating cascading effects that can take down the entire storage infrastructure. Specifically, you could lose all but one node of a cluster and still handle any encryption, decryption or data key generation requests.

          <b>Scalability</b>
          While the amount of data usually only increases, the load on a large-scale storage system may vary significantly from time to time. MinIO's AIStor KMS supports dynamic cluster resizing and nodes can be added or removed at any point without incurring any downtime.

          <b>Multi-Tenancy</b>
          Large-scale storage infrastructures are often used by many applications and teams across the entire organization. Isolating teams and groups into their own namespaces is a core requirement. MinIO's AIStor KMS supports namespacing in the form of enclaves. Each tenant can be assigned its own enclave which is completely independent and isolated from all other enclaves on the KMS cluster.

          <b>Predictable Behavior</b>
          MinIO's AIStor KMS is designed to be easily managed, providing operators with the ability to comprehend its state intuitively. Due to its simple design, MinIO's AIStor KMS is significantly easier to operate than similar solutions that rely on more complex consensus algorithms like Raft, or Paxos.

          <b>Consistent and Performant</b>
          The responsiveness of the KMS for GET/PUT operations directly influences the overall efficiency and speed of the storage system. MinIO's AIStor KMS nodes don’t have to coordinate when handling such requests from the storage system. Therefore, the performance of a MinIO's AIStor KMS cluster increases linearly with the number of nodes. Further, MinIO's AIStor KMS supports request pipelining to handle hundreds of thousands of cryptographic operations per node and second.

          <b>Simplicity</b>
          Operating a KMS cluster does not require expertise in cryptography or distributed systems. Everything can be done from the AIStor Console.

          <a href="https://resources.min.io/request-a-demo-marketplaces/">Request a demo today</a>
        displayName: Minio AIStor Key management System (KMS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - MinIO
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-key-manager-operator@sha256:7a998ea17ef537ec5e998e1627b065ae04ab01b7988196da1cdd1763e05a9924
        version: 2025.8.19175300
      entries:
      - name: minio-key-manager-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-key-manager-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-key-manager-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-key-manager-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-key-manager-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-key-manager-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: https://stakater.com
    name: template-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: template-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-07-04T00:40:36Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.0.1
          operatorframework.io/suggested-namespace: template-operator
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:6ef2956b297795a9d0c83119e71309549975df3ae5652e6a98cef035ace5d51a
        - registry.connect.redhat.com/stakater/template-operator@sha256:ad8312417e4e9621b74d3ad03020516840feef9c4b52c4f0d894dbb6dc6f906a
        version: 0.0.1
      entries:
      - name: template-operator.v0.0.1
        version: 0.0.1
      name: release-0.0
    - currentCSV: template-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-08-29T06:41:42Z"
          description: Operator to distribute plain kubernetes manifests, helm charts
            and, copy secrets and configmaps to multiple namespaces
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.5
          operatorframework.io/suggested-namespace: template-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterTemplateInstance is the Schema for the clustertemplateinstances
              API
            displayName: Cluster Template Instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateGroupInstance is the Schema for the TemplateGroupInstances
              API
            displayName: Template Group Instance
            kind: TemplateGroupInstance
            name: templategroupinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.templates.stakater.com
            version: v1alpha1
          - description: TemplateInstance is the Schema for the templatesinstance
              API
            displayName: Template Instance
            kind: TemplateInstance
            name: templateinstances.tenantoperator.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.templates.stakater.com
            version: v1alpha1
          - description: Template is the Schema for the templates API
            displayName: Template
            kind: Template
            name: templates.tenantoperator.stakater.com
            version: v1alpha1
        description: '"Operator to distribute plain kubernetes manifests, helm charts
          and, copy secrets and configmaps to multiple namespaces"'
        displayName: Template Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - templates
        - secrets
        - configmaps
        - distribution
        - stakater
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maturity: alpha
        provider:
          name: Stakater
          url: https://stakater.com
        relatedImages:
        - registry.connect.redhat.com/stakater/template-operator-bundle@sha256:8831faee6979acfb5bd031b64cbc09fb90e9ea4f48bbff15e9d39ccdf6bf3d95
        - registry.connect.redhat.com/stakater/template-operator@sha256:88f722baf22ac07704503f098f2b0ea7d76815a1ab0552474b1c9ba6639d8142
        version: 0.1.5
      entries:
      - name: template-operator.v0.1.5
        version: 0.1.5
      name: release-0.1
    defaultChannel: release-0.1
    packageName: template-operator
    provider:
      name: Stakater
      url: https://stakater.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-athena-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-athena-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "athena.services.k8s.aws/v1alpha1",
                "kind": "WorkGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
          createdAt: "2025-09-19T21:06:02Z"
          description: AWS Athena controller is a service controller for managing
            Athena resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PreparedStatement represents the state of an AWS athena PreparedStatement
              resource.
            displayName: PreparedStatement
            kind: PreparedStatement
            name: preparedstatements.athena.services.k8s.aws
            version: v1alpha1
          - description: WorkGroup represents the state of an AWS athena WorkGroup
              resource.
            displayName: WorkGroup
            kind: WorkGroup
            name: workgroups.athena.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Athena resources in AWS from within your Kubernetes cluster.

          **About Amazon Athena**

          Amazon Athena is an interactive query service that makes it easy to analyze data directly in Amazon Simple Storage Service (Amazon S3) using standard SQL. With a few actions in the AWS Management Console, you can point Athena at your data stored in Amazon S3 and begin using standard SQL to run ad-hoc queries and get results in seconds.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Athena
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - athena
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Athena Developer Resources
          url: https://aws.amazon.com/athena/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: athena maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/athena-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-athena-controller@sha256:8cff9258583f8acbdd26d263a61ffa11792cf9f4c1d65410bb4c71f9f7aaf178
        version: 1.1.1
      entries:
      - name: ack-athena-controller.v1.1.1
        version: 1.1.1
      - name: ack-athena-controller.v1.1.0
        version: 1.1.0
      - name: ack-athena-controller.v1.0.14
        version: 1.0.14
      - name: ack-athena-controller.v1.0.13
        version: 1.0.13
      - name: ack-athena-controller.v1.0.12
        version: 1.0.12
      - name: ack-athena-controller.v1.0.11
        version: 1.0.11
      - name: ack-athena-controller.v1.0.10
        version: 1.0.10
      - name: ack-athena-controller.v1.0.9
        version: 1.0.9
      - name: ack-athena-controller.v1.0.8
        version: 1.0.8
      - name: ack-athena-controller.v1.0.7
        version: 1.0.7
      - name: ack-athena-controller.v1.0.6
        version: 1.0.6
      - name: ack-athena-controller.v1.0.5
        version: 1.0.5
      - name: ack-athena-controller.v1.0.4
        version: 1.0.4
      - name: ack-athena-controller.v1.0.1
        version: 1.0.1
      - name: ack-athena-controller.v1.0.0
        version: 1.0.0
      - name: ack-athena-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-athena-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Perficient Inc
      provider-url: http://www.perficient.com
    name: devopsinabox
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devopsinabox.v0.0.45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "DevSecOpsManager",
                "metadata": {
                  "name": "devsecopsmanager-sample"
                },
                "spec": {
                  "config": {
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "forwarding": {
                    "fluentd": {
                      "affinity": {},
                      "caFile": "files/default-openshift-logging-fluentd.crt",
                      "image": "registry.redhat.io/openshift4/ose-logging-fluentd:v4.4",
                      "keyFile": "files/default-openshift-logging-fluentd.key",
                      "loglevel": "warn",
                      "nodeSelector": {},
                      "passphrase": "",
                      "persistence": {
                        "accessMode": "ReadWriteOnce",
                        "enabled": false,
                        "size": "10Gi",
                        "storageClass": ""
                      },
                      "port": 24224,
                      "replicas": 2,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "512Mi"
                        }
                      },
                      "sharedkey": "splunkforwarding",
                      "ssl": true,
                      "tolerations": [],
                      "updateStrategy": {
                        "type": "RollingUpdate"
                      }
                    },
                    "preserveDefault": true
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "nameOverride": "",
                  "prometheus": {
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "ManagedNamespace",
                "metadata": {
                  "name": "managednamespace-sample"
                },
                "spec": {
                  "config": {
                    "repoBranch": "master",
                    "repoPath": "",
                    "repoUrl": ""
                  },
                  "container": {
                    "password": "",
                    "registry": "https://quay.io",
                    "username": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "password": "",
                    "server": "https://gitlab.com",
                    "token": "",
                    "username": ""
                  },
                  "manager": {
                    "namespace": ""
                  },
                  "mirrors": {
                    "maven": ""
                  },
                  "nameOverride": ""
                }
              },
              {
                "apiVersion": "apimc.com.perficient/v1alpha1",
                "kind": "SpringBootApp",
                "metadata": {
                  "name": "springbootapp-sample"
                },
                "spec": {
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "config": {
                    "repo": ""
                  },
                  "container": {
                    "image": "placeholder",
                    "registry": ""
                  },
                  "fullnameOverride": "",
                  "git": {
                    "branch": "master",
                    "module": "",
                    "url": ""
                  },
                  "management": {
                    "port": 8081
                  },
                  "maven": {
                    "mirror": ""
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "resources": {},
                  "service": {
                    "port": 8080
                  },
                  "serviceAccount": {
                    "annotations": {}
                  },
                  "version": {
                    "branch": "",
                    "image": "placeholder",
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration,Developer Tools,Application Runtime
          containerImage: quay.io/mrethers/devopsinabox-operator:v0.0.45
          createdAt: 2021-03-03 10:51:09Z
          description: Perficient's DevSecOps Manager For Kubernetes Applications
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/prft-rh/devopsinabox-operator
          support: Matthieu Rethers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevSecOpsManager
            name: devsecopsmanagers.apimc.com.perficient
            version: v1alpha1
          - kind: ManagedNamespace
            name: managednamespaces.apimc.com.perficient
            version: v1alpha1
          - kind: SpringBootApp
            name: springbootapps.apimc.com.perficient
            version: v1alpha1
        description: "The devops-in-a-box Operator manages the entire lifecycle of
          your applications in Kubernetes. It encapsulates all the best devsecops
          practices gathered by Perficient over the years working on many large projects.
          Use it to migrate existing applications with minimum changes or deploy new
          ones with little to no knowledge of Kubernetes.\n\nOnce installed, the Devops-in-a-box
          operator provides the following features:\n\n* **Create/Destroy**: Easily
          launch a new DevSecOpsManager instance to bootstrap your cluster and start
          deploying new applications with total confidence.\n\n* **Gitops Simplified**:
          Configure your cluster and individual applications through Git and watch
          your changes propagate instantly. Just point to where your code is and let
          the operator do the rest with preconfigured cloud-native CI/CD pipelines.\n\n*
          **Production-Ready**: Automatically register your applications with Openshift
          logging and monitoring systems and enjoy special additions for Spring Boot
          like custom dashboards and searchable logs.\n\n*Contact us on Github for
          installation and usage instruction*\n\n### Other Supported Features\n\n**Blue/Green,
          Canary, A/B Deployments**\n\nRun multiple versions of your applications
          simultaneously and configure traffic rules to target specific groups of
          users using Service Mesh.\n\n**Updates via automated operations**\n\nGet
          the latest enhancements automatically through the Operator Lifecycle Manager
          and upgrade all your running applications at once. "
        displayName: Devops-in-a-box
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devops
        - devsecops
        - spring
        - boot
        - java
        - gitops
        - cicd
        - argo
        links:
        - name: Devops-in-a-box Operator
          url: https://github.com/prft-rh/devopsinabox-operator
        maintainers:
        - email: matthieu.rethers@perficient.com
          name: Matthieu Rethers
        maturity: alpha
        provider:
          name: Perficient Inc
          url: http://www.perficient.com
        relatedImages:
        - quay.io/mrethers/devopsinabox-operator:v0.0.45
        - quay.io/openshift-community-operators/devopsinabox@sha256:1c2ad76102151a56ecd753d9b86fd9f155d64b912067738d1858d33a02557928
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.45
      entries:
      - name: devopsinabox.v0.0.45
        version: 0.0.45
      name: alpha
    defaultChannel: alpha
    packageName: devopsinabox
    provider:
      name: Perficient Inc
      url: http://www.perficient.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Arkmq-org
      provider-url: https://arkmq-org.io
    name: activemq-artemis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: activemq-artemis-operator.v2.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "broker.amq.io/v1alpha1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "jolokiaAgentEnabled": false,
                    "journalType": "nio",
                    "managementRBACEnabled": true,
                    "messageMigration": false,
                    "persistenceEnabled": false,
                    "requireLogin": false,
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v1beta1",
                "kind": "ActiveMQArtemisSecurity",
                "metadata": {
                  "name": "ex-prop"
                },
                "spec": {
                  "loginModules": {
                    "propertiesLoginModules": [
                      {
                        "name": "prop-module",
                        "users": [
                          {
                            "name": "bob",
                            "roles": [
                              "root"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "securityDomains": {
                    "brokerDomain": {
                      "loginModules": [
                        {
                          "flag": "sufficient",
                          "name": "prop-module"
                        }
                      ],
                      "name": "activemq"
                    }
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha1",
                "kind": "ActiveMQArtemisScaledown",
                "metadata": {
                  "name": "ex-aaoscaledown"
                },
                "spec": {
                  "localOnly": true
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha2",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha3",
                "kind": "ActiveMQArtemisAddress",
                "metadata": {
                  "name": "ex-aaoaddress"
                },
                "spec": {
                  "addressName": "myAddress0",
                  "queueName": "myQueue0",
                  "routingType": "anycast"
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha4",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "broker.amq.io/v2alpha5",
                "kind": "ActiveMQArtemis",
                "metadata": {
                  "name": "ex-aao"
                },
                "spec": {
                  "deploymentPlan": {
                    "image": "placeholder",
                    "size": 2
                  }
                },
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/arkmq-org/activemq-artemis-operator:2.0.5
          createdAt: "2025-04-04 14:35:00"
          description: An operator for managing the Apache ActiveMQ Artemis message
            broker
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/arkmq-org/activemq-artemis-operator
          support: arkmq-org
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Adding and removing addresses using custom resource definitions
            displayName: ActiveMQ Artemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemisAddress is the Schema for the activemqartemisaddresses
              API
            displayName: Active MQArtemis Address
            kind: ActiveMQArtemisAddress
            name: activemqartemisaddresses.broker.amq.io
            version: v2alpha3
          - description: A stateful deployment of one or more brokers
            displayName: ActiveMQ Artemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha1
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha2
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha3
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha4
          - description: ActiveMQArtemis is the Schema for the activemqartemises API
            displayName: Active MQArtemis
            kind: ActiveMQArtemis
            name: activemqartemises.broker.amq.io
            version: v2alpha5
          - description: Provides internal message migration on clustered broker scaledown
            displayName: ActiveMQ Artemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v1beta1
          - description: ActiveMQArtemisScaledown is the Schema for the activemqartemisscaledowns
              API
            displayName: Active MQArtemis Scaledown
            kind: ActiveMQArtemisScaledown
            name: activemqartemisscaledowns.broker.amq.io
            version: v2alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1alpha1
          - description: Security configuration for the broker
            displayName: ActiveMQ Artemis Security
            kind: ActiveMQArtemisSecurity
            name: activemqartemissecurities.broker.amq.io
            version: v1beta1
        description: ArkMQ Operator manages ActiveMQ Artemis messaging broker in the
          cloud
        displayName: ArkMQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - messaging
        - cloud
        - kubernetes
        - activemq
        - artemis
        - integration
        - operator
        links:
        - name: Documentation
          url: https://arkmq-org.io/docs
        - name: Repository
          url: https://github.com/arkmq-org/activemq-artemis-operator
        maintainers:
        - email: dbruscin@redhat.com
          name: Domenico Francesco Bruscino
        - email: gtully@redhat.com
          name: Gary Tully
        - email: hgao@redhat.com
          name: Howard Gao
        - email: rkieley@redhat.com
          name: Roderick Kieley
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Arkmq-org
          url: https://arkmq-org.io
        relatedImages:
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:52f337cd5d7d5d199eebf4fc722f8c5279f652506f009efbc0f8f6de88e18718
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:1fd99a46758cd29fce846d0d590021758b2a7a3efedb817bb205359caa7d4f67
        - quay.io/community-operator-pipeline-prod/activemq-artemis-operator@sha256:1e3c6ff2d2612c3e85d40b1974b5a3c4f603b77353752389fa588728bd4263c4
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4bba8828e72fd7c9d38c22ba1c2a4f7130817a4e7eac8fb63c13e86465a8e1e7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:3bbfb6177e4839087f615a01ee54f9e06a95acaddbb083e11851b6af9315dbf0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ffa0d7cb4068df84319776dbfb66210297da1b8466ccd5451318fc3de79736c3
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b293c1269166cc6181f68b8d803fdfcf20fdbd5df2787ed0453724b22e298fb2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:9bbb915855f29ea16950a3fd50d0b9bac25f82cba7d197579e04d9c33a5b02c4
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.4
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:b6f61188379def8c796c738549c099bd40bcd9bb44f5c30682c5edd136347aca
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:ddaf00914e3a412d21ec3a814e24f3d4cc382b5c4d6c21e9b1b94dde16ddf194
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4628c2591ec06caf21a3467f6e3faeaede7fa402b4f137d943b25b4dd4ad343
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:5e768aefade6d18cd65cb1eb5ee233844f9f7e7b52631f93a0982b67d382239d
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2895b2cf6292c951a0cdde78e755514f5d7cb43ce9f7b6e7dc381d8dfeec31ab
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:da95bff69deb63545d9782ed20cbdd17c1a79c6b16165580bf96a1d5191e6a59
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:2d155bffd246f26f516a6075ec170babc1373c72bb664cba4e3c58521ffd3558
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:37ab7354273f4dcd19bad6936f10e92220d98b8c0266002221477a0e2d146c81
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:9e1fefc9a94d72a7b5a3f4e2d53ebd23e5ab221df17a5afde42ce973277838f6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:de82344daa0caeabc6559656da0af285436a687e8f371744c0c51bdbc309167d
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:56bae938a6b4cd87f9f5284931f440f9b9d116cda3b63b13f83a4b1c6c641959
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:0.2.1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:70d4c799c7d905fac27457d63e662ea72bca9e899a692e425609e4eaa45129e6
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:a4a0d0453f0761b9e114d0d4cc7c14091a2a9c30790f18ffa6bf1e365d5083fe
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:b71d1a57c8b3820a8133cf7a25a6edde7b29e1c4d32d7c30325183415a8bacd1
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:8f592b2454d092c82b5f61b4e04c3aff173db246539b3c06e15a71960d3bd235
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e2438670831bad39bb08147bbebe67ef9465899fb7750f1a5a185569f99d5fb5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:f25db980c3229e66cf48690aa175f6f93e3dce3974c0e70ab23af6462e7d1893
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:6f2da9ecd8f5021964d0dfa85144830efcfbd0462d18f18f036971e5f17b3bb6
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.2
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:81590ec33b2f00775d3c6596f610887b670de7c0b89148b09eab7584f5b0b0ec
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:49f34122ab46ab5d7ead094b6e40817b62c75b6a809f67db07d68bce750da51e
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f2c853ff91809282c04d5b1381ceae366518bf07b16649a61aa6c60c47f10714
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:42230ce9d11a1949a5bb73b6ee218b1581f585bcd553485ee438e1c5b50fcf8b
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:8fbdd1ba728058d5f437186b9055ca28596140068ebc69ecff4fa52717b19a82
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.7
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:f10983c68aff358e285d89d827c27b3487a3ee3b8e121cdaa295bf5fa8501145
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:fc9b9a4c7550d84fd90335dd4f9e0d8d39c09136c6034b28490a368368cb11f1
        - quay.io/arkmq-org/activemq-artemis-operator:2.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:264b33a9aad90d32a41258904f35092a141d76dcef0c115d414475444100aace
        - quay.io/arkmq-org/activemq-artemis-broker-init:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:decb3932ebb5b848b90a4a23b083f0cb5237eec2844095d8c58a6798da8bb999
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.0
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes@sha256:4be60cfa2a300b2eef9b5f10ea756bd3f55d75fbd8890232d907f8d64080449b
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:88a0098a628fe43a9b422f8ebb774086fb65ed389a195e2779482a4a63447e28
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:93168fd8cd4fb32f041795afb8fb0e21cf81081d6cc463a44356e153dc1ad032
        - quay.io/arkmq-org/activemq-artemis-broker-kubernetes:1.0.5
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:719178251893def955e444e706ff07042ea98d294671315d95da504c1f47de24
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:634f3d8fd620607b56e0360a191d23a9cb8a32ba0ba1f24fe20791ec52343586
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.10
        - quay.io/arkmq-org/activemq-artemis-broker-init:0.2.2
        - quay.io/arkmq-org/activemq-artemis-broker-init@sha256:e438bcacf099a92b143793e88c00c6ac0afb31effe4b0eed49a935c9265ad943
        version: 2.0.5
      entries:
      - name: activemq-artemis-operator.v2.0.5
        version: 2.0.5
      name: stable
    defaultChannel: stable
    packageName: activemq-artemis-operator
    provider:
      name: Arkmq-org
      url: https://arkmq-org.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-pipes-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-pipes-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pipes.services.k8s.aws/v1alpha1",
                "kind": "Pipe",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
          createdAt: "2025-09-19T21:15:33Z"
          description: AWS EventBridge Pipes controller is a service controller for
            managing Pipes resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pipe represents the state of an AWS pipes Pipe resource.
            displayName: Pipe
            kind: Pipe
            name: pipes.pipes.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge Pipes resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge Pipes**

          Amazon EventBridge Pipes connects sources to targets. Pipes are intended for point-to-point integrations between supported [sources](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-source.html) and [targets](https://docs.aws.amazon.com/eventbridge/latest/userguide/eb-pipes-event-target.html), with support for advanced transformations and [enrichment](https://docs.aws.amazon.com/eventbridge/latest/userguide/pipes-enrichment.html). It reduces the need for specialized knowledge and integration code when developing event-driven architectures, fostering consistency across your company’s applications. To set up a pipe, you choose the source, add optional filtering, define optional enrichment, and choose the target for the event data.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge Pipes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pipes
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Pipes Developer Resources
          url: https://aws.amazon.com/eventbridge/pipes/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: pipes maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-pipes-controller@sha256:e46b540219ca42c3b6c08ff99afabf09daea2dca5a335cda6013343df2f35878
        - public.ecr.aws/aws-controllers-k8s/pipes-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-pipes-controller.v1.1.1
        version: 1.1.1
      - name: ack-pipes-controller.v1.1.0
        version: 1.1.0
      - name: ack-pipes-controller.v1.0.25
        version: 1.0.25
      - name: ack-pipes-controller.v1.0.24
        version: 1.0.24
      - name: ack-pipes-controller.v1.0.23
        version: 1.0.23
      - name: ack-pipes-controller.v1.0.22
        version: 1.0.22
      - name: ack-pipes-controller.v1.0.21
        version: 1.0.21
      - name: ack-pipes-controller.v1.0.20
        version: 1.0.20
      - name: ack-pipes-controller.v1.0.18
        version: 1.0.18
      - name: ack-pipes-controller.v1.0.17
        version: 1.0.17
      - name: ack-pipes-controller.v1.0.16
        version: 1.0.16
      - name: ack-pipes-controller.v1.0.15
        version: 1.0.15
      - name: ack-pipes-controller.v1.0.14
        version: 1.0.14
      - name: ack-pipes-controller.v1.0.13
        version: 1.0.13
      - name: ack-pipes-controller.v1.0.12
        version: 1.0.12
      - name: ack-pipes-controller.v1.0.11
        version: 1.0.11
      - name: ack-pipes-controller.v1.0.10
        version: 1.0.10
      - name: ack-pipes-controller.v1.0.9
        version: 1.0.9
      - name: ack-pipes-controller.v1.0.8
        version: 1.0.8
      - name: ack-pipes-controller.v1.0.6
        version: 1.0.6
      - name: ack-pipes-controller.v1.0.3
        version: 1.0.3
      - name: ack-pipes-controller.v1.0.2
        version: 1.0.2
      - name: ack-pipes-controller.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-pipes-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: logic-operator-rhel8
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logic-operator-rhel8.v1.36.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          createdAt: "2025-05-26T14:10:15Z"
          description: OpenShift Serverless Logic Kubernetes Operator for deploying
            workflow applications based on the CNCF Serverless Workflow specification
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-serverless-logic
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubesmarts/kie-tools
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          OpenShift Serverless Logic Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification v0.8](https://serverlessworkflow.io/).
        displayName: OpenShift Serverless Logic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        - workflows
        links:
        - name: Documentation
          url: https://docs.openshift.com/serverless/
        - name: Source Repository
          url: https://github.com/kiegroup/kogito-serverless-operator/
        maintainers:
        - email: serverless-logic-team@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/logic-rhel8-operator@sha256:8d3682448ebdac3aeabb2d23842b7e67a252b95f959c408af805037f9728fd3c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4564ca3dc5bac80d6faddaf94c817fbbc270698a9399d8a21ee1005d85ceda56
        - registry.redhat.io/openshift-serverless-1/logic-operator-bundle@sha256:5fff2717f7b08df2c90a2be7bfb36c27e13be188d23546497ed9ce266f1c03f4
        version: 1.36.0
      entries:
      - name: logic-operator-rhel8.v1.36.0
        version: 1.36.0
      - name: logic-operator-rhel8.v1.35.0
        version: 1.35.0
      - name: logic-operator-rhel8.v1.34.0
        version: 1.34.0
      - name: logic-operator-rhel8.v1.33.0
        version: 1.33.0
      name: alpha
    defaultChannel: alpha
    packageName: logic-operator-rhel8
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: openshift-builds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-builds-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-01-29T08:58:54Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.2
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:6214708f3b517cb901af9f9bd0502ba54f8e5183d0d909a3b559deee6394026e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3fcd8e2bf0bcb8ff8c93a87af2c59a3bcae7be8792f9d3236c9b5bbd9b6db3b2
        - registry.redhat.io/ubi9/buildah@sha256:27d837f9bc69ad3c3651cf3315e2501b137c11baa553d9d46140e5cf7fa7873a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:3523e7c2215e156081e5f05443884d65efad9b7b9ca81a35416695020231916a
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:53d8cf8bf5c1103860ffc4110e2a422371761e76ee79edfef8680e947407c5ce
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:0783132b14ed8050664810f37e7ad6388fefbc586c45ecc7ba69dd5041a55c9f
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:0b1dd35f55181f34775188bc71b3ff8dff66c70acfcfabad19353a02e19351f7
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:04b8bfc0dc6f260ef367ca1cce2bcc84032bc8be17b4a595be12a344813a2632
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6e44758fd4a7b34814ac7c71e52d93633781d39f7897ecab98a090859b487b4f
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:d4b3f451becb7a42341c98dd5b9fc5d2b582c93cac2728070ef4bb83b64abb4c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:e8582857566020c71653228355cbadc2c1cda41e36fc767c204a4edb05e7f42a
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:a4ba591c57bbc815c93023f4a080213be32581ba6a1c2d94b0ecbc5c90b07a1e
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:1e9a54212690f7930d2b681db87a6f79f5a2acc5c982734d4b283a2472b59db5
        version: 1.3.2
      deprecation:
        message: '''builds-1.3'' has reached the end of maintenance support. Please
          upgrade to ''builds-1.5''.'
      entries:
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: builds-1.3
    - currentCSV: openshift-builds-operator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2024-11-15T22:28:01Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenShiftBuild describes the desired state of Builds for
              OpenShift, and the status of all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I) and Buildah. You can create and apply build resources,
          view logs of build runs, \nand manage builds in your OpenShift Container
          Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:adc4a31d63b7044db7094fddcb1a4b235d848e842257f5baf22b6ead03fd6c91
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:ed3fdd1e85f5c5434de6ec0feaec8c7fe4e680ca6c64258287f18ec2886bb71f
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:f4b935c16dcbbc0393733437747497c1d24f2cc782c40c49738e5f37c037b3e7
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:6251cb303774e29d137a37d6527875c371fe1d3f6a5fa0383e80f0e9ef5c02be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:5e87ba403f9e9fff3002a8275aba0d6c36aed54fb4754d90fa0516ad989f187a
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:9d12e6cbf26e31a1e4458c99a972ed36125d759c5ec57773f203cd8dd0985de7
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:1c7c24111f1709d7648156873ced09462638aa4b3cf7e7aee13297e51bcdfb3e
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:1a6ca51c2c68950af605ff367356ab6c0ef7d1b2fd6fb8f00aa665196ff7af84
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:4e9c152023b2db04ee5d55f102490294b393bf78baa3aa8917f85f8547fbe3c8
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:f87c2069105121927fb0c0a2e37f8f72bc6cb5e035fa8d22f7b41ce0e86c0631
        - registry.redhat.io/ubi9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:538416bc1384eb987e8a7e80b57eda3feddaf4260b196709a6e9a9125e3af39e
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:c60ca44275297d2ee8c499d9d7fb4056d104f4bd5dec8a02e7388241ac466c44
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:aadba00eadb3ac46d321a14498fc3dffd5be9afbf698d80adddde0a7a1681fc4
        version: 1.4.1
      entries:
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      name: builds-1.4
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      name: builds-1.5
    - currentCSV: openshift-builds-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "OpenShiftBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "sharedResource": {
                    "state": "Enabled"
                  },
                  "shipwright": {
                    "build": {
                      "state": "Enabled"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.shipwright.io/v1alpha1",
                "kind": "ShipwrightBuild",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "targetNamespace": "openshift-builds"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/openshift-builds/openshift-builds-operator-rhel9
          createdAt: "2025-08-14T12:12:15Z"
          description: Builds for Red Hat OpenShift is a framework for building container
            images on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-builds
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["openshiftbuilds.operator.openshift.io","shipwrightbuilds.operator.shipwright.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/shipwright-io/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              OpenShiftBuild describes the desired state of Builds for OpenShift, and the status of
              all deployed components.
            displayName: Open Shift Build
            kind: OpenShiftBuild
            name: openshiftbuilds.operator.openshift.io
            version: v1alpha1
          - description: ShipwrightBuild represents the deployment of Shipwright's
              build controller on a Kubernetes cluster.
            displayName: Shipwright Build
            kind: ShipwrightBuild
            name: shipwrightbuilds.operator.shipwright.io
            version: v1alpha1
          required:
          - kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
        description: "Builds for Red Hat OpenShift is an extensible build framework
          based on the Shipwright project, \nwhich you can use to build container
          images on an OpenShift Container Platform cluster. \nYou can build container
          images from source code and Dockerfile by using image build tools, \nsuch
          as Source-to-Image (S2I), Buildah, and Buildpacks. You can create and apply
          build resources, view logs of build runs, \nand manage builds in your OpenShift
          Container Platform namespaces.\nRead more: [https://shipwright.io](https://shipwright.io)\n\n##
          Features\n\n* Standard Kubernetes-native API for building container images
          from source code and Dockerfile\n\n* Support for Source-to-Image (S2I) and
          Buildah build strategies\n\n* Extensibility with your own custom build strategies\n\n*
          Execution of builds from source code in a local directory\n\n* Shipwright
          CLI for creating and viewing logs, and managing builds on the cluster\n\n*
          Integrated user experience with the Developer perspective of the OpenShift
          Container Platform web console\n"
        displayName: Builds for Red Hat OpenShift Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - build
        - shipwright
        - tekton
        - cicd
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/builds_for_red_hat_openshift/1.4
        - name: Builds for Openshift
          url: https://github.com/redhat-openshift-builds/operator
        maintainers:
        - email: openshift-builds@redhat.com
          name: Red Hat OpenShift Builds Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift-builds/openshift-builds-waiters-rhel9@sha256:753e24a1cd57ca22552447363aed48f064b525b49e52f774f3e25f41d8dcb329
        - registry.redhat.io/openshift-builds/openshift-builds-image-bundler-rhel9@sha256:38560dea7627632d8b24fa670c1d7ac1218cf86975329908786a460f76de558d
        - registry.redhat.io/openshift-builds/openshift-builds-operator-bundle@sha256:4103bbce41d2f86071f49260e74ae89fd7a556deab7adea686cbe69477f17b58
        - registry.redhat.io/ubi9/buildah@sha256:8ea2b81b32f70b9502d88722a5f69de8c0cf5efa6f30df041873546e9f9438cb
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-webhook-rhel9@sha256:1d5a0bfe0013ab7602efe95be8a814be1e5586728fdf0e198c97739cc12f1d87
        - registry.redhat.io/openshift-builds/openshift-builds-webhook-rhel9@sha256:f4a75c115b5dc8853845f666f3bd063b2736e01e428a605fb5cd0f02736ecce8
        - registry.redhat.io/openshift-builds/openshift-builds-controller-rhel9@sha256:e2a7a42cd5916c05d95d3fb0decc34d0840420532270422aa7f6b7bdd94de1f2
        - registry.redhat.io/openshift-builds/openshift-builds-shared-resource-rhel9@sha256:504179e8ee9f88a23f1d4c28c7683ab7395f4c1c4868765bf3c1406669fd52fc
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c1a84feb88d53e93bdcf2cd5f76e2cbb18541c8b0e2979c132a888d6c280b664
        - registry.redhat.io/openshift-builds/openshift-builds-git-cloner-rhel9@sha256:bcae84267f582278591d93aecf94c8d7fb1376521d9b5208201c959a0fe56774
        - registry.redhat.io/openshift-builds/openshift-builds-image-processing-rhel9@sha256:4dc123fa703006662efcc5569a843a043bb31cf69b6868d6a9b6321c66bfca18
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ce7050192d9850ed4c35cb3295ccdeec1aa593e27e3a3da6d6aca97cde782abe
        - registry.redhat.io/openshift-builds/openshift-builds-rhel9-operator@sha256:ddcfcebd5d788d11f756939f07215b2e3507a1d8fe704a1052e15bb4ebe592e8
        version: 1.5.1
      entries:
      - name: openshift-builds-operator.v1.5.1
        version: 1.5.1
      - name: openshift-builds-operator.v1.5.0
        version: 1.5.0
      - name: openshift-builds-operator.v1.4.1
        version: 1.4.1
      - name: openshift-builds-operator.v1.4.0
        version: 1.4.0
      - deprecation:
          message: '''openshift-builds-operator.v1.3.2'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.2
        version: 1.3.2
      - deprecation:
          message: '''openshift-builds-operator.v1.3.1'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.1
        version: 1.3.1
      - deprecation:
          message: '''openshift-builds-operator.v1.3.0'' has reached the end of maintenance
            support. Please upgrade to ''openshift-builds-operator.v1.5.1''.'
        name: openshift-builds-operator.v1.3.0
        version: 1.3.0
      name: latest
    defaultChannel: latest
    packageName: openshift-builds-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Provider Name
      provider-url: https://your.domain
    name: smilecdr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smilecdr-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.smilecdr.com/v1alpha1",
                "kind": "Smilecdr",
                "metadata": {
                  "name": "smilecdr"
                },
                "spec": {
                  "labels": {
                    "application": "smilecdr",
                    "client": "smilecdr",
                    "env": "dev",
                    "version": "one"
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Smilecdr
            name: smilecdrs.operator.smilecdr.com
            version: v1alpha1
        description: Smilecdr Operator description. TODO.
        displayName: Smilecdr Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - smilecdr-operator
        links:
        - name: Smilecdr Operator
          url: https://smilecdr-operator.domain
        maintainers:
        - email: your@email.com
          name: Maintainer Name
        maturity: alpha
        provider:
          name: Provider Name
          url: https://your.domain
        relatedImages:
        - quay.io/jibran_shaukat/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - quay.io/jibran_shaukat/smilecdr@sha256:cd6b2cec6f5b9ad9e0f0434532f8fb00b6919fcd840e686d96ce08874f2154db
        - registry.connect.redhat.com/smilecdr/smilecdr@sha256:c44f5661a598aa18c4c10372918a45edb0f26d7d2772c47a55c66e4ff874cb5f
        version: 0.0.1
      entries:
      - name: smilecdr-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: smilecdr-operator
    provider:
      name: Provider Name
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: MariaDB Corporation
      provider-url: ""
    name: mariadb-enterprise-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-enterprise-operator.v25.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "suspend": false
                  },
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "key": "dsn"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "name": "my-logical-database"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "table": "*",
                  "username": "my-user"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "adminPasswordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "maxscale-admin"
                    },
                    "adminUsername": "mariadb-enterprise-operator",
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "maxscale-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "guiKubernetesService": {
                    "type": "ClusterIP"
                  },
                  "kubernetesService": {
                    "type": "ClusterIP"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "metrics": {
                    "enabled": true
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 2,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    }
                  ]
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "0 0 * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "sqljob"
                },
                "spec": {
                  "database": "mariadb",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS orders (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  customer_id INT,\n  order_date DATE,\n  total_amount_cents INT\n);\nINSERT INTO orders (customer_id, order_date, total_amount_cents) VALUES\n  (101, '2024-05-15', 5000),\n  (102, '2024-05-14', 7525),\n  (103, '2024-05-13', 10050),\n  (104, '2024-05-12', 2575);",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "enterprise.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "name": "my-user",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
          createdAt: "2025-08-28T11:35:03Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["MariaDB Enterprise"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: MariaDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API. It is used to define
              backup jobs and its storage.
            displayName: Backup
            kind: Backup
            name: backups.enterprise.mariadb.com
            version: v1alpha1
          - description: Connection is the Schema for the connections API. It is used
              to configure connection strings for the applications connecting to MariaDB.
            displayName: Connection
            kind: Connection
            name: connections.enterprise.mariadb.com
            version: v1alpha1
          - description: Database is the Schema for the databases API. It is used
              to define a logical database as if you were running a 'CREATE DATABASE'
              statement.
            displayName: Database
            kind: Database
            name: databases.enterprise.mariadb.com
            version: v1alpha1
          - description: Grant is the Schema for the grants API. It is used to define
              grants as if you were running a 'GRANT' statement.
            displayName: Grant
            kind: Grant
            name: grants.enterprise.mariadb.com
            version: v1alpha1
          - description: MariaDB is the Schema for the mariadbs API. It is used to
              define MariaDB clusters.
            displayName: Maria DB
            kind: MariaDB
            name: mariadbs.enterprise.mariadb.com
            version: v1alpha1
          - description: MaxScale is the Schema for the maxscales API. It is used
              to define MaxScale clusters.
            displayName: Max Scale
            kind: MaxScale
            name: maxscales.enterprise.mariadb.com
            version: v1alpha1
          - description: PhysicalBackup is the Schema for the physicalbackups API.
              It is used to define physical backup jobs and its storage.
            displayName: Physical Backup
            kind: PhysicalBackup
            name: physicalbackups.enterprise.mariadb.com
            version: v1alpha1
          - description: Restore is the Schema for the restores API. It is used to
              define restore jobs and its restoration source.
            displayName: Restore
            kind: Restore
            name: restores.enterprise.mariadb.com
            version: v1alpha1
          - description: SqlJob is the Schema for the sqljobs API. It is used to run
              sql scripts as jobs.
            displayName: Sql Job
            kind: SqlJob
            name: sqljobs.enterprise.mariadb.com
            version: v1alpha1
          - description: User is the Schema for the users API.  It is used to define
              grants as if you were running a 'CREATE USER' statement.
            displayName: User
            kind: User
            name: users.enterprise.mariadb.com
            version: v1alpha1
        description: |
          MariaDB Enterprise Operator for Openshift is an operator that manages MariaDB Enterprise operands declaratively via Custom Resources, with support for the following features:

          * Provision and Configure MariaDB and MaxScale Declaratively: Define MariaDB Enterprise Server and MaxScale clusters in YAML manifests and deploy them with ease in Kubernetes.
          * High Availability with Galera: Ensure availability with MariaDB Enterprise Cluster, providing synchronous multi-master replication.
          * Query and Connection-Based Routing with MaxScale: MaxScale provides query routing and connection load balancing for improved application performance.
          * Cluster-Aware Rolling Updates: Perform rolling updates on MariaDB and MaxScale clusters, ensuring zero-downtime upgrades with no disruptions to your applications.
          * Flexible Storage Configuration and Volume Expansion: Easily configure storage for MariaDB instances, including the ability to expand volumes as needed.
          * Physical Backups based on [mariadb-backup](https://mariadb.com/docs/server/server-usage/backup-and-restore/mariadb-backup/full-backup-and-restore-with-mariadb-backup) and [Kubernetes VolumeSnapshots](https://kubernetes.io/docs/concepts/storage/volume-snapshots/). By leveraging the [BACKUP STAGE](https://mariadb.com/docs/server/reference/sql-statements/administrative-sql-statements/backup-commands/backup-stage) feature, backups are taken without long read locks or service interruptions.
          * Logical Backups based on [mariadb-dump](https://mariadb.com/docs/server/clients-and-utilities/backup-restore-and-import-clients/mariadb-dump).
          * Backup Management: Take, restore, and schedule backups with multiple storage types supported: S3, PVCs, Kubernetes volumes and VolumeSnapshots..
          * Policy-Driven Backup Retention: Implement backup retention policies with bzip2 and gzip compression.
          * Target Recovery Time: Restore your database to the closest available backup based on a specified recovery time.
          * Bootstrap New Instances: Initialize new MariaDB instances from backups, S3, PVCs or VolumeSnapshots to quickly spin up new clusters.
          * TLS Certificate Management: Issue, configure, and rotate TLS certificates and Certificate Authorities (CAs) for secure connections.
          * Advanced TLS Support: customize certificate lifetime, private key algorithm and TLS version.
          * Native Integration with cert-manager: Leverage [cert-manager](https://cert-manager.io/docs/), the de-facto standard for managing certificates in Kubernetes, to enable issuance with private CAs, public CAs and HashiCorp Vault.
          * Prometheus Metrics: Expose metrics using the MariaDB and MaxScale Prometheus exporters.
          * Native Integration with prometheus-operator: Leverage [prometheus-operator](https://github.com/prometheus-operator/prometheus-operator) to scrape metrics from MariaDB and MaxScale instances.
          * Declarative User and Database Management: Manage users, grants, and logical databases in a declarative manner using Kubernetes resources.
          * Secure, immutable and lightweight images based on Red Hat UBI, available for multiple architectires (amd64, arm64 and ppc64le).
          * [Operator certified ](https://catalog.redhat.com/en/software/container-stacks/detail/65789bcbe17f1b31944acb1d#overview) by Red Hat.

          ## Installation

          In order to use this product, you will need a [MariaDB Enterprise subscription](https://mariadb.com/products/enterprise/).

          To be able to pull MariaDB Enterprise images, you will need to configure the MariaDB registry credentials in Openshift using the [MariaDB customer download token](https://customers.mariadb.com/downloads/token/) as described below:

          - Extract your [Openshift global pull secret](https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/images/managing-images#images-update-global-pull-secret_using-image-pull-secrets):
            ```bash
            oc extract secret/pull-secret -n openshift-config --confirm
            ```

          - Login in the MariaDB registry providing the customer download token as password:
            ```bash
            oc registry login \
              --registry="docker.mariadb.com" \
              --auth-basic="<email>:<customer-download-token>" \
              --to=.dockerconfigjson
            ```

          - Update the global pull secret:
            ```bash
            oc set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson
            ```

          Alternatively, you can also create a `Secret` named `mariadb-enterprise` in the namespace where are you installing the operator by using the following command:
            ```bash
            oc create secret docker-registry mariadb-enterprise \
              --docker-server=docker.mariadb.com \
              --docker-username=<email> \
              --docker-password=<customer-download-token> \
              -n <operator-namespace>
            ```

          At this point, you can proceed to install the MariaDB Enterprise Operator, either via the Openshift console or by creating a `Subscription` object as described in the [Openshift installation docs](https://mariadb.com/kb/en/openshift/).

          Refer to the [quickstart guide](https://mariadb.com/kb/en/quickstart/), the [examples catalog](https://mariadb.com/kb/en/examples-catalog/) and the [API reference](https://mariadb.com/kb/en/api-reference/) for getting started with the Custom Resources offered by this operator.
        displayName: MariaDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - database
        - operator
        - mariadb-enterprise-operator
        - database
        links:
        - name: MariaDB Enterprise
          url: https://mariadb.com/products/enterprise/
        - name: Documentation
          url: https://mariadb.com/kb/en/mariadb-enterprise-operator/
        maintainers:
        - email: operator@mariadb.com
          name: MariaDB Corporation
        - email: support@mariadb.com
          name: Support
        - email: sales-EMEA@mariadb.com
          name: Sales EMEA
        - email: sales-APAC@mariadb.com
          name: Sales APAC
        - email: sales-nam@mariadb.com
          name: Sales Americas
        maturity: beta
        minKubeVersion: 1.29.0
        provider:
          name: MariaDB Corporation
        relatedImages:
        - mariadb/mariadb-prometheus-exporter-ubi@sha256:f072175198cad050e80617d13e63f415bacdfe47e2dd177fde7283a96ee56c07
        - mariadb/maxscale-prometheus-exporter-ubi@sha256:cc17c1554795ad01b4d8db799f6aa90e00e41ede431f09d30eec64d9dd4479d9
        - registry.connect.redhat.com/mariadb/mariadb-enterprise-operator-bundle@sha256:73e7f243a6676d559068bf8ab80cc0ffe0db58152675603ca8bd360334439f26
        - docker.mariadb.com/enterprise-server@sha256:9872721d35ebe8d820b1aa9c6dd83b363a7b2df9af11a65bff6b80ff4144460a
        - docker.mariadb.com/mariadb-enterprise-operator@sha256:1effbbfa7632be92c17ef61c18a25f010b187840beec42c31d2410d0209e6985
        - docker.mariadb.com/maxscale@sha256:d9e04a7a406353dab2051e190be0c32ed20b780ed71e56e55cedd4ccb61af416
        version: 25.8.1-1
      entries:
      - name: mariadb-enterprise-operator.v25.8.1-1
        version: 25.8.1-1
      - name: mariadb-enterprise-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-enterprise-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mariadb-enterprise-operator
    provider:
      name: MariaDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconAdmission",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-admission",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconadmission",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-kac"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "admission_controller"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-sidecar-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconcontainer",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-container-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconDeployment",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcondeployment-sample",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falcon-deployment",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-deployment"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconImageAnalyzer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "falcon-operator",
                    "app.kubernetes.io/instance": "falconimageanalyzer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "falconimageanalyzer",
                    "app.kubernetes.io/part-of": "falcon-operator"
                  },
                  "name": "falcon-image-analyzer"
                },
                "spec": {
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "labels": {
                    "crowdstrike.com/component": "sample",
                    "crowdstrike.com/created-by": "falcon-operator",
                    "crowdstrike.com/instance": "falcon-node-sensor",
                    "crowdstrike.com/managed-by": "kustomize",
                    "crowdstrike.com/name": "falconnodesensor",
                    "crowdstrike.com/part-of": "Falcon",
                    "crowdstrike.com/provider": "crowdstrike"
                  },
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security,Monitoring
          containerImage: registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
          createdAt: "2025-06-03T22:34:20Z"
          description: Falcon Operator installs CrowdStrike Falcon Sensors on the
            cluster
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: falcon-operator
          operators.openshift.io/valid-subscription: Not required to use this operator
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: registry.connect.redhat.com/crowdstrike/falcon-operator
          support: CrowdStrike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconAdmission is the Schema for the falconadmissions API
            displayName: Falcon Admission
            kind: FalconAdmission
            name: falconadmissions.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconDeployment is the Schema for the falcondeployments
              API
            displayName: Falcon Deployment
            kind: FalconDeployment
            name: falcondeployments.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconImageAnalyzer is the Schema for the falconImageAnalyzers
              API
            displayName: Falcon Image Analyzer
            kind: FalconImageAnalyzer
            name: falconimageanalyzers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          The CrowdStrike Falcon Operator installs CrowdStrike Falcon custom resources on a Kubernetes cluster.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon to the cluster. The operator exposes custom resources that allow you to protect your Kubernetes clusters when deployed.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--                                             | :--                                                    |
          | [FalconAdmission](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/admission/README.md) | Manages installation of Falcon Admission Controller on the cluster |
          | [FalconImageAnalyzer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/imageanalyzer/README.md) | Manages installation of Falcon Image Assessment at Runtime on the cluster |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |
          | [FalconDeployment](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/falcondeployment/README.md)       | Deploys FalconAdmission, FalconImageAnalyzer, FalconContainer, and FalconNodeSensor CRs from a single manifest |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: CrowdStrike Falcon Platform - Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - registry.connect.redhat.com/crowdstrike/falcon-operator-bundle@sha256:df0dcd31867bb1c3c4321db4412a744fd1b3ba8faddb509d191fab92407b6677
        - registry.connect.redhat.com/crowdstrike/falcon-operator@sha256:ffe00e9e22e209d1e08f29c3d257bbd51937fcbde68a1ed93b542a26dd436dee
        - registry.crowdstrike.com/falcon-container/us-1/release/falcon-sensor@sha256:92d222cd604c9f2c9e9f2c9b402b9fb0858fcb04dac38c15e25125978f43c99b
        - registry.crowdstrike.com/falcon-imageanalyzer/us-1/release/falcon-imageanalyzer@sha256:f08603648658041ad52ccf0cfd058da40b385904830030aa82dc50609aed7ceb
        - registry.crowdstrike.com/falcon-kac/us-1/release/falcon-kac@sha256:5049128dcfe9565f39fc6e28fb94f5bf05fb7c51cf787fc6fe5c63a2b04f800d
        - registry.crowdstrike.com/falcon-sensor/us-1/release/falcon-sensor@sha256:ed4f3fe8d9f5f491600f18d4440dc1d8e3a07d4b9c08a1f17382a23a3a5cf523
        version: 1.8.0
      entries:
      - name: falcon-operator.v1.8.0
        version: 1.8.0
      - name: falcon-operator.v1.7.0
        version: 1.7.0
      - name: falcon-operator.v1.6.0
        version: 1.6.0
      - name: falcon-operator.v1.5.0
        version: 1.5.0
      - name: falcon-operator.v1.4.0
        version: 1.4.0
      - name: falcon-operator.v1.3.1
        version: 1.3.1
      - name: falcon-operator.v1.3.0
        version: 1.3.0
      - name: falcon-operator.v1.2.0
        version: 1.2.0
      - name: falcon-operator.v1.1.0
        version: 1.1.0
      - name: falcon-operator.v1.0.1
        version: 1.0.1
      - name: falcon-operator.v1.0.0
        version: 1.0.0
      name: certified-1.0
    defaultChannel: certified-1.0
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-operator.v1.1.3-v2.6.4.final
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-mem"
                },
                "spec": {
                  "configuration": {
                    "persistence": "mem"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:1.1.3
          createdAt: "2024-09-17"
          description: Deploy and manage Apicurio Registry on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents an Apicurio Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, in-memory
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry on Kubernetes.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Apicurio Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:1.1.3
        - quay.io/community-operator-pipeline-prod/apicurio-registry@sha256:7087ec1377c2612edf3b8a76f6e115d66c3ec2b4b20f5d8374638d6bf0037e93
        version: 1.1.3-v2.6.4.final
      entries:
      - name: apicurio-registry-operator.v1.1.3-v2.6.4.final
        version: 1.1.3-v2.6.4.final
      - name: apicurio-registry-operator.v1.1.2-v2.5.11.final
        version: 1.1.2-v2.5.11.final
      - name: apicurio-registry-operator.v1.1.1-v2.5.9.final
        version: 1.1.1-v2.5.9.final
      - name: apicurio-registry-operator.v1.1.0-v2.4.12.final
        version: 1.1.0-v2.4.12.final
      - name: apicurio-registry-operator.v1.0.0-v2.0.0.final
        version: 1.0.0-v2.0.0.final
      name: 2.x
    defaultChannel: 2.x
    packageName: apicurio-registry
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-client-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-client-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
          createdAt: "2025-09-09T14:14:54Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:028120f2212738917f49e0330b60be5f79a208ce2974bea22b71f533aa6ce3be
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:ffcc0806ec2afb8b40d75c70ef2a6d98bb887879f5b3f573e900861316ce9e41
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:ca10a842900d6515f714a17916a32dfd5a2909543e4d339acefc7e3e41b8d7ac
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:53e67b715878df163997fef55505e76ea9289eb8d7f82021650552951f7fddbe
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:31d145c7074ad6cd85684b5191a3033533bf2149124c7ead5367871da5ed0dd8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:645d4fd09628f7ddcfc765e4ddacaa79fa6a7c35c03292ca2f6e317d48765d6b
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:1711a0aabd28937aac1424adebe96619a48b91f7f888f0588117f80891001da4
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:ef4276b80cacae825d4831de88d3fe6ba7cc324464087b7b28734bd2b8de3315
        version: 4.17.13-rhodf
      entries:
      - name: ocs-client-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-client-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-client-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-client-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-client-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-client-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-client-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-client-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-client-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-client-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-client-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-client-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-client-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-client-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-client-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-client-console"]'
          containerImage: registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
          createdAt: "2025-09-11T15:23:25Z"
          description: OpenShift Data Foundation client operator enables consumption
            of storage services from a remote centralized OpenShift Data Foundation
            provider cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["storageclaims.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/red-hat-storage/ocs-client-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageClaim is the Schema for the storageclaims API
            displayName: Storage Claim
            kind: StorageClaim
            name: storageclaims.ocs.openshift.io
            version: v1alpha1
          - description: StorageClient is the Schema for the storageclients API
            displayName: Storage Client
            kind: StorageClient
            name: storageclients.ocs.openshift.io
            version: v1alpha1
        description: OpenShift Data Foundation client operator enables consumption
          of storage services from a remote centralized OpenShift Data Foundation
          provider cluster.
        displayName: OpenShift Data Foundation Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: OCS Client Operator
          url: https://github.com/red-hat-storage/ocs-client-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:693a60d6a1162659213ae554b1c6ee200b0300a7da72c01415044cf507acab49
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6fee56851da3a87b9f7f477fc59e3385e118dc0b398861a8f13459c998a5af77
        - registry.redhat.io/odf4/ocs-client-console-rhel9@sha256:60a91d6e4fe7e0ab999eab2224ef09fd0681e74920b23b306c7a9893bbb29cb8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/odf4/ocs-client-operator-bundle@sha256:725eabb38876b93ab7693d75013d0292756a2efc00b8608f5716f31f8c324a53
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c11a599bfd09b9b22888e03eabdbe1054d4049718a591d2a45bf0ed89800111e
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:80c01633a6bfca44df37ac38f764035be6998fce965c8709e4f6d54d17eeaa7b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:82a4b481028195cfd75471656fe40c1cbbade4c7b2d170d6bf9422d9d3b84fa7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/ocs-client-rhel9-operator@sha256:03cbd69362f36f71a7c4de2ca6bc4c5016e073692cc514040756fe9ee72b885c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        version: 4.18.11-rhodf
      entries:
      - name: ocs-client-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-client-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-client-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-client-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-client-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-client-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-client-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-client-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-client-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-client-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-client-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-client-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-client-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-sso
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-sso.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Sso",
                "metadata": {
                  "name": "sso-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "kcos-keycloak": {
                    "auth": {
                      "baseURL": "https://keycloak.kcos-sso.svc.cluster.local:8443",
                      "enabled": true,
                      "sessionCookie": "AT"
                    },
                    "authproxy": {
                      "image": "prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044",
                      "initContainer": {
                        "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                      }
                    },
                    "crsCleanupJob": {
                      "clientExec": "oc",
                      "image": "prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "realm": {
                      "accessTokenLifespan": 266400,
                      "accessTokenLifespanForImplicitFlow": 266400,
                      "admin": {
                        "password": "admin"
                      },
                      "offlineSessionIdleTimeout": 2592000,
                      "offlineSessionMaxLifespan": 5184000,
                      "registrationAllowed": true,
                      "ssoSessionIdleTimeout": 266400,
                      "ssoSessionMaxLifespan": 266400
                    }
                  },
                  "subcharts": {
                    "enable": {
                      "kcos-keycloak": true,
                      "keycloak-operator": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:57Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Sso
            name: ssoes.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS SSO Operator
        displayName: Keysight KCOS SSO Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Sso
          url: https://keysight-kcos-sso.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-auth-proxy@sha256:a777ee5116280fbeb1e1342445fef5135f9ca61e983a3e2821f9ca4b2038b044
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-sso@sha256:298417fac2fc053bdc352f77b59b14e39eadd66c871cfb547304a517489f294e
        - prod.harbor.keysight.digital/loadcore/wap-ose-cli@sha256:34dc5f3dcd5938b32acaba60ef6e89a774d1106f3fa832370516ce41174f7f35
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-sso-bundle@sha256:f2c5a8abd1a9ec7665edf1e6206a6212cb752082b09532e4c218bc0990e48287
        version: 0.1.21
      entries:
      - name: keysight-kcos-sso.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-sso
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: https://trilio.io/
    name: tvo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tvo-operator.6.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample"
                },
                "spec": {
                  "crdType": "TVOControlPlane",
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "",
                    "my_ip": {
                      "host_interface": null
                    },
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "creator"
                  },
                  "database": {
                    "common": {
                      "host": "EDIT_DB_HOSTNAME",
                      "root_password": "EDIT_PASSWORD",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "EDIT_PASSWORD",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "EDIT_PASSWORD",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "",
                      "auth_url": "",
                      "ca_cert": "CERT MULTIPLE LINES\n",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "",
                      "keystone_auth_port": "",
                      "keystone_auth_protocol": "",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8784/v2",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8784/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "identity": null,
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "ENDPOINT_PROTOCOL_INTERNAL://ENDPOINT_HOST_INTERNAL:8781/v1/$(tenant_id)s",
                      "password": "EDIT_PASSWORD",
                      "public_auth_host": "",
                      "public_endpoint": "ENDPOINT_PROTOCOL_PUBLIC://ENDPOINT_HOST_PUBLIC:8781/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "5672"
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "S3_BT1",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY1",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET1",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL1",
                      "s3_region_name": "REGION1",
                      "s3_secret_key": "SECRETKEY1",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "S3_BT2",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "NFS_BT3",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "11.30.1.10:/mnt/share"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trilio-openstack
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:3d80cad27a1043fb61735e7a6fb156c98b269cfcc68f8b9712713a7d4f6ad6bd
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:b62229854368726cbb3ac494622afb9574845699a31b400aa4244ee6571364d1
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:50b7f0a5f3fdd98cfb0294d664cc01ffd1e5359be66a3da14f5552e9cb3bae2c
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        version: 6.1.1
      entries:
      - name: tvo-operator.6.1.1
        version: 6.1.1
      name: 6.0.x
    - currentCSV: tvo-operator.6.1.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOControlPlane",
                "metadata": {
                  "name": "tvocontrolplane-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "common": {
                    "dmapi_workers": "16",
                    "memcached_servers": "memcached-0.memcached.openstack.svc:11212",
                    "nova_group_id": "42436",
                    "nova_user_id": "42436",
                    "oslo_enable_proxy_headers_parsing": false,
                    "trustee_role": "member"
                  },
                  "crdType": "TVOControlPlane",
                  "database": {
                    "common": {
                      "host": "openstack.openstack.svc",
                      "port": 3306,
                      "root_password": "openstack",
                      "root_user_name": "root"
                    },
                    "datamover_api": {
                      "database": "dmapi",
                      "password": "dsw21aSyH5gHH6XE72k11Y3t3C308jY5rP02ndwJW",
                      "user": "dmapi"
                    },
                    "wlm_api": {
                      "database": "workloadmgr",
                      "password": "dfeeYb72OwUzX5YuRKcVjxXJaELV8YT0jZczIQwys0v",
                      "user": "workloadmgr"
                    }
                  },
                  "helm3_hook": true,
                  "images": {
                    "triliovault_datamover_api": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_datamover_db_sync": "registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cloud_trust": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_cron": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_db_sync": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_scheduler": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_wlm_workloads": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "keystone": {
                    "common": {
                      "admin_role_name": "admin",
                      "auth_uri": "https://keystone-internal.openstack.svc:5000/v3",
                      "auth_url": "https://keystone-internal.openstack.svc:5000",
                      "cloud_admin_domain_name": "Default",
                      "cloud_admin_project_name": "admin",
                      "cloud_admin_user_name": "admin",
                      "cloud_admin_user_password": "openstack",
                      "is_self_signed_ssl_cert": true,
                      "keystone_auth_host": "keystone-internal.openstack.svc",
                      "keystone_auth_port": "5000",
                      "keystone_auth_protocol": "https",
                      "region_name": "regionOne",
                      "service_project_domain_name": "Default",
                      "service_project_name": "service"
                    },
                    "datamover_api": {
                      "internal_endpoint": "https://triliovault-datamover-internal.trilio-openstack.svc:8784/v2",
                      "password": "defajgh48zgt88wFzzhwEt3xHGeCOdExYbLGQN2XRm",
                      "public_auth_host": "triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-datamover-public-trilio-openstack.apps.trilio.trilio.bos2/v2",
                      "service_desc": "TrilioVault Datamover Api Service",
                      "service_name": "dmapi",
                      "service_type": "datamover",
                      "user": "dmapi"
                    },
                    "keystone_interface": "internal",
                    "wlm_api": {
                      "internal_endpoint": "https://triliovault-wlm-internal.trilio-openstack.svc:8781/v1/$(tenant_id)s",
                      "password": "dsdeJl8axy0gv14fURwUUxy0qUOQbugUUOfXikz6j",
                      "public_auth_host": "triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2",
                      "public_endpoint": "https://triliovault-wlm-public-trilio-openstack.apps.trilio.trilio.bos2/v1/$(tenant_id)s",
                      "service_desc": "TrilioVault Workloadmgr Api Service",
                      "service_name": "TrilioVaultWLM",
                      "service_type": "workloads",
                      "user": "triliovault"
                    }
                  },
                  "labels": {
                    "datamover_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_api": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_cron": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_scheduler": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    },
                    "wlm_workloads": {
                      "node_selector_key": "trilio-control-plane",
                      "node_selector_value": "enabled"
                    }
                  },
                  "pod": {
                    "replicas": {
                      "triliovault_datamover_api": 3,
                      "triliovault_object_store": 3,
                      "triliovault_wlm_api": 3,
                      "triliovault_wlm_cron": 1,
                      "triliovault_wlm_scheduler": 3,
                      "triliovault_wlm_workloads": 3
                    },
                    "security_context": {
                      "triliovault": {
                        "container": {
                          "triliovault_datamover_api": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 0,
                            "runAsUser": 63630
                          },
                          "triliovault_datamover_api_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 63630,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_api": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_cron": {
                            "allowPrivilegeEscalation": false,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_init": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 0
                          },
                          "triliovault_wlm_scheduler": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "triliovault_wlm_workloads": {
                            "privileged": true,
                            "readOnlyRootFilesystem": false,
                            "runAsUser": 42436
                          },
                          "wlm_cloud_trust_creation": {
                            "readOnlyRootFilesystem": false,
                            "runAsGroup": 42436,
                            "runAsUser": 42436
                          }
                        }
                      }
                    }
                  },
                  "rabbitmq": {
                    "common": {
                      "admin_password": "EDIT_PASSWORD",
                      "admin_user": "admin",
                      "driver": "messagingv2",
                      "host": "EDIT_HOST",
                      "port": "15671",
                      "ssl": true
                    },
                    "datamover_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://dmapi:${PASSWORD}@${RABBITMQ_HOST}:5671/dmapi?ssl=1",
                      "user": "dmapi",
                      "vhost": "dmapi"
                    },
                    "wlm_api": {
                      "password": "EDIT_PASSWORD",
                      "transport_url": "rabbit://workloadmgr:${PASSWORD}@${RABBITMQ_HOST}:5671/workloadmgr?ssl=1",
                      "user": "workloadmgr",
                      "vhost": "workloadmgr"
                    }
                  },
                  "triliovault_backup_targets": [
                    {
                      "backup_target_name": "BT1_S3",
                      "backup_target_type": "s3",
                      "is_default": false,
                      "s3_access_key": "",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "",
                      "s3_region_name": "",
                      "s3_secret_key": "",
                      "s3_self_signed_cert": false,
                      "s3_signature_version": "default",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "amazon_s3"
                    },
                    {
                      "backup_target_name": "BT2_Other_S3",
                      "backup_target_type": "s3",
                      "is_default": true,
                      "s3_access_key": "ACCESSKEY2",
                      "s3_auth_version": "DEFAULT",
                      "s3_bucket": "BUCKET2",
                      "s3_bucket_object_lock_enabled": false,
                      "s3_endpoint_url": "ENDPOINT_URL2",
                      "s3_region_name": "REGION2",
                      "s3_secret_key": "SECRETKEY2",
                      "s3_self_signed_cert": true,
                      "s3_signature_version": "default",
                      "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N\n",
                      "s3_ssl_enabled": true,
                      "s3_ssl_verify": true,
                      "s3_type": "other_s3"
                    },
                    {
                      "backup_target_name": "BT3_NFS",
                      "backup_target_type": "nfs",
                      "is_default": false,
                      "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10",
                      "nfs_shares": "192.168.2.3:/nfs/share"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tvo.trilio.io/v1",
                "kind": "TVOBackupTarget",
                "metadata": {
                  "name": "tvobackuptarget-sample",
                  "namespace": "trilio-openstack"
                },
                "spec": {
                  "crdType": "TVOBackupTarget",
                  "images": {
                    "triliovault_wlm_api": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd",
                    "triliovault_object_store": "registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd"
                  },
                  "triliovault_backup_target": {
                    "backup_target_name": "S3_BT4",
                    "backup_target_type": "s3",
                    "is_default": false,
                    "s3_type": "other_s3",
                    "s3_access_key": "",
                    "s3_secret_key": "",
                    "s3_region_name": "",
                    "s3_bucket": "",
                    "s3_endpoint_url": "",
                    "s3_signature_version": "default",
                    "s3_auth_version": "DEFAULT",
                    "s3_bucket_object_lock_enabled": false,
                    "s3_ssl_enabled": true,
                    "s3_ssl_verify": true,
                    "s3_self_signed_cert": true,
                    "s3_ssl_ca_cert": "SSL_CERT_LINE_1\nSSL_CERT_LINE_2\nSSL_CERT_LINE_N",
                    "nfs_shares": "192.168.2.30:/nfs/share",
                    "nfs_options": "nolock,soft,timeo=600,intr,lookupcache=none,nfsvers=3,retrans=10"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
          createdAt: "2025-01-14T11:20:30Z"
          description: Cloud-Native Data Protection for OpenStack
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: tvo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trilioData/triliovault-cfg-scripts
          support: Trilio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trilio for OpenStack Control Plane. Includes all control
              plane components like Datamover Api, Trilio object store and Workloadmgr
              services.
            displayName: Trilio for OpenStack Control Plane
            kind: TVOControlPlane
            name: tvocontrolplanes.tvo.trilio.io
            version: v1
          - description: Trilio for OpenStack Backup Target. This CR is responsible
              to add a new backup target.
            displayName: Trilio for OpenStack Backup Target
            kind: TVOBackupTarget
            name: tvobackuptargets.tvo.trilio.io
            version: v1
        description: |-
          Trilio for OpenStack provides scalable, native, agentless, multi-tenant backup and recovery solutions, ensuring seamless protection of workloads with flexible storage options for dynamic cloud environments.

            * OpenStack Backup and Recovery: Recover any OpenStack application quickly in the event of data loss or misconfiguration. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Such as One-Click-Restore to entire project recovery on different clusters.
            * Disaster Recovery: Recover any OpenStack application quickly in the event of a disaster. Trilio enables easy restoration of entire workloads that reflect the clouds last best-known state. Trilio makes OpenStack disaster recovery easy.

            * Migration: A tenant or administrator can capture an application and its data and non-disruptively migrate that exact point-in-time to another OpenStack cloud. The need to migrate OpenStack workloads can be driven by economics, security or test/dev.

            * Network Functions Virtualization backup and recovery: Business critical services like NFV can quickly be configured to any state, assuring the optimal quality of service. Trilio provides recovery from data damage and allows restoration of entire VNFs, selected VNFs, or individual data items using a point-in-time working copy. The platform also allows recovery from a disaster by restoring the OpenStack network topology and populating it with a working copy of the entire NFV.

          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        - Backup
        - Recovery
        - RHOSO Backup
        - Disaster Recovery
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/openstack-backup-and-recovery/
        - name: Operator Documentation
          url: https://docs.trilio.io/openstack
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        provider:
          name: Trilio
          url: https://trilio.io/
        relatedImages:
        - registry.connect.redhat.com/trilio/trilio-wlm@sha256:32d355d9998fdb7489ccd37534f235d5d161484222cbb4e99de862d1c5a083dd
        - registry.connect.redhat.com/trilio/tvo-operator-rhoso@sha256:29d46db99d97fe28758c0efd2acf27203ce9958734b82cf37c3a9063b5159506
        - registry.connect.redhat.com/trilio/trilio-datamover-api-rhoso@sha256:7d3f9bdde36d68a38e1e45007d222da2e3194bb212637d39fa1aa352f633cf82
        - registry.connect.redhat.com/trilio/trilio-openstack-operator-rhoso@sha256:e282bb056a1780e5170db3e63efdf5099a5ac5240ad44162bc7af83cac893ac7
        version: 6.1.9
      entries:
      - name: tvo-operator.6.1.9
        version: 6.1.9
      name: 6.1.x
    defaultChannel: 6.0.x
    packageName: tvo-operator
    provider:
      name: Trilio
      url: https://trilio.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hbase-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hbase-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
          description: Stackable Operator for Apache HBase
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hbase-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HbaseClusterSpec via `CustomResource`
            displayName: hbaseclusters.hbase.stackable.tech
            kind: HbaseCluster
            name: hbaseclusters.hbase.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache HBase
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hbase
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hbase-operator@sha256:3d6d1cff4d20cbf5264971b702bc461ba2ab0612fa6869148a12ad0401eb914b
        - registry.connect.redhat.com/stackable/stackable-apache-hbase-operator@sha256:3864333176fb7c28e26f5c1f56f082cdefe1ec09974e01f908a731342a567d61
        version: 25.7.0
      entries:
      - name: hbase-operator.v25.7.0
        version: 25.7.0
      - name: hbase-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hbase-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/streamshub
    name: amq-streams-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: alpha
    - currentCSV: amq-streams-console.v2.8.0-13-0.1738265629.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
          createdAt: "2024-11-12T15:34:06Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: amq-streams-console.v2.8.0-13
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.8
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:1b1b80710885b43600e22ff2a24930b5cab06cd91fad439ad8a282c00b4564b9
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:38b80d81e54d59667c1f7dfa4da4125d5c7430816da130ddc84678a4e3cff6cf
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:366d60ff48fb9ba32dc73d92a775046d7b1eb1df21c93e31605372c7858f16ca
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:b83c3263e77904ce1c0f95b7bfab969605c97297a5598fd794d64daf508c52c0
        version: 2.8.0-13+0.1738265629.p
      entries:
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.8.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.9.x
    - currentCSV: amq-streams-console.v2.9.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
          createdAt: "2025-09-22T10:41:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0-0 <2.9.2-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/2.9
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:839368767bec735cd12f002874917c23f58550fe21df976858bdc92946985ec6
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:b7444b70a5dc4543a9976c08a619ff6978c9ff86ac7085d300b4870b7c6b7851
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:cca278607293683a675ff1ff33cb54a8e0d030388932db410c8b174034a5b2b6
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:5889a18a5517e770e51962afab9c9955580770a1760c0f2c845cde88404c9d8d
        version: 2.9.2-2
      entries:
      - name: amq-streams-console.v2.9.2-2
        version: 2.9.2-2
      - name: amq-streams-console.v2.9.2-1
        version: 2.9.2-1
      - name: amq-streams-console.v2.9.1-7
        version: 2.9.1-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-2.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-console.v3.0.1-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"console.streamshub.github.com/v1alpha1",
                "kind":"Console",
                "metadata":{
                  "name":"example-console"
                },
                "spec":{
                  "hostname": "<console-hostname-for-ingress>",
                  "kafkaClusters": [
                    {
                      "name": "<strimzi-kafka-resource-name>",
                      "namespace": "<strimzi-kafka-resource-namespace>",
                      "listener": "<strimzi-kafka-listener-name>",
                      "credentials": {
                        "kafkaUser": {
                          "name": "<strimzi-kafka-user-name-for-connection>"
                        }
                      }
                    }
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "example-console": {
                "description": "Example Console instance with reference to on-cluster Strimzi Kafka resource"
              }
            }
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
          createdAt: "2025-09-24T12:01:17Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-3'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          repository: https://github.com/streamshub/console
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Console deployment
            displayName: Console
            kind: Console
            name: consoles.console.streamshub.github.com
            version: v1alpha1
        description: |-
          The Red Hat streams for Apache Kafka Console provides a web-based user interface tool for monitoring Apache Kafka® instances within an OpenShift cluster.

          It features a user-friendly way to view Kafka topics and consumer groups, facilitating the searching and filtering of streamed messages. The console also offers insights into Kafka broker disk usage, helping administrators monitor and optimize resource utilization. By simplifying complex Kafka operations, the streams for Apache Kafka Console enhances the efficiency and effectiveness of data streaming management within OpenShift environments.
        displayName: Streams for Apache Kafka Console
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        - web
        - console
        - ui
        - user interface
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        provider:
          name: Red Hat
          url: https://github.com/streamshub
        relatedImages:
        - registry.redhat.io/amq-streams/console-api-rhel9@sha256:7f3638288c8a354ba9feceeb14e990ed7a09e3ef245d6fd288cd0bd8ad6c2c75
        - registry.redhat.io/amq-streams/console-operator-bundle@sha256:405eca4deeb4b02827e37b7a9115f73c1ff52dcbd851ed973d3a878634057ee2
        - registry.redhat.io/amq-streams/console-rhel9-operator@sha256:622cdec705ee48b0d2c3faa251dff65e0dcde3c2557fbf8408b1dc69ed1cc90c
        - registry.redhat.io/amq-streams/console-ui-rhel9@sha256:b0f8e0b895e1a4667da521ddacfc476c7fa2b9b0983044bd139978181499c774
        version: 3.0.1-3
      entries:
      - name: amq-streams-console.v3.0.1-3
        version: 3.0.1-3
      - name: amq-streams-console.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-console.v3.0.0-10
        version: 3.0.0-10
      - name: amq-streams-console.v3.0.0-7
        version: 3.0.0-7
      - name: amq-streams-console.v2.9.1-5
        version: 2.9.1-5
      - name: amq-streams-console.v2.9.1-2
        version: 2.9.1-2
      - name: amq-streams-console.v2.9.0-8
        version: 2.9.0-8
      - name: amq-streams-console.v2.9.0-6
        version: 2.9.0-6
      - name: amq-streams-console.v2.9.0-5
        version: 2.9.0-5
      - name: amq-streams-console.v2.9.0-3
        version: 2.9.0-3
      - name: amq-streams-console.v2.8.0-13
        version: 2.8.0-13
      - name: amq-streams-console.v2.8.0-13-0.1738265629.p
        version: 2.8.0-13+0.1738265629.p
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-console
    provider:
      name: Red Hat
      url: https://github.com/streamshub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzounstructured-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
          createdAt: "2025-07-08T09:47:55Z"
          description: Kubernetes Operator for Altair Graph Studio Unstructured
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: Altair Graph Studio Unstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Altair Graph Studio Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Altair Graph Studio Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v3.1.2/README.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-microservices-leader@sha256:0637603c00a99533d57e16009516ec73d40703ae151e631fc281f51f1f0ca93b
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:ea88a3090291f3742ccd4641785c19a80f4d3bd485a4f9c362a7645f73896db3
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-worker@sha256:964633d6b16ddc517a09ac86f3e3698f5a161a72cfa50ea9ba4e301fa631e347
        - registry.connect.redhat.com/cambridgesemantics/unstructured-operator@sha256:633a8992e35601d99c0d690dd71af314605c223b57c2038fa7090be06977b2f9
        version: 3.1.200
      entries:
      - name: anzounstructured-operator.v3.1.200
        version: 3.1.200
      - name: anzounstructured-operator.v3.1.104
        version: 3.1.104
      - name: anzounstructured-operator.v3.1.103
        version: 3.1.103
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.13-xed86eae
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"prometheus":{"serviceMonitor":{"enabled":false}},"mode":"cluster-pool"},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/hubble-export-stdout@sha256:f1b1475a3c89950481264f5f8ca63eea3b911a45dc83fdf15122d695a0fba9b3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-enterprise@sha256:1e1d3e6c199546b2398f4166ff64210c04075bb407c8f07c5be8d3db08560d4a
        - quay.io/isovalent/hubble-rbac@sha256:03f3304c157024356a6de13c3d91b4506b8e54d587920d80ebcc76cbcfeaeda6
        - quay.io/isovalent/hubble-relay@sha256:9145b03acc954efde54e1ab9da878462526c5e820b83a92e32cc24291f784d33
        - quay.io/isovalent/operator-generic@sha256:1cc0b0d2e71564ad8759be326c66ea13b4955b9441fb67e906da534e1b38aefa
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:ecd26b74a01f2b547ddaed4d32d35f8f5e09c378d5c1fc6cfa63f0adf659ac2b
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/cilium@sha256:6887537789603946543aaa370a3ec592287a235231bf16aa386e091823285515
        - quay.io/isovalent/clustermesh-apiserver@sha256:eb3ee3d2cb07e6c5966e4bc2e50615949303be5d5128e156d5c0dbe8404dc093
        - quay.io/isovalent/hubble-export-fluentd@sha256:05cb13f533f2c89b98a47cd58badda07add54077cd22283ca5028cb39f2e4486
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:cc1f0e6b27ccda8a843dfb7fa4a8fef052ba29716646557815677c5d1bea2a1a
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/isovalent/hubble-export-s3@sha256:ce4f71f7f13301133b7d471b87fb81108b2c2da7fc29865ef96bc8431b2be72b
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:3dbb71004362a70683a5ee9f0916bca75c0d1c351d144677fd9d85bcb2ca8bf7
        - quay.io/isovalent/hubble-ui-enterprise-backend@sha256:e892cba6dafcdf6704df3135a0b38d11894a5f276ace40683e74cb62b9c311d8
        - quay.io/isovalent/hubble-enterprise-operator@sha256:0d5e6a5d9c07780c18e5aa81d8f1c95c399bff2d36a44fa5f61ac89a788ca83b
        - quay.io/isovalent/hubble-ui-enterprise@sha256:c9921c65892a5293a4f1bef787636d6ae58cec1af65da06862a47e3450afdc6e
        - quay.io/isovalent/hubble-enterprise-metadata@sha256:72c7aa3961f14df413569cd156820a45553bb24d393b11246a18bf8b6553e022
        version: 1.10.13+xed86eae
      entries:
      - name: cilium.v1.10.13-xed86eae
        version: 1.10.13+xed86eae
      - name: cilium.v1.10.12-xb203fae
        version: 1.10.12+xb203fae
      - name: cilium.v1.10.11-x10ae342
        version: 1.10.11+x10ae342
      name: "1.10"
    - currentCSV: cilium.v1.11.21-xd0d07b1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium Enterprise - eBPF-based Networking, Security, and Observability
        displayName: Cilium Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Enterprise Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Cilium Enterprise
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.23.0@sha256:2fbdad86bb5244075f31487262d790f6fa75cdd993f868b952252905c118621c
        - quay.io/isovalent/hubble-ui-enterprise:v0.23.0@sha256:02635d1392cca7e09ab7797f69147880068ba7e3b0d865be97eeab33edd646a0
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:89d480ae76480829660d4cccfede74717faeb7e62a4a48b932750e41a945783d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:d3148bf3b9765467936ee067b25b5ef3683d3e55260f517c7fe271ac0e19cb47
        - quay.io/coreos/etcd:v3.4.13@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-relay:v1.11.21-cee.1@sha256:5933dd1f8111d6bca0c68db9247385b5d215d5b613095bafab8244098f2c5a26
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - cgr.dev/chainguard/oauth2-proxy:7.4.0@sha256:54857c70321004d3e0b8b8d7208ce5422c1f71c18db5be02d2543f187ea8e8e2
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/operator-generic:v1.11.21-cee.1@sha256:25b668c1febe1d1a70b63dfe926be4e55fbf77fd186e9df7581e41d7e9813469
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/clustermesh-apiserver:v1.11.21-cee.1@sha256:e16d8375ed8a1c3a872047b3aacf3469bde43552f1a51ee3e159ef2e9fea3e97
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:372511f6fa3523dadfe9128caf77520cf74ce24df4a19e1583ee197375e06014
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/cilium:v1.11.21-cee.1@sha256:f41241a15cd96d26727aa1ec413ace643728ddbe42f7fb7b234dbda22bf3950b
        - quay.io/cilium/certgen:v0.1.5@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        version: 1.11.21+xd0d07b1
      entries:
      - name: cilium.v1.11.21-xd0d07b1
        version: 1.11.21+xd0d07b1
      - name: cilium.v1.11.20-x0f6d966
        version: 1.11.20+x0f6d966
      name: "1.11"
    - currentCSV: cilium.v1.12.21-x3861686
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.14 <1.12.21'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise-operator:v1.10.5@sha256:31b18a1ea7002936f04dd070549e0c9a02f8a33394666fc5b0523ab1350c6120
        - quay.io/cilium/startup-script:62093c5c233ea914bfa26a10ba41f8780d9b737f@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/isovalent/hubble-relay:v1.12.21-cee.1@sha256:9d8dadc3f330acf0fa169c428d5fdf1d138d0294065eec5c628f6647106fe8b6
        - quay.io/isovalent/hubble-enterprise:v1.10.5@sha256:340700d4e28ded2aa9c50ff9f938996bb43faa84e0b2ba38ec79452648d7e5d7
        - quay.io/isovalent/hubble-ui-enterprise:v0.28.1@sha256:11dfba0e4fcf43415aa6e8b3b6e98a5c4f468ec67b377328ad48b92037bbfbb0
        - quay.io/isovalent/hubble-export-s3:v1.0.0@sha256:89c178c0a010166c68819ab5de32ab59b64a88aa23a05d8e4f78adb1a4b6c3d3
        - quay.io/isovalent/hubble-rbac:v1.3.0@sha256:4c106dd133861a678eb24cf98b35d5023cccca0834c954c49b68514c366aedf0
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-export-fluentd:v1.5.1@sha256:eaf42a84821509b69e2ef2b658796062e79a90e63045a98689da3da99a12e492
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:b6e40d76f0925e76f0a5b181be3f6c62728fa3e4aaef979caf9ea245f4ed03da
        - quay.io/isovalent/cilium:v1.12.21-cee.1@sha256:3a462308cf12b572db987c03d190ee33691ccb04775d01beb192703a52f8d47b
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-export-stdout:v1.0.3@sha256:bef600089c1671fce11bb4bcb98b81f25e65152a1a2f13b64ea637cfc5779c11
        - quay.io/isovalent/hubble-ui-enterprise-backend:v0.28.1@sha256:515aeb8166dd6995b04d4cd3333c94a740182943ba4e90a38871ee89c0b64e92
        - quay.io/isovalent/clustermesh-apiserver:v1.12.21-cee.1@sha256:3f6549409966086a49e338c7829b9141a73ef2ebeebaac91db430aeb3ccd2e22
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a71b6fe33997db924669253992436e1e02d88769bb97376c0871848c2d330fed
        - quay.io/isovalent/operator-generic:v1.12.21-cee.1@sha256:4e19cda799a747b69f53b5e5a860ba256399d18221d5a4bbf243142ba1c69bb8
        - quay.io/cilium/certgen:v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:f95c9e8002d8e381d604752b1d59b75595badbdb0b11237ac1cac4ecd04f4800
        version: 1.12.21+x3861686
      entries:
      - name: cilium.v1.12.21-x3861686
        version: 1.12.21+x3861686
      - name: cilium.v1.12.19-xca684e0
        version: 1.12.19+xca684e0
      - name: cilium.v1.12.18-xcb1adcf
        version: 1.12.18+xcb1adcf
      - name: cilium.v1.12.17-xae21c14
        version: 1.12.17+xae21c14
      - name: cilium.v1.12.16-x38e0b62
        version: 1.12.16+x38e0b62
      - name: cilium.v1.12.15-x8ca64a3
        version: 1.12.15+x8ca64a3
      - name: cilium.v1.12.14-xc86b260
        version: 1.12.14+xc86b260
      - name: cilium.v1.12.13-x2b5cec9
        version: 1.12.13+x2b5cec9
      - name: cilium.v1.12.12-xbd7b835
        version: 1.12.12+xbd7b835
      - name: cilium.v1.12.11-xffdea76
        version: 1.12.11+xffdea76
      - name: cilium.v1.12.10-x4a0e9e7
        version: 1.12.10+x4a0e9e7
      - name: cilium.v1.12.9-x4be66e4
        version: 1.12.9+x4be66e4
      - name: cilium.v1.12.8-x87b90a9
        version: 1.12.8+x87b90a9
      - name: cilium.v1.12.7-x53a2aab
        version: 1.12.7+x53a2aab
      name: "1.12"
    - currentCSV: cilium-enterprise.v1.13.21-x96449fb
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.7 <1.13.21+x96449fb'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-ui-enterprise:v1.0.3@sha256:55bf787c96561f4a460e2bfe5966421e2d85f3f9a38ac2487c6f6486c97cd5af
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:1074f45a625df6ea412a9ffef3d4f4e53f3107ca45bb1de9bb030c47ce6518d9
        - quay.io/coreos/etcd:v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-export-fluentd:v1.11.0@sha256:50cd2e4c2d04e8befcc2ae1a1463f07cba01edb9b546fb455fe7976311165385
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-relay:v1.13.21-cee.2@sha256:aaf5f959a76d6b517e627d2d0b09dab86ae942616819a9a95f057ed4f45ea480
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/cilium/certgen:v0.1.13@sha256:01802e6a153a9473b06ebade7ee5730f8f2c6cc8db8768508161da3cdd778641
        - quay.io/isovalent/cilium:v1.13.21-cee.2@sha256:f760cd85fa64eb8a197ec282644a8db630745a8a2854ee08aa2662cf67188db5
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:b04e75456359b50b18e75464da5d7b467fdfeb058fc08ce3a10f057ba5f60107
        - quay.io/isovalent/operator-generic:v1.13.21-cee.2@sha256:6f2d0d575e3f74395c959833992cd5fbb6971f4e292df1b7cf45be8b07be5e7a
        - quay.io/cilium/cilium-etcd-operator:v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.0.3@sha256:6756d2b808c38efce4765f90641d26aa2e0e79329ecc378a123cd3dc2c439640
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/hubble-rbac:v1.5.6@sha256:739ca45ea09829b0fa8a6198901b34eb5254df55be071bc854bbd37cae12a557
        - quay.io/isovalent/clustermesh-apiserver:v1.13.21-cee.2@sha256:980b0235fd16817ecc4457ad4317a4601993b76df2e81b07a53f6f437c2705af
        version: 1.13.21+x96449fb
      entries:
      - name: cilium-enterprise.v1.13.21-x96449fb
        version: 1.13.21+x96449fb
      - name: cilium-enterprise.v1.13.20-x5b70650
        version: 1.13.20+x5b70650
      - name: cilium-enterprise.v1.13.19-xdb27142
        version: 1.13.19+xdb27142
      - name: cilium-enterprise.v1.13.18-x0c21a94
        version: 1.13.18+x0c21a94
      - name: cilium-enterprise.v1.13.17-xc8f905f
        version: 1.13.17+xc8f905f
      - name: cilium-enterprise.v1.13.16-xcfe62ed
        version: 1.13.16+xcfe62ed
      - name: cilium-enterprise.v1.13.15-x1dfd26f
        version: 1.13.15+x1dfd26f
      - name: cilium.v1.13.14-xb70933f
        version: 1.13.14+xb70933f
      - name: cilium.v1.13.12-x453f6ad
        version: 1.13.12+x453f6ad
      - name: cilium.v1.13.11-xf6d6935
        version: 1.13.11+xf6d6935
      - name: cilium.v1.13.10-x9d57fc7
        version: 1.13.10+x9d57fc7
      - name: cilium.v1.13.9-xe8a359b
        version: 1.13.9+xe8a359b
      - name: cilium.v1.13.8-x96e1ca9
        version: 1.13.8+x96e1ca9
      - name: cilium.v1.13.7-x7320e7d
        version: 1.13.7+x7320e7d
      - name: cilium.v1.13.6-xd3714ef
        version: 1.13.6+xd3714ef
      - name: cilium.v1.13.5-x8386cae
        version: 1.13.5+x8386cae
      - name: cilium.v1.13.4-xc8b22ea
        version: 1.13.4+xc8b22ea
      - name: cilium.v1.13.3-x69e9561
        version: 1.13.3+x69e9561
      - name: cilium.v1.13.2-x4b1c6c8
        version: 1.13.2+x4b1c6c8
      name: "1.13"
    - currentCSV: cilium-enterprise.v1.14.22-x6d86c88
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.22+x6d86c88'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-enterprise-metadata:current@sha256:71cbb44a41b60ab64c6eae99f819796868a57234a5ff210598bd23b73b3b02c7
        - quay.io/isovalent/hubble-rbac:v1.5.2@sha256:29309837e2246a1b6849d7a09d01ceebfdb43fa0ade973d14584f0e52b856b95
        - quay.io/cilium/cilium-envoy:v1.32.5-1743506100-0821ef0acdf9f824d47d34e02932be522b3e7233@sha256:305c8b7da56ef1b64e4deb0642acf93a3944f2712e40c99a40bb005e6dc9761b
        - quay.io/isovalent/clustermesh-apiserver:v1.14.22-cee.1@sha256:588e2b173ab409288eda2d404571184562d3d9fd12e2d9afe7282a5a7d79113a
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:148b1029eac14049e2ba84d4d75be1ca43ef2d89951adbdeddec392ad7c9fe1e
        - quay.io/isovalent/hubble-relay:v1.14.22-cee.1@sha256:d7ffc880d2e42e0876417ec26115b0e9155ad9a2e7f47c02c434746afa42b29a
        - quay.io/isovalent/operator-generic:v1.14.22-cee.1@sha256:a997fe24ff0b7f7df3ee12748413ae53d6e07c803162b0db4d812b916b0cfc74
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:0543e615d8a48ba6aaf90a8de0b0a79d37689c64bfe71f91c340455d9a359426
        - quay.io/coreos/etcd:v3.5.4@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/isovalent/hubble-ui-enterprise:v1.2.2@sha256:3e03a58f6eea5d2df195f42ee2817f7b28fa20bbe140f5086937eee8da1b8d56
        - quay.io/isovalent/hubble-enterprise:v1.11.6@sha256:eb60d59d1fb1b136855829274b31805c4206f8d050c3bf068e63a552576c1b38
        - quay.io/isovalent/hubble-ui-enterprise-backend:v1.2.2@sha256:4258f7e0bff24ff1e0b214f62b37097032b6426ba28edd3dfc4f451a34de0110
        - quay.io/isovalent/hubble-export-s3:v1.0.2@sha256:518daea1074d260d824c3e03a93ab6b2a7a5f4c22b8e07938770add8257f4a52
        - quay.io/isovalent/oauth2-proxy:7.6.0@sha256:fc7046dee1a204163e68a6d9a42fec200b02129493a7d166f899dd12e2763364
        - quay.io/isovalent/hubble-export-stdout:v1.0.4@sha256:a0ff77efa890478f18f5deeba2cd1e2dfd6c5c77981adac57b14d7734de71ff1
        - quay.io/isovalent/hubble-export-fluentd:v1.8.1@sha256:ce0cce63a8473d5ae73872fc65d82c54f2875511bb03f0bc8ee8480148c43ee1
        - quay.io/isovalent/cilium:v1.14.22-cee.1@sha256:0e664a5a9e0edba38ab1a25eb7fa50861bb15f69ed0bbe3f3c792bd909240da3
        - quay.io/isovalent/hubble-enterprise-operator:v1.11.6@sha256:38253c2e17a8b2e88a226f0976c9cf4a00445212e2886e6f07cbb3a52757324f
        - quay.io/cilium/startup-script:c54c7edeab7fde4da68e59acd319ab24af242c3f@sha256:a9d0f451353f1b2f29ed3868e230ecfae10150eab9fd521de67adb3c8ca9c33a
        - quay.io/cilium/certgen:v0.1.17@sha256:07959f3125dae659ab9036c14df046e2536f7a4f6c92dabda837388b0344dc74
        version: 1.14.22+x6d86c88
      entries:
      - name: cilium-enterprise.v1.14.22-x6d86c88
        version: 1.14.22+x6d86c88
      - name: cilium-enterprise.v1.14.21-xe0c9553
        version: 1.14.21+xe0c9553
      - name: cilium-enterprise.v1.14.20-xe01e743
        version: 1.14.20+xe01e743
      - name: cilium-enterprise.v1.14.19-xc1b9aff
        version: 1.14.19+xc1b9aff
      name: "1.14"
    - currentCSV: cilium-enterprise.v1.15.21-x2661a01
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.21+x2661a01'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/oauth2-proxy:7.12.0-20250819@sha256:fe8f76728c36536d498b2a5c11172cb18dc30bbb555df2a6d14c036b29cb9f10
        - quay.io/isovalent/operator-generic-ubi:v1.15.21-cee.1@sha256:97daac94735e497e1e897beaa01eb50866abace1aa9a3919fc57971e45f538fe
        - quay.io/isovalent/cilium-envoy-ubi:v1.33.4-1752151664-7c2edb0b44cf95f326d628b837fcdd845102ba68@sha256:e81fdde48c3d5cd93111e1d51d9f8554adb32ee937601a9124354e0894933b99
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.15.21-cee.1@sha256:36d00eb9f8ed45a3d8892942cfae097782dd6509889ef7d1a451f9b0334d17e1
        - quay.io/isovalent/hubble-relay-ubi:v1.15.21-cee.1@sha256:8461f720ca95a9c543a298bee1c1e56241f56d5426030c13afed8262f2f5c5ff
        - quay.io/isovalent/certgen-ubi:v0.1.19@sha256:c32e96b0972c9bcd3896945c46ea23e330c9dea407afee3a637daae637296f4c
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:bd1a4d47ef48a6eb84c49d979e8f5ce867af5932828dfd3233fa9f78c2eac1f4
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:a93fdee9e8d05ebe65e9dd8682fa7bc99c54f9d9f26d2cc8519f529255805da8
        - quay.io/isovalent/hubble-ui-backend-ubi:v1.3.7@sha256:a4b5e7a032c1bbcd969a6873bc12935f8b6aa4c2d3a2ed87f5494fcfcdfd02d4
        - quay.io/isovalent/hubble-ui-frontend-ubi:v1.3.7@sha256:f2b4e935abe3eb94f82dab41c9d3d929fca71d82fa39a6c8002683375b5303a4
        - quay.io/isovalent/startup-script-ubi:1755531540-60ee83e@sha256:2e1d5048cbfcd3c77f732da8203a6c792cbf19897b5cf137b9ad23a6a267f6b2
        - quay.io/isovalent/cilium:v1.15.21-cee.1@sha256:39ef22d467783481eccb376fcb5d6bc254ed8f4b3b25032121498db40278f0b8
        version: 1.15.21+x2661a01
      entries:
      - name: cilium-enterprise.v1.15.21-x2661a01
        version: 1.15.21+x2661a01
      - name: cilium-enterprise.v1.15.20-x6df37bb
        version: 1.15.20+x6df37bb
      - name: cilium-enterprise.v1.15.19-x0be49ab
        version: 1.15.19+x0be49ab
      - name: cilium-enterprise.v1.15.18-xeeaa70b
        version: 1.15.18+xeeaa70b
      - name: cilium-enterprise.v1.15.17-xdfd9482
        version: 1.15.17+xdfd9482
      - name: cilium-enterprise.v1.15.16-xc7e31ca
        version: 1.15.16+xc7e31ca
      - name: cilium-enterprise.v1.15.15-xf37f22c
        version: 1.15.15+xf37f22c
      - name: cilium-enterprise.v1.15.14-x0e512e3
        version: 1.15.14+x0e512e3
      - name: cilium-enterprise.v1.15.13-x134fa29
        version: 1.15.13+x134fa29
      name: "1.15"
    - currentCSV: cilium-enterprise.v1.16.15-x6419e26
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.16.15+x6419e26'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/isovalent
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Enterprise for Cilium - eBPF-based Networking, Security,
          and Observability
        displayName: Isovalent Enterprise for Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Isovalent Enterprise for Cilium Homepage
          url: https://isovalent.com/
        maintainers:
        - email: maintainer@isovalent.com
          name: Isovalent
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/isovalent/hubble-relay-ubi:v1.16.15-cee.1@sha256:c992d549627e8f9180c272e7d96ff3783de184796633ed5d0b6fe2cf7e74055d
        - quay.io/isovalent/certgen-ubi:v1.16.15-cee.1@sha256:e3f1fa5ee59887d3a15c43947a853c8a3f4fc650ed72bcd891b277037ea16a9d
        - registry.connect.redhat.com/isovalent/cilium-ee-olm-metadata@sha256:9d825d79fffd39a1149b262d48d08616795e5cfdb4c4bc40f5f192242f71042d
        - quay.io/isovalent/cilium-envoy-ubi:v1.16.15-cee.1@sha256:843e6fdf6845f3d5f841dbd2dce75224b16c9d34c2b0ce7f329d9666466bfc6b
        - quay.io/isovalent/operator-generic-ubi:v1.16.15-cee.1@sha256:970aa156b6f5c390b6835a626190e18656f02db6a5cf27aaa1ac9c7872f4b15a
        - quay.io/isovalent/cilium-ubi:v1.16.15-cee.1@sha256:1d7bbbcb613ef20cd1719bd77473e71c198960ee5d8e563aec9c65cfd3c2efe2
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.16.15-cee.1@sha256:f7573ab4145d306051b5d6fa7ff1af395fc71935f5204839ab1c9f574aee17ad
        - quay.io/isovalent/startup-script-ubi:v1.16.15-cee.1@sha256:5d59a46be44c4d899576e69e0f20e148245e17d905a0769a73fd01f2432d1bab
        - registry.connect.redhat.com/isovalent/cilium-ee-olm@sha256:ec8a17956e75e91b60e2999d10e67ef76f9b4cfd754db415aaf387d21690da87
        version: 1.16.15+x6419e26
      entries:
      - name: cilium-enterprise.v1.16.15-x6419e26
        version: 1.16.15+x6419e26
      - name: cilium-enterprise.v1.16.13-xb2de075
        version: 1.16.13+xb2de075
      - name: cilium-enterprise.v1.16.12-xd373871
        version: 1.16.12+xd373871
      - name: cilium-enterprise.v1.16.11-xbd1cd23
        version: 1.16.11+xbd1cd23
      - name: cilium-enterprise.v1.16.10-x89ab4b0
        version: 1.16.10+x89ab4b0
      - name: cilium-enterprise.v1.16.9-x8ec4137
        version: 1.16.9+x8ec4137
      - name: cilium-enterprise.v1.16.8-yd2b50b4
        version: 1.16.8+yd2b50b4
      - name: cilium-enterprise.v1.16.7-x4867421
        version: 1.16.7+x4867421
      - name: cilium-enterprise.v1.16.6-x7e6eb1c
        version: 1.16.6+x7e6eb1c
      name: "1.16"
    defaultChannel: "1.16"
    packageName: cilium-enterprise
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: nexus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexus-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexus",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexus-sample"
                },
                "spec": {
                  "secret": "nexus-secret",
                  "url": "https://nexus-example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRepository",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrepository-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrepository",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrepository-sample"
                },
                "spec": {
                  "go": {
                    "proxy": {
                      "cleanup": {
                        "policyNames": [
                          "cleanup-policy"
                        ]
                      },
                      "httpClient": {
                        "autoBlock": true,
                        "blocked": true,
                        "connection": {
                          "retries": 1440,
                          "timeout": 1440,
                          "userAgentSuffix": "user-agent-suffix"
                        }
                      },
                      "name": "go-proxy",
                      "negativeCache": {
                        "enabled": true,
                        "timeToLive": 1440
                      },
                      "online": true,
                      "proxy": {
                        "contentMaxAge": 1440,
                        "metadataMaxAge": 1440,
                        "remoteUrl": "https://test.com"
                      },
                      "routingRule": "routing-rule",
                      "storage": {
                        "blobStoreName": "blob-store-name",
                        "strictContentTypeValidation": true
                      }
                    }
                  },
                  "nexusRef": {
                    "name": "nexus-sample"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexusrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexusrole",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexusrole-sample"
                },
                "spec": {
                  "description": "test-role",
                  "id": "test-role",
                  "name": "test-role",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "privileges": [
                    "nx-blobstores-all"
                  ]
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "NexusUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "nexus-operator",
                    "app.kubernetes.io/instance": "nexususer-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nexususer",
                    "app.kubernetes.io/part-of": "nexus-operator"
                  },
                  "name": "nexususer-sample"
                },
                "spec": {
                  "email": "test@gmail.com",
                  "firstName": "John1",
                  "id": "test-user",
                  "lastName": "Doe",
                  "nexusRef": {
                    "kind": "Nexus",
                    "name": "nexus-sample"
                  },
                  "roles": [
                    "nx-admin"
                  ],
                  "secret": "$user-secret:password"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/nexus-operator:3.0.0
          createdAt: "2023-12-11T12:54:50Z"
          description: An Operator for managing Nexus entities.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-nexus-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus is the Schema for the nexus API.
            displayName: Nexus
            kind: Nexus
            name: nexuses.edp.epam.com
            version: v1alpha1
          - description: NexusRepository is the Schema for the nexusrepositories API.
            displayName: Nexus Repository
            kind: NexusRepository
            name: nexusrepositories.edp.epam.com
            version: v1alpha1
          - description: NexusRole is the Schema for the nexusroles API.
            displayName: Nexus Role
            kind: NexusRole
            name: nexusroles.edp.epam.com
            version: v1alpha1
          - description: NexusUser is the Schema for the nexususers API.
            displayName: Nexus User
            kind: NexusUser
            name: nexususers.edp.epam.com
            version: v1alpha1
        description: |
          Nexus Operator is an operator that is responsible for establishing
          a connection to provided NexusRepository Server, reconciling Nexus entities (groups,
          users, repositories, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Nexus Repository.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  nexus-secret
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Nexus` with Nexus instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1alpha1
              kind: Nexus
              metadata:
                name: nexus-sample
              spec:
                secret: nexus-secret                # Secret name
                url: https://nexus.example.com      # Nexus URL
              ```

              Wait for the `.status` field with  `status.connected: true`
        displayName: EDP Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - edp
        - nexus
        - operator
        - artifact
        - storage
        links:
        - name: Nexus Operator
          url: https://github.com/epam/edp-nexus-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/nexus-operator:3.0.0
        - quay.io/openshift-community-operators/nexus-operator@sha256:277b94985c0dc982abb76b5dd8c93b37305ccca32c000d67552d540d4918cbba
        version: 3.0.0
      entries:
      - name: nexus-operator.v3.0.0
        version: 3.0.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Portworx by PureStorage
      provider-url: ""
    name: portworx-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: alpha
    - currentCSV: portworx-operator.v25.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageCluster",
                "metadata": {
                  "name": "portworx",
                  "namespace": "test-operator",
                  "annotations": {
                    "portworx.io/is-openshift": "true"
                  }
                },
                "spec": {}
              },
              {
                "apiVersion": "core.libopenstorage.org/v1",
                "kind": "StorageNode",
                "metadata": {
                  "name": "example",
                  "namespace": "test-operator"
                },
                "spec": {}
              }
            ]
          capabilities: Auto Pilot
          categories: Storage
          certified: "true"
          console.openshift.io/plugins: '["portworx"]'
          containerImage: registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
          createdAt: "2025-09-06T19:38:43Z"
          description: Cloud native storage solution for production workloads
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "core.libopenstorage.org/v1",
              "kind": "StorageCluster",
              "metadata": {
                "name": "portworx",
                "annotations": {
                  "portworx.io/is-openshift": "true"
                }
              }
            }
          support: Portworx, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Storage Cluster installs Portworx in the cluster. It has
              all the necessary configurations to setup and update a Portworx cluster.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.core.libopenstorage.org
            version: v1
          - description: Do not create Storage Node as it is internally created by
              the operator. It represents the status of a Portworx node.
            displayName: Storage Node
            kind: StorageNode
            name: storagenodes.core.libopenstorage.org
            version: v1
        description: |
          Portworx-Enterprise is the most widely-used and reliable cloud-native
          storage solution for production workloads and provides high-availability,
          data protection and security for containerized applications.

          Portworx Enterprise enables you to migrate entire applications, including
          data, between clusters in a single data center or cloud, or between clouds,
          with a single kubectl command.

          The cloud native storage and data management platform that enterprises trust
          to manage data in containers now has an operator which simplifies the install,
          configuration, upgrades and manages the Portworx Enterprise cluster lifecycle.

          Learn more about the Portworx Enterprise
          [the data platform for Kubernetes](https://portworx.com/products/introduction)

          To learn more about the platform features, please visit our
          [product features page](https://portworx.com/products/features)

          ### About Portworx

          Portworx is the solution for running stateful containers in production,
          designed with DevOps in mind. With Portworx, users can manage any database
          or stateful service on any infrastructure using any container scheduler,
          including Kubernetes, Mesosphere DC/OS, and Docker Swarm. Portworx solves
          the five most common problems DevOps teams encounter when running stateful
          services in production: persistence, high availability, data automation,
          security, and support for multiple data stores and infrastructure.

          ### How to install StorageCluster

          To customize your cluster's configuration (specification), use the
          [Spec Generator](https://central.portworx.com/) from PX-Central.

          ### Prerequisite

          Ensure ports 17001-17020 on worker nodes are reachable from master and other worker nodes.

          ### Tutorials

          * [Portworx Enterprise on Openshift](https://portworx.com/openshift)

          * [Stateful applications on Kubernetes](https://docs.portworx.com/portworx-enterprise/concepts/kubernetes-storage-101/applications)

          * [Portworx Enterprise on Kubernetes](https://docs.portworx.com/portworx-install-with-kubernetes)

          * [Kafka on Kubernetes](https://portworx.com/kafka-kubernetes)

          * [Elastisearch on Kubernetes](https://portworx.com/elasticsearch-kubernetes)

          * [PostgreSQL on Kubernetes](https://portworx.com/postgres-kubernetes/)

          * [MongoDB on Kubernetes](https://portworx.com/mongodb-kubernetes/)

          * [Cassandra on Kubernetes](https://portworx.com/cassandra-kubernetes/)

          * [Kubernetes backup and recovery](https://portworx.com/kubernetes-backup/)

          * [Disaster Recovery for Kubernetes](https://portworx.com/kubernetes-disaster-recovery/)

          ### Uninstall

          Deleting the StorageCluster object for Portworx cluster does not stop Portworx
          service running on the nodes, to avoid application downtime.

          To uninstall Portworx completely without wiping the data, you should add the
          following delete strategy to the StorageCluster spec:
          ```
          spec:
            deleteStrategy:
              type: Uninstall
          ```
          **Caution:** To uninstall Portworx and **wipe all the data**, you should use the following
          delete strategy:
          ```
          spec:
            deleteStrategy:
              type: UninstallAndWipe
          ```
        displayName: Portworx Enterprise
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - portworx
        - persistent storage
        - storage
        - cloud native
        - open source
        links:
        - name: Product Features
          url: https://portworx.com/products/features
        - name: Documentation
          url: https://docs.portworx.com/portworx-enterprise/platform/openshift
        - name: Support
          url: https://docs.portworx.com/portworx-enterprise/support/contact-support
        - name: Source Code
          url: https://github.com/libopenstorage/operator
        maintainers:
        - email: support@portworx.com
          name: Portworx
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Portworx by PureStorage
        relatedImages:
        - registry.connect.redhat.com/portworx/portworx-certified-bundle@sha256:79c17136650561ffe5d92d9f677d21de8ecf567e3749063d7c4790ed4ec78750
        - registry.connect.redhat.com/portworx/openstorage-operator@sha256:fcfaf7eb123e463c5f3bb1f20fa30e667ea5fe555a2b0734688a28e0d237ce24
        version: 25.3.1
      entries:
      - name: portworx-operator.v25.3.1
        version: 25.3.1
      - name: portworx-operator.v25.3.0
        version: 25.3.0
      - name: portworx-operator.v25.2.2
        version: 25.2.2
      - name: portworx-operator.v25.2.1
        version: 25.2.1
      - name: portworx-operator.v25.2.0
        version: 25.2.0
      - name: portworx-operator.v24.2.4
        version: 24.2.4
      - name: portworx-operator.v24.2.3
        version: 24.2.3
      - name: portworx-operator.v24.2.2
        version: 24.2.2
      - name: portworx-operator.v24.2.1
        version: 24.2.1
      - name: portworx-operator.v24.2.0
        version: 24.2.0
      - name: portworx-operator.v24.1.3
        version: 24.1.3
      - name: portworx-operator.v24.1.2
        version: 24.1.2
      - name: portworx-operator.v24.1.1
        version: 24.1.1
      - name: portworx-operator.v24.1.0
        version: 24.1.0
      - name: portworx-operator.v23.10.5
        version: 23.10.5
      - name: portworx-operator.v23.10.4
        version: 23.10.4
      - name: portworx-operator.v23.10.3
        version: 23.10.3
      - name: portworx-operator.v23.10.2
        version: 23.10.2
      - name: portworx-operator.v23.10.1
        version: 23.10.1
      - name: portworx-operator.v23.7.0
        version: 23.7.0
      - name: portworx-operator.v23.5.1
        version: 23.5.1
      - name: portworx-operator.v23.5.0
        version: 23.5.0
      - name: portworx-operator.v23.4.0
        version: 23.4.0
      - name: portworx-operator.v23.3.1
        version: 23.3.1
      - name: portworx-operator.v23.3.0
        version: 23.3.0
      - name: portworx-operator.v1.10.5
        version: 1.10.5
      - name: portworx-operator.v1.10.4
        version: 1.10.4
      - name: portworx-operator.v1.10.3
        version: 1.10.3
      - name: portworx-operator.v1.10.2
        version: 1.10.2
      - name: portworx-operator.v1.10.1
        version: 1.10.1
      - name: portworx-operator.v1.10.0
        version: 1.10.0
      - name: portworx-operator.v1.9.1
        version: 1.9.1
      - name: portworx-operator.v1.9.0
        version: 1.9.0
      - name: portworx-operator.v1.8.1
        version: 1.8.1
      - name: portworx-operator.v1.8.0
        version: 1.8.0
      - name: portworx-operator.v1.7.1
        version: 1.7.1
      - name: portworx-operator.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: portworx-certified
    provider:
      name: Portworx by PureStorage
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kinesis-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kinesis-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kinesis.services.k8s.aws/v1alpha1",
                "kind": "Stream",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
          createdAt: "2025-09-19T20:56:43Z"
          description: AWS Kinesis controller is a service controller for managing
            Kinesis resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Stream represents the state of an AWS kinesis Stream resource.
            displayName: Stream
            kind: Stream
            name: streams.kinesis.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kinesis resources in AWS from within your Kubernetes cluster.

          **About Amazon Kinesis**

          You can use Amazon Kinesis Data Streams to collect and process large streams of data records in real time. You can create data-processing applications, known as Kinesis Data Streams applications. A typical Kinesis Data Streams application reads data from a data stream as data records. These applications can use the Kinesis Client Library, and they can run on Amazon EC2 instances.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kinesis
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kinesis
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kinesis Developer Resources
          url: https://aws.amazon.com/kinesis/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kinesis maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kinesis-controller@sha256:de760dbe730e38cd9fadebe5fdb69299c04512e6a70f567ce678fc0bedc53e9c
        - public.ecr.aws/aws-controllers-k8s/kinesis-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kinesis-controller.v1.1.1
        version: 1.1.1
      - name: ack-kinesis-controller.v1.1.0
        version: 1.1.0
      - name: ack-kinesis-controller.v1.0.16
        version: 1.0.16
      - name: ack-kinesis-controller.v1.0.15
        version: 1.0.15
      - name: ack-kinesis-controller.v1.0.14
        version: 1.0.14
      - name: ack-kinesis-controller.v1.0.13
        version: 1.0.13
      - name: ack-kinesis-controller.v1.0.12
        version: 1.0.12
      - name: ack-kinesis-controller.v1.0.11
        version: 1.0.11
      - name: ack-kinesis-controller.v1.0.10
        version: 1.0.10
      - name: ack-kinesis-controller.v1.0.9
        version: 1.0.9
      - name: ack-kinesis-controller.v1.0.8
        version: 1.0.8
      - name: ack-kinesis-controller.v1.0.7
        version: 1.0.7
      - name: ack-kinesis-controller.v1.0.6
        version: 1.0.6
      - name: ack-kinesis-controller.v1.0.5
        version: 1.0.5
      - name: ack-kinesis-controller.v1.0.4
        version: 1.0.4
      - name: ack-kinesis-controller.v1.0.3
        version: 1.0.3
      - name: ack-kinesis-controller.v1.0.2
        version: 1.0.2
      - name: ack-kinesis-controller.v1.0.1
        version: 1.0.1
      - name: ack-kinesis-controller.v1.0.0
        version: 1.0.0
      - name: ack-kinesis-controller.v0.0.19
        version: 0.0.19
      - name: ack-kinesis-controller.v0.0.18
        version: 0.0.18
      - name: ack-kinesis-controller.v0.0.17
        version: 0.0.17
      - name: ack-kinesis-controller.v0.0.16
        version: 0.0.16
      - name: ack-kinesis-controller.v0.0.15
        version: 0.0.15
      - name: ack-kinesis-controller.v0.0.10
        version: 0.0.10
      - name: ack-kinesis-controller.v0.0.9
        version: 0.0.9
      - name: ack-kinesis-controller.v0.0.8
        version: 0.0.8
      - name: ack-kinesis-controller.v0.0.6
        version: 0.0.6
      - name: ack-kinesis-controller.v0.0.5
        version: 0.0.5
      - name: ack-kinesis-controller.v0.0.4
        version: 0.0.4
      - name: ack-kinesis-controller.v0.0.3
        version: 0.0.3
      - name: ack-kinesis-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-kinesis-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: jaeger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jaeger-operator.v1.65.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "jaegertracing.io/v1",
                "kind": "Jaeger",
                "metadata": {
                  "name": "my-jaeger"
                },
                "spec": {
                  "strategy": "allInOne"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing
          certified: "false"
          containerImage: quay.io/jaegertracing/jaeger-operator:1.65.0
          createdAt: "2025-01-22T20:40:19Z"
          description: Provides tracing, monitoring and troubleshooting for microservices-based
            distributed systems
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jaegertracing/jaeger-operator
          support: Jaeger Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Jaeger is the Schema for the jaegers API
            displayName: Jaeger
            kind: Jaeger
            name: jaegers.jaegertracing.io
            version: v1
        description: |-
          Jaeger, inspired by [Dapper](https://research.google.com/pubs/pub36356.html) and [OpenZipkin](http://zipkin.io/), is a distributed tracing system released as open source by Uber Technologies. It is used for monitoring and troubleshooting microservices-based distributed systems.
          ### Core capabilities
          Jaeger is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed context propagation
          * Distributed transaction monitoring
          * Root cause analysis
          * Service dependency analysis
          * Performance / latency optimization
          * OpenTracing compatible data model
          * Multiple storage backends: Badger, Cassandra, Elasticsearch, Memory.
          ### Operator features
          * **Multiple modes** - Supports `allInOne`, `production` and `streaming` [modes of deployment](https://www.jaegertracing.io/docs/latest/operator/#deployment-strategies).
          * **Configuration** - The Operator manages [configuration information](https://www.jaegertracing.io/docs/latest/operator/#configuring-the-custom-resource) when installing Jaeger instances.
          * **Storage** - [Configure storage](https://www.jaegertracing.io/docs/latest/operator/#storage-options) used by Jaeger. By default, `memory` is used. Other options include `badger`, `cassandra` or `elasticsearch`. On OpenShift, the operator can delegate creation of an Elasticsearch cluster to the Elasticsearch Operator if deployed.
          * **Agent** - can be deployed as [sidecar](https://www.jaegertracing.io/docs/latest/operator/#auto-injecting-jaeger-agent-sidecars) (default) and/or [daemonset](https://www.jaegertracing.io/docs/latest/operator/#installing-the-agent-as-daemonset).
          * **UI** - Optionally setup ingress (Kubernetes) or secure route (OpenShift) to provide [access to the Jaeger UI](https://www.jaegertracing.io/docs/latest/operator/#accessing-the-jaeger-console-ui).
          ### Before you start
          1. Ensure that the appropriate storage solution, that will be used by the Jaeger instance, is available and configured.
          2. If intending to deploy an Elasticsearch cluster via the Jaeger custom resource, then the Elasticsearch Operator must first be installed.
          ### Troubleshooting
          * https://www.jaegertracing.io/docs/latest/troubleshooting/
        displayName: Community Jaeger Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tracing
        - monitoring
        - troubleshooting
        links:
        - name: Jaeger Operator Source Code
          url: https://github.com/jaegertracing/jaeger-operator
        maintainers:
        - email: jaeger-tracing@googlegroups.com
          name: Jaeger Google Group
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: CNCF
        relatedImages:
        - quay.io/jaegertracing/jaeger-operator:1.65.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/jaeger@sha256:907dfe1d01f73c8a385da3661ddbe03db31e0b9ad6f5ab2336460402d0d915a0
        version: 1.65.0
      entries:
      - name: jaeger-operator.v1.65.0
        version: 1.65.0
      - name: jaeger-operator.v1.62.0
        version: 1.62.0
      - name: jaeger-operator.v1.61.0
        version: 1.61.0
      - name: jaeger-operator.v1.60.0
        version: 1.60.0
      - name: jaeger-operator.v1.59.0
        version: 1.59.0
      - name: jaeger-operator.v1.57.0
        version: 1.57.0
      - name: jaeger-operator.v1.56.0
        version: 1.56.0
      - name: jaeger-operator.v1.55.0
        version: 1.55.0
      - name: jaeger-operator.v1.53.0
        version: 1.53.0
      - name: jaeger-operator.v1.52.0
        version: 1.52.0
      - name: jaeger-operator.v1.51.0
        version: 1.51.0
      - name: jaeger-operator.v1.49.0
        version: 1.49.0
      - name: jaeger-operator.v1.48.1
        version: 1.48.1
      - name: jaeger-operator.v1.47.0
        version: 1.47.0
      - name: jaeger-operator.v1.46.0
        version: 1.46.0
      - name: jaeger-operator.v1.45.0
        version: 1.45.0
      - name: jaeger-operator.v1.44.0
        version: 1.44.0
      - name: jaeger-operator.v1.43.0
        version: 1.43.0
      - name: jaeger-operator.v1.42.0
        version: 1.42.0
      - name: jaeger-operator.v1.41.1
        version: 1.41.1
      - name: jaeger-operator.v1.41.0
        version: 1.41.0
      - name: jaeger-operator.v1.40.0
        version: 1.40.0
      - name: jaeger-operator.v1.39.0
        version: 1.39.0
      - name: jaeger-operator.v1.38.1
        version: 1.38.1
      - name: jaeger-operator.v1.38.0
        version: 1.38.0
      - name: jaeger-operator.v1.37.0
        version: 1.37.0
      - name: jaeger-operator.v1.36.0
        version: 1.36.0
      - name: jaeger-operator.v1.35.0
        version: 1.35.0
      - name: jaeger-operator.v1.34.1
        version: 1.34.1
      - name: jaeger-operator.v1.34.0
        version: 1.34.0
      - name: jaeger-operator.v1.33.0
        version: 1.33.0
      - name: jaeger-operator.v1.32.0
        version: 1.32.0
      - name: jaeger-operator.v1.31.0
        version: 1.31.0
      - name: jaeger-operator.v1.30.0
        version: 1.30.0
      - name: jaeger-operator.v1.29.1
        version: 1.29.1
      - name: jaeger-operator.v1.25.0
        version: 1.25.0
      - name: jaeger-operator.v1.24.0
        version: 1.24.0
      name: stable
    defaultChannel: stable
    packageName: jaeger
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/trustification/trusted-profile-analyzer-operator
    name: rhtpa-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable
    - currentCSV: rhtpa-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhtpa.io/v1",
                "kind": "TrustedProfileAnalyzer",
                "metadata": {
                  "name": "trustedprofileanalyzer-sample"
                },
                "spec": {
                  "appDomain": "change-me",
                  "collector": {},
                  "database": {},
                  "image": {
                    "fullName": "registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e",
                    "pullPolicy": "IfNotPresent"
                  },
                  "infrastructure": {
                    "port": 9010
                  },
                  "ingress": {},
                  "metrics": {
                    "enabled": false
                  },
                  "modules": {
                    "createDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "createImporters": {
                      "enabled": false,
                      "image": {},
                      "importers": {}
                    },
                    "importer": {
                      "concurrency": 1,
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {},
                      "workingDirectory": {
                        "size": "32Gi"
                      }
                    },
                    "migrateDatabase": {
                      "enabled": false,
                      "image": {}
                    },
                    "server": {
                      "enabled": true,
                      "image": {},
                      "infrastructure": {},
                      "ingress": {},
                      "metrics": {},
                      "replicas": 1,
                      "resources": {
                        "requests": {
                          "cpu": 1,
                          "memory": "8Gi"
                        }
                      },
                      "rust": {},
                      "tracing": {}
                    }
                  },
                  "oidc": {
                    "clients": {
                      "cli": {
                        "clientSecret": null
                      },
                      "frontend": {}
                    }
                  },
                  "openshift": {
                    "useServiceCa": true
                  },
                  "partOf": "trustify",
                  "replicas": 1,
                  "rust": {},
                  "storage": {},
                  "tls": {},
                  "tracing": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
          createdAt: "2025-09-19T15:04:16Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Trusted Profile Analyzer"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trustification/trusted-profile-analyzer-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: TrustedProfileAnalyzer
            name: trustedprofileanalyzers.rhtpa.io
            version: v1
        description: |
          [Tech Preview]
          Red Hat Trusted Profile Analyzer (RHTPA) provides a single source of truth for your software inventory across all your environments. You can verify the integrity of your in-house and third-party applications against their respective Software Bill of Materials (SBOM) document.  Trusted Profile Analyzer provides vulnerability, component and license transparency, strengthens your software supply chain security, and enables an accurate, up-to-date understanding of your software inventory.

          Features and Benefits:

          Enhanced Software Supply Chain Security: Offers critical transparency and integrity verification across the entire software lifecycle.

          Reduced Operational Risk: Proactively identifies and helps remediate vulnerabilities minimizing risk and compliance violations.

          Improved Compliance and Auditability: Simplifies the process of meeting regulatory requirements for software transparency and integrity by providing verifiable documentation and continuous monitoring.

          Faster Remediation and Incident Response: Provides targeted, actionable insights to prioritize and accelerate the patching and remediation of vulnerabilities.

          Increased Developer Productivity & Trust: Integrates security into existing workflows without significant overhead, allowing developers to build and deploy with greater confidence in the integrity of their code and components.

          Centralized Visibility and Control: Consolidates disparate security information into a single pane of glass, offering a holistic view of your application security posture.

          Red Hat Trusted Profile Analyzer is a crucial tool for any organization running applications on OpenShift that needs to confidently understand, verify, and secure the software they deploy, especially in an era of complex software supply chains and increasing regulatory scrutiny.

          [Docs](https://docs.redhat.com/en/documentation/red_hat_trusted_profile_analyzer/2.1).
        displayName: Red Hat Trusted Profile Analyzer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rhtpa
        - trusted-profile-analyzer
        links:
        - name: Trusted Profile Analyzer Operator
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        maintainers:
        - email: rhtpa-support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.29.0
        provider:
          name: Red Hat
          url: https://github.com/trustification/trusted-profile-analyzer-operator
        relatedImages:
        - registry.redhat.io/rhtpa/rhtpa-rhel9-operator@sha256:618810ffac4f2d4962693dc3537318e973fd507a98deb1260a7dd8f0153aff38
        - registry.redhat.io/rhtpa/rhtpa-trustification-service-rhel9@sha256:d5cf4a5bff94b59197f668a63d29591e3bc92ee89402edc70039e592d75cb84e
        - registry.redhat.io/rhtpa/rhtpa-operator-bundle@sha256:dc12800b318b9d358809e389e75d4d12a9f02015bfa0961ab17654ba8b0f5ff3
        version: 1.0.2
      entries:
      - name: rhtpa-operator.v1.0.2
        version: 1.0.2
      - name: rhtpa-operator.v1.0.1
        version: 1.0.1
      - name: rhtpa-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable
    packageName: rhtpa-operator
    provider:
      name: Red Hat
      url: https://github.com/trustification/trusted-profile-analyzer-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq7-interconnect-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.10.x
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: 1.2.0
    - currentCSV: amq7-interconnect-operator.v1.10.20
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"interconnectedcloud.github.io/v1alpha1","kind":"Interconnect","metadata":{"name":"amq-interconnect"},"spec":{"deploymentPlan":{"size":2,"role":"interior","placement":"Any"}}}]'
          capabilities: Basic Install
          categories: Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
          createdAt: "2019-06-28T22:00:00Z"
          description: Layer 7 Networking
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: https://github.com/interconnectedcloud/qdr-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of AMQ Interconnect
            displayName: Red Hat Integration - AMQ Interconnect
            kind: Interconnect
            name: interconnects.interconnectedcloud.github.io
            version: v1alpha1
        description: |2

          AMQ Interconnect is a lightweight [AMQP 1.0](https://www.amqp.org/) message router for building large, highly resilient messaging networks for hybrid cloud and IoT/edge deployments. AMQ Interconnect transparently learns the addresses of messaging endpoints (such as clients, servers, and message brokers) and flexibly routes messages between them.

          ### Core Capabilities

          * High throughput, low latency, shortest path message forwarding based on Layer 7 address routing mechanisms

          * `Interior` mode deployments for any arbitrary topology of geographically-distributed and interconnected routers

          * `Edge` mode deployments for extremely large scale device endpoint connectivity

          * Automatic message traffic rerouting when the network topology changes (resiliency without restrictions)

          * Flexible addressing schemes and delivery semantics (anycast, multicast, closest, balanced)

          * Integrated management with full support for the draft AMQP management specification

          * Full-featured security capabilities for authentication, authorization, and policy-based resource access control

          ### Operator Features

          * **Flexible deployment plans** - Configurable deployment plans are available for `interior` and `edge` mode scenarios. These plans include all dependent resources

          * **Placement directives** - Directives are provided to control how the pods should be scheduled

          * **Connectivity configuration defaults** - Configuration defaults are automatically generated for listeners, connectors, and SSL/TLS setup

          * **Exposes the service** - Integrated management of OpenShift Routes for exposed listener services for client, inter-router, and edge communications

          * **Security certificate management** - Certificates are created and managed through integration with AMQ Certificate Manager

          ### Troubleshooting

          After deploying Interconnect, check any of the following to verify that it is operating correctly:

          * The Interconnect instance

          * The Deployment (or DaemonSet) instance

          * An individual pod for the Deployment (or DaemonSet)

          * A Route created for exposed services

          In addition, use `qdstat` commands to verify connectivity. This can be done using `oc exec` against an interconnect pod, for example `oc exec -it <podname> -- qdstat -c`
          to list connections.

          ### Documentation

          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/products/red_hat_amq#interconnect).
        displayName: Red Hat Integration - AMQ Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq#interconnect
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9
        - registry.redhat.io/amq7/amq-interconnect-operator-metadata@sha256:d7b2c9cc8181eb372c787bb29323dc2d4cd1bb30ddb9e4bb15081a7a61e516a0
        - registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43
        version: 1.10.20
      entries:
      - name: amq7-interconnect-operator.v1.10.20
        version: 1.10.20
      - name: amq7-interconnect-operator.v1.10.19
        version: 1.10.19
      - name: amq7-interconnect-operator.v1.10.18
        version: 1.10.18
      - name: amq7-interconnect-operator.v1.10.17
        version: 1.10.17
      - name: amq7-interconnect-operator.v1.10.16
        version: 1.10.16
      - name: amq7-interconnect-operator.v1.10.15
        version: 1.10.15
      - name: amq7-interconnect-operator.v1.10.14
        version: 1.10.14
      - name: amq7-interconnect-operator.v1.10.13
        version: 1.10.13
      - name: amq7-interconnect-operator.v1.10.12
        version: 1.10.12
      - name: amq7-interconnect-operator.v1.10.11
        version: 1.10.11
      - name: amq7-interconnect-operator.v1.10.10
        version: 1.10.10
      - name: amq7-interconnect-operator.v1.10.9
        version: 1.10.9
      - name: amq7-interconnect-operator.v1.10.8
        version: 1.10.8
      - name: amq7-interconnect-operator.v1.10.7
        version: 1.10.7
      - name: amq7-interconnect-operator.v1.10.6
        version: 1.10.6
      - name: amq7-interconnect-operator.v1.10.5
        version: 1.10.5
      - name: amq7-interconnect-operator.v1.10.4
        version: 1.10.4
      - name: amq7-interconnect-operator.v1.10.3
        version: 1.10.3
      - name: amq7-interconnect-operator.v1.10.2
        version: 1.10.2
      name: latest
    defaultChannel: 1.10.x
    packageName: amq7-interconnect-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Nokia
      provider-url: https://docs.eda.dev
    name: eda-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eda-openshift-operator.v4.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:a5645cb55c3e6c22655c6c8606b9bd51a49db44aecc3fe52e0eb80cca83d17c4
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:c43a7fa7475db83011cc08b728982511fc0279a259e1aacc8ddde9e15f947725
        version: 4.0.0
      entries:
      - name: eda-openshift-operator.v4.0.0
        version: 4.0.0
      name: alpha
    - currentCSV: eda-openshift-operator.v4.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.eda.nokia.com/v1",
                "kind": "ConnectControllerConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "operator"
                  },
                  "name": "connectcontrollerconfig-sample",
                  "namespace": "eda-connect-k8s"
                },
                "spec": {
                  "connectPluginName": "",
                  "controllerImagePullSecret": "",
                  "edaNamespace": "eda",
                  "heartBeat": "",
                  "logLevel": "",
                  "skipTLSVerify": false,
                  "tlsCertificateData": "",
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
          createdAt: "2025-06-10T15:47:27Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: eda-connect-k8s
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["connectnetworkattachmentstatuses.connect.eda.nokia.com",
            connectnetworkdefinitions.connect.eda.nokia.com]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConnectNetworkAttachmentStatus is the Schema for the connectnetworkattachmentstatuses
              API
            displayName: Connect NetworkAttachment Status
            kind: ConnectNetworkAttachmentStatus
            name: connectnetworkattachmentstatuses.connect.eda.nokia.com
            version: v1
          - description: ConnectNetworkDefinition is the Schema for the connectnetworkdefinitions
              API
            displayName: Connect Network Definition
            kind: ConnectNetworkDefinition
            name: connectnetworkdefinitions.connect.eda.nokia.com
            version: v1
          - description: ConnectControllerConfig is the Schema for the connectcontrollerconfigs
              API
            displayName: Connect Controller Config
            kind: ConnectControllerConfig
            name: connectcontrollerconfigs.config.eda.nokia.com
            version: v1
        description: "K8s operator to deploy eda connect controller to configure EDA
          Connect.\eb"
        displayName: eda-openshift-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - EDA Connect
        - K8s Controller
        - EDA K8s Controller Operator
        links:
        - name: EDA Connect K8s Controller Operator
          url: https://docs.eda.dev/connect/openshift-plugin/
        maintainers:
        - email: support@nokia.com
          name: nokia
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Nokia
          url: https://docs.eda.dev
        relatedImages:
        - ghcr.io/nokia-eda/eda-connect-k8s-operator@sha256:ec264b52d581b8fd76e737b9811ab894957a4edf2f627f90bd9fe707890e6bf5
        - registry.connect.redhat.com/nokia-ni/eda-openshift-operator@sha256:19e133c41b8f947aa83afba0417d8db3570cc2e1368813996fa5d135fdca4e04
        version: 4.0.1
      entries:
      - name: eda-openshift-operator.v4.0.1
        version: 4.0.1
      name: stable
    defaultChannel: alpha
    packageName: eda-k8s-operator
    provider:
      name: Nokia
      url: https://docs.eda.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-application-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-application-gateway-operator.v25.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMApplicationGateway",
                "metadata": {
                  "name": "iag-instance"
                },
                "spec": {
                  "configuration": [
                    {
                      "dataKey": "config",
                      "name": "test-config",
                      "type": "configmap"
                    },
                    {
                      "discoveryEndpoint": "https://isam.mmfa.ibm.com/mga/sps/oauth/oauth20/metadata/test",
                      "postData": [
                        {
                          "name": "redirect_uris",
                          "values": [
                            "https://isam.mmfa.ibm.com/pkmsoidc"
                          ]
                        },
                        {
                          "name": "client_name",
                          "value": "OperatorTest"
                        },
                        {
                          "name": "enforce_pkce",
                          "value": "false"
                        },
                        {
                          "name": "all_users_entitled",
                          "value": "true"
                        },
                        {
                          "name": "consent_action",
                          "value": "never_prompt"
                        }
                      ],
                      "secret": "oidc-client",
                      "type": "oidc_registration"
                    },
                    {
                      "headers": [
                        {
                          "name": "Authorization",
                          "secretKey": "value",
                          "type": "secret",
                          "value": "githubsecret"
                        }
                      ],
                      "type": "web",
                      "url": "https://raw.github.com/ibm-security/iag-config/master/test/sample1.yaml"
                    },
                    {
                      "type": "literal",
                      "value": "version: \"22.07\"\n\nserver:\n  local_applications:\n    cred_viewer:\n      path_segment: creds\n      enable_html: true\n\nidentity:\n  oidc:\n    discovery_endpoint: \"https://iag-dev.ibmcloudsecurity.com/oidc/endpoint/default/.well-known/openid-configuration\"\n    client_id: 1cbfe647-9e5f-4d99-8e05-8ed1c862eb47\n    client_secret: uPP8rM7N0e\n"
                    }
                  ],
                  "deployment": {
                    "image": "icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9",
                    "imagePullPolicy": "Always",
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "lang": "C",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 8,
                      "periodSeconds": 9,
                      "successThreshold": 7,
                      "timeoutSeconds": 1
                    },
                    "readinessProbe": {
                      "failureThreshold": 2,
                      "initialDelaySeconds": 7,
                      "periodSeconds": 8,
                      "successThreshold": 4,
                      "timeoutSeconds": 5
                    },
                    "serviceAccountName": "iag"
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
          createdAt: "2025-02-05T21:15:37Z"
          description: The IBM Application Gateway operator manages the lifecycle
            of IBM Application Gateway containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/ibm-application-gateway-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMApplicationGateway is the Schema for the ibmapplicationgateways
              API
            displayName: IBMApplication Gateway
            kind: IBMApplicationGateway
            name: ibmapplicationgateways.ibm.com
            version: v1
        description: "The [IBM Application Gateway (IAG)](https://ibm.biz/ibm-app-gateway)
          image provides a containerized secure Web Reverse proxy which is designed
          to sit in front of your application, seamlessly adding authentication and
          authorization protection to your application.\n\nAn IAG instance deployed
          on Kubernetes can be a complex deployment.  In particular the configuration
          can be defined externally in one or more locations, and changes to this
          configuration may require all instances to be reloaded for the changes to
          take effect. \n\nThe internal Kubernetes deployment controller does not
          have any knowledge of how an IBM Application Gateway instance should behave
          when the configuration changes. As such this IBM Application Gateway specific
          Kubernetes operator is available to be deployed. Once deployed the operator
          can manage the lifecycle of the IBM Application Gateway instances.\n\nSee
          the project [Readme](https://github.com/IBM-Security/ibm-application-gateway-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Application Gateway Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Application Gateway product documentation
          url: https://ibm.biz/ibm-app-gateway
        maintainers:
        - email: isamdev@au1.ibm.com
          name: IBM Application Gateway Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - icr.io/ibmappgateway/ibm-application-gateway@sha256:1144b79a2930fe942ce627b0abcc3bed2f0b0770ae8313cb77ffe6772511c1f9
        - registry.connect.redhat.com/ibm/ibm-application-gateway-operator-bundle@sha256:e4ece7113fa7056980507408ac29d60676cf51c20d71a058a1accf765a6ea7e3
        - icr.io/ibmappgateway/ibm-application-gateway-operator@sha256:6a16d4ef802a866942f504b11a948565245d303f322944200c1a1269fe7c371c
        version: 25.2.0
      entries:
      - name: ibm-application-gateway-operator.v25.2.0
        version: 25.2.0
      - name: ibm-application-gateway-operator.v24.4.0
        version: 24.4.0
      name: stable
    defaultChannel: stable
    packageName: ibm-application-gateway-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchservice-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opensearchservice.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
          createdAt: "2025-09-19T18:05:50Z"
          description: AWS OpenSearch Service controller is a service controller for
            managing OpenSearch Service resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS opensearchservice Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.opensearchservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage OpenSearch Service resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearch Service**

          Amazon OpenSearch Service (successor to Amazon Elasticsearch Service) is a managed service that makes it easy to deploy, operate, and scale OpenSearch clusters in the AWS Cloud. Amazon OpenSearch Service supports OpenSearch and legacy Elasticsearch OSS. When you create a cluster, you have the option of which search engine to use. OpenSearch Service offers broad compatibility with Elasticsearch OSS 7.10, the final open source version of the software. For information about what changed with the recent service renaming, and the actions you might need to take, see [Amazon OpenSearch Service - Summary of changes](https://docs.aws.amazon.com/opensearch-service/latest/developerguide/rename.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearch Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearch Service Developer Resources
          url: https://aws.amazon.com/opensearch-service/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearch service maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchservice-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchservice-controller@sha256:699f483cf96822576e3b79b690f256e2dbf4ca8d4426daf4298c980aefdeba4f
        version: 1.1.1
      entries:
      - name: ack-opensearchservice-controller.v1.1.1
        version: 1.1.1
      - name: ack-opensearchservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-opensearchservice-controller.v1.0.15
        version: 1.0.15
      - name: ack-opensearchservice-controller.v1.0.12
        version: 1.0.12
      - name: ack-opensearchservice-controller.v1.0.10
        version: 1.0.10
      - name: ack-opensearchservice-controller.v1.0.9
        version: 1.0.9
      - name: ack-opensearchservice-controller.v1.0.8
        version: 1.0.8
      - name: ack-opensearchservice-controller.v1.0.7
        version: 1.0.7
      - name: ack-opensearchservice-controller.v1.0.6
        version: 1.0.6
      - name: ack-opensearchservice-controller.v1.0.5
        version: 1.0.5
      - name: ack-opensearchservice-controller.v1.0.4
        version: 1.0.4
      - name: ack-opensearchservice-controller.v1.0.2
        version: 1.0.2
      - name: ack-opensearchservice-controller.v1.0.1
        version: 1.0.1
      - name: ack-opensearchservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-opensearchservice-controller.v0.0.28
        version: 0.0.28
      - name: ack-opensearchservice-controller.v0.0.27
        version: 0.0.27
      - name: ack-opensearchservice-controller.v0.0.26
        version: 0.0.26
      - name: ack-opensearchservice-controller.v0.0.24
        version: 0.0.24
      - name: ack-opensearchservice-controller.v0.0.23
        version: 0.0.23
      - name: ack-opensearchservice-controller.v0.0.19
        version: 0.0.19
      - name: ack-opensearchservice-controller.v0.0.17
        version: 0.0.17
      - name: ack-opensearchservice-controller.v0.0.15
        version: 0.0.15
      - name: ack-opensearchservice-controller.v0.0.14
        version: 0.0.14
      - name: ack-opensearchservice-controller.v0.0.13
        version: 0.0.13
      - name: ack-opensearchservice-controller.v0.0.12
        version: 0.0.12
      - name: ack-opensearchservice-controller.v0.0.11
        version: 0.0.11
      - name: ack-opensearchservice-controller.v0.0.10
        version: 0.0.10
      - name: ack-opensearchservice-controller.v0.0.9
        version: 0.0.9
      - name: ack-opensearchservice-controller.v0.0.8
        version: 0.0.8
      - name: ack-opensearchservice-controller.v0.0.7
        version: 0.0.7
      - name: ack-opensearchservice-controller.v0.0.6
        version: 0.0.6
      - name: ack-opensearchservice-controller.v0.0.5
        version: 0.0.5
      - name: ack-opensearchservice-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-xtradb-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "docker.io/percona/percona-xtradb-cluster:8.0.42-33.1",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "docker.io/percona/haproxy:2.8.15",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "docker.io/percona/proxysql2:2.7.3",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "docker.io/percona/fluentbit:4.0.1",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.3.1",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "docker.io/percona/percona-xtrabackup:8.0.35-34.1",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-xtradb-cluster-operator:1.18.0
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <v1.18.0
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            clustercheck: Y2x1c3RlcmNoZWNrcGFzc3dvcmQ=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ```
          ###  Release Highlights
          * **PMM3 support**

          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**

          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**

          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**

          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.

        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - database
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/percona-xtradb-cluster-operator@sha256:d0c9d03130774d961f62032f684c494313787354186d028e9709e2463f69431c
        - docker.io/percona/percona-xtradb-cluster-operator:1.18.0
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator.v1.17.0
        version: 1.17.0
      - name: percona-xtradb-cluster-operator.v1.9.0
        version: 1.9.0
      name: stable
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: ""
    name: gitops-primer
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitops-primer.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "primer.gitops.io/v1alpha1",
                "kind": "Export",
                "metadata": {
                  "name": "primer"
                },
                "spec": {
                  "method": "download"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/migtools/gitops-primer:v0.0.11
          operatorframework.io/suggested-namespace: gitops-primer-system
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Export is the Schema for the exports API
            displayName: Export
            kind: Export
            name: exports.primer.gitops.io
            version: v1alpha1
        description: Export and normalize Kubernetes YAML files for GitOps
        displayName: gitops-primer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - operator
        links:
        - name: Gitops Primer
          url: https://github.com/migtools/gitops-primer
        maturity: alpha
        provider:
          name: Konveyor
        relatedImages:
        - quay.io/migtools/gitops-primer:v0.0.11
        - quay.io/openshift-community-operators/gitops-primer@sha256:97af48469275f1b3a669f8e24a97a80ce9897fc5b12643a4f298b82dbe26e20b
        - quay.io/openshift/origin-oauth-proxy:4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6d57bfd91fac9b68eb72d27226bc297472ceb136c996628b845ecc54a48b31cb
        - quay.io/migtools/gitops-primer-downloader:v0.0.11
        - quay.io/migtools/gitops-primer-export:v0.0.11
        - quay.io/migtools/gitops-primer-webhook:v0.0.11
        version: 0.0.11
      entries:
      - name: gitops-primer.v0.0.11
        version: 0.0.11
      - name: gitops-primer.v0.0.10
        version: 0.0.10
      - name: gitops-primer.v0.0.9
        version: 0.0.9
      - name: gitops-primer.v0.0.8
        version: 0.0.8
      - name: gitops-primer.v0.0.7
        version: 0.0.7
      - name: gitops-primer.v0.0.6
        version: 0.0.6
      - name: gitops-primer.v0.0.5
        version: 0.0.5
      - name: gitops-primer.v0.0.4
        version: 0.0.4
      - name: gitops-primer.v0.0.3
        version: 0.0.3
      - name: gitops-primer.v0.0.2
        version: 0.0.2
      - name: gitops-primer.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: gitops-primer
    provider:
      name: Konveyor
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: watcher-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: watcher-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "Watcher",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcher",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcher"
                },
                "spec": {
                  "apiServiceTemplate": {
                    "tls": {
                      "caBundleSecretName": "combined-ca-bundle"
                    }
                  },
                  "databaseInstance": "openstack"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherAPI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcher-operator"
                  },
                  "name": "watcherapi-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherApplier",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherapplier-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherapplier",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherapplier-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              },
              {
                "apiVersion": "watcher.openstack.org/v1beta1",
                "kind": "WatcherDecisionEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "watcher-operator",
                    "app.kubernetes.io/instance": "watcherdecisionengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "watcherdecisionengine",
                    "app.kubernetes.io/part-of": "watcher-operator"
                  },
                  "name": "watcherdecisionengine-sample"
                },
                "spec": {
                  "memcachedInstance": "memcached",
                  "secret": "watcher",
                  "serviceAccount": "watcher-watcher"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-09-02T13:32:31Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WatcherAPI is the Schema for the watcherapis API
            displayName: Watcher API
            kind: WatcherAPI
            name: watcherapis.watcher.openstack.org
            version: v1beta1
          - description: WatcherApplier is the Schema for the watcherappliers API
            displayName: Watcher Applier
            kind: WatcherApplier
            name: watcherappliers.watcher.openstack.org
            version: v1beta1
          - description: WatcherDecisionEngine is the Schema for the watcherdecisionengines
              API
            displayName: Watcher Decision Engine
            kind: WatcherDecisionEngine
            name: watcherdecisionengines.watcher.openstack.org
            version: v1beta1
          - description: Watcher is the Schema for the watchers API
            displayName: Watcher
            kind: Watcher
            name: watchers.watcher.openstack.org
            version: v1beta1
        description: The Watcher Operator project
        displayName: Watcher Operator (Technology Preview)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Watcher Operator
          url: https://github.com/openstack-k8s-operators/watcher-operator
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-watcher-decision-engine-rhel9@sha256:c759d66cf0986c0036be91c008cbd059b4412c94160ed82dd12cd4e00d0685a9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:eed8234b0fe47083ab941f513709f778bfcb80e32747b71a65fd5f62328a5ca0
        - registry.redhat.io/rhoso-operators/watcher-operator-bundle@sha256:01ab69b9c33a91cb194ad644bb64884b358af78467e2bb766ef373e4c36e795d
        - registry.redhat.io/rhoso-operators/watcher-rhel9-operator@sha256:63c341a3e20a522c601098502a8fd8f1164d23c773ae8f55782c43ff4b6c8c18
        - registry.redhat.io/rhoso/openstack-watcher-api-rhel9@sha256:e97369e3a47bb882f1920c6af3f6345a842af45c58f0f5c5df9ac2933cf3098a
        - registry.redhat.io/rhoso/openstack-watcher-applier-rhel9@sha256:c92a82f901613991926367bb8aab4f875c27a3e4a5282d92847a8463d98cd9e3
        version: 1.0.14
      entries:
      - name: watcher-operator.v1.0.14
        version: 1.0.14
      - name: watcher-operator.v1.0.13
        version: 1.0.13
      - name: watcher-operator.v1.0.12
        version: 1.0.12
      - name: watcher-operator.v1.0.11
        version: 1.0.11
      - name: watcher-operator.v1.0.10
        version: 1.0.10
      - name: watcher-operator.v1.0.9
        version: 1.0.9
      - name: watcher-operator.v1.0.8
        version: 1.0.8
      - name: watcher-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: watcher-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-bridge-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-bridge-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
          createdAt: 2021-11-02 11:11 UTC
          description: Enhance OCP using Red Hat Quay container registry
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n```\n\n_Note: Until the operator is running, new requests for builds
          will fail since the webserver the MutatingWebhookConfiguration invokes is
          not available and a proper is response is required in order for the object
          to be persisted in etcd._\n\n#### Create the QuayIntegration Custom Resource\n\nFinally,
          to complete the integration between OpenShift and Quay, a `QuayIntegration`
          custom resource needs to be created.\n\nThe following is an example of a
          basic definition of a `QuayIntegration` resource associated from the associated
          CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind: QuayIntegration\nmetadata:\n
          \ name: example-quayintegration\nspec:\n  clusterID: openshift\n  credentialsSecretName:
          <NAMESPACE>/<SECRET>\n  quayHostname: https://<QUAY_URL>\n```\n\nThe _clusterID_
          is a value which should be unique across the entire ecosystem. This value
          is optional and defaults to `openshift`.\n\nThe _credentialsSecretName_
          is a NamespacedName value of the secret containing the token that was previously
          created.\n\nNote: If Quay is using self signed certificates, the property
          `insecureRegistry: true`\n\nOrganizations within Quay should be created
          for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4478b8a78bdde23e0d40476f8ad9e4bf0628fba7603a390e8de7e6a319abc1ec
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:501fe9b003b5066b3936def0ef4de05abd1a12e795e718d6c9a62d2c6c804c2e
        version: 3.4.7
      entries:
      - name: quay-bridge-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-bridge-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "my-quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecretName": "<NAMESPACE>/<SECRET>",
                  "quayHostname": "https://<QUAY_URL>",
                  "insecureRegistry": false,
                  "whitelistNamespaces": ["default"]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
          createdAt: 2021-09-15 23:09 UTC
          description: Enhance OCP using Red Hat Quay container registry
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/redhat-cop/quay-openshift-registry-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a connected Quay instance.
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.redhatcop.redhat.io
            version: v1alpha1
        description: "## Overview\n\nThe OpenShift Container Platform contains an
          in cluster image registry for pushing and pulling images that either produced
          within the cluster or are references from external sources. Quay is a highly
          available, enterprise grade image registry with enhanced role based access
          control (RBAC) features. The goal of this operator is to support replicating
          the features found within OpenShift intgrated image registry with Quay.\n\n##
          Functionality\n\nThe following capabilities are enabled through this operator:\n\n*
          Synchronization of OpenShift namespaces as Quay organizations\n    * Creation
          of Robot accounts for each default namespace service account\n    * Creation
          of Secrets for each created Robot Account\n        * Associate Robot Secret
          to Service Account as Mountable and Image Pull Secret\n    * Synchronization
          of OpenShift ImageStreams as Quay Repositories\n* Automatically rewrite
          new Builds making use of ImageStreams to output to Quay\n* Automatically
          import ImageStream tag once build completes\n\n## Prerequisites\n\nThe following
          requirements must be satisfied prior to setting and configuring the integration
          between OpenShift and Quay:\n\n* A Red Hat Quay environment with a user
          with superuser permissions\n* A Red Hat OpenShift Container Platform environment
          (at least version 4.0) with cluster-admin permissions on an OpenShift cluster
          \n*\tOpenShift Command Line tool (CLI)\n\n## Setup and Configuration\n\nThe
          setup process involves steps within both Quay and OpenShift as described
          below:\n\n### Quay Setup\n\nThe integration between OpenShift and Quay makes
          extensive use of the Quay RESTFul API. To allow for other methods aside
          from user credentials to be used for authenticating against the API, an
          application is required to be created. When an application is created, an
          OAuth token is generated which can be used by the operator in OpenShift.
          Applications are created within a Quay organization. While an existing organization
          can be utilized, it is recommended that a dedicated organization be used.\n\n1.\tLogin
          to Quay as a user with superuser access and select the organization for
          which the external application will be configured within.\n2.\tOn the left
          hand side of the page, select **Applications**\n3.\tCreate a new application
          by selecting the **Create New Application** button and entering a name for
          the application\n4.\tAfter the new application has been created, select
          the application that was created\n5.\tOn the lefthand navigation bar, select
          the **Generate Token** button to create a new OAuth2 token\n6.\tSelect all
          of the checkboxes to grant the necessary access that is needed for the integration\n7.\tReview
          the permissions that will be assigned and then select the **Authorize Application**
          button\n8.\tTake note of the generated Access Token as it will be needed
          in the following section\n\n### OpenShift Setup\n\nTo complete the process
          to begin to integrate OpenShift and Quay, several steps are required. Before
          beginning, ensure that you have the OpenShift Command Line tool installed
          and that you are logged into OpenShift as a cluster administrator. \n\n####
          Create A Secret for the Quay OAuth Token\n\nThe Operator will use the previously
          obtained Access Token to communicate with Quay. Store this token within
          OpenShift as a secret.\n\nExecute the following command to create a secret
          called `quay-integration` with a key called `token` containing the access
          token:\n\n```\n$ oc create secret generic quay-integration --from-literal=token=<access_token>\n```\n\nThis
          token will be referenced in the final section.\n\n#### MutatingWebhookConfiguration
          Support\n\nSupport for dynamic interception of API requests that are performed
          as part of OpenShift’s typical build process is facilitated through a [MutatingWebhookConfiguration](https://docs.openshift.com/container-platform/3.11/architecture/additional_concepts/dynamic_admission_controllers.html).
          A MutatingWebhookConfiguration allows for invoking an API running within
          a project on OpenShift when certain API requests are received. In particular,
          to support Quay’s integration with OpenShift, any new Build requests should
          be intercepted so that the output can be modified to target Quay instead
          of OpenShift’s integrated registry. \n\nKubernetes requires that the webhook
          endpoint is secured via SSL using a certificate that makes use of the certificate
          authority for the cluster. Fortunately, OpenShift provides support for generating
          a certificate signed by the cluster. A script is available in the documentation
          called `webhook-create-signed-cert.sh` that automates the process for requesting
          a new certificate.\n\nFirst, create a `Service` for the Operator's webhook
          endpoint:\n```yaml\napiVersion: v1\nkind: Service\nmetadata:\n  labels:\n
          \   name: quay-bridge-operator\n  name: quay-bridge-operator\n  namespace:
          openshift-operators\nspec:\n  ports:\n    - name: https\n      port: 443\n
          \     protocol: TCP\n      targetPort: 8443\n  selector:\n    name: quay-bridge-operator\n
          \ sessionAffinity: None\n  type: ClusterIP\n```\n\nExecute the following
          command to request the certificate:\n\n```\n./webhook-create-signed-cert.sh
          --namespace openshift-operators --secret quay-bridge-operator-webhook-certs
          --service quay-bridge-operator\n```\n\nThe result will place the newly created
          private key and certificate within a secret within the secret specified.
          The secret will be mounted into the appropriate located within the operator
          as declared in the Deployment of the Operator. \n\nOnce the secret has been
          created, focus can shift to the `MutatingWebhookConfiguration` (see below).
          The majority of the content should be in place to apply to the cluster.
          The one section that requires configuration is the caBundle property. This
          refers to the Certificate Authority (CA) for the OpenShift environment.
          The OpenShift CA is available as a _ConfigMap_ within the `kube-system`
          namespace.\n\nExecute the following command to retrieve the CA and format
          the result as a single line so that it can be entered into the MutatingWebhookConfiguration
          resource:\n\n```\noc get configmap -n kube-system extension-apiserver-authentication
          -o=jsonpath='{.data.client-ca-file}' | base64 | tr -d '\\n'\n```\n\nReplace
          the `${CA_BUNDLE}` variable in the following `MutatingWebhookConfiguration`
          YAML and create it:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\nkind:
          MutatingWebhookConfiguration \nmetadata:\n  name: quay-bridge-operator\nwebhooks:\n
          \ - name: quayintegration.redhatcop.redhat.io\n    clientConfig: \n      service:\n
          \       namespace: openshift-operators\n        name: quay-bridge-operator\n
          \       path: \"/admissionwebhook\"\n      caBundle: \"${CA_BUNDLE}\"\n
          \   rules: \n    - operations:  [ \"CREATE\" ]\n      apiGroups: [ \"build.openshift.io\"
          ]\n      apiVersions: [\"v1\" ]\n      resources: [ \"builds\" ]\n    failurePolicy:
          Fail\n    sideEffects: None\n    admissionReviewVersions: [v1beta1]\n    matchPolicy:
          Exact\n    timeoutSeconds: 30  \n```\n\n_Note: Until the operator is running,
          new requests for builds will fail since the webserver the MutatingWebhookConfiguration
          invokes is not available and a proper is response is required in order for
          the object to be persisted in etcd._\n\n#### Create the QuayIntegration
          Custom Resource\n\nFinally, to complete the integration between OpenShift
          and Quay, a `QuayIntegration` custom resource needs to be created.\n\nThe
          following is an example of a basic definition of a `QuayIntegration` resource
          associated from the associated CRD.\n\n```\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          QuayIntegration\nmetadata:\n  name: example-quayintegration\nspec:\n  clusterID:
          openshift\n  credentialsSecretName: <NAMESPACE>/<SECRET>\n  quayHostname:
          https://<QUAY_URL>\n```\n\nThe _clusterID_ is a value which should be unique
          across the entire ecosystem. This value is optional and defaults to `openshift`.\n\nThe
          _credentialsSecretName_ is a NamespacedName value of the secret containing
          the token that was previously created.\n\nNote: If Quay is using self signed
          certificates, the property `insecureRegistry: true`\n\nOrganizations within
          Quay should be created for the related namespaces in OpenShift."
        displayName: Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Source Code
          url: https://github.com/redhat-cop/quay-openshift-registry-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:6c2ab763239f6e5a2a72890120bf1f20f6277850cec59a7e11cea2b6a06f12e8
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:9149df5ffc83dce6593072b7f72805e07e700b84527e45b4b7486b9c38f70731
        version: 3.5.7
      entries:
      - name: quay-bridge-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-bridge-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.10.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:8f30d4c7c491e9a804a8abf80570d1e1924cd0d787c6abb727331541c435d5ff
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:33dc7673155f01728225221768f6b90d87cc7609dcda46b394dba2421c136a23
        version: 3.10.15
      entries:
      - name: quay-bridge-operator.v3.10.15
        version: 3.10.15
      - name: quay-bridge-operator.v3.10.14
        version: 3.10.14
      - name: quay-bridge-operator.v3.10.13
        version: 3.10.13
      - name: quay-bridge-operator.v3.10.12
        version: 3.10.12
      - name: quay-bridge-operator.v3.10.11
        version: 3.10.11
      - name: quay-bridge-operator.v3.10.10
        version: 3.10.10
      - name: quay-bridge-operator.v3.10.9
        version: 3.10.9
      - name: quay-bridge-operator.v3.10.8
        version: 3.10.8
      - name: quay-bridge-operator.v3.10.7
        version: 3.10.7
      - name: quay-bridge-operator.v3.10.6
        version: 3.10.6
      - name: quay-bridge-operator.v3.10.5
        version: 3.10.5
      - name: quay-bridge-operator.v3.10.4
        version: 3.10.4
      - name: quay-bridge-operator.v3.10.3
        version: 3.10.3
      - name: quay-bridge-operator.v3.10.2
        version: 3.10.2
      - name: quay-bridge-operator.v3.10.1
        version: 3.10.1
      - name: quay-bridge-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-bridge-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.11.13
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:2b6b1a0b56cfb348ed4035b13d31501ff5757c101ed21027b330a7852576b891
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:511f8471db70e8d5fd9d0b547c73f1416bbac6f0fe6cdeb6c59a9ef8bd845ac2
        version: 3.11.13
      entries:
      - name: quay-bridge-operator.v3.11.13
        version: 3.11.13
      - name: quay-bridge-operator.v3.11.12
        version: 3.11.12
      - name: quay-bridge-operator.v3.11.11
        version: 3.11.11
      - name: quay-bridge-operator.v3.11.10
        version: 3.11.10
      - name: quay-bridge-operator.v3.11.9
        version: 3.11.9
      - name: quay-bridge-operator.v3.11.8
        version: 3.11.8
      - name: quay-bridge-operator.v3.11.7
        version: 3.11.7
      - name: quay-bridge-operator.v3.11.6
        version: 3.11.6
      - name: quay-bridge-operator.v3.11.5
        version: 3.11.5
      - name: quay-bridge-operator.v3.11.4
        version: 3.11.4
      - name: quay-bridge-operator.v3.11.3
        version: 3.11.3
      - name: quay-bridge-operator.v3.11.2
        version: 3.11.2
      - name: quay-bridge-operator.v3.11.1
        version: 3.11.1
      - name: quay-bridge-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-bridge-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.12.12
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:4c4c98b6cecd51c6644eb79b85789b7ebbcdc662edbb66ce435ea3b5e58d9273
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:c6e04de3ad3b8c0ded95146d090e120bffed7f9dc5e74332a183fe1c2e45d16d
        version: 3.12.12
      entries:
      - name: quay-bridge-operator.v3.12.12
        version: 3.12.12
      - name: quay-bridge-operator.v3.12.11
        version: 3.12.11
      - name: quay-bridge-operator.v3.12.10
        version: 3.12.10
      - name: quay-bridge-operator.v3.12.9
        version: 3.12.9
      - name: quay-bridge-operator.v3.12.8
        version: 3.12.8
      - name: quay-bridge-operator.v3.12.7
        version: 3.12.7
      - name: quay-bridge-operator.v3.12.6
        version: 3.12.6
      - name: quay-bridge-operator.v3.12.5
        version: 3.12.5
      - name: quay-bridge-operator.v3.12.4
        version: 3.12.4
      - name: quay-bridge-operator.v3.12.3
        version: 3.12.3
      - name: quay-bridge-operator.v3.12.2
        version: 3.12.2
      - name: quay-bridge-operator.v3.12.1
        version: 3.12.1
      - name: quay-bridge-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-bridge-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.13.8
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:b8aa74c5b2635a6d3413eeac6fc0f6eb9a2e459089aaabbf626ea3ef92a67ce0
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:e8ee1944d8e0e9d515a6e63a3e39901c304abd421fcad2abbf9a3c43c7ca6243
        version: 3.13.8
      entries:
      - name: quay-bridge-operator.v3.13.8
        version: 3.13.8
      - name: quay-bridge-operator.v3.13.7
        version: 3.13.7
      - name: quay-bridge-operator.v3.13.6
        version: 3.13.6
      - name: quay-bridge-operator.v3.13.5
        version: 3.13.5
      - name: quay-bridge-operator.v3.13.4
        version: 3.13.4
      - name: quay-bridge-operator.v3.13.3
        version: 3.13.3
      - name: quay-bridge-operator.v3.13.2
        version: 3.13.2
      - name: quay-bridge-operator.v3.13.1
        version: 3.13.1
      - name: quay-bridge-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-bridge-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.14.5
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:73d91324ccc995e9a6db525f194bee340c17dd624ce520c767ea8fcf2346f786
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b12f0dfb8fdeae3f521e41fcf60ef118bdfd08ad7f68bd87c5494b067b7c7c58
        version: 3.14.5
      entries:
      - name: quay-bridge-operator.v3.14.5
        version: 3.14.5
      - name: quay-bridge-operator.v3.14.4
        version: 3.14.4
      - name: quay-bridge-operator.v3.14.3
        version: 3.14.3
      - name: quay-bridge-operator.v3.14.2
        version: 3.14.2
      - name: quay-bridge-operator.v3.14.1
        version: 3.14.1
      - name: quay-bridge-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-bridge-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.15.2
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:b2ac4a48c80209d7bebe8b1b2d066571f99a9f47b78604e47a8300d027941b66
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:d29abe14abc67d8836cc6f7656556a0a7fa8f4a8eea3f78e40265b339e7430f5
        version: 3.15.2
      entries:
      - name: quay-bridge-operator.v3.15.2
        version: 3.15.2
      - name: quay-bridge-operator.v3.15.1
        version: 3.15.1
      - name: quay-bridge-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-bridge-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:0ad7af440c6e00fb578ef40e6d600964324fd94ce7835a28d0ec355b7415093f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:99dc6b5af811bb33892c17334dd4e1a7eddc6982d49a793b1789f505e00f49b1
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:f7c4ed28a549fe7414312bd22c2ae140f1f28d30656fb9fb0c8ade05a8f82088
        version: 3.6.10
      entries:
      - name: quay-bridge-operator.v3.6.10
        version: 3.6.10
      - name: quay-bridge-operator.v3.6.9
        version: 3.6.9
      - name: quay-bridge-operator.v3.6.8
        version: 3.6.8
      - name: quay-bridge-operator.v3.6.7
        version: 3.6.7
      - name: quay-bridge-operator.v3.6.6
        version: 3.6.6
      - name: quay-bridge-operator.v3.6.5
        version: 3.6.5
      - name: quay-bridge-operator.v3.6.4
        version: 3.6.4
      - name: quay-bridge-operator.v3.6.2
        version: 3.6.2
      - name: quay-bridge-operator.v3.6.1
        version: 3.6.1
      - name: quay-bridge-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-bridge-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.7.14
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c59956d70416fab0c6d7b3ecd8c9cfa3cf9415f75cfdfe2214f44b0b4aced594
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:6a0dd80dc48ad0aaa576b3de78b2b265e29a3c10e4996aa6b8f129592fe20f09
        version: 3.7.14
      entries:
      - name: quay-bridge-operator.v3.7.14
        version: 3.7.14
      - name: quay-bridge-operator.v3.7.13
        version: 3.7.13
      - name: quay-bridge-operator.v3.7.12
        version: 3.7.12
      - name: quay-bridge-operator.v3.7.11
        version: 3.7.11
      - name: quay-bridge-operator.v3.7.10
        version: 3.7.10
      - name: quay-bridge-operator.v3.7.9
        version: 3.7.9
      - name: quay-bridge-operator.v3.7.8
        version: 3.7.8
      - name: quay-bridge-operator.v3.7.7
        version: 3.7.7
      - name: quay-bridge-operator.v3.7.6
        version: 3.7.6
      - name: quay-bridge-operator.v3.7.5
        version: 3.7.5
      - name: quay-bridge-operator.v3.7.4
        version: 3.7.4
      - name: quay-bridge-operator.v3.7.3
        version: 3.7.3
      - name: quay-bridge-operator.v3.7.2
        version: 3.7.2
      - name: quay-bridge-operator.v3.7.1
        version: 3.7.1
      - name: quay-bridge-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: quay-bridge-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.8.15
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:f8544a546dd6d209355c3a1cbe35138f6910491fc4584b401634431da4414ec0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:975f287513039b2244894b9031e590f665d48f9af480dfdc4b85d4f96dd4bc37
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:7cace73b03bb02202a04b6026eb9f7323610fd0ccd5ba18a6d00fd66577a96b4
        version: 3.8.15
      entries:
      - name: quay-bridge-operator.v3.8.15
        version: 3.8.15
      - name: quay-bridge-operator.v3.8.14
        version: 3.8.14
      - name: quay-bridge-operator.v3.8.13
        version: 3.8.13
      - name: quay-bridge-operator.v3.8.12
        version: 3.8.12
      - name: quay-bridge-operator.v3.8.11
        version: 3.8.11
      - name: quay-bridge-operator.v3.8.10
        version: 3.8.10
      - name: quay-bridge-operator.v3.8.9
        version: 3.8.9
      - name: quay-bridge-operator.v3.8.8
        version: 3.8.8
      - name: quay-bridge-operator.v3.8.7
        version: 3.8.7
      - name: quay-bridge-operator.v3.8.6
        version: 3.8.6
      - name: quay-bridge-operator.v3.8.5
        version: 3.8.5
      - name: quay-bridge-operator.v3.8.4
        version: 3.8.4
      - name: quay-bridge-operator.v3.8.3
        version: 3.8.3
      - name: quay-bridge-operator.v3.8.2
        version: 3.8.2
      - name: quay-bridge-operator.v3.8.1
        version: 3.8.1
      - name: quay-bridge-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-bridge-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayIntegration",
                "metadata": {
                  "name": "quay"
                },
                "spec": {
                  "clusterID": "openshift",
                  "credentialsSecret": {
                    "name": "quay-credentials",
                    "namespace": "openshift-operators"
                  },
                  "insecureRegistry": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          containerImage: registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
          createdAt: 2021-07-09 14:07 UTC
          description: Enhance OCP using Red Hat Red Hat container registry
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          quay-version: 3.9.16
          repository: https://github.com/quay/quay-bridge-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QuayIntegration is the Schema for the quayintegrations API
            displayName: Quay Integration
            kind: QuayIntegration
            name: quayintegrations.quay.redhat.com
            version: v1
        description: Enhance OCP using Red Hat Quay container registry
        displayName: Red Hat Quay Bridge Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - registry
        - containers
        - quay
        links:
        - name: Quay Bridge Operator
          url: http://quay.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/quay/quay-bridge-operator-bundle@sha256:c173a99026499700c08092f8418c79f39527d9b914690b199b41ed781c279de1
        - registry.redhat.io/quay/quay-bridge-operator-rhel8@sha256:875e0b63fdb7d44879650eb2f788ea705d753520841a5d875428f0b3ca70a36c
        version: 3.9.16
      entries:
      - name: quay-bridge-operator.v3.9.16
        version: 3.9.16
      - name: quay-bridge-operator.v3.9.15
        version: 3.9.15
      - name: quay-bridge-operator.v3.9.14
        version: 3.9.14
      - name: quay-bridge-operator.v3.9.13
        version: 3.9.13
      - name: quay-bridge-operator.v3.9.12
        version: 3.9.12
      - name: quay-bridge-operator.v3.9.11
        version: 3.9.11
      - name: quay-bridge-operator.v3.9.10
        version: 3.9.10
      - name: quay-bridge-operator.v3.9.9
        version: 3.9.9
      - name: quay-bridge-operator.v3.9.8
        version: 3.9.8
      - name: quay-bridge-operator.v3.9.7
        version: 3.9.7
      - name: quay-bridge-operator.v3.9.6
        version: 3.9.6
      - name: quay-bridge-operator.v3.9.5
        version: 3.9.5
      - name: quay-bridge-operator.v3.9.4
        version: 3.9.4
      - name: quay-bridge-operator.v3.9.3
        version: 3.9.3
      - name: quay-bridge-operator.v3.9.2
        version: 3.9.2
      - name: quay-bridge-operator.v3.9.1
        version: 3.9.1
      - name: quay-bridge-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-bridge-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-logging.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Logging",
                "metadata": {
                  "name": "logging-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "clusterDomain": "cluster.local",
                  "elasticsearch": {
                    "storage": "16Gi",
                    "tag": "7.17.6-3747-35"
                  },
                  "environment": "openshift",
                  "fluentbit": {
                    "runOnMaster": false,
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "grafana": {
                    "dashboards": [
                      "wap",
                      "load-core"
                    ],
                    "dataSources": [
                      "wap",
                      "load-core"
                    ]
                  },
                  "image": {
                    "elasticInitImage": "rhel8/toolbox@sha256:a753c24cb8b03066d494f010ad654d23fb0559f14d7121542298da3e0d48340d",
                    "elasticInitRepository": "registry.access.redhat.com",
                    "fbImage": "fluent-bit:1.8.15-3747-20",
                    "fbRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "createClusterRoles": true,
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:33:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Logging
            name: loggings.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Logging Operator
        displayName: Keysight WAP Logging Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Logging
          url: https://keysight-wap-logging.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-logging@sha256:f1e3b3ad84cd0f96c92cdaf4609e29257035cb982a761e832dae2095e636c721
        - registry.connect.redhat.com/loadcore/keysight-wap-logging-bundle@sha256:2aeeefa7493eed8960004904567894244418982209e1b3a40488b759463edcdb
        version: 0.1.21
      entries:
      - name: keysight-wap-logging.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-logging
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stakater
      provider-url: github.com/stakater-ab/hibernation-operator
    name: hibernation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hibernation-operator.v0.1.102
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-09-03T09:23:19Z"
          description: Operator to engage and disengage hibernation on k8s resources
            using cron expressions
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <0.1.102
          operatorframework.io/suggested-namespace: hibernation-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterResourceSupervisor is the Schema for the resourcesupervisors
              API
            displayName: Cluster Resource Supervisor
            kind: ClusterResourceSupervisor
            name: clusterresourcesupervisors.hibernation.stakater.com
            version: v1beta1
        description: Schedule k8s resources using cron expressions via Hibernation
          Operator
        displayName: Hibernation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - redhat
        - k8s
        - stakater
        - hibernation
        - operator
        links:
        - name: Stakater
          url: https://stakater.com
        - name: Support
          url: https://docs.stakater.com/content/sre/support/support.html
        maintainers:
        - email: hello@stakater.com
          name: StakaterAB
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Stakater
          url: github.com/stakater-ab/hibernation-operator
        relatedImages:
        - registry.connect.redhat.com/stakater/hibernation-operator-bundle@sha256:fdef8f42c5425b9607915439daf3e5e2b77705cd220a083a00b42ee2e122d894
        - registry.connect.redhat.com/stakater/hibernation-operator@sha256:80c6d45a2ca70f55809620a82c309abb1e8272a0515f9c1a2905b3302c1467a1
        version: 0.1.102
      entries:
      - name: hibernation-operator.v0.1.102
        version: 0.1.102
      name: release-0.1
    defaultChannel: release-0.1
    packageName: hibernation-operator
    provider:
      name: Stakater
      url: github.com/stakater-ab/hibernation-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-commonui-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-commonui-operator.v4.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "CommonWebUI",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "example-commonwebui",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator"
                  },
                  "name": "example-commonwebui"
                },
                "spec": {
                  "commonWebUIConfig": {
                    "cpuLimits": "300",
                    "cpuMemory": "256",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.8.0",
                    "ingressPath": "/common-nav",
                    "landingPage": "",
                    "requestLimits": "300",
                    "requestMemory": "256",
                    "serviceName": "common-web-ui"
                  },
                  "globalUIConfig": {
                    "cloudPakVersion": "3.6.0",
                    "defaultAdminUser": "admin",
                    "defaultAuth": "",
                    "enterpriseLDAP": "",
                    "enterpriseSAML": "",
                    "osAuth": "",
                    "sessionPollingInterval": 5000
                  },
                  "operatorVersion": "4.8.0",
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "300m",
                      "memory": "256Mi",
                      "ephemeral-storage": "251Mi"
                    }
                  },
                  "version": "4.8.0"
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "common-web-ui-config",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "default": "true",
                    "name": "common-web-ui-config"
                  },
                  "name": "common-web-ui-config"
                },
                "spec": {
                  "about": {
                    "copyright": "© 2018, 2020 IBM. All rights reserved.",
                    "licenses": [
                      "yq, version 3.3.0, MIT+GPL",
                      "MongoDB, version 4.0.16 Community Edition, SSPL",
                      "Ansible: © 2017 Red Hat, Inc., http://www.redhat.com; © Henry Graham (hzgraham) \u003cHenry.Graham@mail.wvu.edu\u003e",
                      "calico-bird: © 1998–2008, Martin Mares \u003cmj@ucw.cz\u003e; © 1998–2000 Pavel Machek \u003cpavel@ucw.cz\u003e;",
                      "© 1998–2008, Ondrej Filip \u003cfeela@network.cz\u003e; © 2009–2013,  CZ.NIC z.s.p.o.",
                      "chrony: © Richard P. Curnow  1997-2003, GPL v2",
                      "collectd, © 2017-2018, version 5.7.2, GPL v2, \u003chttps://github.com/collectd/collectd/tree/collectd-5.7.2\u003e",
                      "crudini: © Pádraig Brady \u003cP@draigBrady.com\u003e",
                      "Galera-3: © 2007–2014 Codership Oy \u003cinfo@codership.com\u003e",
                      "glusterfs: © 2010–2013+ James Shubin \u003chttps://ttboj.wordpress.com/\u003e",
                      "haproxy: © 2000–2013  Willy Tarreau \u003cw@1wt.eu\u003e",
                      "heketi v6.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi v8.0.0: © 2015 The heketi Authors, GPL v2",
                      "heketi-master/apps.app.go: © 2015 The heketi Authors",
                      "heketi-master/client/api/go-client/backup.go: © 2016 The heketi Authors",
                      "heketi-master/doc/man/heketi-cli.8: © 2016 The heketi Authors",
                      "heketi-master/extras/docker/gluster/gluster-setup.sh: © 2016 Red Hat, Inc. \u003chttp://www.redhat.com\u003e",
                      "ieee-data: © 2013 Luciano Bello \u003cluciano@debian.org\u003e",
                      "javax.mail: © 2017 Oracle and/or its affiliates. All rights reserved.",
                      "keepalived: © 2001-2017 Alexandre Cassen \u003cacassen@gmail.com\u003e",
                      "libonig2: © 2006–2008 Max Kellermann \u003cmax@duempel.org\u003e; © 2014–2015 Jörg Frings-Fürst \u003cdebian@jff-webhosting.net\u003e",
                      "libtomcrypt: © 2004 Sam Hocevar \u003csam@hocevar.net\u003e, GPL v2",
                      "mariadb-common: © 2018 MariaDB. All rights reserved.",
                      "mariaDB: © 2018 MariaDB. All rights reserved. \u003chttps://mariadb.com/\u003e",
                      "mariadb-server: © 2018 MariaDB. All rights reserved.",
                      "minitar: © 2004 Mauricio Julio Fernandez Pradier and Austin Ziegler",
                      "MongoDB: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "nvmi-cli: © 1989, 1991 Free Software Foundation, Inc., GPL v2",
                      "OpenJDK: © 2018 Oracle Corporation and/or its affiliates",
                      "openshift-mariadb-galera: © 2007 Free Software Foundation, Inc. \u003chttp://fsf.org/\u003e",
                      "percona-xtrabackup: © 2006–2018 Percona LLC.",
                      "pwgen: © Christian Thöing \u003cc.thoeing@web.de\u003e",
                      "rdoc: © 2001–2003 Dave Thomas, The Pragmatic Programmers",
                      "readline: © Chet Ramey \u003cchet.ramey@case.edu\u003e",
                      "John the Ripper password cracker: © 1996–2013 by Solar Designer \u003csolar@openwall.com\u003e",
                      "spdx-exceptions: © 2018 SPDX Workgroup a Linux Foundation Project. All rights reserved.",
                      "socat: © 2001–2010 Gerhard Rieger",
                      "sshpass: © 2006, 2008 Lingnu Open Source Consulting Ltd. \u003chttp://www.lingnu.com\u003e",
                      "timelimit: © 2001, 2007 - 2010  Peter Pentchev, GPL v2",
                      "ua-parser-js: © 2012-2018 Faisal Salman \u003cf@faisalman.com\u003e, GPL v2",
                      "ubuntu-cloud-keyring: © 2010 Michael Vogt \u003cmichael.vogt@canonical.com\u003e",
                      "unboundid-ldapsdk: © 2015 UnboundID. The LDAP SDK for Java is developed by UnboundID. \u003cinfo@unboundid.com\u003e",
                      "xmpp4r: © Lucas Nussbaum \u003clucas@lucas-nussbaum.net\u003e, Stephan Maka \u003cstephan@spaceboyz.net\u003e, and others.",
                      "module-assistant: © 2003-2008 Eduard Bloch \u003cblade@debian.org\u003e, version 0.11.8, GPL v2;",
                      "© 2009 Cyril Brulebois \u003ckibi@debian.org\u003e, version 0.11.8, GPL v2; © 2013-2018 Andreas Beckmann \u003canbe@debian.org\u003e, version 0.11.8, GPL v2",
                      "module-init-tools: © 2011 ProFUSION embedded systems, version 22, GPL v2",
                      "thin: © 2017 Marc-Andre Cournoyer \u003cmacournoyer@gmail.com\u003e, version 1.7.2, GPL v2",
                      "gosu, © 1999-2014, version 1.1, GPL v3",
                      "mercurial (Python), © 2006-2018 ,version v4.5.3, GPL v2",
                      "garden-runc, © 2015-Present CloudFoundry.org Foundation, Inc. All Rights Reserved, version 1.17.0, GPLv2",
                      "libtomcrypt0, © 2003-2007 Tom St Denis \u003ctomstdenis@gmail.com\u003e, version 1.17-7, GPLv2",
                      "console-setup-min, © 1999,2000,2001,2002,2003,2006,2007,2008,2009,2010,2011 Anton Zinoviev, \u003canton@lml.bas.bg\u003e,version 1.108, GPLv2",
                      "dracut, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "dracut-core, © 2009 Harald Hoyer \u003charald@redhat.com\u003e, version 044+3-3, GPLv2",
                      "g++, version 5.4.0-6ubuntu, GPL v2",
                      "libstdc++6, version 5.4.0-6ubuntu, GPL v3",
                      "libstdc++-5-dev, version 5.4.0-6ubuntu, GPL v3",
                      "docker-engine-selinux, version 3b5fac4, GPLv2",
                      "unorm, version 1.5.0, GPL v2",
                      "psmisc, version 22.20, GPL v2",
                      "lvm2-devel, version 2.0.2, GPL v2",
                      "nfs-utils, version 1.3, GPL v2",
                      "popt-static, version 1.13, GPL v2",
                      "sysvinit-tools, version 2.88, GPL v2",
                      "stunnel, version 5.53, GPL v2",
                      "stunnel, version 5.39, GPL v2",
                      "LVM2, version 2.02.180-10.el7_6.2, GPL v2",
                      "sysdig, version 2c43237, GPL",
                      "chisels, version 9722dbc, GPL",
                      "MongoDB, version 4.0.12, SSPL",
                      "ffi (Ruby Gem), 1.11.1, GPL",
                      "inotify-tools, v3.14, GPL v2",
                      "logrotate, v3.8.6, GPL v2",
                      "checker-qual, version 2.0.0, GPLv2",
                      "ocrad-bower, v1.0.0, GPL v3",
                      "Grafana, v7.5.12, AGPL",
                      "IBM Semeru Runtime Open Edition binaries, GPL v2",
                      "checker-compat-qual 2.5.5, GPL-2.0",
                      "javax.annotation-api 1.3.1, GPL-2.0"
                    ],
                    "logoUrl": "IBM Cloud Pak | Administration Hub",
                    "logoAltText": "IBM Cloud Pak | Administration Hub"
                  },
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog"
                    ],
                    "docUrlMapping": "http://ibm.biz/cpcs_adminui",
                    "logoAltText": "IBM Cloud Pak Administration Hub",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpack-logo.svg",
                    "logoWidth": "190px"
                  },
                  "login": {
                    "loginDialog": {
                      "acceptText": "Your acceptance text here",
                      "dialogText": "You must set your dialog for this environment",
                      "enable": false,
                      "headerText": "Header text here"
                    },
                    "logoAltText": "Cloud Pak",
                    "logoHeight": "47px",
                    "logoUrl": "/common-nav/api/graphics/logincloudpak.svg",
                    "logoWidth": "190px"
                  },
                  "navItems": [
                    {
                      "id": "providers",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms"
                    }
                  ]
                }
              },
              {
                "apiVersion": "foundation.ibm.com/v1",
                "kind": "NavConfiguration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "icp4i",
                    "app.kubernetes.io/managed-by": "ibm-commonui-operator",
                    "app.kubernetes.io/name": "ibm-commonui-operator",
                    "name": "icp4i"
                  },
                  "name": "icp4i"
                },
                "spec": {
                  "header": {
                    "disabledItems": [
                      "createResource",
                      "catalog",
                      "bookmark"
                    ],
                    "logoAltText": "Cloud Pak for Integration",
                    "logoUrl": "/common-nav/graphics/ibm-cloudpak-integration.svg"
                  },
                  "navItems": [
                    {
                      "detectionServiceName": true,
                      "id": "metering",
                      "label": "Metering",
                      "serviceId": "metering-ui",
                      "serviceName": "metering-ui",
                      "url": "/metering/dashboard?ace_config={ 'showClusterData': false }\u0026dashboard=cpi.icp.main"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "monitoring",
                      "isAuthorized": [
                        "Administrator",
                        "ClusterAdministrator",
                        "Operator"
                      ],
                      "label": "Monitoring",
                      "serviceId": "monitoring-ui",
                      "serviceName": "ibm-monitoring-grafana",
                      "target": "_blank",
                      "url": "/grafana"
                    },
                    {
                      "id": "id-access",
                      "label": "Identity and access",
                      "serviceId": "webui-nav"
                    },
                    {
                      "id": "providers",
                      "parentId": "id-access",
                      "label": "Identity providers",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/realms?useNav=icp4i"
                    },
                    {
                      "id": "teams-ids",
                      "parentId": "id-access",
                      "label": "Teams and service IDs",
                      "serviceId": "webui-nav",
                      "url": "/common-nav/identity-access/teams?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "logging",
                      "label": "Logging",
                      "serviceId": "kibana",
                      "serviceName": "kibana",
                      "target": "_blank",
                      "url": "/kibana"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "releases",
                      "label": "Helm Releases",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/instances?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "repos",
                      "label": "Helm Repositories",
                      "serviceId": "catalog-ui",
                      "serviceName": "catalog-ui",
                      "url": "/catalog/repositories?useNav=icp4i"
                    },
                    {
                      "detectionServiceName": true,
                      "id": "licensing",
                      "label": "Licensing",
                      "serviceId": "ibm-license-service-reporter",
                      "url": "/license-service-reporter",
                      "isAuthorized": [
                        "ClusterAdministrator",
                        "CloudPakAdministrator"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operators.ibm.com/v1alpha1",
                "kind": "SwitcherItem",
                "metadata": {
                  "name": "example-switcheritem"
                },
                "spec": {
                  "cloudPakInfo": {
                    "display": "Administration",
                    "label": "administration-hub",
                    "landingPage": "/common-nav/dashboard",
                    "logoURL": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
          description: The IBM Common Web UI delivers the common header API and the
            identity and access pages for IBM Cloud Platform Common Services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.8.0
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-commonui-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            displayName: CommonWebUI service
            kind: CommonWebUI
            name: commonwebuis.operators.ibm.com
            version: v1alpha1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: NavConfiguration
            name: navconfigurations.foundation.ibm.com
            version: v1
          - description: |
              Documentation For additional details regarding install parameters check: https://ibm.biz/icpfs39install.
               License By installing this product you accept the license terms https://ibm.biz/icpfs39license.'
            kind: SwitcherItem
            name: switcheritems.operators.ibm.com
            version: v1alpha1
        description: "**Important:** Do not install this operator directly. Only install
          this operator using the IBM Common Services Operator. For more information
          about installing this operator and other Common Services operators, see
          [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally,
          you can exit this panel and navigate to the IBM Common Services tile in
          OperatorHub to learn more about the operator.\n\n If you are using this
          operator as part of an IBM Cloud Pak, see the documentation for that IBM
          Cloud Pak to learn more about how to install and use the operator service.
          For more information about IBM Cloud Paks, see [IBM Cloud Paks that use
          Common Services](http://ibm.biz/cpcs_cloudpaks).\n\n You can use the ibm-commonui-operator
          to install the Common Web UI service for the IBM Cloud Platform Common Services
          and access the Common Web UI console. You can use the Common Web UI console
          to access information and features from other IBM Cloud Platform Common
          Services or IBM Cloud Paks that you install. \n\nFor more information about
          the available IBM Cloud Platform Common Services, see the [IBM Knowledge
          Center](http://ibm.biz/cpcsdocs). \n## Supported platforms \n\n Red Hat
          OpenShift Container Platform 4.2 or newer installed on one of the following
          platforms: \n\n- Linux x86_64 \n- Linux on Power (ppc64le) \n- Linux on
          IBM Z and LinuxONE \n## Prerequisites\n\n The Common Web UI service has
          dependencies on other IBM Cloud Platform Common Services. Before you install
          this operator, you need to first install the operator dependencies and prerequisites:
          \n For the list of operator dependencies, see the IBM Knowledge Center [Common
          Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).
          \n For the list of prerequisites for installing the operator, see the IBM
          Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).
          \n## Documentation \n\n To install the operator with the IBM Common Services
          Operator follow the the installation and configuration instructions within
          the IBM Knowledge Center. \n- If you are using the operator as part of an
          IBM Cloud Pak, see the documentation for that IBM Cloud Pak, for a list
          of IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          \n- If you are using the operator with an IBM Containerized Software, see
          the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall)."
        displayName: Ibm Common UI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        - Web Console
        - Common Services
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-commonui-operator
        maintainers:
        - email: siddarth.patel@ibm.com
          name: Siddarth Patel
        - email: sgrube@us.ibm.com
          name: Steve Grube
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-commonui-operator@sha256:87788afbc24851deb2c9c3f14b773bab174b2a2f02ac2fa0dd656a97433ffca2
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        version: 4.8.0
      entries:
      - name: ibm-commonui-operator.v4.8.0
        version: 4.8.0
      name: v4.7
    defaultChannel: v4.7
    packageName: ibm-commonui-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/dns-operator:v0.6.0
          createdAt: "2024-09-13T09:55:39Z"
          description: A Kubernetes Operator to manage the lifecycle of DNS resources
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: A Kubernetes Operator to manage the lifecycle of DNS resources
        displayName: DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: DNS Operator
          url: https://github.com/Kuadrant/dns-operator
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dns-operator@sha256:d5eadbf006aade69ac18156c891614da899d0779ba9e8fbaf755a256c46f055c
        - quay.io/kuadrant/dns-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: dns-operator.v0.6.0
        version: 0.6.0
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-custom-metrics-autoscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: custom-metrics-autoscaler.v2.17.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "CloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "cloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "cloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "cluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "eventing.keda.sh/v1alpha1",
                "kind": "ClusterCloudEventSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "keda",
                    "app.kubernetes.io/instance": "clustercloudeventsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clustercloudeventsource",
                    "app.kubernetes.io/part-of": "keda"
                  },
                  "name": "clustercloudeventsource-sample"
                },
                "spec": {
                  "clusterName": "clustercluster-sample",
                  "destination": {
                    "http": {
                      "uri": "http://foo.bar"
                    }
                  }
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "openshift-keda"
                },
                "spec": {
                  "admissionWebhooks": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "operator": {
                    "logEncoder": "console",
                    "logLevel": "info"
                  },
                  "serviceAccount": null,
                  "watchNamespace": ""
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "example-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "example-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
          createdAt: 20 Aug 2025, 21:31
          description: Custom Metrics Autoscaler Operator, an event-driven autoscaler
            based upon KEDA
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.1 <2.17.2-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-keda
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/custom-metrics-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudEventSource
            name: cloudeventsources.eventing.keda.sh
            version: v1alpha1
          - kind: ClusterCloudEventSource
            name: clustercloudeventsources.eventing.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
        description: |
          ## About the managed application
          Custom Metrics Autoscaler for OpenShift is an event driven autoscaler based upon KEDA.  Custom Metrics Autoscaler can monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed the metrics from those sources into the Kubernetes horizontal pod autoscaler.  With Custom Metrics Autoscaler, you can have event driven and serverless scale of deployments within any Kubernetes cluster.
          ## About this Operator
          The Custom Metrics Autoscaler Operator deploys and manages installation of KEDA Controller in the cluster. Install this operator and follow installation instructions on how to install Custom Metrics Autoscaler in your cluster.

          ## Prerequisites for enabling this Operator
          ## How to install Custom Metrics Autoscaler in the cluster
          The installation of Custom Metrics Autoscaler is triggered by the creation of `KedaController` resource. Please refer to the [KedaController Spec](https://github.com/openshift/custom-metrics-autoscaler-operator/blob/main/README.md#the-kedacontroller-custom-resource) for more details on available options.

          Only a resource named `keda` in the namespace where the operator is installed will trigger the installation, reconfiguration or removal of the KEDA Controller resource.

          There should be only one KEDA Controller in the cluster.

          The operator will behave in this manner whether it is installed with the `AllNamespaces` or `OwnNamespace` install mode. While the operator more closely matches the `OwnNamespace` semantics, `AllNamespaces` is a supported installation mode to allow it to be installed to namespaces with existing `OperatorGroups` which require that installation mode, such as `openshift-operators`.
        displayName: Custom Metrics Autoscaler
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Custom Metrics Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/cma/nodes-cma-autoscaling-custom.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9@sha256:ace9358fbd2d81a6e16bfa049c763ca16acf6cc2a59cdc0b6282e5e0dfc902cc
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-operator-bundle@sha256:aeae57b2fc706b782a3e3905e5c1825fe831f96f5b1c9478e1f2b217307a892b
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-adapter-rhel9@sha256:0c8ae9323f927346e168ba829f6c7679af6b2dd359c8d0a020dd0ae3954481ce
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-admission-webhooks-rhel9@sha256:75631a21de057844e9ae45ae2496a22e16f3a32a6fd5f6cb5f90e173e3be2180
        - registry.redhat.io/custom-metrics-autoscaler/custom-metrics-autoscaler-rhel9-operator@sha256:9a359ae6612d2f22fe983653418143873cb9939f83dca6f008ec3cb8f3d53990
        version: 2.17.2-1
      entries:
      - name: custom-metrics-autoscaler.v2.17.2-1
        version: 2.17.2-1
      - name: custom-metrics-autoscaler.v2.15.1-6
        version: 2.15.1-6
      - name: custom-metrics-autoscaler.v2.15.1-4
        version: 2.15.1-4
      - name: custom-metrics-autoscaler.v2.14.1-467
        version: 2.14.1-467
      - name: custom-metrics-autoscaler.v2.14.1-454
        version: 2.14.1-454
      - name: custom-metrics-autoscaler.v2.13.1-427
        version: 2.13.1-427
      - name: custom-metrics-autoscaler.v2.12.1-394
        version: 2.12.1-394
      - name: custom-metrics-autoscaler.v2.11.2-322
        version: 2.11.2-322
      - name: custom-metrics-autoscaler.v2.11.2-311
        version: 2.11.2-311
      - name: custom-metrics-autoscaler.v2.10.1-267
        version: 2.10.1-267
      name: stable
    defaultChannel: stable
    packageName: openshift-custom-metrics-autoscaler-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataDirect Networks
      provider-url: https://www.ddn.com
    name: vmstore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vmstore-csi-driver-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.tintri.com/v1alpha1",
                "kind": "VmstoreCsi",
                "metadata": {
                  "name": "vmstorecsi-sample"
                },
                "spec": {
                  "config": {
                    "debug": true,
                    "vmstore_map": {
                      "vmstore2": {
                        "dataIp": "\u003cdataIp\u003e",
                        "managementAddress": "\u003cmanagementAddress\u003e",
                        "mountPoint": "/tintri-csi",
                        "password": "tintri99",
                        "username": "admin",
                        "vmstoreFS": "/tintri/csi"
                      }
                    }
                  },
                  "csi_snapshotter": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                    "resources": null,
                    "tag": "v6.3.2",
                    "timeout": "120m"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "tintri/vmstore-openshift-file-driver",
                    "tag": "v1.1.0"
                  },
                  "priorityClassName": "system-cluster-critical",
                  "provisioner": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                    "resources": null,
                    "tag": "v4.0.1",
                    "timeout": "120m"
                  },
                  "registrar": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                    "resources": null,
                    "tag": "v2.10.1"
                  },
                  "resizer": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.k8s.io/sig-storage/csi-resizer",
                    "resources": null,
                    "tag": "v1.12.0"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "512m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "64Mi"
                    }
                  },
                  "secretName": "vmstore-csi-file-driver-config",
                  "volumeNamePrefix": "pvc-vmstore"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-13T21:26:54Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: VmstoreCsi
            name: vmstorecsis.charts.tintri.com
            version: v1alpha1
        description: "A Container Storage Interface (CSI) Driver for Tintri Vmstore
          filesystems. This operator eases the installation\nand management of the
          Vmstore CSI Driver in an OpenShift or other Operator Lifecycle Manager enabled
          cluster.\n\n\n## Prerequisites\nMake sure that the top level folder where
          the CSI volumes will be created exists on VMStore and is shared over NFS.\nFor
          example `tintri/csi`.\n## Usage\n\n### Dynamically provisioned volumes\n\nFor
          dynamic volume provisioning, the administrator needs to set up a _StorageClass_
          in the PV yaml (examples/nginx-dynamic-volume.yaml for this example) pointing
          to the driver.\nFor dynamically provisioned volumes, Kubernetes generates
          volume name automatically (for example `pvc-vmstore-cfc67950-fe3c-11e8-a3ca-005056b857f8`).\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\n```\n\n#### Example\n\nRun
          Nginx pod with dynamically provisioned volume:\n\n```bash\noc apply -f examples/nginx-dynamic-volume.yaml\n\n#
          to delete this pod:\noc delete -f examples/nginx-dynamic-volume.yaml\n```\n\n###
          Static (pre-provisioned) volumes\n\nThe driver can use already existing
          Vmstore filesystem,\nin this case, _StorageClass_, _PersistentVolume_ and
          _PersistentVolumeClaim_ should be configured.\n\n#### _StorageClass_ configuration\n\n```yaml\napiVersion:
          storage.k8s.io/v1\nkind: StorageClass\nmetadata:\n  name: vmstore-csi-file-driver-sc-nginx-persistent\nprovisioner:
          vmstore.csi.tintri.com\nallowVolumeExpansion: true\nallowedTopologies:\n-
          matchLabelExpressions:\n  - key: topology.vmstore.csi.tintri.com/zone\n
          \   values:\n    - zone-1\nmountOptions:                         # list
          of options for `mount -o ...` command\n#  - noatime                          #\n```\n\n####
          _PersistentVolume_ configuration\n\n```yaml\napiVersion: v1\nkind: PersistentVolume\nmetadata:\n
          \ name: vmstore-csi-file-driver-pv-nginx-persistent\n  labels:\n    name:
          vmstore-csi-file-driver-pv-nginx-persistent\nspec:\n  storageClassName:
          vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n    - ReadWriteOnce\n
          \ capacity:\n    storage: 1Gi\n  csi:\n    driver: vmstore.csi.tintri.com\n
          \   volumeHandle: vmstore1/172.30.230.128:/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img\n```\n\n####
          _PersistentVolumeClaim_ (pointing to created _PersistentVolume_)\n\n```yaml\napiVersion:
          v1\nkind: PersistentVolumeClaim\nmetadata:\n  name: vmstore-csi-file-driver-pvc-nginx-persistent\nspec:\n
          \ storageClassName: vmstore-csi-file-driver-sc-nginx-persistent\n  accessModes:\n
          \   - ReadWriteOnce\n  resources:\n    requests:\n      storage: 1Gi\n  selector:\n
          \   matchLabels:\n      # to create 1-1 relationship for pod - persistent
          volume use unique labels\n      name: vmstore-csi-file-driver-pv-nginx-persistent\n```\n\n###
          Topology configuration\nIn order to configure CSI driver with kubernetes
          topology, use the `zone` parameter in driver config or storageClass parameters.
          Example config file with zones:\n  ```bash\n  vmstore_map:\n    vmstore1:\n
          \     mountPoint: /tintri-csi                            # mountpoint on
          the host where the tintri fs will be mounted\n      dataIp: 172.30.200.37\n
          \     managementAddress: tafzq23.tintri.com\n      vmstoreFS: /tintri/fs01\n
          \     zone: zone-1\n      username: admin\n      password: password\n\n
          \   vmstore2:\n      mountPoint: /tintri-csi                            #
          mountpoint on the host where the tintri fs will be mounted\n      dataIp:
          10.136.40.71\n      managementAddress: tafbm05.tintri.com\n      vmstoreFS:
          /tintri/csi\n      zone: zone-2\n      username: admin\n      password:
          password\n  ```\n\nThis will assign volumes to be created on Vmstore that
          correspond with the zones requested by allowedTopologies values.\n\n\n###
          Defaults/Configuration/Parameter options\n\n| Default  | Config               |
          \ Parameter    |  Desc                             |\n|----------|----------------------|---------------|-----------------------------------|\n|
          \  -      | dataIp               |   dataIp      | Data address used to
          mount Vmstore filesystem |\n| -    | mountPoint       |    -      | Mount
          point inside of the driver container where Vmstore filesystem will be mounted|\n|
          \  -      | username             |      -        | Vmstore API username
          |\n|   -      | password             |      -        | Vmstore API password
          |\n|   -      | zone                 |      -        | Zone to match topology.vmstore.csi.tintri.com/zone
          \           |\n|   -      | vmstoreFS            |   vmstoreFS   | Path
          to filesystem on Vmstore |\n|   -      | managementAddress    |      -        |
          FQDN for Vmstore management operations |\n|   -      | defaultMountOptions
          \ |   mountOptions| Mount options |\n|   0777   | mountPointPermissions|
          \  mountPointPermissions | File permissions to be set on mount point |\n\n####
          Example\n\nRun nginx server using PersistentVolume.\n\n**Note:** Pre-configured
          volume file should exist on the Vmstore:\n`/tintri/csi/vmstorefs-mnt/persistent/volumefile-persistent.img`.\n\n```bash\noc
          apply -f nginx-persistent-volume.yaml\n\n# to delete this pod and volume:\noc
          delete -f nginx-persistent-volume.yaml\n```\n\n## Snapshots\nTo use CSI
          snapshots, the snapshot CRDs along with the csi-snapshotter must be installed.\n```bash\noc
          apply -f deploy/kubernetes/snapshots/\n```\n\nAfter that the snapshot class
          for VMStore CSI must be created\n\n```bash\noc apply -f examples/snapshot-class.yaml\n```\n\nNow
          a snapshot can be created, for example\n```bash\noc apply -f examples/snapshot-from-dynamic.yaml\n```\n\nWe
          can create a volume using the snapshot\n```bash\noc apply -f examples/nginx-from-snapshot.yaml\n```\n\n##
          Troubleshooting\n### Driver logs\nTo collect all driver related logs, you
          can use the `oc logs` command.\nAll in on command:\n```bash\nmkdir vmstore-csi-logs\nfor
          name in $(oc get pod -owide | grep vmstore | awk '{print $1}'); do oc logs
          $name --all-containers > vmstore-csi-logs/$name; done\n```\n\nTo get logs
          from all containers of a single pod \n```bash\noc logs <pod_name> --all-containers\n```\n\nLogs
          from a single container of a pod\n```bash\noc logs <pod_name> -c driver\n```\n\n####
          Driver secret\n```bash\noc get secret vmstore-csi-file-driver-config -o
          json | jq '.data | map_values(@base64d)' > vmstore-csi-logs/vmstore-csi-file-driver-config\n```\n\n####
          PV/PVC/Pod data\n```bash\noc get pvc > vmstore-csi-logs/pvcs\noc get pv
          > vmstore-csi-logs/pvs\noc get pod > vmstore-csi-logs/pods\n```\n\n####
          Extended info about a PV/PVC/Pod:\n```bash\noc describe pvc <pvc_name> >
          vmstore-csi-logs/<pvc_name>\noc describe pv <pv_name> > vmstore-csi-logs/<pv_name>\noc
          describe pod <pod_name> > vmstore-csi-logs/<pod_name>\n```"
        displayName: Vmstore CSI file driver
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - CSI
        - Vmstore
        - storage
        - Tintri
        - DDN
        links:
        - name: DDN vmstore Filesystem
          url: https://www.ddn.com/products/lustre-file-system-vmstore/
        maintainers:
        - email: akhodos@tintri.com
          name: Alexey Khodos
        - email: rrethinam@tintri.com
          name: Ranjith Rethinam
        maturity: alpha
        provider:
          name: DataDirect Networks
          url: https://www.ddn.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/ddn/vmstore-csi-file-driver-operator@sha256:64df1c504a9d70bd2b8b34520c8d431a92bb54702d601a57a2a989c6dd982462
        - tintri/vmstore-csi-driver-operator-controller@sha256:749ceff2bd46c71d81338510c1cc5866d39c66345f971d3093f986ea0417311d
        version: 1.1.0
      entries:
      - name: vmstore-csi-driver-operator.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vmstore-csi-driver-operator
    provider:
      name: DataDirect Networks
      url: https://www.ddn.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox Inc.
      provider-url: https://accuknox.com
    name: accuknox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: accuknox-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
          createdAt: "2024-02-09T17:06:30Z"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: public.ecr.aws/k9v9d5v2/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: alpha
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:b6fae25e229fb7fca757872fc0460642d52c4f9906f776afac50c54a69b05b68
        - public.ecr.aws/k9v9d5v2/agents-operator@sha256:656d6bac4371adef3e86658011ba68ae2f98514f6ddaea0667d4269eb88dd02e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - public.ecr.aws/k9v9d5v2/discovery-engine-hardening@sha256:3eb28d44ef93e6ba331e9085e64475458e1c234017c77dc30b8bc802def434c3
        - public.ecr.aws/k9v9d5v2/discovery-engine-sumengine@sha256:4ca78f90b28b38badb0520f148ce702a3a6cd6b02bb62e8b1a6782a5b82e0868
        - public.ecr.aws/k9v9d5v2/policy-enforcement-agent@sha256:f004fc3572a4d66da1bc0db9e05a279d761a42d6ed62f95de7a2bb42ce71e17b
        - public.ecr.aws/k9v9d5v2/feeder-service@sha256:f5bc340d9d3e34804dd54a03701d272e0ce5c04160d1aa41984cb3d877eb5b76
        - public.ecr.aws/k9v9d5v2/shared-informer-agent@sha256:6d18d13312d4309b6e32b18c063d5b26f15fbf057fda2b9e6b7e8ebb1a5885e9
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:042623e1faacc9fc22222a1c3e546a561abf7fc05d41ae9c915a69795f6bec7e
        - public.ecr.aws/k9v9d5v2/discovery-engine-discover@sha256:052efbe1064503853adab497415b9407f51578e1456931a8a0e60353aa53e0c2
        - public.ecr.aws/k9v9d5v2/discovery-engine-offloader@sha256:3c2291310e30e35a931d2af389f7ade656ce86f609d2e42ecdabf1000764695c
        version: 0.2.1
      entries:
      - name: accuknox-operator.v0.2.1
        version: 0.2.1
      name: alpha
    - currentCSV: accuknox-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.accuknox.com/v1",
                "kind": "AccuknoxConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "accuknox-operator",
                    "app.kubernetes.io/instance": "accuknoxconfig",
                    "app.kubernetes.io/name": "accuknoxconfig-sample",
                    "app.kubernetes.io/part-of": "accuknox-operator"
                  },
                  "name": "accuknoxconfig-sample",
                  "namespace": "accuknox-agents"
                },
                "spec": {
                  "ppsHost": "PLEASE_FILL_IN",
                  "knoxGateway": "PLEASE_FILL_IN",
                  "spireHost": "PLEASE_FILL_IN"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Security
          containerImage: public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
          createdAt: "2024-08-23T04:25:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: accuknox-agents
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: docker.io/accuknox/accuknox-operator
          support: Accuknox Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccuknoxConfig is the Schema for the accuknoxconfigs API
            kind: AccuknoxConfig
            name: accuknoxconfigs.operator.accuknox.com
            version: v1
        description: |2-

          The AccuKnox Operator brings native K8s support for deploying and managing AccuKnox Components on a Kubernetes cluster.
          The AccuKnox Operator helps to easily onboard a k8s cluster to the AccuKnox SaaS.

          AccuKnox Operator deploys and manages the components:

          1. Discovery Engine

          2. Policy Enforcement Agent

          3. Shared Informer Agent

          4. Feeder Service

          5. Agents Operator

          ## Prerequisites for using this Operator

          1. Get AccuKnox licence - https://app.accuknox.com/sign-up

          2. After Successful Sign-up into AccuKnox Application.

             Please naviagate to Settings >> Manager Cluster >> Onboard Cluster >>

             After successful onboarding, Copy the join token created by AccuKnox

          3. Create a secret using jointoken by executing the below command:

          ```
          oc create secret generic accuknox-jointoken --from-literal=token=<joinToken> -n accuknox-agents
          ```

          ## Deploy AccuKnox Operator using AccuknoxConfig Custom Resource

          All the information to configure the AccuKnox Operator will be provided via Accuknox during Cluster Onboarding steps.

          ### AccuknoxConfig Spec

          | Spec | Description |
          | ---- | ----------- |
          | spireHost | "spire host url" |
          | ppsHost   |  "pps host url" |
          | knoxGateway | "knox gateway url" |
          | joinTokenSecretName | "join token secret name if it's different than accuknox-jointoken" |
          | joinTokenSecretKey | "key name used to store the jointoken with the secret" |

          *Example*

          ```
          apiVersion: operator.accuknox.com/v1
          kind: AccuknoxConfig
          metadata:
            name: accuknoxconfig-default
            namespace: accuknox-agents
          spec:
            ppsHost: <ppsHost url>
            knoxGateway: <knox gateway url>
            spireHost: <spire host url>
          ```
        displayName: Accuknox Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CNAPP
        - CSPM
        - CWPP
        - Cloud Native
        - Runtime Security
        - Zero-Trust
        maturity: stable
        provider:
          name: Accuknox Inc.
          url: https://accuknox.com
        relatedImages:
        - docker.io/accuknox/shared-informer-agent@sha256:9f53d7dba9569487c42c6cf9e88988a1adb4c0971742e8b521700dce2d22b8cc
        - public.ecr.aws/k9v9d5v2/accuknox-operator@sha256:a935ad806e82760bb838bada30d3ed1a54c8c61b95fe5c6109f48985c6faf4d7
        - docker.io/accuknox/discovery-engine-hardening@sha256:e596aaaabaec94e0abb4f92ad9266dc010c0e88a89d600e8750a360099cae3cc
        - docker.io/accuknox/discovery-engine-offloader@sha256:59cc594009de89a574cd569e701b6765d59a6899a03ad07eee1c025331df4331
        - docker.io/accuknox/discovery-engine-sumengine@sha256:f7927128b86fdfb635377bd599df06f0eb9b9f0d1b068a0047468e480855443e
        - docker.io/accuknox/policy-enforcement-agent@sha256:a0364343afeac4537589bfc247f64fffd7aac16f8afe07ce050bc58b2af248fb
        - registry.connect.redhat.com/kubearmor/k9v9d5v2@sha256:67b2b8783a04460019057cf826a74c18747c8c64ee7fe8ce83c8d0b7cd87cb36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        - docker.io/accuknox/agents-operator@sha256:b0ae2728f15ebae904efb778cd29215ccf07ffca945a59deb5484f6c19c1b21f
        - docker.io/accuknox/discovery-engine-discover@sha256:e2d37bfbd0e9d23c129c143eb7c9aab1c3ed55e9571b44888eea2b842f8a2361
        - docker.io/accuknox/feeder-service@sha256:295e897b6f27ce61c45fb0179b734ef6d4c9e91864bcc5d2884a2d1dd4db81b6
        version: 0.7.1
      entries:
      - name: accuknox-operator.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: accuknox-operator-certified
    provider:
      name: Accuknox Inc.
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: resource-locker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: resource-locker-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "ResourceLocker",
                "metadata": {
                  "name": "combined"
                },
                "spec": {
                  "patches": [
                    {
                      "id": "patch1",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0).metadata.name }}: {{ (index . 1).metadata.name }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    },
                    {
                      "id": "patch2",
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 0) }}: {{ (index . 1) }}\n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "Namespace",
                          "name": "resource-locker-test"
                        },
                        {
                          "apiVersion": "v1",
                          "fieldPath": "$.metadata.uid",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "resource-locker-test"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "resource-locker-test"
                      }
                    }
                  ],
                  "resources": [
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-same-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "podSelector": {}
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    },
                    {
                      "object": {
                        "apiVersion": "networking.k8s.io/v1",
                        "kind": "NetworkPolicy",
                        "metadata": {
                          "name": "allow-from-default-namespace",
                          "namespace": "resource-locker-test"
                        },
                        "spec": {
                          "ingress": [
                            {
                              "from": [
                                {
                                  "namespaceSelector": {
                                    "matchLabels": {
                                      "name": "default"
                                    }
                                  }
                                }
                              ]
                            }
                          ],
                          "podSelector": {}
                        }
                      }
                    }
                  ],
                  "serviceAccountRef": {
                    "name": "default"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
          createdAt: "2021-11-30T19:38:40Z"
          description: This operator provides a facility to lock resources and / or
            patches into the state described by its custom resource preventing any
            drifts.
          operatorframework.io/suggested-namespace: resource-locker-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/resource-locker-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResourceLocker Represents the intetion to create and enforce
              a set of resources and/or patches.
            displayName: Resource Locker
            kind: ResourceLocker
            name: resourcelockers.redhatcop.redhat.io
            version: v1alpha1
        description: |
          The resource locker operator allows you to specify a set of configurations that the operator will "keep in place" (lock) preventing any drifts.
          Two types of configurations may be specified:
          * Resources. This will instruct the operator to create and enforce the specified resource. In this case the operator "owns" the created resources.
          * Patches to resources. This will instruct the operator to patch- and enforce the change on- a pre-existing resource. In this case the operator does not "own" the resource.
          Locked resources are defined with the `ResourceLocker` CRD. Here is the high-level structure of this CRD:
          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: ResourceLocker
          metadata:
            name: test-simple-resource
          spec:
            resources:
            - object:
                apiVersion: v1
            ...
            patches:
            - targetObjectRef:
              ...
              patchTemplate: |
                metadata:
                  annotations:
                    ciao: hello
            ...
            serviceAccountRef:
              name: default
          ```
          It contains:
          * `resources`: representing an array of resources
          * `patches`: representing an array of patches
          * `serviceAccountRef`: a reference to a service account defined in the same namespace as the ResourceLocker CR, that will be used to create the resources and apply the patches. If not specified the service account will be defaulted to: `default`
          For each ResourceLocker a manager is dynamically allocated. For each resource and patch a controller with the needed watches is created and associated with the previously created manager.
          **Note**: creating a `ResourceLocker` CR will not work without due preparation (of the service account in particular). Follow these [instructions](https://github.com/redhat-cop/resource-locker-operator/blob/master/EXAMPLES.md) for a step by step walkthrough of the needed prerequisites to make this operator work.
        displayName: Resource Locker Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - resource_management
        - enforcing
        - patches
        - gitops
        links:
        - name: Resource Locker Operator
          url: https://github.com/redhat-cop/resource-locker-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/resource-locker-operator:latest
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/resource-locker-operator@sha256:9cc08639be2af4b72a9b816420d87dfa290a168fb83edc947a04b3e06cdb675b
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/resource-locker-operator@sha256:2176db02f561c24012c5179495c89c3819500dff21e89a62052d465eb3314b44
        version: 1.1.3
      entries:
      - name: resource-locker-operator.v1.1.3
        version: 1.1.3
      - name: resource-locker-operator.v1.1.2
        version: 1.1.2
      - name: resource-locker-operator.v1.1.1
        version: 1.1.1
      - name: resource-locker-operator.v1.1.0
        version: 1.1.0
      - name: resource-locker-operator.v1.0.9
        version: 1.0.9
      name: alpha
    defaultChannel: alpha
    packageName: resource-locker-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lib-bucket-provisioner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lib-bucket-provisioner.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Big Data
          certified: "false"
          containerImage: quay.io/noobaa/pause
          createdAt: "2020-07-01T06:28:48Z"
          description: Library for the dynamic provisioning of object store buckets
            to be used by object store providers.
          repository: https://github.com/kube-object-storage/lib-bucket-provisioner
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: Claim a bucket just like claiming a PV. Automate you app
              bucket provisioning by creating OBC with your app deployment. A secret
              and configmap (name=claim) will be created with access details for the
              app pods.
            displayName: ObjectBucketClaim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
          - description: Used under-the-hood. Created per ObjectBucketClaim and keeps
              provisioning information
            displayName: ObjectBucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
        description: |
          ### CRD-only Operator

          This operator package is **CRD-only** and the operator is a no-op operator.

          Instead, bucket provisioners using this library are using these CRD's and using CSV [required-crds](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/Documentation/design/building-your-csv.md#required-crds) them so that OLM can install it as a dependency.

          ### Generic Bucket Provisioning

          Kubernetes natively supports dynamic provisioning for many types of file and block storage, but lacks support for object bucket provisioning.

          This repo is a placeholder for an object store bucket provisioning library, very similar to the Kubernetes [sig-storage-lib-external-provisioner](https://github.com/kubernetes-sigs/sig-storage-lib-external-provisioner/blob/master/controller/controller.go) library.

          ### Known Provisioners
          - https://github.com/noobaa/noobaa-operator (NooBaa)
          - https://github.com/rook/rook (Rook-Ceph)
          - https://github.com/yard-turkey/aws-s3-provisioner (AWS-S3)
        displayName: lib-bucket-provisioner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - object
        - bucket
        - storage
        - cloud
        - s3
        links:
        - name: Github
          url: https://github.com/kube-object-storage/lib-bucket-provisioner
        maintainers:
        - email: jcope@redhat.com
          name: Jon Cope
        - email: jvance@redhat.com
          name: Jeff Vance
        - email: gmargali@redhat.com
          name: Guy Margalit
        - email: dzaken@redhat.com
          name: Danny Zaken
        - email: nbecker@redhat.com
          name: Nimrod Becker
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/lib-bucket-provisioner@sha256:132ecc88db79fb8b117a7a7974ca8f704600bac489923d918b78fd00378d5a5f
        - quay.io/noobaa/pause
        version: 2.0.0
      entries:
      - name: lib-bucket-provisioner.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: lib-bucket-provisioner
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ClusterPulse
      provider-url: ""
    name: clusterpulse
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterpulse.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.clusterpulse.io/v1alpha1",
                "kind": "ClusterPulse",
                "metadata": {
                  "name": "clusterpulse-prod"
                },
                "spec": {
                  "api": {
                    "enabled": true,
                    "imagePullPolicy": "Always",
                    "logging": {
                      "json": true,
                      "level": "INFO"
                    },
                    "oauth": {
                      "enabled": true,
                      "headerEmail": "X-Forwarded-Email",
                      "headerUser": "X-Forwarded-User"
                    },
                    "rbac": {
                      "auditEnabled": true,
                      "cacheTTL": 0,
                      "defaultDeny": true,
                      "enabled": true,
                      "groupSyncInterval": 300
                    },
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "security": {
                      "allowedOrigins": [
                        "*"
                      ],
                      "sessionExpireMinutes": 1440,
                      "trustedHosts": [
                        "*"
                      ]
                    },
                    "server": {
                      "host": "0.0.0.0",
                      "port": 8080,
                      "reload": false,
                      "workers": 1
                    }
                  },
                  "clusterEngine": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    },
                    "timing": {
                      "cacheTTL": "600",
                      "fullScanInterval": "300",
                      "reconciliationInterval": "30"
                    }
                  },
                  "frontend": {
                    "config": {
                      "features": {
                        "devTools": false,
                        "mockData": false,
                        "nodeDetails": true,
                        "operatorDetails": true,
                        "registryStatus": true
                      },
                      "refreshInterval": 30000,
                      "theme": {
                        "defaultTheme": "auto"
                      }
                    },
                    "enabled": true,
                    "oauth": {
                      "enabled": true,
                      "image": "quay.io/openshift/origin-oauth-proxy:4.17"
                    },
                    "replicas": 3,
                    "route": {
                      "enabled": true,
                      "tls": {
                        "termination": "reencrypt"
                      }
                    }
                  },
                  "policyEngine": {
                    "batch": {
                      "cacheClearBatchSize": 500,
                      "redisPipelineBatchSize": 1000,
                      "redisScanBatchSize": 100
                    },
                    "cache": {
                      "groupCacheTTL": 300,
                      "policyCacheTTL": 300
                    },
                    "enabled": true,
                    "limits": {
                      "maxPoliciesPerUser": 100,
                      "policyValidationInterval": 300
                    },
                    "logging": {
                      "level": "INFO"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "enabled": true,
                    "volumePermissions": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "ClusterConnection",
                "metadata": {
                  "name": "example-cluster"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "cluster-credentials"
                  },
                  "displayName": "Example Production Cluster",
                  "endpoint": "https://api.example.openshift.com:6443",
                  "monitoring": {
                    "interval": 30,
                    "timeout": 10
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "MonitorAccessPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "access": {
                    "effect": "Allow",
                    "enabled": true
                  },
                  "identity": {
                    "priority": 100,
                    "subjects": {
                      "groups": [
                        "default-group"
                      ],
                      "users": [
                        "default-user"
                      ]
                    }
                  },
                  "operations": {
                    "audit": {
                      "logAccess": false,
                      "requireReason": false
                    }
                  },
                  "scope": {
                    "clusters": {
                      "default": "none",
                      "rules": [
                        {
                          "permissions": {
                            "view": true,
                            "viewAuditInfo": true,
                            "viewCosts": true,
                            "viewMetadata": true,
                            "viewMetrics": true,
                            "viewSecrets": true,
                            "viewSensitive": true
                          },
                          "resources": {
                            "namespaces": {
                              "visibility": "all"
                            },
                            "nodes": {
                              "filters": {
                                "hideMasters": false
                              },
                              "visibility": "all"
                            },
                            "operators": {
                              "visibility": "none"
                            },
                            "pods": {
                              "visibility": "all"
                            }
                          },
                          "selector": {
                            "matchPattern": ".*"
                          }
                        }
                      ]
                    }
                  }
                }
              },
              {
                "apiVersion": "clusterpulse.io/v1alpha1",
                "kind": "RegistryConnection",
                "metadata": {
                  "name": "dockerhub",
                  "namespace": "clusterpulse"
                },
                "spec": {
                  "displayName": "Docker Hub",
                  "endpoint": "https://registry-1.docker.io",
                  "healthCheckPaths": [
                    "/v2/"
                  ],
                  "insecure": false,
                  "labels": {
                    "environment": "production",
                    "provider": "dockerhub"
                  },
                  "monitoring": {
                    "checkCatalog": false,
                    "interval": 60,
                    "maxCatalogEntries": 100,
                    "timeout": 10
                  },
                  "skipTLSVerify": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: docker.io/clusterpulse/clusterpulse-operator:0.1.0
          createdAt: "2025-09-26T21:09:27Z"
          description: Provides OpenShift cluster health monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/ClusterPulse/operator
          support: None
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a connection to a monitored Kubernetes/OpenShift
              cluster
            displayName: Cluster Connection
            kind: ClusterConnection
            name: clusterconnections.clusterpulse.io
            version: v1alpha1
          - description: Manages the deployment and configuration of ClusterPulse
              monitoring platform
            displayName: ClusterPulse Instance
            kind: ClusterPulse
            name: clusterpulses.charts.clusterpulse.io
            version: v1alpha1
          - description: Defines fine-grained access control policies for cluster
              monitoring
            displayName: Monitor Access Policy
            kind: MonitorAccessPolicy
            name: monitoraccesspolicies.clusterpulse.io
            version: v1alpha1
          - description: Defines a connection to a container registry for monitoring
            displayName: Registry Connection
            kind: RegistryConnection
            name: registryconnections.clusterpulse.io
            version: v1alpha1
        description: |
          ## About ClusterPulse
          ClusterPulse is a comprehensive multi-cluster monitoring and management platform designed for OpenShift environments. It provides centralized visibility, policy-based access control, and real-time monitoring across multiple clusters through a unified dashboard. For more information, visit [https://github.com/ClusterPulse/operator](https://github.com/ClusterPulse/operator).

          ## Features and Capabilities
          - **Multi-Cluster Monitoring**: Connect and monitor multiple OpenShift/Kubernetes clusters from a single interface
          - **Policy-Based Access Control**: Fine-grained access policies using MonitorAccessPolicy CRDs
          - **Real-Time Metrics**: Live cluster health, node status, and resource utilization tracking
          - **Registry Monitoring**: Monitor connected container registries
          - **OAuth Integration**: Seamless OpenShift OAuth integration for authentication
          - **Redis-Backed Storage**: Scalable data storage using Redis for metrics and state
        displayName: ClusterPulse
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - clusterpulse
        links:
        - name: ClusterPulse
          url: https://github.com/ClusterPulse/operator
        maintainers:
        - email: asmoore@redhat.com
          name: Asa Moore
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: ClusterPulse
        relatedImages:
        - docker.io/clusterpulse/clusterpulse-operator:0.1.0
        - quay.io/community-operator-pipeline-prod/clusterpulse:0.1.0
        version: 0.1.0
      entries:
      - name: clusterpulse.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: clusterpulse
    provider:
      name: ClusterPulse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: geunwookim
      provider-url: ""
    name: ncn-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ncn-operator.v0.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-1"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-19T01:21:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: jwahn@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:61685ecb519d17160149a5dc16ae28d6fd96264652a6f19fef861addfa4ec6f4
        - quay.io/rasen708/ncn-operator:v0.2.2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.2.2
      entries:
      - name: ncn-operator.v0.2.2
        version: 0.2.2
      - name: ncn-operator.v0.2.1
        version: 0.2.1
      - name: ncn-operator.v0.2.0
        version: 0.2.0
      name: beta
    - currentCSV: ncn-operator.v0.9.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.gw.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
          createdAt: "2024-08-01T05:45:38Z"
          description: Provides the ncn extend
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.gw.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/geunwookim/ncn-operator
          support: geunwookim
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.gw.com
            version: v1
          - description: Label is the Schema for the labels API
            displayName: Node Label
            kind: Label
            name: labels.cmmd.gw.com
            version: v1
        description: This operator is for ncn extend
        displayName: ncn-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ncn
        - memory
        - extend
        links:
        - name: NCN Operator
          url: https://ncn-operator.domain
        maintainers:
        - email: geunwookim@gmail.com
          name: geunwookim
        maturity: betav1
        provider:
          name: geunwookim
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ncn-operator@sha256:a99a5ca46c5bd3065905588f4d0e3cdc6409d435abdc02793be88e77c1a8f0f5
        - quay.io/rasen708/ncn-operator@sha256:a09963745bd9c1419532d42d91237b7496c7a78e6a1c460a8e714ea1a0ce4d45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        version: 0.9.8
      entries:
      - name: ncn-operator.v0.9.8
        version: 0.9.8
      - name: ncn-operator.v0.9.2
        version: 0.9.2
      - name: ncn-operator.v0.8.1
        version: 0.8.1
      - name: ncn-operator.v0.5.6
        version: 0.5.6
      - name: ncn-operator.v0.4.3
        version: 0.4.3
      - name: ncn-operator.v0.4.2
        version: 0.4.2
      name: betav1
    defaultChannel: betav1
    packageName: ncn-operator
    provider:
      name: geunwookim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-engine.v2.7.5
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T22:32:21Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.7.5'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:2025e424c11c6fe4a4395945c2e7d124d07cde43bb8d8cbe70286fe6ae345754
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:4fa34ab0aa8854eb5e942f2ed315041bf2026ce0bed86f57125878af0c73dc8b
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:b124242a55b1b2a2258cb12094617a0e009cf1cf58a60ab1b42d48d4ea2514ac
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:e2e16ff6213c0274126c9d64d495c0abbe17379e841ab730de66520908f1a190
        - registry.redhat.io/multicluster-engine/cluster-api-rhel9@sha256:cd8d6e0056fb0436275c12cc6b42f11c0ccd06729c7108e964b03d6c48d063bd
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:90c19d2aaa50e7cc9184f2f220b568c34c9fdcfdcba314da555885c999078057
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:8c420720a48ea7c3aa096ae1929dcaa979c39e7865c5d43cba899e294745fc94
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:2cbd8c470dc4b263fd11952aa792094a7550f4b99e10f22186d51621bc8d3b3f
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:2e3fe5e3b39721cc77b7d32b5682300eb60b1bc6940b8e83662b458d88489140
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:3defab1abc7395c9f9c22912ec6825be11b120a1416404eb0599e91f7f9c679a
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:816712bb3ccb58b705a85e5f4a55e519316fd8b9d7d9295d1042420ce44a4bad
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:081151ca906b0f1afbcd3ee9c6d9e5bfe234494223c660ef376aa4004b0508f4
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:15c2873703afcc74e41ae27dbb4761ad19c4bd2594ca6de82d3b6c02752e0ce3
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:a6736c7cac00f78c0e8877549ae2b74b7207cc348f24f175133db49fd38f240a
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:2216b015943eb8d9f25bf4ac37fbad2b7215a5c0003bd7c3b3d97b3b0a5ba708
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:a24e9855a11d879d5308e29af6e3b2947365b7e466e7fa12dc57997140753604
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:4a0f1d3e696310f7a5bccbb358c538023a28f5bbc23b5743d10fd61b452d45d0
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:e6a5e71f0aee8a682179cfd8f85b45a2ee04a1de680fc03dabdb659a2d2c0396
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:b82b1837dfc24a367f68741dece6d19c89c576706f006ee8d5123a0c6782c492
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:fedc699c0f786fb303d3956769f9c5fca1473ec4bf014ec87c5a4d40e34dcdb3
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:9d1be5eb4fa357fe6bd97a52a53b3eeb537c431daf7af8a30efc3e0b6c5d5d66
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:555fcfce226e3b059f8f911d44a48578b898e085b6ad2575479feee71ec8baaf
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:7369bc79a283ff8b34c3189ea83e55f88a00d619cc78f8dad03d302dcb34a550
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:c0fb7985620d5bd0dfadd57cb75664ee6777b8dac5a3463967e9b649e9eb836c
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:0684bcde2c018ae893bd71be8df3448afa09d199415e83285d2c740cff66192d
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:8e41d489e1def9b540ddd6387a1cc151b24e91a5fa44a74fbbaa0d1978588d57
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:9e5c66ffdf39ae923ecccf79be538b883f2bdc8019169d55057565f077997d67
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:041044c08a1d526b349b5ad5fc4c4457f3632b1353530ed925c65ef2cce8e9c5
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:44b07a18116ad827c23ca5a98b9fe8ef74a99abe4787cd01d848de42df44178d
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:2a78ab192a554e78f3fc367c47e571f96b650ee9fb5ba0d06674f340f4320d94
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:45571dbb55bd213400eec6d2065b40615185bf3e52efdf495ab73201ec94faff
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:00d415071f4c10575f9e5068868a7f80b0c0fb9c8614da86bc515923bed136a3
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:1c176688f13052783662115d5cabe8d614b8d9f79507dc2bdd4be1604e2eecc0
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:b66f2f4b5e0d77d501f9cc9da483f976f3c0eaf0ca72a5e9c9435c92d58e1388
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:7d2ad8e800339c15efa5d5954106e01151d34135b25e9a23e30f0027ee23f193
        version: 2.7.5
      entries:
      - name: multicluster-engine.v2.7.5
        version: 2.7.5
      - name: multicluster-engine.v2.7.4
        version: 2.7.4
      - name: multicluster-engine.v2.7.3
        version: 2.7.3
      - name: multicluster-engine.v2.7.2
        version: 2.7.2
      - name: multicluster-engine.v2.7.1
        version: 2.7.1
      - name: multicluster-engine.v2.7.0
        version: 2.7.0
      name: stable-2.7
    - currentCSV: multicluster-engine.v2.8.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T13:08:30Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0 <2.8.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational
              components for managing multiple Kubernetes-based clusters. The deployment
              of the multicluster engine components is determined based on the configuration
              that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:aa2754d35c8460a2879d4d523ebf32351d36e8fd7886335b5c0d2ca4befb8e8e
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:fb2b8cef9916f387d26e65d9b53c7567ae1433d43d16e879ddb582ddc248eaa3
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:c5ac848e3b7eb88cdbdbf0b5f4aa96698d852b3a8118cf7146eb7e1948afaeba
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:3788b2674dda0b2cb4e9ca21bdce0b9e44702d72458b2396ef0acdad37fa866b
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:5ad96dad74c6f4967bd5fd397d6697e01b3b25bb3b8bae123a38e96964ddfa17
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:83a82f9558c406899e573308947c15b1f36de454006bb0e2ac44709f0c2c28d7
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:ee014c576263e19798df98460ebb5f1a21cb6249f09fe8a267b2dfa98b07d1eb
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:7ca7e87615e12b9d7fa75befd05f4f45bdc4bf8ef1b576c69f36a23c66c90cd6
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:ea2ba066abd170e1087558c7b701faeb1be2e1667943dc14440f19c2ee5b0c30
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:261b99b93d98825ade0a0d3b9c4e465a785f35d2a310dedd4bbd6e4aa1f33ef9
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:ec58463239eaa9bd15b31cbcfea0a34fd650201bac6120aa7ddaf72734451e06
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:8379827635dfc4b51e04145d8048f7dcf9c6956ea1a22371bd22f948ae01b183
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:ab0898d513b9b5ca2f5ca5879adc4e73da3232d515c0647a2a0ec28792264db1
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:0d1dfcb06eaa26de74994891371a5657aded7a59972f006f8c4a3859369877f3
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:f869555beec0875c74f30a88428ac9641029dfed0365d45b23d5cae469cf7d0f
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:f694b6e9e76941b11cc8a15153d9005cffe9bbbf58d0962f8ff546e90a63d93c
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:bee8f0c047cf8dbfa3f28712ce91bea55e926b2edb26c394f18a2d780e142083
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:ada8b480a86d0a9e3ecc58f8b1cb53dbd7452b2b9183b3ba2b15a6a58c4436d1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:7c86c89ff57091d4112783d82d1add47f2d56dba584418d5664c198ba048e0c2
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:5ea86c20383a6932c593e020fd3b7b1a2a8e666bc374bff8b4bca0ee2d4ecc33
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:806b6c956f7bad824f0fe9acc3b534964e8f5f3399a1813f75511bc0f3904414
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:e5e459e665afc14a1540e300015222b7c1ac896e2bee889283d67c3e19ca01ac
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:d4d11d0aeb3e1c9ff2a80e769d92dce509a0202a2ef1c9d2430030acdbf85bce
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:eaa77ae0f1a1f3d32e269e73eb29a73d11012d5e920af1011d1842bd904d7293
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:c3b62ab1234257d30f41fa1ab33cb319321777a18de8ed751be1fcd48b204f2e
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:b9afa4d5328f17c7c3ce1dc105d177ecacf3447092d51a50f250a4b0ea692e2c
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:18b0e86305afe2eae3c26e637755fa9e8ab2abf30bc15e5428467a9a81ec99b0
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:a7f63077799a62fe6ece76a12346a87b5cbae966ab69df6ec6d951259dbf0e4a
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:f708f3e1007ed2b0e420e51d8b7ff9dd9bccaea9234755f4375f5514018211b4
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:f809a8774b0e3f891af46931dd4d2c11b95da4e7f5a8aee1814152d9a468d357
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:988eb4aa01ba0d7e6728b23d043e4c0babe872ac23b6520f4daac714619a360b
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:5461643a61850d95409d989e81c304053372f1dc5dbba291f3f3386de9ecab1c
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:c6cd578123c1c86fcb3305ecca117543106f943cbcc4baf9065683988b2e10ff
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:cf457b5a13cdd54c45a7b88f0f150a04e756bfac727c4365cb12f113d904937d
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:0392c81a404be343d658dbc64153cbc46a64d0adf69aa8024c3f727351102d75
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:318ca0c7e9cd9a7336178519cadbf21eaeb474cd4a8eb5cf411a640d1c77372e
        version: 2.8.3
      entries:
      - name: multicluster-engine.v2.8.3
        version: 2.8.3
      - name: multicluster-engine.v2.8.2
        version: 2.8.2
      - name: multicluster-engine.v2.8.1
        version: 2.8.1
      - name: multicluster-engine.v2.8.0
        version: 2.8.0
      name: stable-2.8
    - currentCSV: multicluster-engine.v2.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "multicluster.openshift.io/v1", "kind": "MultiClusterEngine",
            "metadata": {"name": "multiclusterengine"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-30T19:22:43Z"
          description: Foundational components for central management of multiple
            OpenShift Container Platform and Kubernetes clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.8.0 <2.9.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"multicluster.openshift.io/v1",
            "kind":"MultiClusterEngine","metadata":{"name":"engine"},"spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-engine
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/open-cluster-management/backplane-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MultiClusterEngine defines the configuration for an instance
              of a multicluster engine, a central point providing the foundational components for managing multiple Kubernetes-based clusters. The deployment of the multicluster engine components is
              determined based on the configuration that is defined in this resource.
            displayName: MultiCluster Engine
            kind: MultiClusterEngine
            name: multiclusterengines.multicluster.openshift.io
            version: v1
        description: "Multicluster engine for Kubernetes provides the foundational
          components that are necessary for the centralized management of multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the engine to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can use the APIs that are provided by
          the engine to distribute configuration based on placement policy. Placement
          policy is a significant part of creating sophisticated multicluster management
          applications because you can select the applicable clusters.\n## Installing\nInstall
          the multicluster engine for Kubernetes operator by following instructions
          that display when you click the `Install` button. After installing the operator,
          create an instance of the `MultiClusterEngine` resource to install the engine
          components that provide the management APIs. Note: If you are using the
          engine to manage non-OpenShift Container Platform 4.x clusters, you need
          to create a Kubernetes `Secret` resource that contains your OpenShift Container
          Platform pull secret. Specify this `Secret` in the `MultiClusterEngine`
          resource, as described in the install documentation.\nYou can find additional
          installation guidance in the [documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/cluster_mce_overview#mce-install-intro). "
        displayName: multicluster engine for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mce
        - multicluster-engine
        - multiclusterengine
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/clusters/index
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/multicluster-engine/cluster-proxy-addon-rhel9@sha256:8d189b565e4c3c434988aa4ff0ecad4c5f86b2a94120325c142ff111b25facde
        - registry.redhat.io/multicluster-engine/provider-credential-controller-rhel9@sha256:cd9dc89b933b394b08fe05ab5f0bbecea727bfb2c29770f11ec1db4aa5740b3c
        - registry.redhat.io/multicluster-engine/hive-rhel9@sha256:bf343592bf2f4399b5ac61709f56528435990e903812e264cbef845ff4433231
        - registry.redhat.io/multicluster-engine/mce-capi-webhook-config-rhel9@sha256:7f1f70813e6d753b4acb04c7e9d7acb0047ad91f8a9a10e0fe2fff6c4411c8b4
        - registry.redhat.io/multicluster-engine/placement-rhel9@sha256:ad382d5eb347c448c46495a0a17503d789db4eb1096ad71a588b316b235e437f
        - registry.redhat.io/multicluster-engine/cluster-curator-controller-rhel9@sha256:902c5b3b4a4e8cb0a14fa5aba8451ec02830d5167ff1b91323ded62255983cd9
        - registry.redhat.io/multicluster-engine/work-rhel9@sha256:edbb174cb69dd31339a53542d24c4fd00552e41fa09d7e2553baa7a1ad8dd810
        - registry.redhat.io/multicluster-engine/clusterclaims-controller-rhel9@sha256:25ec9a0cc4ac92d4bed75c38fa37ef8d5ff9478cc475049b303d74eb58687a81
        - registry.redhat.io/multicluster-engine/assisted-image-service-rhel9@sha256:6b95560a6a6dc1ddd41c76f275f83baec52ec9d79de04bfe54feaed02db123a3
        - registry.redhat.io/multicluster-engine/must-gather-rhel9@sha256:a41cafc65e30deb073e051542702e258901b552ad00e42eacab05d149d6f0565
        - registry.redhat.io/multicluster-engine/addon-manager-rhel9@sha256:ee75e805f0f9d73f073ce870f5d46962c0033d63158017841d17410febfc2c57
        - registry.redhat.io/multicluster-engine/cluster-proxy-rhel9@sha256:36a6000b6482966f8bbf86e45c25335c066736ae1f8695a8b68e6b76b7c491ac
        - registry.redhat.io/openshift4/ose-cluster-api-rhel9@sha256:9a72b8081a57846d942981ced07da86ad8e197309a1bc3c9d53876528b58f8ec
        - registry.redhat.io/multicluster-engine/backplane-rhel9-operator@sha256:37087ad8678f3e8ec8e2b2d035d3a12087ce9e0c993194a04a9874fd07e80d91
        - registry.redhat.io/multicluster-engine/kube-rbac-proxy-mce-rhel9@sha256:1b218f0aa0ea1f494094e3bc783ef40968cbff780e22de186d578a1d64c33bf8
        - registry.redhat.io/multicluster-engine/cluster-image-set-controller-rhel9@sha256:72a4ac73bb5e7797329e1856f785751d48815a64e82350a5d694bb5a8e121956
        - registry.redhat.io/multicluster-engine/assisted-service-8-rhel8@sha256:3c450cb6a3ba6f2e3028856f14967803c1a691d95b8f54bd9d29d93691d6e795
        - registry.redhat.io/openshift4/ose-baremetal-cluster-api-controllers-rhel9@sha256:f855d5ce17d3ac4560afc73c35d03f045f84551fdccc00312c890288a889b9ce
        - registry.redhat.io/multicluster-engine/cluster-api-provider-agent-rhel9@sha256:f3d65f68538e94260fad54f145433a36366f8c88f9c9f80b7fc5545ba62e1773
        - registry.redhat.io/multicluster-engine/image-based-install-rhel9@sha256:9555fd5f577f32a1d8ebe25f0d03e453c8ea925e4a12dbbe1015c58d3c8528b0
        - registry.redhat.io/multicluster-engine/capoa-control-plane-rhel9@sha256:b510a25a6370ca6b8d2f74165814ebfdae0ce9898cc343e73fab6344eebf2bed
        - registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:1cc6fddaf9e55637e90d5c3e598467df8967b8a362f8b8d46286b256bad6dd3f
        - registry.redhat.io/multicluster-engine/managed-serviceaccount-rhel9@sha256:ea57247cf11f6ab966893ce4e6d2948232a4314588af2ca33c1d9689da7482dd
        - registry.redhat.io/multicluster-engine/assisted-service-9-rhel9@sha256:57699ecb529cb46843de847deef3e3757ca252910249d8deac16fcb00c01b636
        - registry.redhat.io/multicluster-engine/hypershift-addon-rhel9-operator@sha256:8901e984b3b47f0650cddc1c36bdd89af22390f3d4a562d562fde7774d8ba6db
        - registry.redhat.io/multicluster-engine/registration-operator-rhel9@sha256:d068d73f41d8f8ecbec36752d0a432913bfbe2580f489abefe513c06233e80c9
        - registry.redhat.io/multicluster-engine/hypershift-cli-rhel9@sha256:5e6594464bf7ddc2a66f127f645d93ad35d8cd95dd70544fa7fcec791f67f45c
        - registry.redhat.io/openshift4/ose-aws-cluster-api-controllers-rhel9@sha256:cf13b0658271cd0dbd1ce6dcc350e64c5459ef02ba4dc88a0031de9d6f26ffd1
        - registry.redhat.io/multicluster-engine/assisted-installer-controller-rhel9@sha256:ac43feaf079fa0d250d53cd9685fb68677d888dd091730faa32757dcdf2c95e4
        - registry.redhat.io/multicluster-engine/cluster-api-provider-kubevirt-rhel9@sha256:8076db8ccd5578cb5aba2d66f53790563d3f42b93a6cfd14c7094d3a25707982
        - registry.redhat.io/multicluster-engine/hypershift-rhel9-operator@sha256:f8c3898e29f8c0a20c3be92bc4ee5a9443b9fc8218db95ba541fe3e57a89c40d
        - registry.redhat.io/multicluster-engine/assisted-installer-rhel9@sha256:0a9c0ccb6a75273474fb04b785aa8fc58aec2f9ffbff8c232d359e14892bbae2
        - registry.redhat.io/multicluster-engine/registration-rhel9@sha256:6a75df76fb0d9516b13fa6b079bc6c4d49dc6774f0d4f31343abfb8d70edcc23
        - registry.redhat.io/multicluster-engine/console-mce-rhel9@sha256:cecec1f7d7f0d37cf17fb25ccf34602728f2a09c61af7e212967b4ee28d2dec3
        - registry.redhat.io/multicluster-engine/managedcluster-import-controller-rhel9@sha256:1bbd134d44839258172e013fe84bfd174d36d6b7232e2cc8a3d8556998181d9d
        - registry.redhat.io/multicluster-engine/clusterlifecycle-state-metrics-rhel9@sha256:ce0fd7bdc73cd2c84d1a9c5af528916488ca8116dd05d83dacf72f2af4248e65
        - registry.redhat.io/multicluster-engine/discovery-rhel9@sha256:a91276fcf5a589ef11f01b39df7d33ceb58ebcb47f95dfab772114bce7c1b4a8
        - registry.redhat.io/multicluster-engine/multicloud-manager-rhel9@sha256:8db1386d12d40dfbc80054c023897e5bfc0a987f6d07792b0f9f50f59f1a68d8
        - registry.redhat.io/multicluster-engine/capoa-bootstrap-rhel9@sha256:a8b1278a2df9b17cee66813715995385bb3539af05e4df9b5bf7d9ccd508c313
        - registry.redhat.io/multicluster-engine/mce-operator-bundle@sha256:4d9d003842e3ffc386f4e07ce7a81cfa85e676031fb6fdc96aefa9654244974a
        - registry.redhat.io/rhel8/postgresql-12@sha256:4773a29755ea826c0d3b58a8b8297c1d7388019516195cae9bd815f28ec4bca0
        version: 2.9.0
      entries:
      - name: multicluster-engine.v2.9.0
        version: 2.9.0
      name: stable-2.9
    defaultChannel: stable-2.9
    packageName: multicluster-engine
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: datagrid
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datagrid-operator.v8.2.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "adminAuth": {
                    "secretName": "basic-auth"
                  },
                  "clusterName": "example-infinispan",
                  "name": "mycache"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC batch-cache"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
          createdAt: "2021-08-24T17:39:44Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Restore API schema.
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Interoperability:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Data Grid Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request `0.25` (limit `0.50`) CPUs, 512MiB of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
          Red Hat Data Grid is a distributed, in-memory data store that increases application performance and delivers open-source capabilities to handle demanding use cases.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2/html/data_grid_operator_8.2_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.2
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:1435aa60e8ab4d25c86daae68dcb93d14999f50d7c1e11a3c5726a86d3d30453
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4d8057709958ef60bcabdcec52f05eae8df121fd3f4762c3218abca38130f638
        - registry.redhat.io/datagrid/datagrid-8-openj9-11-rhel8@sha256:c8607d6538a7528e27e6ecb138e0a93a03b3c9fa1f8a9a5aad3fd9f4a6814448
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:a7641fc5f4c7aed4b73d7d98ccb58ef64cf14ab16d72378f56dba3ad4b8714c6
        version: 8.2.8
      entries:
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.2.x
    - currentCSV: datagrid-operator.v8.3.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
          createdAt: "2022-11-03T08:40:57Z"
          description: Create and manage Red Hat Data Grid clusters.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3/html/data_grid_operator_8.3_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.3
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:4d9d7221b029866837add5f3cb52dd0b7956f57996ce2d8b76344ce2de678716
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:7dffd170b9b7d021b1615e6d746cbdfcbd819a1306ecdbaf042bcaf2202f7daa
        version: 8.3.9
      entries:
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.3.x
    - currentCSV: datagrid-operator.v8.4.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create cache --template=org.infinispan.DIST_SYNC mycache\nput --cache=mycache hello world\nput --cache=mycache hola mundo\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "templateName": "org.infinispan.DIST_SYNC"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
          createdAt: 2024-05-30T08:05:19 EDT
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4/html/data_grid_operator_8.4_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.4
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:fef9789befa5796a9bec23023c9dfda27810846dc4406b126e0c344a656611a2
        - registry.redhat.io/datagrid/datagrid-8-rhel8-operator@sha256:e5315682869de72514e3e027bfb02bf0bb64919453ca738c7ecfa1e34634ac2b
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:111d1f4ab8f006386102d1bec4cd6a41b446d9ae401a43c821b2f318c8ba74c8
        version: 8.4.16
      entries:
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: 8.4.x
    - currentCSV: datagrid-operator.v8.5.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infinispan.org/v1",
                "kind": "Infinispan",
                "metadata": {
                  "name": "example-infinispan"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "example-backup"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Batch",
                "metadata": {
                  "name": "example-batch-inline"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "config": "create counter --concurrency-level=1 --initial-value=5 --storage=VOLATILE --type=weak batch-counter\n"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "example-cache"
                },
                "spec": {
                  "clusterName": "example-infinispan",
                  "name": "mycache",
                  "template": "\u003cdistributed-cache /\u003e"
                }
              },
              {
                "apiVersion": "infinispan.org/v2alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "example-restore"
                },
                "spec": {
                  "cluster": "example-infinispan",
                  "container": {
                    "cpu": "1000m",
                    "extraJvmOpts": "-Djava.property=me",
                    "memory": "1Gi"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
          createdAt: 2025-08-29T05:44:37 UTC
          description: Create and manage Red Hat Data Grid clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Data Grid", "Red Hat
            Runtimes", "Red Hat Integration", "Red Hat Process Automation"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/infinispan/infinispan-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is the Schema for the backups API
            displayName: Backup
            kind: Backup
            name: backups.infinispan.org
            version: v2alpha1
          - description: Batch is the Schema for the batches API
            displayName: Batch
            kind: Batch
            name: batches.infinispan.org
            version: v2alpha1
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.infinispan.org
            version: v2alpha1
          - description: Infinispan is the Schema for the infinispans API
            displayName: Infinispan Cluster
            kind: Infinispan
            name: infinispans.infinispan.org
            version: v1
          - description: Restore is the Schema for the restores API
            displayName: Restore
            kind: Restore
            name: restores.infinispan.org
            version: v2alpha1
        description: |
          Red Hat Data Grid is a distributed, in-memory data store built from tried and trusted open-source technology.

          ### Data Grid
          * **Schemaless structure:** Store objects in key/value pairs.
          * **Grid storage:** Distribute and replicate data across clusters.
          * **Elasticity:** Scale to meet workload demands without service disruption.
          * **Polyglot access:** Read, write, and query from multiple client languages with different endpoints.
          * **Continuous availability:** Create a fault-tolerant caching service that guarantees business continuity.

          ### Operator capabilities
          * Built-in intelligence to automate Data Grid cluster deployment.
          * Infinispan CR for service configuration.
          * Cross site configuration and management.
          * Deployment of Grafana and Prometheus resources.
          * Cache CR for fully configurable caches.
          * Batch CR for scripting bulk resource creation.
          * REST and Hot Rod endpoints available at port `11222`.
          * Default application user: `developer`. Infinispan Operator generates credentials in an authentication secret at startup.
          * Data Grid pods request 1Gi of memory and 1Gi of ReadWriteOnce persistent storage. Data Grid Operator lets you adjust resource allocation to suit your requirements.
        displayName: Data Grid
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - infinispan
        - key value
        - database
        - datagrid
        - open source
        links:
        - name: Release information
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5/html/data_grid_operator_8.5_release_notes/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_data_grid/8.5
        - name: Operator source code
          url: https://github.com/infinispan/infinispan-operator
        - name: Product page
          url: https://www.redhat.com/en/technologies/jboss-middleware/data-grid
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:24cf79efa89bbfa495665b1711a27531b7ab0813695a89b2c38c96c2e792eb24
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:0d368459c98c5788dd55d2add5b4b7d438b9a8c814ea8aee2222ea956699cbb6
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:532dca6facb584c1261c0480c499af9f857c70dc390e0d46e46afec2561468af
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:f59ae21773dfc207454a709ac5f7aad2be6a6a20ac88da26c379edeed54537b8
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:6e1766784b684561666134d41dca5331b80d50e63695fe7b1a1d6d4cdb73f0bd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:ae852e4679978725bb7d24c1bdfe25b40bd883130ed04a36d4fca49dbea024ef
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:ad4f91f984b4b898012963d52e6ebcd166ec03e76e26eba3191ecb87f51fb18d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:54bc26a0fddc160a4e0d1a813aec78abb3c9cb77db5039b74c23de759faa67f5
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9d9e466b7c367d0406406550c74bc790aa68f3f875ea4f652eb28f2e25a05458
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:4b9bf39d2ff14bcfbd4628a399093703d87a2b2b9c8e74c3d7cfdff4b5167fa4
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:424b30dccd773f69c079b35a42e8c3e339246c5d490e14e3fcfb8bd8144a9160
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:5656cd188d054551489d45be996266d03e3faa8814b8bfc0f360cf0dceb83cbe
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bcb040bf640691568a93a4c71c8072c37ffb76b4296ade83186f279862b7f197
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:985193caf9bc3698c4d85aeedd1cfea09bccda881636d67885387b2a83c65ebd
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:9223b4b5c1c866a91033dac40f30a4caae0f6946cdb64935a3fdfffdb9ede660
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:c25e3152d6a6b6d5cd86d311bf567e594b9948440785209cd8d7986076a5a10d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:d522a20ca6909bd464f85967d8b86fe42b83f773478a80785f9a85a0cb87dfed
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:48cf13e71953cea6f62354289f0d8ece7a8a481ee0bb2c76eb9352d438db9ca3
        - registry.redhat.io/datagrid/datagrid-8-rhel9-operator@sha256:d719d778b0848588a7e61d3c3fa689763f51a6533af4cf9e1c738c4ed8fac760
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:879f3dcc24ffe795472779f777d955666ba4041c17e717d5e165cd7f12d7955a
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:c15d4fd827f554291c14851f187947f4109c0e20257fa2f1e141b32669569358
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:4dbe1b7a1a36990c0a31d5a66f3cc4ca2954c9e1ec3b87431113a4111530277d
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:78a6bed6fcb6962e5689b92806789b72a66db21b7c607b392a573c98593f2c09
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:9bc0b730570ab55fbeca83075949f4caa96ed5c37637420a59cd4a7db01ea840
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:b56269cb849afca34e7f82cc8e01a44bb32af6126f24db28019a58e556d4cb3c
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:11bfe5c8109d26164cff428232b2001ab8df464f14a382661b936543336e2960
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:bf7760520f5bea151a531d538d525c4591c47166995aefb4d77fef138de11544
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:554b831a0803d55a4e3b7022011b5a5f6d2b4b65009065a9b28bc8e55fc3ad46
        - registry.redhat.io/datagrid/datagrid-8-rhel9@sha256:a711c50e489995fb60e5dec23ce286360743feda5066e4a0861d448aa4f2e903
        - registry.redhat.io/datagrid/datagrid-8-prod-operator-bundle@sha256:d7a4f9b8e5eb9bf6d11e00898eb168e86636f0633959eeb2b36b2d6e5e505f71
        version: 8.5.11
      entries:
      - name: datagrid-operator.v8.5.11
        version: 8.5.11
      - name: datagrid-operator.v8.5.10
        version: 8.5.10
      - name: datagrid-operator.v8.5.9
        version: 8.5.9
      - name: datagrid-operator.v8.5.8
        version: 8.5.8
      - name: datagrid-operator.v8.5.7
        version: 8.5.7
      - name: datagrid-operator.v8.5.6
        version: 8.5.6
      - name: datagrid-operator.v8.5.5
        version: 8.5.5
      - name: datagrid-operator.v8.5.4
        version: 8.5.4
      - name: datagrid-operator.v8.5.3
        version: 8.5.3
      - name: datagrid-operator.v8.5.2
        version: 8.5.2
      - name: datagrid-operator.v8.5.1
        version: 8.5.1
      - name: datagrid-operator.v8.5.0
        version: 8.5.0
      - name: datagrid-operator.v8.4.16
        version: 8.4.16
      - name: datagrid-operator.v8.4.15
        version: 8.4.15
      - name: datagrid-operator.v8.4.14
        version: 8.4.14
      - name: datagrid-operator.v8.4.13
        version: 8.4.13
      - name: datagrid-operator.v8.4.12
        version: 8.4.12
      - name: datagrid-operator.v8.4.11
        version: 8.4.11
      - name: datagrid-operator.v8.4.10
        version: 8.4.9
      - name: datagrid-operator.v8.4.8
        version: 8.4.8
      - name: datagrid-operator.v8.4.7
        version: 8.4.7
      - name: datagrid-operator.v8.4.6
        version: 8.4.6
      - name: datagrid-operator.v8.4.5
        version: 8.4.5
      - name: datagrid-operator.v8.4.4
        version: 8.4.4
      - name: datagrid-operator.v8.4.3
        version: 8.4.3
      - name: datagrid-operator.v8.4.2
        version: 8.4.2
      - name: datagrid-operator.v8.4.1
        version: 8.4.1
      - name: datagrid-operator.v8.4.0
        version: 8.4.0
      - name: datagrid-operator.v8.3.9
        version: 8.3.9
      - name: datagrid-operator.v8.3.8
        version: 8.3.8
      - name: datagrid-operator.v8.3.7
        version: 8.3.7
      - name: datagrid-operator.v8.3.6
        version: 8.3.6
      - name: datagrid-operator.v8.3.5
        version: 8.3.5
      - name: datagrid-operator.v8.3.4
        version: 8.3.4
      - name: datagrid-operator.v8.3.3
        version: 8.3.3
      - name: datagrid-operator.v8.3.2
        version: 8.3.2
      - name: datagrid-operator.v8.3.1
        version: 8.3.1
      - name: datagrid-operator.v8.3.0
        version: 8.3.0
      - name: datagrid-operator.v8.2.8
        version: 8.2.8
      - name: datagrid-operator.v8.2.7
        version: 8.2.7
      - name: datagrid-operator.v8.2.6
        version: 8.2.6
      - name: datagrid-operator.v8.2.5
        version: 8.2.5
      - name: datagrid-operator.v8.2.4
        version: 8.2.4
      name: stable
    defaultChannel: stable
    packageName: datagrid
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-05-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/joget-dx8-operator-rhmp/support?utm_source=openshift_console
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle-rhmp@sha256:1f3d005224eada1b71d04f21d07a7c4fde159725c7a44d967d996152c802bede
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        version: 0.0.10
      entries:
      - name: joget-dx8-openshift-operator.v0.0.10
        version: 0.0.10
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator-rhmp
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "psp": false,
                  "serviceAccount": "default",
                  "controller": {
                    "enabled": true,
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": ""
                      }
                    },
                    "pvc": {
                      "enabled": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "path": "/",
                          "annotations": {
                            "ingress.kubernetes.io/protocol": "https"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {
                        "ingress.kubernetes.io/protocol": "https"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {}
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer",
                    "priorityClassName": "",
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      }
                    ],
                    "resources": {}
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true
                    },
                    "svc": {
                      "type": "NodePort",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "path": "/",
                      "annotations": {},
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "affinity": {},
                    "tolerations": [],
                    "nodeSelector": {}
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "image": "registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2",
                      "schedule": "0 0 * * *",
                      "priorityClassName": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130",
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "tolerations": [],
                      "nodeSelector": {}
                    }
                  },
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "resources": {},
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/neuvector-certified-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.3.4 of the NeuVector Certified Operator deploys version 4.4.3 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Grant Service Account Access to the Privileged SCC

                   oc -n neuvector adm policy add-scc-to-user privileged -z default

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          **Add NeuVector license from NeuVector WebUI->setting**


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/scanner@sha256:42004e13b348bb43b58ae8538c983a70b237e3aca870758ac629fb4fc43e8130
        - registry.access.redhat.com/ubi8@sha256:228824aa581f3b31bf79411f8448b798291c667a37155bdea61cfa128b2833f2
        - registry.connect.redhat.com/neuvector/controller@sha256:acc943568520282061547b310c50080dc32924a531bbb543d959b281eb6664f4
        - registry.connect.redhat.com/neuvector/enforcer@sha256:be1d299ec3532d55444681a882e9fd1238949da6598987308e2800ed26291b7a
        - registry.connect.redhat.com/neuvector/manager@sha256:8245cd179d11fff279690c5a422f46b4ff252ed72b9b998f0c9b5453bc7d8fc0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle-marketplace@sha256:7864f9f1cf3ad1d551b698b2e803302d207704b448b15f6c0013cbd9e854593c
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:6155ac3c95aa6471c9857e0a6f5e5d19e6fc1d8cc7d014108d7d4a10d14f96bb
        version: 1.3.4
      entries:
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator-rhmp
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: vault-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-secrets-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultConnection",
                "metadata": {
                  "name": "vaultconnection-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "address": "http://vault.vault.svc.cluster.local:8200"
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultAuth",
                "metadata": {
                  "name": "vaultauth-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultConnectionRef": "vaultconnection-sample",
                  "method": "kubernetes",
                  "mount": "kubernetes",
                  "kubernetes": {
                    "role": "sample",
                    "serviceAccount": "default"
                  }
                }
              },
              {
                "apiVersion": "secrets.hashicorp.com/v1beta1",
                "kind": "VaultStaticSecret",
                "metadata": {
                  "name": "vaultstaticsecret-sample",
                  "namespace": "tenant-1"
                },
                "spec": {
                  "vaultAuthRef": "vaultauth-sample",
                  "mount": "kvv2",
                  "type": "kv-v2",
                  "path": "secret",
                  "refreshAfter": "5s",
                  "destination": {
                    "create": true,
                    "name": "secret1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
          createdAt: "2025-09-26T21:50:51Z"
          description: The Vault Secrets Operator (VSO) allows Pods to consume Vault
            secrets natively from Kubernetes Secrets.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hashicorp/vault-secrets-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CSISecrets
            name: csisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: HCPAuth is the Schema for the hcpauths API
            displayName: HCPAuth
            kind: HCPAuth
            name: hcpauths.secrets.hashicorp.com
            version: v1beta1
          - description: HCPVaultSecretsApp is the Schema for the hcpvaultsecretsapps
              API
            displayName: HCPVault Secrets App
            kind: HCPVaultSecretsApp
            name: hcpvaultsecretsapps.secrets.hashicorp.com
            version: v1beta1
          - description: SecretTransformation is the Schema for the secrettransformations
              API
            displayName: Secret Transformation
            kind: SecretTransformation
            name: secrettransformations.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuthGlobal is the Schema for the vaultauthglobals API
            displayName: Vault Auth Global
            kind: VaultAuthGlobal
            name: vaultauthglobals.secrets.hashicorp.com
            version: v1beta1
          - description: VaultAuth is the Schema for the vaultauths API
            displayName: Vault Auth
            kind: VaultAuth
            name: vaultauths.secrets.hashicorp.com
            version: v1beta1
          - description: VaultConnection is the Schema for the vaultconnections API
            displayName: Vault Connection
            kind: VaultConnection
            name: vaultconnections.secrets.hashicorp.com
            version: v1beta1
          - description: VaultDynamicSecret is the Schema for the vaultdynamicsecrets
              API
            displayName: Vault Dynamic Secret
            kind: VaultDynamicSecret
            name: vaultdynamicsecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultPKISecret is the Schema for the vaultpkisecrets API
            displayName: Vault PKISecret
            kind: VaultPKISecret
            name: vaultpkisecrets.secrets.hashicorp.com
            version: v1beta1
          - description: VaultStaticSecret is the Schema for the vaultstaticsecrets
              API
            displayName: Vault Static Secret
            kind: VaultStaticSecret
            name: vaultstaticsecrets.secrets.hashicorp.com
            version: v1beta1
        description: |-
          The Vault Secrets Operator (VSO) allows Pods to consume Vault secrets
          natively from Kubernetes Secrets.

          ## Overview

          The Vault Secrets Operator operates by watching for changes to its supported set of Custom Resource Definitions (CRD).
          Each CRD provides the specification required to allow the *Operator* to synchronize a Vault Secrets to a Kubernetes Secret.
          The *Operator* writes the *source* Vault secret data directly to the *destination* Kubernetes Secret, ensuring that any
          changes made to the *source* are replicated to the *destination* over its lifetime. In this way, an application only needs
          to have access to the *destination* secret in order to make use of the secret data contained within.

          See the developer docs for more info [here](https://developer.hashicorp.com/vault/docs/platform/k8s/vso), including
          [examples](https://developer.hashicorp.com/vault/docs/platform/k8s/vso/examples) and a
          [tutorial](https://developer.hashicorp.com/vault/tutorials/kubernetes/vault-secrets-operator)
          for getting started.
        displayName: Vault Secrets Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vault
        - security
        - secrets
        links:
        - name: Vault Secrets Operator
          url: https://github.com/hashicorp/vault-secrets-operator
        - name: Documentation
          url: https://developer.hashicorp.com/vault/docs/platform/k8s/vso
        maintainers:
        - email: support@hashicorp.com
          name: HashiCorp Support
        maturity: stable
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator-bundle@sha256:e284d0839795f763ca53402c2b050fa9ffedebccd0f3832fe7259368c3ad32f2
        - registry.connect.redhat.com/hashicorp/vault-secrets-operator@sha256:84ac7003c4cb90c35f1e296585bb58187a86d64f972133e2661839bdb532ddb8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:503d5ee0fbcd5bd3492785f4246e19cbdb92960be0d05f3c6f123a32fe0fe84d
        version: 1.0.1
      entries:
      - name: vault-secrets-operator.v1.0.1
        version: 1.0.1
      - name: vault-secrets-operator.v1.0.0
        version: 1.0.0
      - name: vault-secrets-operator.v0.10.0
        version: 0.10.0
      - name: vault-secrets-operator.v0.9.1
        version: 0.9.1
      - name: vault-secrets-operator.v0.9.0
        version: 0.9.0
      - name: vault-secrets-operator.v0.8.1
        version: 0.8.1
      - name: vault-secrets-operator.v0.8.0
        version: 0.8.0
      - name: vault-secrets-operator.v0.7.0
        version: 0.7.0
      - name: vault-secrets-operator.v0.6.0
        version: 0.6.0
      - name: vault-secrets-operator.v0.5.2
        version: 0.5.2
      - name: vault-secrets-operator.v0.5.1
        version: 0.5.1
      - name: vault-secrets-operator.v0.5.0
        version: 0.5.0
      - name: vault-secrets-operator.v0.4.3
        version: 0.4.3
      name: stable
    defaultChannel: stable
    packageName: vault-secrets-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "s3.services.k8s.aws/v1alpha1",
                "kind": "Bucket",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
          createdAt: "2025-09-19T19:23:45Z"
          description: AWS S3 controller is a service controller for managing S3 resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bucket represents the state of an AWS s3 Bucket resource.
            displayName: Bucket
            kind: Bucket
            name: buckets.s3.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Storage Service (S3) resources in AWS from within your Kubernetes cluster.

          **About Amazon S3**

          Amazon Simple Storage Service (Amazon S3) is an object storage service that offers industry-leading scalability, data availability, security, and performance. This means customers of all sizes and industries can use it to store and protect any amount of data for a range of use cases, such as data lakes, websites, mobile applications, backup and restore, archive, enterprise applications, IoT devices, and big data analytics. Amazon S3 provides easy-to-use management features so you can organize your data and configure finely-tuned access controls to meet your specific business, organizational, and compliance requirements. Amazon S3 is designed for 99.999999999% (11 9s) of durability, and stores data for millions of applications for companies all around the world.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Developer Resources
          url: https://aws.amazon.com/s3/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/s3-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-s3-controller@sha256:058a37a9a2fe6a42bbdddabb3fc7495450f1ebbd7f46194866c1491c40ce1a9f
        version: 1.1.1
      entries:
      - name: ack-s3-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3-controller.v1.0.35
        version: 1.0.35
      - name: ack-s3-controller.v1.0.33
        version: 1.0.33
      - name: ack-s3-controller.v1.0.32
        version: 1.0.32
      - name: ack-s3-controller.v1.0.31
        version: 1.0.31
      - name: ack-s3-controller.v1.0.30
        version: 1.0.30
      - name: ack-s3-controller.v1.0.29
        version: 1.0.29
      - name: ack-s3-controller.v1.0.28
        version: 1.0.28
      - name: ack-s3-controller.v1.0.27
        version: 1.0.27
      - name: ack-s3-controller.v1.0.26
        version: 1.0.26
      - name: ack-s3-controller.v1.0.25
        version: 1.0.25
      - name: ack-s3-controller.v1.0.24
        version: 1.0.24
      - name: ack-s3-controller.v1.0.23
        version: 1.0.23
      - name: ack-s3-controller.v1.0.21
        version: 1.0.21
      - name: ack-s3-controller.v1.0.20
        version: 1.0.20
      - name: ack-s3-controller.v1.0.19
        version: 1.0.19
      - name: ack-s3-controller.v1.0.18
        version: 1.0.18
      - name: ack-s3-controller.v1.0.17
        version: 1.0.17
      - name: ack-s3-controller.v1.0.16
        version: 1.0.16
      - name: ack-s3-controller.v1.0.15
        version: 1.0.15
      - name: ack-s3-controller.v1.0.14
        version: 1.0.14
      - name: ack-s3-controller.v1.0.13
        version: 1.0.13
      - name: ack-s3-controller.v1.0.12
        version: 1.0.12
      - name: ack-s3-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3-controller.v1.0.9
        version: 1.0.9
      - name: ack-s3-controller.v1.0.7
        version: 1.0.7
      - name: ack-s3-controller.v1.0.6
        version: 1.0.6
      - name: ack-s3-controller.v1.0.5
        version: 1.0.5
      - name: ack-s3-controller.v1.0.4
        version: 1.0.4
      - name: ack-s3-controller.v1.0.3
        version: 1.0.3
      - name: ack-s3-controller.v1.0.2
        version: 1.0.2
      - name: ack-s3-controller.v1.0.1
        version: 1.0.1
      - name: ack-s3-controller.v1.0.0
        version: 1.0.0
      - name: ack-s3-controller.v0.1.8
        version: 0.1.8
      - name: ack-s3-controller.v0.1.7
        version: 0.1.7
      - name: ack-s3-controller.v0.1.6
        version: 0.1.6
      - name: ack-s3-controller.v0.1.5
        version: 0.1.5
      - name: ack-s3-controller.v0.1.4
        version: 0.1.4
      - name: ack-s3-controller.v0.1.3
        version: 0.1.3
      - name: ack-s3-controller.v0.1.2
        version: 0.1.2
      - name: ack-s3-controller.v0.1.1
        version: 0.1.1
      - name: ack-s3-controller.v0.1.0
        version: 0.1.0
      - name: ack-s3-controller.v0.0.20
        version: 0.0.20
      - name: ack-s3-controller.v0.0.19
        version: 0.0.19
      - name: ack-s3-controller.v0.0.18
        version: 0.0.18
      - name: ack-s3-controller.v0.0.17
        version: 0.0.17
      - name: ack-s3-controller.v0.0.16
        version: 0.0.16
      - name: ack-s3-controller.v0.0.15
        version: 0.0.15
      - name: ack-s3-controller.v0.0.14
        version: 0.0.14
      - name: ack-s3-controller.v0.0.13
        version: 0.0.13
      - name: ack-s3-controller.v0.0.12
        version: 0.0.12
      - name: ack-s3-controller.v0.0.11
        version: 0.0.11
      - name: ack-s3-controller.v0.0.10
        version: 0.0.10
      - name: ack-s3-controller.v0.0.9
        version: 0.0.9
      - name: ack-s3-controller.v0.0.8
        version: 0.0.8
      - name: ack-s3-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ssm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ssm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
          createdAt: "2025-09-19T20:57:07Z"
          description: AWS SSM controller is a service controller for managing SSM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Document represents the state of an AWS ssm Document resource.
            displayName: Document
            kind: Document
            name: documents.ssm.services.k8s.aws
            version: v1alpha1
          - description: PatchBaseline represents the state of an AWS ssm PatchBaseline
              resource.
            displayName: PatchBaseline
            kind: PatchBaseline
            name: patchbaselines.ssm.services.k8s.aws
            version: v1alpha1
          - description: ResourceDataSync represents the state of an AWS ssm ResourceDataSync
              resource.
            displayName: ResourceDataSync
            kind: ResourceDataSync
            name: resourcedatasyncs.ssm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SSM resources in AWS from within your Kubernetes cluster.

          **About Amazon SSM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon SSM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ssm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SSM Developer Resources
          url: https://aws.amazon.com/SSM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ssm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ssm-controller@sha256:922fd6daa660d2250bad8a3c54733cceae845624a14651cb20e049e29bf90e1c
        - public.ecr.aws/aws-controllers-k8s/ssm-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-ssm-controller.v1.1.1
        version: 1.1.1
      - name: ack-ssm-controller.v1.1.0
        version: 1.1.0
      - name: ack-ssm-controller.v1.0.12
        version: 1.0.12
      - name: ack-ssm-controller.v1.0.10
        version: 1.0.10
      - name: ack-ssm-controller.v1.0.9
        version: 1.0.9
      - name: ack-ssm-controller.v1.0.8
        version: 1.0.8
      - name: ack-ssm-controller.v1.0.7
        version: 1.0.7
      - name: ack-ssm-controller.v1.0.6
        version: 1.0.6
      - name: ack-ssm-controller.v1.0.5
        version: 1.0.5
      - name: ack-ssm-controller.v1.0.4
        version: 1.0.4
      - name: ack-ssm-controller.v1.0.3
        version: 1.0.3
      - name: ack-ssm-controller.v1.0.2
        version: 1.0.2
      - name: ack-ssm-controller.v1.0.1
        version: 1.0.1
      - name: ack-ssm-controller.v1.0.0
        version: 1.0.0
      - name: ack-ssm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ssm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: devopstales
      provider-url: devopstales.github.io
    name: community-trivy-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-trivy-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "trivy-operator.devopstales.io/v1",
                "kind": "NamespaceScanner",
                "metadata": {
                  "name": "trivy-operator-main-config"
                },
                "spec": {
                  "crontab": "*/5 * * * *",
                  "namespace_selector": "trivy-scan",
                  "clusterWide": "False",
                  "policyreport": "False",
                  "registry": [
                    {
                      "name": "docker.io",
                      "user": "",
                      "password": ""
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: devopstales/trivy-operator:2.4.1
          createdAt: "2022-07-11T15:00:00Z"
          description: Trivy Operator for scheduled imagescans and an Admission Control.
          repository: https://github.com/devopstales/trivy-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Confifuration for trivy-operator.
            displayName: NamespaceScanner
            kind: NamespaceScanner
            name: namespace-scanners.trivy-operator.devopstales.io
            version: v1
          - description: |
              VulnerabilityReport summarizes vulnerabilities in application dependencies and operating system packages
              built into container images.
            displayName: VulnerabilityReport
            kind: VulnerabilityReport
            name: vulnerabilityreports.trivy-operator.devopstales.io
            version: v1
        description: |
          Trivy Operator for scheduled imagescans and an Admission Control.

          ### Add Github token to extend your Github API Rate Limit
          The Operator use Github to download Trivy database. If your Github API Rate Limit Exceeded add your Github token to the operator deployment as GITHUB_TOKEN variable.

          ```yaml
          ...
          env:
          - name: GITHUB_TOKEN
            value: "githubToken"
          ```

          ### Scheduled Image scans
          Default every 5 minutes execute a scan script. It will get image list from all namespaces with the label `trivy-scan=true`, and then scan this images with trivy, finally we will get metrics on `http://[pod-ip]:9115/metrics`

          You can chane the schedule of the image scanning by setting the `crontab` variable in the namespace-scanner object.


          ### Cache Scan results
          From the 2.4 version you can use an external redis to store scanning results and speed up the scanning process.

          ```bash
          ...
          env:
          - name: REDIS_ENABLED
            value: "True"
          - name: REDIS_BACKEND
            value: redis://redis-cache:6379
          ```

          ### Persist Trivy database cache
          To Persist Trivy cache you need to create a pvc and mount to the Deployment under `/home/trivy-operator/trivy-cache`

          ```yaml
          ...
            volumeMounts:
            - name: cache
              mountPath: "/home/trivy-operator/trivy-cache"
          volumes:
          - name: cache
            persistentVolumeClaim:
              claimName: trivy-cache
          ```

          ### Trivy Image Validator
          The admission controller function can be configured as a ValidatingWebhook in a k8s cluster. Kubernetes will send requests to the admission server when a Pod creation is initiated. The admission controller checks the image using trivy if it is in a namespace with the lable `trivy-operator-validation=true`.

          To enable the admission controller you need to set the `clusterWide` variable to `true` in the namespace-scanner object.

          Enable the scheduled scan and the admission controller for a specific namespace:
          ```bash
          kubectl label namespaces test-apps trivy-scan=true
          # or
          kubectl label namespaces test-apps trivy-operator-validation=true
          ```

          You can define policy to the Admission Controller, by adding annotation to the pod trough the deployment:

          ```bash
          spec:
            ...
            template:
              metadata:
                annotations:
                  trivy.security.devopstales.io/medium: "5"
                  trivy.security.devopstales.io/low: "10"
                  trivy.security.devopstales.io/critical: "2"
          ...
          ```

          ### Monitorng
          Trivy operator has a prometheus endpoint wher yo can get the resoults of the namespace scans and the adission controller.

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep so_vulnerabilities

          # HELP so_vulnerabilities Container vulnerabilities
          # TYPE so_vulnerabilities gauge
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="MEDIUM"} 93
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="HIGH"} 76
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:1.18",severity="CRITICAL"} 25
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="UNKNOWN"} 0
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="LOW"} 23
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="MEDIUM"} 88
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="HIGH"} 60
          so_vulnerabilities{exported_namespace="trivytest",image="docker.io/library/nginx:latest",severity="CRITICAL"} 8
          ```

          ```bash
          curl -s http://10.43.179.39:9115/metrics | grep ac_vulnerabilities

          # HELP ac_vulnerabilities Admission Controller vulnerabilities
          # TYPE ac_vulnerabilities gauge
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="UNKNOWN"} 0.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="LOW"} 83.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="MEDIUM"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="HIGH"} 6.0
          ac_vulnerabilities{exported_namespace="trivytest",image="nginxinc/nginx-unprivileged:latest",severity="CRITICAL"} 4.0
          ```

          ### Policy Report reation
          The PolicyReport object is a protopype object probosed by the Kubernetes policy work group. The Policy Report Custom Resource Definition (CRD) can be used as a common way to provide policy results to Kubernetes cluster administrators and users, using native tools.

          Add the PolicyReport CRDs to your cluster (v1alpha2):
          ```bash
          kubectl create -f https://github.com/kubernetes-sigs/wg-policy-prototypes/raw/master/policy-report/crd/v1alpha2/wgpolicyk8s.io_policyreports.yaml
          ```

          You can enable the Policy Report creatin by setting the `policyreport` to `True` in the NamespaceScanner object.

          This objects can be visualized by the Policy Reporter UI. The Policy Reporter UI is a monitoring and Observability Tool for the PolicyReport CRD with an optional UI. It is created by Kyverno. The main goal was a tool to visualize the resoluts of the Kyverno policies, but because it uses the PolicyReports CRD it can visualize the resoults of the trivy-operator scans.
        displayName: Community Trivy Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trivy
        - security
        links:
        - name: Documentation
          url: https://devopstales.github.io/trivy-operator
        - name: Blog
          url: https://devopstales.github.io
        maintainers:
        - email: devopstales@protonmail.com
          name: devopstales
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: devopstales
          url: devopstales.github.io
        relatedImages:
        - devopstales/trivy-operator:2.4.1
        - quay.io/openshift-community-operators/community-trivy-operator@sha256:e3652e68e1e569d7f40432fd8cbb7bfdc6738a116ecc1011c71a36f7ca5663c6
        version: 2.4.0
      entries:
      - name: community-trivy-operator.v2.4.0
        version: 2.4.0
      - name: community-trivy-operator.v2.3.0
        version: 2.3.0
      - name: community-trivy-operator.v2.2.0
        version: 2.2.0
      - name: community-trivy-operator.v2.1.1
        version: 2.1.1
      name: stable
    defaultChannel: stable
    packageName: community-trivy-operator
    provider:
      name: devopstales
      url: devopstales.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: kubernetes-nmstate-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-nmstate-operator.4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nmstate.io/v1",
                "kind": "NMState",
                "metadata": {
                  "name": "nmstate"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
          createdAt: "2022-02-21 08:46:16"
          description: |
            Kubernetes NMState is a declaritive means of configuring NetworkManager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-nmstate
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/kubernetes-nmstate
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an NMState deployment.
            displayName: NMState
            kind: NMState
            name: nmstates.nmstate.io
            version: v1
        description: A Kubernetes Operator to install Kubernetes NMState
        displayName: Kubernetes NMState Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - nmstate
        - networking
        - NetworkManager
        links:
        - name: Kubernetes Nmstate Operator
          url: https://github.com/nmstate/kubernetes-nmstate
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: GA
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:3b538de0bbaa43d2368402ea1c0eaf767c8b9e5a28e112a309148a630dd3c052
        - registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:5b61484c6a2bbc377435093bd9abb31df980ece470325cae92f3106963b6482c
        - registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:ced3644f560047b419c286c11bc74239a5873d7d7903461b848e722884730e97
        - registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:af8ccc7a075f11bf6f31310830704e3a27b23587094f56e518b1f264c92b0fe4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: kubernetes-nmstate-operator.4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: kubernetes-nmstate-operator.4.18.0-202508271223
        version: 4.18.0-202508271223
      - name: kubernetes-nmstate-operator.4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: kubernetes-nmstate-operator.4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: kubernetes-nmstate-operator.4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: kubernetes-nmstate-operator.4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: kubernetes-nmstate-operator.4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: kubernetes-nmstate-operator.4.18.0-202505280403
        version: 4.18.0-202505280403
      - name: kubernetes-nmstate-operator.4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: kubernetes-nmstate-operator.4.18.0-202505061007
        version: 4.18.0-202505061007
      - name: kubernetes-nmstate-operator.4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: kubernetes-nmstate-operator.4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: kubernetes-nmstate-operator.4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: kubernetes-nmstate-operator.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: kubernetes-nmstate-operator.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: kubernetes-nmstate-operator.4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: kubernetes-nmstate-operator.4.18.0-202503190933
        version: 4.18.0-202503190933
      - name: kubernetes-nmstate-operator.4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: kubernetes-nmstate-operator.4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: kubernetes-nmstate-operator.4.18.0-202502250802
        version: 4.18.0-202502250802
      - name: kubernetes-nmstate-operator.4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: kubernetes-nmstate-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-observability-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
          createdAt: "2025-06-24T11:30:26Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=..0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cluster-observability-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs",
              "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs",
              "prometheusagents.monitoring.rhobs",
              "perses.perses.dev"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
          support: Cluster Observability (https://issues.redhat.com/projects/COO/)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: Perses is the Schema for the perses API
            displayName: Perses
            kind: Perses
            name: perses.perses.dev
            version: v1alpha1
          - description: A Perses Dashboard
            displayName: Perses Dashboard
            kind: PersesDashboard
            name: persesdashboards.perses.dev
            version: v1alpha1
          - description: A Perses Datasource
            displayName: Perses Datasource
            kind: PersesDatasource
            name: persesdatasources.perses.dev
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
          - description: UIPlugin defines a console plugin for observability
            displayName: UIPlugin
            kind: UIPlugin
            name: uiplugins.observability.openshift.io
            version: v1alpha1
        description: |-
          Cluster Observability Operator is a Go based Kubernetes operator to easily setup and manage various observability tools.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          - Setup console plugins
          - Setup korrel8r
          - Setup Perses
          - Setup Cluster Health Analyzer
          ### Documentation
          - **[Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/cluster_observability_operator/index)**
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Cluster Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e
        - registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf
        - registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0
        - registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8
        - registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d
        - registry.redhat.io/cluster-observability-operator/cluster-observability-operator-bundle@sha256:8c27ac0dfae9e507601dc0a33ea19c8f757e744350a41f41b39c1cb8d60867b2
        - registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552
        - registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3
        - registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed
        - registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5
        - registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd
        - registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c
        - registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1
        - registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223
        - registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404
        version: 1.2.2
      entries:
      - name: cluster-observability-operator.v1.2.2
        version: 1.2.2
      - name: cluster-observability-operator.v1.2.1
        version: 1.2.1
      - name: cluster-observability-operator.v1.2.0
        version: 1.2.0
      - name: cluster-observability-operator.v1.1.1
        version: 1.1.1
      - name: cluster-observability-operator.v1.1.0
        version: 1.1.0
      - name: cluster-observability-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: cluster-observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-diagnostics
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-diagnostics.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Diagnostics",
                "metadata": {
                  "name": "diagnostics-sample"
                },
                "spec": {
                  "agentDiagnosticsService": {
                    "port": 8080
                  },
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "wapdbDiagnosticsService": {
                    "dumpedTables": "[ \"--regexpr cluster_logs_week_*\", \"current_week_cluster_logs\", \"config_metadata_dbs\", \"config_metadata_dbs_id_seq\", \"config_metadata_tags\", \"config_tags\", \"csv_reports_meta\", \"custom_import_handler_dbs\", \"data_point_db_models\", \"db_test_entries\", \"license_server_metadata\", \"operations\", \"result_file_metadata_dbs\", \"result_metadata_dbs\", \"result_metadata_tags\", \"result_tags\", \"session_meta\", \"session_types\", \"sessions\", \"stat_view_query_data\", \"test_run_configs\", \"test_run_infos\", \"test_runs\", \"wireless_globalsettings\", \"wireless_sessionconfigs\", \"wireless_templates\"]",
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:37:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Diagnostics
            name: diagnostics.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Diagnostics Operator
        displayName: Keysight WAP Diagnostics Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Diagnostics
          url: https://keysight-wap-diagnostics.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-diagnostics@sha256:97bf8affa7bf1f3a391f5f036313e7cd8196be2eb8e95debeab06c1b6c964c5e
        - registry.connect.redhat.com/loadcore/keysight-wap-diagnostics-bundle@sha256:4e8400a38a04c61dc20ece1ee7dac1ad8a56ac309621f3225e7389e84e20c870
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-diagnostics.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-diagnostics
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator@sha256:d68db92bcb217d94dc0ad0a1cea625a0a9c9299d886d325a903163f9f887dcd5
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator-bundle@sha256:85e75ec797f70c5807ec1ea25208cf0bb1990daa05eca9fbc82bc38aa4a840ab
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Splunk Inc.
      provider-url: www.splunk.com
    name: splunk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: splunk-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "IndexerCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "LicenseMaster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "licensemaster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "MonitoringConsole",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v3",
                "kind": "Standalone",
                "metadata": {
                  "finalizers": [
                    "enterprise.splunk.com/delete-pvc"
                  ],
                  "name": "standalone-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "clustermanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "IndexerCluster",
                "metadata": {
                  "name": "indexercluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "LicenseManager",
                "metadata": {
                  "name": "licensemanager-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "MonitoringConsole",
                "metadata": {
                  "name": "monitoringconsole-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "SearchHeadCluster",
                "metadata": {
                  "name": "searchheadcluster-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "enterprise.splunk.com/v4",
                "kind": "Standalone",
                "metadata": {
                  "name": "standalone-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Big Data, Logging & Tracing, Monitoring, Security, AI/Machine
            Learning
          containerImage: splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
          createdAt: "2025-07-16T09:03:58Z"
          description: The Splunk Operator for Kubernetes enables you to quickly and
            easily deploy Splunk Enterprise on your choice of private or public cloud
            provider. The Operator simplifies scaling and management of Splunk Enterprise
            by automating administrative workflows using Kubernetes best practices.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=2.4.0 <2.8.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/splunk/splunk-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager is the Schema for the cluster manager API
            displayName: Cluster Manager
            kind: ClusterManager
            name: clustermanagers.enterprise.splunk.com
            version: v4
          - description: ClusterMaster is the Schema for the cluster manager API
            displayName: Cluster Master
            kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v3
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v1
          - kind: ClusterMaster
            name: clustermasters.enterprise.splunk.com
            version: v2
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v3
          - description: IndexerCluster is the Schema for a Splunk Enterprise indexer
              cluster
            displayName: Indexer Cluster
            kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v4
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v1
          - kind: IndexerCluster
            name: indexerclusters.enterprise.splunk.com
            version: v2
          - description: LicenseManager is the Schema for a Splunk Enterprise license
              manager.
            displayName: License Manager
            kind: LicenseManager
            name: licensemanagers.enterprise.splunk.com
            version: v4
          - description: LicenseMaster is the Schema for a Splunk Enterprise license
              master.
            displayName: License Master
            kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v3
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v1
          - kind: LicenseMaster
            name: licensemasters.enterprise.splunk.com
            version: v2
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v3
          - description: MonitoringConsole is the Schema for the monitoringconsole
              API
            displayName: Monitoring Console
            kind: MonitoringConsole
            name: monitoringconsoles.enterprise.splunk.com
            version: v4
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v3
          - description: SearchHeadCluster is the Schema for a Splunk Enterprise search
              head cluster
            displayName: Search Head Cluster
            kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v4
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v1
          - kind: SearchHeadCluster
            name: searchheadclusters.enterprise.splunk.com
            version: v2
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v3
          - description: Standalone is the Schema for a Splunk Enterprise standalone
              instances.
            displayName: Standalone
            kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v4
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v1
          - kind: Standalone
            name: standalones.enterprise.splunk.com
            version: v2
        description: The Splunk Operator for Kubernetes enables you to quickly and
          easily deploy Splunk Enterprise on your choice of private or public cloud
          provider. The Operator simplifies scaling and management of Splunk Enterprise
          by automating administrative workflows using Kubernetes best practices.
        displayName: Splunk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Splunk
        - Enterprise
        - Operator
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://github.com/splunk/splunk-operator/tree/main/docs
        maintainers:
        - email: support@splunk.com
          name: Splunk Operator Maintainers
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Splunk Inc.
          url: www.splunk.com
        relatedImages:
        - splunk/splunk@sha256:99aec4a74ea4c9910b50cc035b4c0c85ed4711a4c9b456028690d0f0c7af2657
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - registry.connect.redhat.com/splunk/splunk-operator-bundle@sha256:433bec6334c4771e140a24ea8753e7bcbf358c4504c942658fd3364915ac86ac
        - splunk/splunk-operator@sha256:53bbd9b1a99c6909ef1976fb12aa780dfe6f86b898ba509d75bf1a63ca15bfe0
        version: 2.8.1
      entries:
      - name: splunk-operator.v2.8.1
        version: 2.8.1
      - name: splunk-operator.v2.8.0
        version: 2.8.0
      - name: splunk-operator.v2.7.1
        version: 2.7.1
      - name: splunk-operator.v2.7.0
        version: 2.7.0
      name: stable
    defaultChannel: stable
    packageName: splunk-operator
    provider:
      name: Splunk Inc.
      url: www.splunk.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-mq-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-mq-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mq.services.k8s.aws/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
          createdAt: "2025-09-19T19:24:03Z"
          description: AWS MQ controller is a service controller for managing MQ resources
            in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker represents the state of an AWS mq Broker resource.
            displayName: Broker
            kind: Broker
            name: brokers.mq.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MQ resources in AWS from within your Kubernetes cluster.

          **About Amazon MQ**

          Amazon MQ is a managed message broker service for Apache ActiveMQ and RabbitMQ that makes it easy to set up and operate message brokers on AWS. Amazon MQ reduces your operational responsibilities by managing the provisioning, setup, and maintenance of message brokers for you. Because Amazon MQ connects to your current applications with industry-standard APIs and protocols, you can easily migrate to AWS without having to rewrite code.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon MQ
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mq
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MQ Developer Resources
          url: https://docs.aws.amazon.com/amazon-mq/latest/developer-guide/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: mq maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/mq-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-mq-controller@sha256:702ffd931cb4baafb1eac49f9fda5588d8b02f3b8c696e477dd626b0efad9f54
        version: 1.1.1
      entries:
      - name: ack-mq-controller.v1.1.1
        version: 1.1.1
      - name: ack-mq-controller.v1.1.0
        version: 1.1.0
      - name: ack-mq-controller.v1.0.13
        version: 1.0.13
      - name: ack-mq-controller.v1.0.11
        version: 1.0.11
      - name: ack-mq-controller.v1.0.10
        version: 1.0.10
      - name: ack-mq-controller.v1.0.9
        version: 1.0.9
      - name: ack-mq-controller.v1.0.8
        version: 1.0.8
      - name: ack-mq-controller.v1.0.7
        version: 1.0.7
      - name: ack-mq-controller.v1.0.6
        version: 1.0.6
      - name: ack-mq-controller.v1.0.5
        version: 1.0.5
      - name: ack-mq-controller.v1.0.2
        version: 1.0.2
      - name: ack-mq-controller.v1.0.1
        version: 1.0.1
      - name: ack-mq-controller.v1.0.0
        version: 1.0.0
      - name: ack-mq-controller.v0.0.36
        version: 0.0.36
      - name: ack-mq-controller.v0.0.35
        version: 0.0.35
      - name: ack-mq-controller.v0.0.34
        version: 0.0.34
      - name: ack-mq-controller.v0.0.31
        version: 0.0.31
      - name: ack-mq-controller.v0.0.29
        version: 0.0.29
      - name: ack-mq-controller.v0.0.27
        version: 0.0.27
      - name: ack-mq-controller.v0.0.26
        version: 0.0.26
      - name: ack-mq-controller.v0.0.25
        version: 0.0.25
      - name: ack-mq-controller.v0.0.24
        version: 0.0.24
      - name: ack-mq-controller.v0.0.23
        version: 0.0.23
      - name: ack-mq-controller.v0.0.22
        version: 0.0.22
      - name: ack-mq-controller.v0.0.21
        version: 0.0.21
      - name: ack-mq-controller.v0.0.20
        version: 0.0.20
      - name: ack-mq-controller.v0.0.19
        version: 0.0.19
      - name: ack-mq-controller.v0.0.18
        version: 0.0.18
      - name: ack-mq-controller.v0.0.17
        version: 0.0.17
      - name: ack-mq-controller.v0.0.16
        version: 0.0.16
      - name: ack-mq-controller.v0.0.15
        version: 0.0.15
      - name: ack-mq-controller.v0.0.14
        version: 0.0.14
      - name: ack-mq-controller.v0.0.13
        version: 0.0.13
      - name: ack-mq-controller.v0.0.12
        version: 0.0.12
      - name: ack-mq-controller.v0.0.11
        version: 0.0.11
      - name: ack-mq-controller.v0.0.10
        version: 0.0.10
      - name: ack-mq-controller.v0.0.9
        version: 0.0.9
      - name: ack-mq-controller.v0.0.8
        version: 0.0.8
      - name: ack-mq-controller.v0.0.7
        version: 0.0.7
      name: alpha
    defaultChannel: alpha
    packageName: ack-mq-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-pipelines-operator-rh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: latest
    - currentCSV: openshift-pipelines-operator-rh.v1.14.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.14.6'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.56.4
          - Tekton Triggers: v0.26.2
          - Pipelines as Code: v0.24.6
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.16.0
          - Tekton Results (tech-preview): v0.9.2

          ## Getting Started
          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/container-platform/OPENSHIFT_DOCS_VERSION/cicd/pipelines/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)
          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:d9972bc64dfa7e260bd9b949bbb85ef5ae532e7c6df8d874d3d3963af1097726
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:ddf0cb64ea57de872515e19de63b2dec5f56738c9d7eaa283a4a489545ef4902
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:82c8a91d521d7e6d81fb462fa2b804f420f52ee062df5cda39545165a2f19e0c
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:f68dc18a4dc6a1886507097551f4a6fce39062d03c6eb86c067d317812b0d34a
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:aa6eb4c7ade9d6add1e4ab94197c2ce72a253fea0d7461698ca0e37854f1dcbe
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:ffa8c6d8fb5bacb59da27f4ebde87385359b3f97a6a023b4a4488ced2d46c574
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:63fc1351191a54d12d4e463a1bb4cc732620cba3000af86ba7f7a579514f5fe4
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:d01ba7ead670d5ab9c1355b90010657988e46b4d1aa0bb68b194eaf4001b2d7a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:ae0260abd2725c2f4bb25c0c07d86290978892a49050e0fd23f466401aba668c
        - registry.redhat.io/ubi8/ubi-minimal@sha256:6825946f10e8c81586fa7815c4f06016af05cfc231d921ec9699402a6655d201
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ce0ea093c522af329acdc85148c2473ea45cd44270c15f8dd32965b0d56f304f
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:db3ff043410d1dd0621fef4483761727f7241d788b136e446fb861953350a2cb
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:2e75096ffe20d1261e8a645fefd32e37295d727aa6a280f10041a44443917469
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8d4450d68ac03d8275e337d2c55692ef35e954400d0c9f041a943d163a8b7680
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:5e7f5a2813a591e7b13fea9fc4e2f599b5ec855a015cd5a4c287c4e178358756
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:6327385e3d3c9df1ea50e803955d980b4cd35fb15593fcfec0d477d91a45021f
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel8@sha256:6e7c7a1ca458e08ad85e1146815ec00a30016f709b45e28b464043e875c2a3e8
        - registry.redhat.io/rhel8/skopeo@sha256:e297b84ba6f1d5eb2c37e2c3b31c9a9a4c39f9590e7d7c32388165245bde4835
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:83913744a1bb7ff23091cfa541b807cfb82b38dde0c37e0f1dd9c17cade99775
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:7a63e7a762b9c7b15057ed9d5bcbea08acb4c8933a300c6ec336824d49a6e937
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:f9fed20cdfebbb226e035cc62f037174f5ecdb3139a5edc7e39f019f6d2b1e98
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:059c08316152c209de13017e322d6fe5eacfde81c3b0039e55b5de8f2f27183a
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:e23c901868229e4f6a4016aa7798f1aba879ff35ffe594f5b0e4d39f1bff9e52
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:b3bedee6084e9d9da549d3e637d2bf176aa365b5818e6b6377525577edd710dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:301c3b793d8068f9e50c855196da91b6c07111ef500b02a91fbb354c9fc1316b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:e2bd929246d62038a24a0685cb3e4873c131c9b7713d6a5268ca09a8614538be
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:bf0c38beefa24ca3d04ebfa6140a1cc5a9a56fba81ae05799279a559093b9033
        - registry.redhat.io/rhel8/postgresql-13@sha256:3903d562425dd564986123280dc076548fb3d24c74138dcd857652f53052e0a7
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:57fb86ba112d79a228cbacc96cd0525c123464bcac956354722e3f2aeff42508
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:9c29f6eff1f7019993446d85d226f04c5614a18351dfb4042a8bcb62276810de
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:71d91c7b80b6d196025f68749bd8518b8b0affd8d5064a7c5c70d125656fbc36
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:2261f4f657f70759f4d0c898e44ffafceb8dbe2144063d50227140463bdbf2c5
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:abf4442b6acc714c2392b86af046d5f32d247ee0c3763bb18c57d0473edfcf02
        version: 1.14.6
      entries:
      - name: openshift-pipelines-operator-rh.v1.14.6
        version: 1.14.6
      - name: openshift-pipelines-operator-rh.v1.14.5
        version: 1.14.5
      - name: openshift-pipelines-operator-rh.v1.14.4
        version: 1.14.4
      - name: openshift-pipelines-operator-rh.v1.14.3
        version: 1.14.3
      name: pipelines-1.14
    - currentCSV: openshift-pipelines-operator-rh.v1.15.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "db_user": "result",
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.15.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.59.6
          - Tekton Triggers: v0.27.0
          - Pipelines as Code: v0.27.2
          - Tekton Chains: v0.20.1
          - Tekton Hub (tech-preview): v1.17.1
          - Tekton Results (tech-preview): v0.10.0
          - Manual Approval Gate (tech-preview): v0.2.2

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.15/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:bf94c350923da706c6e0a427e2d91af3e13e751a12388a4ed75d1c34dd7b3177
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:8edd2f89d7adda6c1a5aae7414f50e0d786ec8443c1db13834ea4924009b6419
        - registry.redhat.io/rhel8/postgresql-13@sha256:a92a579f1aef66ac188d24fd489c456a1a3e311d95dcce652da6b81e28fbf725
        - registry.redhat.io/rhel8/buildah@sha256:aac6629389db17e99894c5bee0da01d4c8065d11d8c6f6e1602f9484290baa70
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:601fcac1abc6c8efa1248c5ca229213d6660708470a1c416d41c6aff0ba4691a
        - registry.redhat.io/ubi8/ubi-minimal@sha256:f729a7f5685ea823e87ffd68aff988f2b8ff8d52126ade4e6de7c68088f28ebd
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:7f00ff1b407d9ea107d3d969988a6279b28bde86943f490a007fa922dd40a1c3
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:eed464a3b4e23f0b2623aa1b923c10232d235f4272f562955926c1748e295d2b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:aca0024cac7a8c41b1ed151eb1e1188f09bc636866f4e03f1b973f90890b37cf
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:f04ea12675e0eb2354635cd2f9e99d6e59619d1c611e0ea95c3ce8222d615322
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:827313c2691fb8833e42418b65b4a3360a11cabc2c92a7d8e7b20183f06cbf68
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:1300d171dbf09c63071ba5586e223491a13acfc6dbab27eff04648aeb788d455
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:312d1f90765d05a86845856c677e1ece4c57e81a9bfe3fd1e76868d624e3d72d
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:322ef2385620661a7914f47c73a6d145ec0c9ce1ccde30645bb66148545ddc96
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:75d6228eb656b8778fb3bb8dc68c686ac1c9b787e5343ae3c2467f8e724af83e
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:7958925cecce1b4b02ac4649e3bfd7fa3b5361c16ab3593700b8da93fdf1c79a
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:538c63c8e7af3cba91526f2f9ac99fb8f6b0f4781db31771c959612c209dd090
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:166789a990df9f6ab04fd3df0a8cd301bc19b9bedde94215cb27306060eb1e2c
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:ac1a6e057a5599fda7e048ff77030e98c92b5d51cac2472c1ef3910964fc17e6
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:bf6cf2e87fb19f7aa9a490b83c16af69834c0721220a643710a1b077959e91ca
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:d1f86458079449c86407ac716c3cbf608cad06de29edeecbde6586f76628daa9
        - registry.redhat.io/ubi8/openjdk-17@sha256:e8cc2e476282b75d89c73057bfa713db22d72bdb2808d62d981a84c33beb2575
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:479e3eed93e0aa13e3489514fe00375a519eb45969d099c490b4acd056ea5adf
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:83eb08bb9579cc19132af282c455bca32a0fd4f5436a35e7ad23b5efa2faa3cf
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:f02af8475c33f6f38a2f90ff1be1fb1b2094860c4d5e300aefdd7eb84b4a529a
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:1db14f23077785108660dde38c98e2e30c8301f8e2e1c594dd4636e0ba29f84f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:61c22119f37e620a3807056cde35bc07c90867765b149294056326b1d11494f1
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:70bb0ebcfe5da45e5c239aa011dc5f1ef48844116d7346bb05433b503107268a
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:7c81ae208a16e17d0102074867add37dfb1207ee95cbbc6c732d88e61bbc59fb
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:e749db81c9d7e11e2802833a2a02e93afc357c83d1a2a827e537262df281a4f7
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:49c346bf7fc69a9796690236dc8597e35a874343b02fb74c5d2f1c05d356e41c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:abac279a0a5ecde57952b6bef634f8ba7527cd8c8956a2ea30e9449174ad52bd
        - registry.redhat.io/rhel8/skopeo@sha256:29465586e92ed04d2878dcdeda2d508c7e8e7539762e10296b52d631929960aa
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:acf8017029f4b2e7a1c3da9bc62c65239d313d24e3012712356a191bea9a2e6f
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:e7b000a2735bbf98ee79bd093f7b703ed09d7c1e11ed2a5b22c4b8ac8d440386
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7152c3c174d25f2a5239456ba677ac5315bfb82bc3b3ac5ee0eb8f9f74fb6632
        version: 1.15.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.15.3
        version: 1.15.3
      - name: openshift-pipelines-operator-rh.v1.15.2
        version: 1.15.2
      - name: openshift-pipelines-operator-rh.v1.15.1
        version: 1.15.1
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.15
    - currentCSV: openshift-pipelines-operator-rh.v1.16.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "clusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "communityClusterTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.16.4'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              ClusterTasks, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.62.9
          - Tekton Triggers: v0.29.2
          - Pipelines as Code: v0.28.2
          - Tekton Chains: v0.22.3
          - Tekton Hub (tech-preview): v1.18.2
          - Tekton Results (tech-preview): v0.12.3
          - Manual Approval Gate (tech-preview): v0.3.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.16/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:8cb87920e6fc8185ead49aafb7799fead6a4a510eafdae0ee29e85248dae3945
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:aeeed543412cde0bd3a9baec965cb99d80f2b6acb77ae6356ed51b4b1262d006
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:b41b888fc1a3b7215e6b71ee97bf64a1816914a8f33e96ad3930a4e2d5b5ffa7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:f9122d71c3862cd2afafc8eb7f1a39c83eee306e5baf38514e6e39215c70a72b
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:8129e62906c68f04075a44c939375a7bafdeaeac673a20a4fa74ffd6285187ac
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:aacdabe243dc573f3161a3d04fbf014b77d9962207305d70fcafce475f7eea93
        - registry.redhat.io/ubi8/openjdk-17@sha256:632e78989471859ca4ed1148b951a911f7e3b6e6516482be20967c4171612c55
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:19b3dd69c1d7e17890b7163aef322c1efbc76fed7f2b1faf26aaa05fd8161184
        - registry.redhat.io/rhel8/postgresql-13@sha256:a308263f5d1071be8ac5c13fd84b40609988ca30ed56c22a8bc6769e107ce09b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:227159c5a93966a43fa3a84cb69536ff0051cc261da068561241261d3837998a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:92042dcacd4ab0f2a78aab03fd91a9a39d7d7366b5398fd797ecb145acf5fce4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:90e2a7ea498da756e09c7e4d19c7fb91d429cc28c57f97c6958958e3466cb58c
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:7edbd11e5a449408d673cae1fc2b9b5bab492ecff2ac6e640c8cd69a9b90ff08
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:7da7f929d2711fc55dda01fe51a474bfa7adf5aaf78be6d65a6d65850958d6e7
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e1fe4a86211cf22e5ad2b23c989d24b69f597db4c3604e84777dbd0c585ec186
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:c6d654c71aba35c3c7f5c1a2ca5b8401e95adf6f649c599a9aff25177c2b3323
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:d2bb3f9f79ef719d8fdafef3a639e3eb8bc199ee56fcfb33e85975d536d2a6ac
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:4e0e4c358faa9cafc3964d6d40ee4aa73f8dfe8b6338dfb8206613d7c5241842
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:856229e9e37683c926e3449658d76f93d910384eff62c2ca192768a653b71533
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:e84c2d80e92815c6bf384e1504a194dec9b796a3aceac2eed555fff241249571
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:f3e6df07fd86acb91037d55a1970e5f8c6ab12a7517752692c224c7687022812
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:b94bd566ae3d9ce5f47273d867164486826b19fca0ce11d1a7d18312413e87ba
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:dcab4a3d519975340ade59a0ca0222104e68acd052a61f2aeb8c736be4b672cb
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:474658c262b539dd1c83ac22bb090b3fdac5de9cffc06587d36dbc6fa4ce208c
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:81a8b903d0fc4b81e6a23fac862d18da4cc18b14cca8213665df37dfaabbc2e9
        - registry.redhat.io/ubi8/ubi-minimal@sha256:a47c89f02b39a98290f88204ed3d162845db0a0c464b319c2596cfd1e94b444e
        - registry.redhat.io/rhel8/buildah@sha256:5c7cd7c9a3d49e8905fc98693f6da605aeafae36bde5622dc78e12f31db3cd59
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:80d46bc5ce8c931af22a97ccbce05b59740e313167c88ad910febcc2f2dc0320
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:549d74a18ec52db91fa949cac9b8e3d943d478d5dec9850d2ba0cb1bb526f578
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:47f094ccd12db7a43c8a91c1a3d2a190bc2b83f1745e7eb8ce097d0c8e1e2c88
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:2fea7d8e29d70c5893ddb6caf2bf1d4144787f5f0de8739351d2097ed09ee751
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:5672f3a375c6a274200b423eb2643c818aba75488dec0b210fec2532814ba51e
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:6acc31f179b08658976cddcb9ef42f377edb29933d1d83a2ecebfd455bb68571
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:d36dbb60c7b5caca7f3be08705f8ab91d7f01abaf8bfee4aee73bdbbb3e7e71a
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel8@sha256:46635d636ea2e9a6ba0760bf8c528028700ee40b0ba8466797034371185b16d0
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:7d92dca9da27c066cb530172e1811493d8c916e67d8068f5ddc471fa3f3f7856
        - registry.redhat.io/rhel8/skopeo@sha256:1dcb4189c92da21ee58b3ca5433615e0208194c8412f7f1fb45fa491179aaf9b
        version: 1.16.4
      entries:
      - name: openshift-pipelines-operator-rh.v1.16.4
        version: 1.16.4
      - name: openshift-pipelines-operator-rh.v1.16.3
        version: 1.16.3
      - name: openshift-pipelines-operator-rh.v1.16.2
        version: 1.16.2
      - name: openshift-pipelines-operator-rh.v1.16.1
        version: 1.16.1
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.16
    - currentCSV: openshift-pipelines-operator-rh.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.16.0 <1.17.2'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.65.7
          - Tekton Triggers: v0.30.1
          - Pipelines as Code: v0.29.1
          - Tekton Chains: v0.23.1
          - Tekton Hub (tech-preview): v1.19.2
          - Tekton Results (tech-preview): v0.13.4
          - Manual Approval Gate (tech-preview): v0.4.1

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.17/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel8@sha256:7abc88bea01dd6513f6e48334f0b98501f909a1dc1ab9e83efeb3c1a834f8f92
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel8@sha256:9668d852123b160e49511242371edc6e941a8e8b8c2ffa60080d91cc6b97d3d7
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel8@sha256:7c710882752f12be3c22bedc8d32096c111e42c5b9b392163f61cf8d72be3ec0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel8@sha256:922b9e86a674a6f3eb82f510569ab9cde97a1ad833c263335cd4164bf4164b66
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel8@sha256:66309ae437c4647f48efe3d6f2b696df83859323700427d0025b94795751a859
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel8@sha256:650391ffd3e4c7b173bf9d718405b2375c6ff2027abc0b7ce4f67e79e7ef4b81
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel8@sha256:41e2f2c7397a8b9e093e18977666e93943734f51780c769c99958ae1d20af53e
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel8@sha256:e673c20fff2fc684f5098282dd6ac6cb265045945249e995b81289532ce8c856
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel8@sha256:a392d2c85baedd7428c579e91d9f8acf8232399b90523b1093dc3447aaf56a16
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel8@sha256:41ec6db921812fd1e80b05d51c224ddbce63bf0c57f02ccf07461d2202e97b90
        - registry.redhat.io/rhel8/skopeo@sha256:7fd65200b4847fc84fba0466519b87f915e28c0e9bc90d3022beecfebffd4af8
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel8@sha256:b5cceead6b8e9c99655c9ba810679b9d6fb978e5ff18d020a46d475c4c26fa29
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel8@sha256:4fabae1312c1aaf8a57bd2de63bd040956faa0c728453f2a4b4002705fba0f0c
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel8@sha256:44da76d75f567c2ba623bae646e1d552d2c0fded1accdde3aab4f4bcdf3adc28
        - registry.redhat.io/openshift-pipelines/pipelines-rhel8-operator@sha256:eab798472b76bd51235198fd3b95189740f8d650ebc6bb01fe653c025c1defaa
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel8@sha256:4224040b00cfd76e85f80aecc5095098564cc62e9c477fe2f7a428b30df7d4d9
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel8@sha256:26f7b695fc1430359ab2bfd2915cf0b7e88b06d8695e26761bdc2ce8935f0fb3
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel8@sha256:0b6ee6cd98cb7e835cfc9544996c6f85288d7252071c100bed853eb593d6befb
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel8@sha256:4a96f1bb7e3a5c7ca62dff2199fa15ef964b129e4e36e9619837c1221538f496
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel8@sha256:4b87782b35e7284a365c49f1c094aae98531ac630db1958294009c041a2d8440
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:60635156d6b4e54529195af3bdd07329dcbe6239757db86e536dbe561aa77247
        - registry.redhat.io/ubi8/ubi-minimal@sha256:c12e67af6a7e15113d76bc72f10bef2045c026c71ec8b7124c8a075458188a83
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel8@sha256:8e52b4b31b0d9ec772abe6e6ef9c6698e89893395f364968216bebc3367da4b2
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel8@sha256:429bc3d6044ca7dab780ff4b5eb91d42d38a29eb7ef428a25c53b7dda49871bf
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel8@sha256:df11cdbab2a1c6b9b800ef124c73f2fb5e3981402b21bc822ccc13e95836197a
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel8@sha256:4240501c076605a01d969328bbb8ad885c598b44ec1dc1c066312491a0eb6715
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel8@sha256:c92ebe828b6ced6c93a3df5f4ccdfba022471729ad08b184de2f7410ecf592cc
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel8@sha256:d851a2dd8ec95bfd7ac6ec5515c8cdadfcbfc6e6ec84b40630b4fe595f978fa5
        - registry.redhat.io/rhel8/buildah@sha256:6d2dcb651ba680cf4ec74331f8349dec43d071d420625a1703370acc8d984e9e
        - registry.redhat.io/rhel8/postgresql-13@sha256:757d8dd6f243717d820846da7144a49913f28f223b74ec305b2215f54720f8a0
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:37109f6de75dda750496f6668164d0dceaf9db5278e8c034708436f39de3276e
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel8@sha256:db27cb4f0c350d99546b0f1bdf02d1ba271a8053a4552d88ff302adca60a3e29
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel8@sha256:61c9e8b1660d8935ee316ecc113e8141daf05149d172685cabd861ae8aeed3eb
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel8@sha256:99f5383d8c6bc046a0bf48fd6c2064a5d055645085248327305f3ee4ae71979c
        - registry.redhat.io/ubi8/openjdk-17@sha256:e5fc9e34382da41b10d4ea566639d31707493abff9e8de81f5accfd66876409e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel8@sha256:84eed8ba1189448e02ffd284c7d913e7b3cb7efe7a8e5b0edf890f7eb17fc439
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel8@sha256:a14270967460883f283ec1ce17ff8b82db1e0777dda4f6d690f15ac38dcbad85
        version: 1.17.2
      entries:
      - name: openshift-pipelines-operator-rh.v1.17.2
        version: 1.17.2
      - name: openshift-pipelines-operator-rh.v1.17.1
        version: 1.17.1
      - name: openshift-pipelines-operator-rh.v1.17.0
        version: 1.17.0
      - name: openshift-pipelines-operator-rh.v1.16.0
        version: 1.16.0
      - name: openshift-pipelines-operator-rh.v1.15.0
        version: 1.15.0
      name: pipelines-1.17
    - currentCSV: openshift-pipelines-operator-rh.v1.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  }
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.17.0 <1.18.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v0.68.1
          - Tekton Triggers: v0.31.0
          - Pipelines as Code: v0.33.2
          - Tekton Chains: v0.24.0
          - Tekton Hub (tech-preview): v1.20.1
          - Tekton Results: v0.14.0
          - Manual Approval Gate (tech-preview): v0.5.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.18/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:0a60ce46c05256c74f770901e0bb79ad89e88ba882b643c1ff7b2a1738463acc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c90a9c3a20f088241980b76986ba4f43b0243fd6d94a826fa7b5888e01c13093
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:80d441ec5d07c5059277a3a9173a35dfc0b8372d6e4d7823514131a94bae3fd7
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:eaad37453a5ebef51c28d4a068f56e9f394e710c72bb86342cd2189001420836
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:fa5c456fc91319c55fac9a011dc7bb32ef85db293c73d74436e903b76ceb4349
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:e63d956a0534f2e726301e00e877c4fe2f34ca4559e59d2df7567c04ded311a3
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:d72842a7c6a1e50a68576c93f60047afec801cf83db64d7d3e34b25bf3c753ec
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:b9675b22bc8759903472c1eb0e03f90c75460f02799d894294c161641cad9414
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:cd28c2c90dd340324b0d8709640cd7636f28286114b0023649b6e21be13e03c2
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:72901cf4e091fe5f6a577144fdffaea8570712e9887a8bd715d5ae5bc346ff1d
        - registry.redhat.io/rhel9/buildah@sha256:6886997da8c92214e99e3e8da47a0f08239172d2544b4edae058d5cb018effab
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:e4d49cf6a2fb8b3585ecfe3b3da3e9de090fb99de3dee2da0601b1f6b071173c
        - registry.redhat.io/ubi9/ubi-minimal@sha256:ac61c96b93894b9169221e87718733354dd3765dd4a62b275893c7ff0d876869
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:8b4bf043e0fb237553503d63b129c432600bb4e65b2096b246b257b404b42be9
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:8ae0ca6411e4fcd5c0605344bc6930cf122ffd80e71199509d58423a8b41de1d
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:6abc903c784bb455e1f2fde45a48afe775239668e1d0dbc1b3cc6afca3c0b477
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:dbc462cb50385dec3a220b5d0caed0e0211c166182026f86e3c51a1351ba0de4
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:cbd89c531953a43b8055bda72f7696be80f57fd1d782567a91d82e0fbcb3d26d
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:3964c3248dfb1006ea9a46295baf7f8cfa71952d6b9863b52f87e8f3d7863b1e
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:e6a7738cdb0e8aa9e1af381273b1c18ec1414b08a66fc851f6e11e5aa7a6706f
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:1b57b64657eca0002145100bae262a19f02eb40068e07994df467e28a1618f90
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:e27a30d471dda4983a5b8192709d283a9d082558f3caf31b30db2214e418b017
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:d055d2b35a663aef3e1aafdbed0b12957867c0670c946ebae66e9c44a768bda2
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:96f5e1922fb9d6dc17e271576dda9bd032bd6dcc04e7496fb15a2f6b5871cd67
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:3696cf724ce7865626ae408d82fd87727c741e4d3eb3475c271ec6d56f0882a4
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:776d34b0ea2e35c11ad4050e6b40cf4b40c84dbe0d5f3e6ad669c494a210ea8e
        - registry.redhat.io/rhel9/postgresql-13@sha256:ecd5a5583db922c517c735797db3fca587c0ef0b410e7dd494ace2267d89cf05
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:fcd5fd10de7334bd59a8da4997f91008ced95c6cde40d56d50f3b3e32c0f19e1
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:16d799503674641d4690822edc9c399c61429ccd31621d295eee4a4d1ef1c076
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:c20ab6fa719506bca942669ea84b113335b05df3ed275257b88be234cade7c8d
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:20e19c02efdecf94a7b16483047e5db81c50dc2a0aa59d4895067e6c7769efcc
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:8fec8e7a8b3e5182c67104de4e3abe32593b682c4a8986d1e9ab835dafb95ff4
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:409c14649d8e7007d4b72b6feffb71dabd26322bf2a7c77f5307f2546fdfe81b
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:b550a8bc2a87d07a5b22a3ec3fb83efd8eb26ca6e80d404f2e453cc9b8434da7
        - registry.redhat.io/ubi9/openjdk-17@sha256:5d619e202695acf76a5f32c07ace6709f5ee5d3999bf977e1bf95ad34197f44a
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:86ea94b12e8923035e3e4f39b6a53b3ab32c5f77d7eea170aa9b34aecc6f8577
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:36a4c6a3273f2371ab5e679f1955221c286ea753e29e6ab8da7c983681a68f02
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e2b4c3b505c4b5aa36f8a5e661389b0de4e799c74c2c7f16a065e5522b438e89
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:2ba6782af0d7e5c4d623de31814f43f4f7c3c38d09b57705021957d334067f87
        - registry.redhat.io/rhel9/skopeo@sha256:d077e35100d6d7b3795f8e3594293dbf8b65c6bc18b909aa555fe56f048b356a
        version: 1.18.1
      entries:
      - name: openshift-pipelines-operator-rh.v1.18.1
        version: 1.18.1
      - name: openshift-pipelines-operator-rh.v1.18.0
        version: 1.18.0
      name: pipelines-1.18
    - currentCSV: openshift-pipelines-operator-rh.v1.19.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.18.0 <1.19.3'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.0.0
          - Tekton Triggers: v0.32.0
          - Pipelines as Code: v0.35.3
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.21.0
          - Tekton Results (tech-preview): v0.15.3
          - Manual Approval Gate (tech-preview): v0.6.0
          - Tekton Pruner (tech-preview): v0.1.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.19/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:1f870e70669ad584db1ee31b1dcd822d8b8d0cc03028c7cb09238fb7abf19a42
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:c9fbb018a70df9a542236ace556fa3343ae350f43c0629f13e525854508f73f9
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:80af5fa2d87e7c405f16575519293ca7fccc49d635b4f3d0d334dea4fbd28721
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:31a40f8095a3f8cd04750810d964572339965ae0b2d357c57c0b671a95e1cdbe
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:8bcaeaa751c61eb927d6353984f81d6a92b4bd4f9a5be9dc620a11b06ff9e6ba
        - registry.redhat.io/ubi9/openjdk-17@sha256:cdbd64ba1b7826d09bc180fa19aa6fbc09ddb5bff5e02b7e1351353bbabebb78
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:2c3074ecc0c74d508020bf1e11667d868150915b55d436a43c2b4fc22815bf48
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:a90a6c269b3fc4c77652d4862878211eee5eca1d8047e561d07e3ef1292e52f5
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:d21cc7e094aa46ba7f6ea717a3d7927da489024a46a6c1224c0b3c5834dcb7a6
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:5f39f9e61ddbce92629e6502bbef737365c8a1bb2dc2ecaf9f9d8092dea844c9
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:aa1b6f374643ed4e89ba194afde8a5138e401c8cdeb45a6bb5df5b608f58358b
        - registry.redhat.io/rhel9/buildah@sha256:a6e04c061b8d261dba85d01dcb64d3f7dba8c0b9765e03f05ec79850a248d75a
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:095e93f8dd31a0985d6fef6a6168da456630febe89ab9e0cb9480170f1036541
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:68d9007c5abb2ad3b7666249a61d9428c2cf07165bd8090a6e6def0b7cf70fa3
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:92b8c92eb0bec102215b97f1d60d76c7a57747aeec9594ce1e0e46f7c27391d6
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:ba596c629fe8b5b70db97923ff11392dc12115c8e7afefee0d48c12440b8dd86
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:6aaabb9af37e46a856371e29f141de6b0fac623128099bb927fc30f8f61e9db4
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:41d669fbe8bb48481890fb9fb52de9d4cac98cb74e6a6b284d046a8d21b06e58
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:3f7e1d83eb1e21860cc2ae0bfe3abe8c9058e92a69698dba9d5871069016e90d
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:d000daec36a6cdbe41a85ee4eac24a712f7628441a412f751baf8d7778626f62
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:90a5eb2c0decb72234cbea5354bb7737cebc53c467229b2c16a1425b4bf408dc
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:b247e9f3a2db32b575999387ef4f91d685a67c1f3950a6e265e1fe14c1fac196
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:28258903330a0715141e4685b3760ff18b0d75da0694985bba7c3345d7808b5f
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:2d2499b7b12aace07c4b5ab4121caad445aab5bc8dd73377171683b36d417d73
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:22cfadc42e793553182acf05a34fb6a7ae0d7055431c2e393699ecfeb171b173
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:25bcdb30f0fe34300c4f61f62513f4853f65fe11bafde58bcf19b7b3c3fb9af7
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:251c87e2c48e7f377bcb399ed64d50dff11de3459629a73a63bac17c353eb16b
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:10c620aa3a9c4bfbca185fb460e2521473a5fe6d518342b89a3dada5cf14e65a
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:b8cdf2a1873a3df1fb8b8f6d58d06235cef28d228abd70d0bb83cdcfbe3bc230
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:7a0c6794e9284209dc131f3087ffc01fd5946e20a9261422808c6dd3938407de
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:fd6f4cd7801d7c20cf5a6e546022136404212da5d9997750e643f6e640b00e3d
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:3b1c6f8acdf17a5185413147e0248baa559378ed79e08cb098c724036020576e
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:70293cb8a80aa548933ff5f502bac89945a5cd08801c4ca3aac08a10dd01f62f
        - registry.redhat.io/ubi9/ubi-minimal@sha256:14f14e03d68f7fd5f2b18a13478b6b127c341b346c86b6e0b886ed2b7573b8e0
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:198cf9e5c9641757a0453dec2747a24b018a9b85e10eb0dfdc7a2e75ed09cf6d
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:363e28a82ed022e507c3976d3a42ee63c053eb7871712c213b12235d8234b7ef
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:c3ba6b9652205407cc4d4d47b4a928fe8fa86ce4427e6b44ee1d53faf3455f21
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:22095c436562ff92771d9349f57d007d82415dae519e3da1de7d51e3f1042a93
        - registry.redhat.io/rhel9/skopeo@sha256:993f799d58c48151ecae7a84e63860f6c0039e8e18efed5ad67e7752f4b7e408
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:13608cfe2b87535776c65728aa40cfaa84ccf63afe64d1f2e799ea0e26819239
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:37e2f1fb8a2b4c7ac3d5390ba68a9bf98a0c83a38567b6a9778764061b3d2bab
        version: 1.19.3
      entries:
      - name: openshift-pipelines-operator-rh.v1.19.3
        version: 1.19.3
      - name: openshift-pipelines-operator-rh.v1.19.2
        version: 1.19.2
      - name: openshift-pipelines-operator-rh.v1.19.1
        version: 1.19.1
      - name: openshift-pipelines-operator-rh.v1.19.0
        version: 1.19.0
      name: pipelines-1.19
    - currentCSV: openshift-pipelines-operator-rh.v1.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "ManualApprovalGate",
                "metadata": {
                  "name": "manual-approval-gate"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "OpenShiftPipelinesAsCode",
                "metadata": {
                  "name": "pipelines-as-code"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonAddon",
                "metadata": {
                  "name": "addon"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonChain",
                "metadata": {
                  "name": "chain"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonConfig",
                "metadata": {
                  "name": "config"
                },
                "spec": {
                  "addon": {
                    "params": [
                      {
                        "name": "pipelineTemplates",
                        "value": "true"
                      },
                      {
                        "name": "resolverTasks",
                        "value": "true"
                      },
                      {
                        "name": "resolverStepActions",
                        "value": "true"
                      },
                      {
                        "name": "communityResolverTasks",
                        "value": "true"
                      }
                    ]
                  },
                  "params": [
                    {
                      "name": "createRbacResource",
                      "value": "true"
                    },
                    {
                      "name": "legacyPipelineRbac",
                      "value": "true"
                    },
                    {
                      "name": "createCABundleConfigMaps",
                      "value": "true"
                    }
                  ],
                  "profile": "all",
                  "pruner": {
                    "keep": 100,
                    "resources": [
                      "pipelinerun",
                      "taskrun"
                    ],
                    "schedule": "0 8 * * *"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonHub",
                "metadata": {
                  "name": "hub"
                },
                "spec": {
                  "api": {
                    "catalogRefreshInterval": "30m"
                  },
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonInstallerSet",
                "metadata": {
                  "name": "foo"
                },
                "spec": {
                  "manifests": []
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPipeline",
                "metadata": {
                  "name": "pipeline"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonPruner",
                "metadata": {
                  "name": "pruner"
                },
                "spec": {
                  "disabled": true,
                  "targetNamespace": "openshift-pipelines"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonResult",
                "metadata": {
                  "name": "result"
                },
                "spec": {
                  "auth_disable": true,
                  "db_enable_auto_migration": true,
                  "db_host": "tekton-results-postgres-service.openshift-pipelines.svc.cluster.local",
                  "db_port": 5432,
                  "log_level": "debug",
                  "logging_pvc_name": "tekton-logs",
                  "logs_api": true,
                  "logs_buffer_size": 32768,
                  "logs_path": "/logs",
                  "logs_type": "File",
                  "prometheus_port": 9090,
                  "server_port": 8080,
                  "targetNamespace": "openshift-pipelines",
                  "tls_hostname_override": "tekton-results-api-service.openshift-pipelines.svc.cluster.local"
                }
              },
              {
                "apiVersion": "operator.tekton.dev/v1alpha1",
                "kind": "TektonTrigger",
                "metadata": {
                  "name": "trigger"
                },
                "spec": {
                  "targetNamespace": "openshift-pipelines"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          console.openshift.io/plugins: '["pipelines-console-plugin"]'
          description: Red Hat OpenShift Pipelines is a cloud-native CI/CD solution
            for building pipelines using Tekton concepts which run natively on OpenShift
            and Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.19.0 <1.20.0'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["tektoninstallersets.operator.tekton.dev",
            "tektonconfigs.operator.tekton.dev","tektonpipelines.operator.tekton.dev","tektontriggers.operator.tekton.dev","tektonaddons.operator.tekton.dev",
            "tektonhubs.operator.tekton.dev", "tektonresults.operator.tekton.dev",
            "tektonchains.operator.tekton.dev", "openshiftpipelinesascodes.operator.tekton.dev",
            "manualapprovalgates.operator.tekton.dev","tektonpruners.operator.tekton.dev"]'
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/tektoncd/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of latest version of Manual Approval
              Gate
            displayName: Manual Approval Gate
            kind: ManualApprovalGate
            name: manualapprovalgates.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of OpenShift
              Pipelines As Code
            displayName: OpenShift Pipelines As Code
            kind: OpenShiftPipelinesAsCode
            name: openshiftpipelinesascodes.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Addons (exclusively
              Tasks, StepActions, consoleyamlsamples, consoleclidownloads and consolequickstarts)
            displayName: Tekton Addons
            kind: TektonAddon
            name: tektonaddons.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Chains
            displayName: Tekton Chains
            kind: TektonChain
            name: tektonchains.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton components,
              TektonPipeline, TektonTrigger and TektonAddon
            displayName: Tekton Configuration
            kind: TektonConfig
            name: tektonconfigs.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Hub
            displayName: Tekton Hub
            kind: TektonHub
            name: tektonhubs.operator.tekton.dev
            version: v1alpha1
          - description: This CustomResourceDefinition (CRD) is used internally by
              the other OpenShift Pipelines CRDs to maintain the lifecycle of OpenShift
              Pipelines Components
            displayName: Tekton Installer Set
            kind: TektonInstallerSet
            name: tektoninstallersets.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pipelines
            displayName: Tekton Pipelines
            kind: TektonPipeline
            name: tektonpipelines.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Pruner
            displayName: Tekton Pruner
            kind: TektonPruner
            name: tektonpruners.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Results
            displayName: Tekton Results
            kind: TektonResult
            name: tektonresults.operator.tekton.dev
            version: v1alpha1
          - description: Represents an installation of latest version of Tekton Triggers
            displayName: Tekton Triggers
            kind: TektonTrigger
            name: tektontriggers.operator.tekton.dev
            version: v1alpha1
        description: |
          Red Hat OpenShift Pipelines is a cloud-native continuous integration and delivery
          (CI/CD) solution for building pipelines using [Tekton](https://tekton.dev).
          Tekton is a flexible Kubernetes-native open-source CI/CD framework, which
          enables automating deployments across multiple platforms (Kubernetes,
          serverless, VMs, etc) by abstracting away the underlying details.

          ## Features

          * Standard CI/CD pipelines definition
          * Build images with Kubernetes tools such as S2I, Buildah, Buildpacks, Kaniko, etc
          * Deploy applications to multiple platforms such as Kubernetes, serverless and VMs
          * Easy to extend and integrate with existing tools
          * Scale pipelines on-demand
          * Portable across any Kubernetes platform
          * Designed for microservices and decentralized team
          * Integrated with OpenShift Developer Console

          ## Installation

          _Red Hat OpenShift Pipelines Operator_ gets installed into a single namespace (openshift-operators) which would then install _Red Hat OpenShift Pipelines_ into the openshift-pipelines namespace. _Red Hat OpenShift Pipelines_ is however cluster-wide and can run pipelines created in any namespace.

          ### Components

          - Tekton Pipelines: v1.3.2
          - Tekton Triggers: v0.33.0
          - Pipelines as Code: v0.37.0
          - Tekton Chains: v0.25.1
          - Tekton Hub (tech-preview): v1.22.3
          - Tekton Results (tech-preview): v0.16.0
          - Manual Approval Gate (tech-preview): v0.7.0
          - Tekton Pruner (tech-preview): v0.2.0

          ## Getting Started

          In order to get familiar with _OpenShift Pipelines_ concepts and create your first pipeline, follow the [OpenShift Pipelines Docs](https://docs.openshift.com/pipelines/1.20/create/creating-applications-with-cicd-pipelines.html).

          ## CLI (tkn)

          Tekton Pipelines cli project provides a CLI (tkn) for interacting with OpenShift Pipelines.

          [Download tkn](/command-line-tools) (links will be available on `command line tools` page after this operator is installed).
        displayName: Red Hat OpenShift Pipelines
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tektoncd
        - openshift
        - build
        - pipeline
        links:
        - name: Tekton Pipeline GitHub Repository
          url: https://github.com/tektoncd/pipeline
        - name: Tekton Operator GitHub Repository
          url: https://github.com/tektoncd/operator
        maintainers:
        - email: pipelines-extcomm@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-pipelines/pipelines-entrypoint-rhel9@sha256:a712bb411d2433b01793dff25785c1e265fa865a52f7c8732c9fb4418ab61757
        - registry.redhat.io/openshift-pipelines/pipelines-git-init-rhel9@sha256:8f0494341244a75546a4b80b065702be04e6d200421aa4e7ec90385c9dabc6ab
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-cli-rhel9@sha256:818bfa4a3a874974b5f7f8bcf2faeb6c05e1795381ee1ceacf89f17b1ec0ef7f
        - registry.redhat.io/openshift-pipelines/pipelines-opc-rhel9@sha256:c1bee4a2bb495de6d4da962b6e7b2447caa51cc36c3005e4083bd31420ef5102
        - registry.redhat.io/rhel9/skopeo@sha256:a65a413f8a2864389a09dc750690d97afbcdc5c70821e0f85e99e8adba7954e0
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-controller-rhel9@sha256:bbd1918660be7b3c78777dcccc87d3b26bf5adfc077844ceffbbdd53696dc0a3
        - registry.redhat.io/openshift-pipelines/pipelines-pruner-controller-rhel9@sha256:5021158a4752a26f8b33adea7da66695b22d2ad155d346d35dde5f971bc1a55b
        - registry.redhat.io/openshift-pipelines/pipelines-operator-webhook-rhel9@sha256:2d496752b19f765f1ea36ec94b873ebd8fa604764a56aa67d3ed94b33e5fe75c
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-watcher-rhel9@sha256:e3e64abef53151f6b15f15bcbff85668167a571e98badbd8a2ca252ff3a5b8dc
        - registry.redhat.io/openshift-pipelines/pipelines-rhel9-operator@sha256:161df929e2bd4c5acff594be31d5458365de5ae0d2189bb2f647c3fc6657fa28
        - registry.redhat.io/openshift-pipelines/pipelines-pipelines-as-code-webhook-rhel9@sha256:ec74a562c18fefe36ec7ae3ba083e4eb76c56b562f916578900f8767ae98e6c2
        - registry.redhat.io/openshift-pipelines/pipelines-results-watcher-rhel9@sha256:9415e889ef98df6d2b2dca1e5d798fb765eeea0d9828e48424374d5267a6b0d1
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-core-interceptors-rhel9@sha256:1b6aecfce45ed3163227b845b212e83385cdbe999edfc2e9387a0c76f3f8d8e9
        - registry.redhat.io/ubi9/openjdk-17@sha256:555c8bcd2596637befd1a6da4aa949c5a2640cf66e09c2dfb932afe37e39095e
        - registry.redhat.io/openshift-pipelines/pipelines-operator-bundle@sha256:c161c515181728ad53e6d15b59444a0f7dc468a1a84a9dea5850a99aa29ff9e0
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-controller-rhel9@sha256:df15eb1f1f5f237bfe481c16ab3d40bf7fb4123f3c930bcfb121e2cc8dc42882
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-controller-rhel9@sha256:c3013c347bca8a520443a44ac2817fbba5744a9e91069ac5f09e91447ebf6b28
        - registry.redhat.io/rhel9/buildah@sha256:18453ab0a62154283aaf4c45efd7543d1c8ecca1aebae46dedc7ddc0f410b232
        - registry.redhat.io/openshift-pipelines/pipelines-serve-tkn-cli-rhel9@sha256:115eea1a5ca87f3b6b72db1df05b7386bf73dc70c9c2e4dc5e1a66700d960c37
        - registry.redhat.io/openshift-pipelines/pipelines-chains-controller-rhel9@sha256:980ed7e81ea55522e1681dc1e9b170b68d16f07061dd1f1ffd142367d59552f1
        - registry.redhat.io/openshift-pipelines/pipelines-webhook-rhel9@sha256:e63298d07ccec5804dfab4e42bcc8163c29cb3ac371c91248777423a4edd31ce
        - registry.redhat.io/openshift-pipelines/pipelines-results-retention-policy-agent-rhel9@sha256:441e1c7124c1674e3d4c38c3e82a4fe5568e76d13c89c04bf009d6b97435de09
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-webhook-rhel9@sha256:036250bcdc6fb95273bf07948dfdd4dce765ab84c803ceca0d0b4abf6ca19dec
        - registry.redhat.io/openshift-pipelines/pipelines-console-plugin-rhel9@sha256:14dab8d87323fc7d112213f7f4db3ee7cf3841747de669ce710ae6a794053457
        - registry.redhat.io/openshift-pipelines/pipelines-hub-api-rhel9@sha256:d56d6b46331b4569077ab1e7a7b609ba01f0a1dd62b695d62f8aa0f32ea1b621
        - registry.redhat.io/openshift-pipelines/pipelines-hub-db-migration-rhel9@sha256:458f865f1b246c5f0682f5956795771a0515b89edfcffbbc0da1a3d3f532645e
        - registry.redhat.io/openshift-pipelines/pipelines-resolvers-rhel9@sha256:96db721939b22572d7d8d8b95dcb5c92fdedb98e9cd0aabb7d17b703a915b07f
        - registry.redhat.io/openshift-pipelines/pipelines-results-api-rhel9@sha256:dd941acd357ae2c0fd7e15fb8d32d7ee6b2847435101dd1e1f107db48af18bf3
        - registry.redhat.io/openshift-pipelines/pipelines-cli-tkn-rhel9@sha256:5683321a5376e63c3ff3932270a1850d5e261f76212020ce867334166f90d2f7
        - registry.redhat.io/source-to-image/source-to-image-rhel9@sha256:bc10e7d6122ede5cf4f2882c200447683255465a1080e826761e90bfb983506e
        - registry.redhat.io/ubi9/ubi-minimal@sha256:2f06ae0e6d3d9c4f610d32c480338eef474867f435d8d28625f2985e8acde6e8
        - registry.redhat.io/openshift-pipelines/pipelines-manual-approval-gate-webhook-rhel9@sha256:3cf990fb2fc422f3527a7c8b43d06ccd4d32bb5487ec8cb2d4a97f0f5d13f170
        - registry.redhat.io/openshift-pipelines/pipelines-nop-rhel9@sha256:c4d05e546c466c57379fff502816012f615be073eb8cdce54677b4ab206a3f5c
        - registry.redhat.io/openshift-pipelines/pipelines-triggers-eventlistenersink-rhel9@sha256:c31af237afb84214bf2a4c06a7636ca071fe97203b41da90d7878109ccd64784
        - registry.redhat.io/openshift-pipelines/pipelines-events-rhel9@sha256:c0f438e6095aa81f3f7d820614e0414cc43374014c1505d7641bc1dba69a3903
        - registry.redhat.io/openshift-pipelines/pipelines-controller-rhel9@sha256:901b5108ceb669c985e75f5fb033c496e9f028863df99ab32651d1b08433a352
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-pipelines/pipelines-sidecarlogresults-rhel9@sha256:7eafffa6e4e04f10c1d4e84e02d51c00eb8649c60f8f9a13da14492f24634891
        - registry.redhat.io/rhel9/postgresql-13@sha256:36647bd6d50d2e0e9c99ecf190bd515577c7ce9476e49d927c4d4065011e3441
        - registry.redhat.io/openshift-pipelines/pipelines-cache-rhel9@sha256:823e018e0329d764ac694f177ce88af9074f364c29a1c17929f9dd0b596d03dd
        - registry.redhat.io/openshift-pipelines/pipelines-operator-proxy-rhel9@sha256:e7998e113f75e8f29602bae210c8b66ed06634b699f1d99c5875ef879220d733
        - registry.redhat.io/openshift-pipelines/pipelines-hub-ui-rhel9@sha256:bc403f93ef6c5526fa30152eab384c4a25bb0bc6ab4158a1e24f637b87a57259
        - registry.redhat.io/openshift-pipelines/pipelines-workingdirinit-rhel9@sha256:a7163a9e1fe891de021b1003558e6e64fe17ed17ab0820ba16bba09aa112401e
        version: 1.20.0
      entries:
      - name: openshift-pipelines-operator-rh.v1.20.0
        version: 1.20.0
      name: pipelines-1.20
    defaultChannel: latest
    packageName: openshift-pipelines-operator-rh
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aikit-operator-rhmp/support?utm_source=openshift_console
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:524d315efa226f97ff26b8668f9297e755c6f0c815b8f3385b317667553f0e62
        - registry.marketplace.redhat.com/rhm/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator-rhmp
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: F5 Networks Inc.
      provider-url: ""
    name: f5-ipam-controller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: f5-ipam-controller-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fic.f5.com/v1",
                "kind": "F5IpamCtlr",
                "metadata": {
                  "name": "f5ipamctlr-sample"
                },
                "spec": {
                  "args": {
                    "infoblox_grid_host": "IP_ADDRESS_OF_SERVER",
                    "infoblox_labels": "CIDR_Labels",
                    "infoblox_netview": "NetView",
                    "infoblox_password": "PASSWORD",
                    "infoblox_username": "USERNAME",
                    "infoblox_wapi_version": "VERSION",
                    "insecure": true,
                    "ip_range": "IP_RANGE_LABELS",
                    "log_level": "DEBUG_OR_INFO",
                    "orchestration": "kubernetes",
                    "provider": "INFOBLOX_OR_F5-IP-PROVIDER"
                  },
                  "image": {
                    "pullPolicy": "Always",
                    "repo": "f5-ipam-controller",
                    "user": "f5networks",
                    "version": "latest"
                  },
                  "namespace": "kube-system",
                  "rbac": {
                    "create": true
                  },
                  "serviceAccount": {
                    "create": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
          createdAt: "2025-08-08T03:48:02Z"
          description: Operator to install F5 IPAM Controller.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/F5Networks/f5-ipam-controller
          support: F5 Operators Team <f5_cis_operators@f5.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This Operator installs F5 IPAM controller(FIC)
            kind: F5IpamCtlr
            name: f5ipamctlrs.fic.f5.com
            version: v1
        description: |
          ## Introduction
          This Operator installs F5 IPAM Controller in your OpenShift Cluster. This enables to configure and deploy FIC using Helm Charts.
          ## F5 IPAM Controller
          The F5 IPAM Controller is a Docker container that runs in an orchestration environment and interfaces with an IPAM system. It allocates IP addresses from an IPAM system’s address pool for hostnames in an orchestration environment. The F5 IPAM Controller watches orchestration-specific resources and consumes the hostnames within each resource.
          ## Documentation
          Refer to F5 IPAM Controller documentation
          - https://github.com/F5Networks/f5-ipam-controller/blob/main/README.md
          ## Prerequisites
          ### Configure Infoblox with network and netview [Provider: Infoblox]
          Refer Infoblox documentation
          ### Create persistent volume and persistent volume claim [Provider: f5-ip-provider]
          Create persistent volume and persistent volume claim
          ```
          oc apply -f https://raw.githubusercontent.com/F5Networks/f5-ipam-controller/main/docs/config_examples/f5-ip-provider/localstorage-pv-pvc-example.yaml
          ```
        displayName: F5 Ipam controller
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipam
        - ip
        - ipaddress
        - controller
        - ipv4
        - ipv6
        - F5
        links:
        - name: Documentation
          url: https://clouddocs.f5.com/containers/latest/userguide/ipam/
        - name: Github Repo
          url: https://github.com/F5Networks/f5-ipam-controller
        maintainers:
        - email: f5_cis_operators@f5.com
          name: F5 Operators Team
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: F5 Networks Inc.
        relatedImages:
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator@sha256:f2ae6f887e68b756204499caef787f6dfb692a72d19d585c448dd48b250752e0
        - registry.connect.redhat.com/f5networks/f5-ipam-controller@sha256:09a358db300d82e00fd3b4e255917df82c02e1bc58425f8cff21cc2c77eba1b6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61d446b8b81cc1545ee805dbd46f921aecb1517c3478bdff654ab9a2a637845a
        - registry.connect.redhat.com/f5networks/f5-ipam-controller-operator-bundle@sha256:6763e1a2be507a17b2ea17f0914d6d2c10ad1196d49b79881f8008867627af8a
        version: 0.0.8
      entries:
      - name: f5-ipam-controller-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: f5-ipam-controller-operator
    provider:
      name: F5 Networks Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:f5f025c7f56643fab9148bb5de8ee34a84d02b78a6abb4b5592306cf54296654
        - registry.connect.redhat.com/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "docker.io/percona/percona-postgresql-operator:2.7.0-ppg17.5.2-postgres",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "docker.io/percona/percona-pgbouncer:1.24.1",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "docker.io/percona/percona-pgbackrest:2.55.0",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:3.2.0",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/percona/percona-postgresql-operator:2.7.0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          olm.skipRange: <v2.7.0
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native

          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - docker.io/percona/percona-postgresql-operator:2.7.0
        - quay.io/community-operator-pipeline-prod/percona-postgresql-operator@sha256:44e687b38a48f68bbbb919e6df0d3b87b75492aca7bd41e508a2aa0e80d024a9
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator.v2.6.1
        version: 2.6.1
      name: stable
    defaultChannel: stable
    packageName: percona-postgresql-operator
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elasticache-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elasticache-controller.v1.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "CacheSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "ReplicationGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elasticache.services.k8s.aws/v1alpha1",
                "kind": "UserGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
          createdAt: "2025-09-19T19:19:33Z"
          description: AWS ElastiCache controller is a service controller for managing
            ElastiCache resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CacheCluster represents the state of an AWS elasticache CacheCluster
              resource.
            displayName: CacheCluster
            kind: CacheCluster
            name: cacheclusters.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheParameterGroup represents the state of an AWS elasticache
              CacheParameterGroup resource.
            displayName: CacheParameterGroup
            kind: CacheParameterGroup
            name: cacheparametergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: CacheSubnetGroup represents the state of an AWS elasticache
              CacheSubnetGroup resource.
            displayName: CacheSubnetGroup
            kind: CacheSubnetGroup
            name: cachesubnetgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ReplicationGroup represents the state of an AWS elasticache
              ReplicationGroup resource.
            displayName: ReplicationGroup
            kind: ReplicationGroup
            name: replicationgroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCache represents the state of an AWS elasticache
              ServerlessCache resource.
            displayName: ServerlessCache
            kind: ServerlessCache
            name: serverlesscaches.elasticache.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCacheSnapshot represents the state of an AWS elasticache
              ServerlessCacheSnapshot resource.
            displayName: ServerlessCacheSnapshot
            kind: ServerlessCacheSnapshot
            name: serverlesscachesnapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS elasticache Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.elasticache.services.k8s.aws
            version: v1alpha1
          - description: UserGroup represents the state of an AWS elasticache UserGroup
              resource.
            displayName: UserGroup
            kind: UserGroup
            name: usergroups.elasticache.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS elasticache User resource.
            displayName: User
            kind: User
            name: users.elasticache.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ElastiCache resources in AWS from within your Kubernetes cluster.

          **About Amazon ElastiCache**

          Amazon ElastiCache allows you to seamlessly set up, run, and scale popular open-source compatible in-memory data stores in the cloud. Build data-intensive apps or boost the performance of your existing databases by retrieving data from high throughput and low latency in-memory data stores. Amazon ElastiCache is a popular choice for real-time use cases like Caching, Session Stores, Gaming, Geospatial Services, Real-Time Analytics, and Queuing.

          Amazon ElastiCache offers fully managed Redis, voted the most loved database by developers in the Stack Overflow 2020 Developer Survey, and Memcached for your most demanding applications that require sub-millisecond response times.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ElastiCache
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticache
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ElastiCache Developer Resources
          url: https://aws.amazon.com/elasticache/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elasticache maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elasticache-controller:1.2.3
        - quay.io/community-operator-pipeline-prod/ack-elasticache-controller@sha256:a489c815e9d3777d775ccd65bbd0ecd5fe7c577de36c504b63cf1620531d4ccb
        version: 1.2.3
      entries:
      - name: ack-elasticache-controller.v1.2.3
        version: 1.2.3
      - name: ack-elasticache-controller.v1.2.2
        version: 1.2.2
      - name: ack-elasticache-controller.v1.2.1
        version: 1.2.1
      - name: ack-elasticache-controller.v1.2.0
        version: 1.2.0
      - name: ack-elasticache-controller.v1.1.0
        version: 1.1.0
      - name: ack-elasticache-controller.v1.0.7
        version: 1.0.7
      - name: ack-elasticache-controller.v1.0.6
        version: 1.0.6
      - name: ack-elasticache-controller.v1.0.5
        version: 1.0.5
      - name: ack-elasticache-controller.v1.0.4
        version: 1.0.4
      - name: ack-elasticache-controller.v1.0.3
        version: 1.0.3
      - name: ack-elasticache-controller.v1.0.2
        version: 1.0.2
      - name: ack-elasticache-controller.v1.0.1
        version: 1.0.1
      - name: ack-elasticache-controller.v1.0.0
        version: 1.0.0
      - name: ack-elasticache-controller.v0.2.6
        version: 0.2.6
      - name: ack-elasticache-controller.v0.2.5
        version: 0.2.5
      - name: ack-elasticache-controller.v0.2.4
        version: 0.2.4
      - name: ack-elasticache-controller.v0.2.3
        version: 0.2.3
      - name: ack-elasticache-controller.v0.2.2
        version: 0.2.2
      - name: ack-elasticache-controller.v0.2.1
        version: 0.2.1
      - name: ack-elasticache-controller.v0.2.0
        version: 0.2.0
      - name: ack-elasticache-controller.v0.1.1
        version: 0.1.1
      - name: ack-elasticache-controller.v0.1.0
        version: 0.1.0
      - name: ack-elasticache-controller.v0.0.31
        version: 0.0.31
      - name: ack-elasticache-controller.v0.0.30
        version: 0.0.30
      - name: ack-elasticache-controller.v0.0.29
        version: 0.0.29
      - name: ack-elasticache-controller.v0.0.28
        version: 0.0.28
      - name: ack-elasticache-controller.v0.0.27
        version: 0.0.27
      - name: ack-elasticache-controller.v0.0.26
        version: 0.0.26
      - name: ack-elasticache-controller.v0.0.25
        version: 0.0.25
      - name: ack-elasticache-controller.v0.0.24
        version: 0.0.24
      - name: ack-elasticache-controller.v0.0.23
        version: 0.0.23
      - name: ack-elasticache-controller.v0.0.22
        version: 0.0.22
      - name: ack-elasticache-controller.v0.0.21
        version: 0.0.21
      - name: ack-elasticache-controller.v0.0.20
        version: 0.0.20
      - name: ack-elasticache-controller.v0.0.19
        version: 0.0.19
      - name: ack-elasticache-controller.v0.0.18
        version: 0.0.18
      - name: ack-elasticache-controller.v0.0.17
        version: 0.0.17
      - name: ack-elasticache-controller.v0.0.16
        version: 0.0.16
      - name: ack-elasticache-controller.v0.0.15
        version: 0.0.15
      - name: ack-elasticache-controller.v0.0.14
        version: 0.0.14
      - name: ack-elasticache-controller.v0.0.13
        version: 0.0.13
      - name: ack-elasticache-controller.v0.0.12
        version: 0.0.12
      - name: ack-elasticache-controller.v0.0.11
        version: 0.0.11
      - name: ack-elasticache-controller.v0.0.10
        version: 0.0.10
      - name: ack-elasticache-controller.v0.0.9
        version: 0.0.9
      - name: ack-elasticache-controller.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: ack-elasticache-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: volsync-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable
    - currentCSV: volsync-product.v0.11.2-0.1749674925.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
          createdAt: 16 Apr 2025, 06:42
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.11.2'
          olm.substitutesFor: volsync-product.v0.11.2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:491e7aea221eb0e1d2992310e53d50db287cbf32f62d090bd842236b603d38bb
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:9da8572c79b75767952cf699e5b1699b9f46d83586aa2d6a17fb23c092713b62
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 0.11.2+0.1749674925.p
      entries:
      - name: volsync-product.v0.11.2
        version: 0.11.2
      - name: volsync-product.v0.11.2-0.1749674925.p
        version: 0.11.2+0.1749674925.p
      - name: volsync-product.v0.11.1
        version: 0.11.1
      - name: volsync-product.v0.11.0
        version: 0.11.0
      name: stable-0.11
    - currentCSV: volsync-product.v0.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
          createdAt: "2025-07-07T13:46:18Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.12.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:7de09e5394985cdec87f5afab65b3343b8a37c0eb7f778774b21933ae69d691c
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:7d1ac69fe3d0fddc66acdfce9361a4fc8dd3f549f66c98450f7a0e4c5cc893a0
        version: 0.12.2
      entries:
      - name: volsync-product.v0.12.2
        version: 0.12.2
      - name: volsync-product.v0.12.1
        version: 0.12.1
      - name: volsync-product.v0.12.1-0.1749677898.p
        version: 0.12.1+0.1749677898.p
      - name: volsync-product.v0.12.0
        version: 0.12.0
      name: stable-0.12
    - currentCSV: volsync-product.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationDestination",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationdestination-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationdestination",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationdestination-sample"
                },
                "spec": {
                  "rsync": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "capacity": "10Gi",
                    "copyMethod": "Snapshot",
                    "serviceType": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "volsync.backube/v1alpha1",
                "kind": "ReplicationSource",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "volsync",
                    "app.kubernetes.io/instance": "replicationsource-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "replicationsource",
                    "app.kubernetes.io/part-of": "volsync"
                  },
                  "name": "replicationsource-sample"
                },
                "spec": {
                  "rsync": {
                    "address": "my.host.com",
                    "copyMethod": "Clone",
                    "sshKeys": "secretRef"
                  },
                  "sourcePVC": "pvcname",
                  "trigger": {
                    "schedule": "0 * * * *"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
          createdAt: "2025-08-27T13:12:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.13.1'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/volsync
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A ReplicationDestination is a VolSync resource that you can
              use to define the destination of a VolSync replication or synchronization.
            displayName: Replication Destination
            kind: ReplicationDestination
            name: replicationdestinations.volsync.backube
            version: v1alpha1
          - description: A ReplicationSource is a VolSync resource that you can use
              to define the source PVC and replication mover type, enabling you to
              replicate or synchronize PVC data to a remote location.
            displayName: Replication Source
            kind: ReplicationSource
            name: replicationsources.volsync.backube
            version: v1alpha1
        description: |-
          Asynchronous volume replication for Kubernetes CSI storage

          VolSync is a Kubernetes operator that performs asynchronous replication of
          persistent volumes within, or across, clusters. VolSync supports replication
          independent of the storage system. This means that replication can be used
          with storage systems that do not natively support replication. Data can also be
          replicated across different types (and vendors) of storage.

          VolSync supports both 1:1 replication relationships as well as 1:many
          relationships. This provides the flexibility to support use cases such as:

          - Disaster recovery
          - Mirroring data to a test environment
          - Data distribution to a set of remote clusters from a central site
          - Migrating between storage vendors (changing the StorageClass of a
            persistent volume claim).
          - Creating periodic data backups

          ### How it works

          You specify the details of how, when, and where to replicate the data
          in a ReplicationSource object in the same namespace as the persistent
          volume claim (PVC).

          You create a ReplicationDestination object at the destination, which
          specifies the target for the replicated data.

          VolSync uses multiple replication methods to replicate data:

          - Rclone-based replication for 1:many data distribution:

            Data is replicated from the source to an intermediate cloud storage
            service, which is [supported by Rclone](https://rclone.org/#providers).
            The destinations retrieve the data from the intermediate location.

          - Restic-based backup of PVC contents:

            Data in a PVC is backed up by using the [restic](https://restic.net/)
            program. This method works well when the deployment configuration of
            the application is already source-controlled, and only the
            preservation of its persistent state is needed.

          - Rsync-based replication for one-to-one data replication:

            Data is replicated directly to a remote location. The replication uses
            the [Rsync](https://rsync.samba.org/) utility over an ssh connection
            to securely and efficiently transfer data.

          ### Documentation

          For documentation about installing and using the VolSync Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [VolSync persistent volume replication service](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync) in the Red Hat Advanced Cluster Management
          documentation.

          For additional information about the VolSync project, including features that might not be supported by Red Hat, see
          the [VolSync community](https://volsync.readthedocs.ioe/) documentation.

          ### Support & Troubleshooting

          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: VolSync
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - csi
        - data
        - disaster recovery
        - replication
        - storage
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/business_continuity/business-cont-overview#volsync
        - name: Source
          url: https://github.com/backube/volsync
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat ACM Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/volsync-operator-bundle@sha256:32cd9b5812506656ff9ff1c2c9fdacde44f179313ea2bfbbc18564fee635ca67
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:50840a92f947e739ef4e7cc1106341511221cd22901fc818f4f53697ef73045b
        version: 0.13.1
      entries:
      - name: volsync-product.v0.13.1
        version: 0.13.1
      - name: volsync-product.v0.13.0
        version: 0.13.0
      name: stable-0.13
    defaultChannel: stable
    packageName: volsync-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/sumologic-kubernetes-collection-helm-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle-marketplace@sha256:28f388286bc4b4981182ff1ab9e927f59744dcb755392907fe65cd3b96a44519
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator-rhmp
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-rest-api-browser-v1
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-rest-api-browser-v1.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rest.loadcore.keysight.com/v1alpha1",
                "kind": "Apibrowser",
                "metadata": {
                  "name": "apibrowser-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-06-29T14:08:05Z"
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Apibrowser
            name: apibrowsers.rest.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Rest API Browser Operator v1
        displayName: Keysight Rest API Browser Operator v1
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Rest Api Browser V1
          url: https://keysight-rest-api-browser-v1.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-rest-api-browser-v1@sha256:3ada6a0fa630acb0fa962a11d2c272ff679dded1106c4693d2c8f5ec9ab414a4
        - registry.connect.redhat.com/loadcore/keysight-rest-api-browser-v1-bundle@sha256:89ef7cf0dc8b27e58a63a82a701894a7cc4d0643bf0c8c8e20c9bb59d1a1b1bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        version: 0.1.11
      entries:
      - name: keysight-rest-api-browser-v1.v0.1.11
        version: 0.1.11
      name: alpha
    defaultChannel: alpha
    packageName: keysight-rest-api-browser-v1
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-bundle@sha256:80ca5ad4634305808ac5a3e8ff656ce6ff69b3c16dccc5e03f42763fa89bc762
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/stolostron/multicluster-global-hub
    name: multicluster-global-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-global-hub-operator.v1.4.0-dev.20250302
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: planning
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/postgresql-16:9.5-1732622748
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb-operator:latest
        - quay.io/stolostron/multicluster-global-hub-manager:v1.4.0
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:857c1bdff46d1a1d5abb51f75a80b7944bb461687b179b6f1622103947b8c3b1
        - quay.io/prometheuscommunity/postgres-exporter:v0.15.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-relations/spicedb:latest
        - quay.io/stolostron/grafana:2.12.0-SNAPSHOT-2024-09-03-21-11-25
        - quay.io/stolostron/multicluster-global-hub-agent:v1.4.0
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api:latest
        - quay.io/stolostron/multicluster-global-hub-operator:v1.4.0
        version: 1.4.0-dev.20250302
      entries:
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250302
        version: 1.4.0-dev.20250302
      - name: multicluster-global-hub-operator.v1.4.0-dev.20250225
        version: 1.4.0-dev.20250225
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240911
        version: 1.3.0-dev.20240911
      - name: multicluster-global-hub-operator.v1.3.0-dev.20240724
        version: 1.3.0-dev.20240724
      - name: multicluster-global-hub-operator.v1.2.0-dev.20240516
        version: 1.2.0-dev.20240516
      - name: multicluster-global-hub-operator.v1.2.0-dev
        version: 1.2.0-dev
      - name: multicluster-global-hub-operator.v1.1.0-dev
        version: 1.1.0-dev
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: multicluster-global-hub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
          createdAt: "2025-06-10T06:58:35Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-operator-globalhub-1-5@sha256:4db1251323bd64930461cb5025da41baffcf5c00c51ea4e49940b7130971b510
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/postgres-exporter-globalhub-1-5@sha256:1108572a8bdafde926254f1ff94c2b22e048706be660b07c793e9f9e038a0f4a
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:d101e2b9ab4d8ca7c87c86ea4e3630dcc82a9c5a78885179c8b16773c8aac587
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/glo-grafana-globalhub-1-5@sha256:73fdf00538e63595b7cc92d484e087ff815579a95d7259c579c6e7dc1beacff6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-agent-globalhub-1-5@sha256:6f7520eac83abec7780f831d8473544d0b8db43fc6e77d59e59317b09eb1f6c1
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/multicluster-global-hub-manager-globalhub-1-5@sha256:01253426d65f1297a1955a8393051e6a4b1df5e5249774fc5189a75f35e7d644
        version: 1.5.0
      entries:
      - name: multicluster-global-hub-operator.v1.5.0
        version: 1.5.0
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-1.5
    - currentCSV: multicluster-global-hub-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-03-05T02:13:54Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.32.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:ecdd5c5fc609128e0af751f87b797b71e4ab0200a9abaaca835ff9a3f293fb10
        - quay.io/stolostron/multicluster-global-hub-operator:latest
        version: 1.1.0
      entries:
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.10
    - currentCSV: multicluster-global-hub-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:latest
          createdAt: "2024-07-03T06:52:02Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.8 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:be099e23851cb4b6d15e9c44c4da8fb6cba74aab005597d305757d1dae8b0122
        - quay.io/stolostron/multicluster-global-hub-operator:v1.2.0
        version: 1.2.1
      entries:
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.11
    - currentCSV: multicluster-global-hub-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
          createdAt: "2024-10-10T05:42:48Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          amq-streams-cluster-operator-v2.6.0-2-f49bf7559-569mw   1/1     Running   0          22m
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kafka-0                                           1/1     Running   0          20m
          kafka-kafka-1                                           1/1     Running   0          20m
          kafka-kafka-2                                           1/1     Running   0          20m
          kafka-zookeeper-0                                       1/1     Running   0          21m
          kafka-zookeeper-1                                       1/1     Running   0          21m
          kafka-zookeeper-2                                       1/1     Running   0          21m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgres-0                      2/2     Running   0          22m
          ```
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-user-workloads/acm-multicluster-glo-tenant/release-globalhub-1-3/multicluster-global-hub-manager-globalhub-1-3@sha256:7656d01a4d7b39c61c3249695d68011d9e0df308627b9efd117686e901281ab1
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        - quay.io/stolostron/inventory-api@sha256:a15e62ee070dba5d74cf336b3b89b467593c5a33cfe3623d922f0af4fd7df392
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:715cb5d0c6c407ec5c0303fc7e6d83fc77ce7e06775e169a475f7c075f1e7000
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:636aea861da80fc36fb27a419a18c5b3fbb1bc072e8ae33c75e1cc47e2c05f75
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:96119f61731cec7eb037c3a4a7c2c8bfdf9ef11fd5b08abc25e3d24ddf342b06
        version: 1.3.0
      entries:
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.12
    - currentCSV: multicluster-global-hub-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "global-hub.open-cluster-management.io/v1alpha1",
                "kind": "ManagedClusterMigration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHubAgent",
                "metadata": {
                  "name": "multiclusterglobalhubagent"
                },
                "spec": {
                  "transportConfigSecretName": "transport-config"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
          createdAt: "2025-02-14T08:24:47Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1alpha4",
            "kind":"MulticlusterGlobalHub","metadata":{"name":"multiclusterglobalhub","namespace":"multicluster-global-hub"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterMigration is a global hub resource that allows
              you to migrate managed clusters from one hub to another
            displayName: Managed Cluster Migration
            kind: ManagedClusterMigration
            name: managedclustermigrations.global-hub.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHubAgent is the Schema for the multiclusterglobalhubagents
              API
            displayName: Multicluster Global Hub Agent
            kind: MulticlusterGlobalHubAgent
            name: multiclusterglobalhubagents.operator.open-cluster-management.io
            version: v1alpha1
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the multiCluster global hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes needs to be installed. You can find the support matrix in the official document.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                    READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-68dcd446f4-pg257                  3/3     Running   0          19m
          kafka-kraft-0                                           1/1     Running   0          20m
          kafka-kraft-1                                           1/1     Running   0          20m
          kafka-kraft-2                                           1/1     Running   0          20m
          multicluster-global-hub-grafana-5b7cfb6876-5rcnn        2/2     Running   0          19m
          multicluster-global-hub-grafana-5b7cfb6876-wq99v        2/2     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-hzfvw        1/1     Running   0          19m
          multicluster-global-hub-manager-7f56d78c49-wbkzl        1/1     Running   0          19m
          multicluster-global-hub-operator-768bd5bd4f-59xjd       1/1     Running   0          22m
          multicluster-global-hub-postgresql-0                    2/2     Running   0          22m
          strimzi-cluster-operator-v0.43.0-6fff49d5b5-5j22r       1/1     Running   0          22m
          ```
          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for
          Kubernetes, see [Multicluser GlobalHub Documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.9/html-single/multicluster_global_hub/index#doc-wrapper) in the Red Hat Advanced Cluster Management
          documentation.

          ## Support & Troubleshooting
          Product support, which includes Support Cases, Product Pages, and Knowledgebase articles, is available when you have
          a [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management)
          subscription.
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-agent@sha256:8dfc33dfed33a85733ee5e02dcd4f05662c75263791e0961ba17cbb3d0c340c9
        - quay.io/stolostron/multicluster-global-hub-manager@sha256:4cd602ff6f8dd951848dc8a95c856004295c9f4710f77d4dafcc0735c1492395
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:085e9fc6cdc6a42df10a38856c8ccff25e5acffaba67ef95da72027e90e507fc
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/multicluster-global-hub-operator@sha256:2ea267aac67763b4f160c4e6876c0226d2fa135ee00614814a59b4de1d91f4c4
        - quay.io/prometheuscommunity/postgres-exporter@sha256:386b12d19eab2a37d7cd8ca8b4c7491cc7a830d9581f49af6c98a393da9605e6
        - quay.io/redhat-services-prod/project-kessel-tenant/kessel-inventory/inventory-api@sha256:6a55a1919ecd92d94e05c12be1b6450956f5fda2ce17af7027cac3806c4970a4
        - quay.io/stolostron/grafana@sha256:719b320a05d10158fadce5ae2fc6e4254a18957a6608b6f81e876cf46d3a6897
        version: 1.4.0
      entries:
      - name: multicluster-global-hub-operator.v1.4.0
        version: 1.4.0
      - name: multicluster-global-hub-operator.v1.3.0
        version: 1.3.0
      - name: multicluster-global-hub-operator.v1.2.1
        version: 1.2.1
      - name: multicluster-global-hub-operator.v1.2.0
        version: 1.2.0
      - name: multicluster-global-hub-operator.v1.1.0
        version: 1.1.0
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.13
    - currentCSV: multicluster-global-hub-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha1",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "storage": {
                    "name": "storage-secret"
                  },
                  "transport": {
                    "name": "transport-secret"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
          createdAt: "2022-08-23T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha1
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.5 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\nmulticluster-global-hub-rbac-77888b7dc4-z9vbp
          \                   1/1     Running   0          10s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.5.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:218905f7902e6ac80366856d6c5ae37cd6d167c81e9521df7c3c3d6619f6c425
        version: 0.5.0
      entries:
      - name: multicluster-global-hub-operator.v0.5.0
        version: 0.5.0
      name: release-2.6
    - currentCSV: multicluster-global-hub-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha2",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "name": "transport-secret"
                      },
                      "postgres": {
                        "name": "storage-secret"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
          createdAt: "2023-02-09T20:00:31Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha2
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n- PostgreSQL needs to be installed
          and you must create database for the multicluster global hub. You must create
          a secret named postgresql-secret that contains the database access credential
          in `open-cluster-management` namespace. You can run the following command
          to create the secret: \n\n```\nkubectl create secret generic storage-secret
          -n open-cluster-management \\\n  --from-literal=database_uri=<postgresql-uri>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)
          to install postgres in `hoh-postgres` namespace and create the secret `storage-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n- Kafka needs to be installed,
          and you must create two sections, `spec` and `status`. You need to create
          a secret named kafka-secret that contains the kafka access information should
          be created in `open-cluster-management` namespace. You can run the following
          command to create the secret: \n\n```\nkubectl create secret generic transport-secret
          -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n  --from-literal=CA=<CA-for-kafka-server>\n```\n >_Note:_ There is a
          sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `transport-secret`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\n## How to Install\n\nInstall the
          Multicluster Global Hub Operator by following the instructions that are
          displayed after you select the Install button. The operator must be installed
          in the same namespace as Red Hat Advanced Cluster Management for Kubernetes
          (RHACM), by default, in `open-cluster-management` namespace.\n\nA pod will
          be created in `open-cluster-management` namespace\n\n```\n$ kubectl get
          pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          19s\n```\n\nThe operator is
          now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installing the operator,
          create an instance of the MulticlusterGlobalHub resource to install the
          Multicluster Global Hub.\n\nThe pods of Multicluster Global Hub are created
          in the `open-cluster-management` namespace:\n\n```\n$ kubectl get pods -n
          open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: alpha
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:09fe022370abfcefcf3abfe8b6debc2f201a0ac8b8dacaf1278b8a6bee2b5273
        - quay.io/stolostron/multicluster-global-hub-operator:v0.6.0
        version: 0.6.0
      entries:
      - name: multicluster-global-hub-operator.v0.6.0
        version: 0.6.0
      name: release-2.7
    - currentCSV: multicluster-global-hub-operator.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha3",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "largeScale": {
                      "kafka": {
                        "transportFormat": "cloudEvents"
                      }
                    },
                    "type": "largeScale"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
          createdAt: "2023-07-19T07:53:07Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub is the Schema for the multiclusterglobalhubs
              API
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha3
        description: "Multicluster Global Hub Operator defines the configuration for
          multicluster global hub installation and upgrade with one custom resource.\n\n##
          Prerequisites\n\n- Red Hat Advanced Cluster Management for Kubernetes (RHACM)
          2.7 or later needs to be installed.\n## How to Install\n\nInstall the Multicluster
          Global Hub Operator by following the instructions that are displayed after
          you select the Install button. The operator must be installed in the same
          namespace as Red Hat Advanced Cluster Management for Kubernetes (RHACM),
          by default, in `open-cluster-management` namespace.\n\nA pod will be created
          in `open-cluster-management` namespace\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME
          \                                                            READY   STATUS
          \   RESTARTS   AGE\nmulticluster-global-hub-operator-5ccbbc878d-98fdp                1/1
          \    Running   0          19s\n```\n\nThe operator is now providing new
          Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`\n\n##
          Using the Multicluster Global Hub Operator\n\nAfter installed the operator,
          create an instance of the MulticlusterGlobalHub resource to instantiate
          the Multicluster Global Hub.\n\nYou need to prepare:\n\n- PostgreSQL needs
          to be installed and you must create database for the multicluster global
          hub. You must create a secret named multicluster-global-hub-storage that
          contains the database access credential in `open-cluster-management` namespace.
          You can run the following command to create the secret: \n\n```\nkubectl
          create secret generic multicluster-global-hub-storage -n open-cluster-management
          \\\n  --from-literal=database_uri=<postgresql-uri> \\\n --from-file=ca.crt=<CA-for-postgres-server>
          \n```\n>_Note:_ There is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/storage)(Note:
          the client version of kubectl must be v1.21+) to install postgres in `hoh-postgres`
          namespace and create the secret `multicluster-global-hub-storage` in namespace
          `open-cluster-management` automatically. To override the secret namespace,
          set `TARGET_NAMESPACE` environment variable to the RHACM installation namespace
          before executing the script.\n\n- Kafka needs to be installed, and you must
          create three topics, `spec` `status` and `event`. You need to create a secret
          named multicluster-global-hub-transport that contains the kafka access information
          should be created in `open-cluster-management` namespace. You can run the
          following command to create the secret: \n\n```\nkubectl create secret generic
          multicluster-global-hub-transport -n open-cluster-management \\\n  --from-literal=bootstrap_server=<kafka-bootstrap-server-address>
          \\\n      --from-file=ca.crt=<CA-cert-for-kafka-server> --from-file=client.crt=<Client-cert-for-kafka-server>
          --from-file=client.key=<Client-key-for-kafka-server> \n```\n >_Note:_ There
          is a sample script available [here](https://github.com/stolostron/multicluster-global-hub/tree/main/operator/config/samples/transport)
          to install kafka in `kafka` namespace and create the secret `multicluster-global-hub-transport`
          in namespace `open-cluster-management` automatically. To override the secret
          namespace, set `TARGET_NAMESPACE` environment variable to the RHACM installation
          namespace before executing the script.\n\nOnce an instance of the MulticlusterGlobalHub
          is created, the following pods are created in the `open-cluster-management`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management\nNAME                                                             READY
          \  STATUS    RESTARTS   AGE\nmulticluster-global-hub-grafana-79f8d59f58-qwcjl
          \                1/1     Running   0          22s\nmulticluster-global-hub-operator-5ccbbc878d-98fdp
          \               1/1     Running   0          56s\nmulticluster-global-hub-manager-7c55644568-mth8f
          \                1/1     Running   0          11s\n```\n"
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: beta
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/stolostron/multicluster-global-hub-operator:v0.7.0
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:5b865231214649649a2a9fb06934f931503f68ed740e150be1c7c8fcbb6f1317
        version: 0.7.0
      entries:
      - name: multicluster-global-hub-operator.v0.7.0
        version: 0.7.0
      name: release-2.8
    - currentCSV: multicluster-global-hub-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1alpha4",
                "kind": "MulticlusterGlobalHub",
                "metadata": {
                  "name": "multiclusterglobalhub"
                },
                "spec": {
                  "dataLayer": {
                    "postgres": {
                      "retention": "18m"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
          createdAt: "2023-10-16T03:07:33Z"
          description: Manages the installation and upgrade of the Multicluster Global
            Hub.
          operatorframework.io/suggested-namespace: multicluster-global-hub
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/multicluster-global-hub
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterGlobalHub defines the configuration for an instance
              of the MultiCluster Global Hub
            displayName: Multicluster Global Hub
            kind: MulticlusterGlobalHub
            name: multiclusterglobalhubs.operator.open-cluster-management.io
            version: v1alpha4
        description: |
          The Multicluster Global Hub Operator contains the components of multicluster global hub. The Operator deploys all of the required components for global multicluster management. The components include `multicluster-global-hub-manager` and `multicluster-global-hub-grafana` in the global hub cluster and `multicluster-global-hub-agent` in the managed hub clusters.
          The Operator also deploys the strimzi kafka and crunchy postgres if you do not bring your own kafka and postgres.

          ## Prerequisites
          - Red Hat Advanced Cluster Management for Kubernetes (RHACM) 2.7 or later needs to be installed.

          ## How to Install
          Install the Multicluster Global Hub Operator by following the instructions that are displayed after you select the Install button.
          A pod will be created in `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          multicluster-global-hub-operator-5ccbbc878d-98fdp                1/1     Running   0          19s
          ```
          The operator is now providing new Custom Resources Definitions: `multiclusterglobalhubs.operator.open-cluster-management.io`

          ## Using the Multicluster Global Hub Operator
          After installed the operator, create an instance of the MulticlusterGlobalHub resource to instantiate the Multicluster Global Hub.
          Once an instance of the MulticlusterGlobalHub is created, the following pods are created in the `multicluster-global-hub` namespace:
          ```
          $ kubectl get pods -n multicluster-global-hub
          NAME                                                             READY   STATUS    RESTARTS   AGE
          kafka-entity-operator-54ff95c7dc-cmmwc                           3/3     Running   0          21h
          kafka-kafka-0                                                    1/1     Running   0          21h
          kafka-kafka-1                                                    1/1     Running   0          21h
          kafka-kafka-2                                                    1/1     Running   0          21h
          kafka-zookeeper-0                                                1/1     Running   0          21h
          kafka-zookeeper-1                                                1/1     Running   0          21h
          kafka-zookeeper-2                                                1/1     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-mp7fm                 2/2     Running   0          21h
          multicluster-global-hub-grafana-579cdf759d-4hpcg                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-5np24                 2/2     Running   0          21h
          multicluster-global-hub-manager-7bdcdd745f-n9b9n                 2/2     Running   0          21h
          multicluster-global-hub-operator-59744b8d75-nrwvd                1/1     Running   0          21h
          multicluster-global-hub-postgres-0                               1/1     Running   0          21h
          strimzi-cluster-operator-v0.36.1-684d7cccb7-zjbf9                1/1     Running   0          21h
          ```

          ## Documentation
          For documentation about installing and using the Multicluster GlobalHub Operator with Red Hat Advanced Cluster Management for Kubernetes, see [Multicluser Global Hub Documentation](https://github.com/stolostron/multicluster-global-hub/tree/main/doc#readme).
        displayName: Multicluster Global Hub Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multicluster-global-hub
        - multiple-hubs
        links:
        - name: Multicluster Global Hub Operator
          url: https://github.com/stolostron/multicluster-global-hub
        maintainers:
        - email: acm-contact@redhat.com
          name: acm-contact
        maturity: stable
        provider:
          name: Red Hat, Inc
          url: https://github.com/stolostron/multicluster-global-hub
        relatedImages:
        - quay.io/openshift-community-operators/multicluster-global-hub-operator@sha256:16245de1e88228c231ab2edb4e9380753eff982bb72febadd00cdaad286a8eb3
        - quay.io/stolostron/multicluster-global-hub-operator:v1.0.0
        version: 1.0.0
      entries:
      - name: multicluster-global-hub-operator.v1.0.0
        version: 1.0.0
      name: release-2.9
    defaultChannel: release-1.5
    packageName: multicluster-global-hub-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/stolostron/multicluster-global-hub
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: mercury-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mercury-operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:1.0.1",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CustomerCreditRating",
                  "expose": ["http"]
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:1.0.1
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:1.0.1
        - quay.io/openshift-community-operators/mercury-operator@sha256:0c253fd002cb389db40b2438d4d1c9e953a84125dbf0f4b6f5e3a6efd87cf3a6
        version: 1.0.1
      entries:
      - name: mercury-operator.1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: dev
    - currentCSV: mercury-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomainInfra",
                "metadata": {
                  "name": "my-example-sdi",
                  "labels": {
                    "app": "examples"
                  }
                },
                "spec": {
                  "kafka": {
                    "replicas": 3,
                    "storage": {
                      "type": "persistent-claim",
                      "size": "100Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-credit-rating",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-credit-rating-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_CREDIT_RATING",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-customer-offer",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/customer-offer-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "CUSTOMER_OFFER",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              },
              {
                "apiVersion": "mercury.redhat.io/v1alpha1",
                "kind": "ServiceDomain",
                "metadata": {
                  "name": "example-party-routing-profile",
                  "labels": {
                    "app": "example"
                  }
                },
                "spec": {
                  "businessImage": "quay.io/ecosystem-appeng/party-routing-profile-example:v1.0.2",
                  "serviceDomainInfra": "my-example-sdi",
                  "type": "PARTY_ROUTING_PROFILE",
                  "expose": {
                    "http": {
                      "api-version": "v1"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-example-sdi": {
                "description": "Example Service Domain Infrastructure"
            },
            {
             "example-customer-credit-rating": {
               "description": "Example Customer Credit Rating Service Domain"
            },
            {
             "example-customer-offer": {
               "description": "Example Customer Offer Service Domain"
            },
            {
             "example-party-routing-profile": {
               "description": "Example Party Routing Profile Service Domain"
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/ecosystem-appeng/mercury-operator:v1.0.2
          description: Deploys and manages BIAN Service Domains with the Mercury framework
          repository: https://github.com/open-accelerators/mercury
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A Service Domain Infra represents a group of Service Domains that interact between themselves and
              share the same infrastructure resources
            displayName: Service Domain Infra
            kind: ServiceDomainInfra
            name: servicedomaininfras.mercury.redhat.io
            version: v1alpha1
          - description: |-
              A Service Domain is a standalone service representing one of the BIAN defined Service Domains.
              e.g. Customer Offer. A Service Domain must belong to a Service Domain Infra
            displayName: Service Domain
            kind: ServiceDomain
            name: servicedomains.mercury.redhat.io
            version: v1alpha1
          required:
          - description: Represents a Kafka cluster
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Integration is the Schema for the Camel K integrations API
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
        description: |-
          # An open platform for a modular, digital bank

          Project Mercury aims at developing a solution approach for the modernization of core banking systems, exploring new technologies and ways of operating in an open hybrid cloud. Creating and enabling the foundation of an open platform for modular, digital banking. Using open-source software and open standards currently available, combined with industry best practices.

          The provided reference implementation can serve as an example. While some resulting artifacts leverage the expertise of developing this solution approach can become reusable templates, repeatable processes, and building blocks of cloud-native components and applications.

          We invite the collaboration of technology experts and experienced business practitioners across the financial services industry to join and participate in co-creation.
        displayName: Mercury Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bian
        - mercury
        links:
        - name: GitHub Repository
          url: https://github.com/open-accelerators/mercury
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/ecosystem-appeng/mercury-operator:v1.0.2
        - quay.io/openshift-community-operators/mercury-operator@sha256:655bc42f0a752377d349d20d14ef9a91a920436c16102f5854187e1c83e80446
        version: 1.0.2
      entries:
      - name: mercury-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: 1.0.x
    packageName: mercury-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:ccd64090b85536ddaf1b81ca07d40279c9f277b1c5911628f11d678b98dc3bad
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:f0e9f64c353fdbee3bf7ec0eaf070c3b75330ef659a881ec4ebc874ba584d085
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:9db455e082454734d2a17811dcfec53d3d60c28df6e7fe8bc979a45b28a35061
        version: 0.5.3
      entries:
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      name: threescale-2.11
    - currentCSV: apicast-operator.v0.6.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:12f7a8e69a1369a51bbcd2b34d4ec8c66dd1fc9fb515349a04e7b202953423e3
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:8650430afd813fd749fcd80ad5f6fd0aeaab3df048e8fe34e6aa72a07ab801fc
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:8c7fe66b8d4044ad965e7b3d149e9d8e3846ac4e55465873f7095a25b297b3df
        version: 0.6.4
      entries:
      - name: apicast-operator.v0.6.4
        version: 0.6.4
      - name: apicast-operator.v0.6.3
        version: 0.6.3
      - name: apicast-operator.v0.6.2
        version: 0.6.2
      - name: apicast-operator.v0.6.1
        version: 0.6.1
      - name: apicast-operator.v0.6.0
        version: 0.6.0
      - name: apicast-operator.v0.5.3
        version: 0.5.3
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-2.12
    - currentCSV: apicast-operator.v0.7.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.7.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.13/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: msoriano+apicastoperator@redhat.com
          name: Miguel Soriano
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:88dc7a289c803248efbe151999707325b1bc0aac2d26827ea1707e46e5c71652
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:ed063f507fa8e2f378d0adbde823c110484d0212d124cc72d67efa987e86444e
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:bde44c0fb9f13f212184b24e0bef45db734a8439a92179a2089e5aebd6b87b63
        version: 0.7.5
      entries:
      - name: apicast-operator.v0.7.5
        version: 0.7.5
      name: threescale-2.13
    - currentCSV: apicast-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          olm.skipRange: '>=0.6.2 <0.8.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator@sha256:21db8c1bde63b5d574b66b46ec3ee95eb05a3c34b9541ca32d5bbf4daf18daf4
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:68b310ed3cfd65db893ba015ef1d5442365201c0ced006c1915e90edb99933ea
        - registry.redhat.io/3scale-amp2/apicast-rhel7-operator-metadata@sha256:bcc5ed9560667c7347f3044c7fa6148e3867c17e93e83595a62794339a5a6b0d
        version: 0.8.1
      entries:
      - name: apicast-operator.v0.8.1
        version: 0.8.1
      name: threescale-2.14
    - currentCSV: apicast-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.8.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.14/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/3scale-amp2/apicast-rhel9-operator@sha256:c4d6acae4e4cc3236c5b8d8dcfb339b072688ab114785d8598bbe68b1f3571a9
        - registry.redhat.io/3scale-amp2/3scale-apicast-operator-bundle@sha256:29fe669a56a69adb537798c91b9359fa3b051da5ffa9f27789a2067a3a72330a
        version: 0.12.5
      entries:
      - name: apicast-operator.v0.12.5
        version: 0.12.5
      - name: apicast-operator.v0.12.4
        version: 0.12.4
      - name: apicast-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: apicast-operator.v0.7.9-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway to be used together with Red Hat
            3scale
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |-
          APIcast is an API gateway to be used together with Red Hat 3scale. This Operator supports standalone deployment of the gateway as opposed to the built-in APIcast as part of a 3scale installation.

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator) for more information.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale APIcast gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/installing-apicast#deploying-apicast-gateway-self-managed-operator
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:ea2229948fea841343fcaa337bb7e2854acf5f563d3ffb0af8bda1c21ec2a91f
        - registry.redhat.io/3scale-mas/apicast-operator-bundle@sha256:d3a608b37aa089ea46e803696ba4957585b875f8a810f517e891f385ad105937
        - registry.redhat.io/3scale-mas/apicast-rhel7-operator@sha256:3fd03155935a1a1946669db9dadc8e82737de46e1fcb0ba92a9815a277ae35aa
        version: 0.7.9-mas
      entries:
      - name: apicast-operator.v0.7.9-mas
        version: 0.7.9-mas
      - name: apicast-operator.v0.7.8-mas
        version: 0.7.8-mas
      - name: apicast-operator.v0.7.7-mas
        version: 0.7.7-mas
      - name: apicast-operator.v0.7.6-mas
        version: 0.7.6-mas
      - name: apicast-operator.v0.7.5-mas
        version: 0.7.5-mas
      - name: apicast-operator.v0.7.4-mas
        version: 0.7.4-mas
      - name: apicast-operator.v0.7.3-mas
        version: 0.7.3-mas
      - name: apicast-operator.v0.7.2-mas
        version: 0.7.2-mas
      - name: apicast-operator.v0.7.1-mas
        version: 0.7.1-mas
      - name: apicast-operator.v0.7.0-mas
        version: 0.7.0-mas
      - name: apicast-operator.v0.6.0-mas
        version: 0.6.0-mas
      - name: apicast-operator.v0.5.2-0.1649688685.p
        version: 0.5.2+0.1649688685.p
      - name: apicast-operator.v0.5.2-0.1646742993.p
        version: 0.5.2+0.1646742993.p
      - name: apicast-operator.v0.5.2
        version: 0.5.2
      - name: apicast-operator.v0.5.1
        version: 0.5.1
      - name: apicast-operator.v0.5.0
        version: 0.5.0
      - name: apicast-operator.v0.5.0-0.1638984223.p
        version: 0.5.0+0.1638984223.p
      - name: apicast-operator.v0.5.0-0.1638885516.p
        version: 0.5.0+0.1638885516.p
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: apicast-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-storage
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-storage.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Storage",
                "metadata": {
                  "name": "storage-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "minio": {
                    "clusterDomain": "cluster.local",
                    "helmKubectlJqImage": {
                      "repository": "docker.io/bskim45/helm-kubectl-jq"
                    },
                    "image": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio",
                      "tag": "RELEASE.2021-12-29T06-49-06Z-3747-37-ocp"
                    },
                    "mcImage": {
                      "repository": "prod.harbor.keysight.digital/loadcore/wap-minio-mc",
                      "tag": "RELEASE.2021-12-29T06-52-55Z-3747-37-ocp"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "securityContext": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:32:20Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Storage
            name: storages.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Storage Operator
        displayName: Keysight WAP Storage Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Storage
          url: https://keysight-wap-storage.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-storage@sha256:9f7710cfe72658cd5456fd54f86ca6be5c9a4de2a01114c0f7531b70712bd981
        - registry.connect.redhat.com/loadcore/keysight-wap-storage-bundle@sha256:0955dbda03bd7307364a1a458ab8d6ac5a8c9feb6150706dcede9676a6513b0d
        version: 0.1.21
      entries:
      - name: keysight-wap-storage.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-storage
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: open-liberty-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: open-liberty-operator.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploy and manage applications running on Liberty
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          This advanced Operator can be used to deploy and manage Open Liberty applications with consistent, production-grade QoS. This operator is based on the [Runtime Component Operator](https://github.com/application-stacks/runtime-component-operator) and provides all of its capabilities in addition to Open Liberty specific features such as gathering traces and dumps (Day-2 operations) and easily configuring and managing the single sign-on information for your Open Liberty applications.

          Open Liberty Operator enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!

          Here are some key features:

          #### Application Lifecyle
          You can deploy your Open Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          You can also easily configure and use a single storage for serviceability related Day-2 operations, such as gatherig server traces and dumps.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Single Sign-On (SSO)
          Open Liberty provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows to easily configure and manage the single sign-on information for your applications.

          #### Exposing metrics to Prometheus
          Expose the Open Liberty application's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `spec.monitoring` field to control configurations such as poll interval and security credentials.

          #### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply add the following configuration (to specify the volume size and the location to persist) to your OpenLibertyApplication CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc/) for more information.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://github.com/OpenLiberty/open-liberty-operator/tree/main/doc
        - name: Open Liberty
          url: https://openliberty.io/
        - name: Open Liberty Container Images
          url: https://github.com/OpenLiberty/ci.docker
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:591fc4b5867dff26df1952653423b33889c0f56ba3f756057694d4c0e396eeb0
        - registry.connect.redhat.com/ibm/open-liberty-operator-controller@sha256:18f188e48725c785557c8edea46f69c464d6f81ab47860a3bb0ec978c20a26db
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: open-liberty-operator.v0.8.1
        version: 0.8.1
      - name: open-liberty-operator.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: open-liberty-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread",
                    "heap"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
          createdAt: "2023-08-28T16:39:10Z"
          description: Deploy and manage containerized Liberty applications
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d45c296d43dc4c4100f38bb736f073e04c8b4d72b68ddf1189e311aff8ad66e3
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:3225c49cf30a146784533496ce478466034d4549ff9721752958bac342fdec77
        version: 1.2.2
      entries:
      - name: open-liberty-operator.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: open-liberty-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
          createdAt: "2024-08-01T20:19:13Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:d68bb66a8133702a198f7be655d7cff5f500bb620c460a81547971e00124ef73
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:b0fa6b0356d83626509289b33dfa669bd7d32542aba986a18d296b85ce63389a
        version: 1.3.3
      entries:
      - name: open-liberty-operator.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: open-liberty-operator.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
          createdAt: "2025-07-18T20:45:06Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty-operator@sha256:67b5b46dfed008dab4ea7fd176d3bc31dae8940d72194a10aebb98b1dea9043e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:fe759127885502dc66423530ff03e37973355c5566a1c27e988b59d71528dd2b
        version: 1.4.4
      entries:
      - name: open-liberty-operator.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: open-liberty-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyPerformanceData",
                "metadata": {
                  "name": "openliberty-performance-data-sample"
                },
                "spec": {
                  "interval": 30,
                  "podName": "Specify_Pod_Name_Here",
                  "timespan": 240
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyApplication",
                "metadata": {
                  "name": "openliberty-app-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyDump",
                "metadata": {
                  "name": "openliberty-dump-sample"
                },
                "spec": {
                  "include": [
                    "thread"
                  ],
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "apps.openliberty.io/v1beta2",
                "kind": "OpenLibertyTrace",
                "metadata": {
                  "name": "openliberty-trace-sample"
                },
                "spec": {
                  "maxFileSize": 20,
                  "maxFiles": 5,
                  "podName": "Specify_Pod_Name_Here",
                  "traceSpecification": "*=info:com.ibm.ws.webcontainer*=all"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
          createdAt: "2025-09-16T03:08:31Z"
          description: Deploy and manage containerized Liberty applications
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/OpenLiberty/open-liberty-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1
          - description: Represents the deployment of an Open Liberty application
            displayName: OpenLibertyApplication
            kind: OpenLibertyApplication
            name: openlibertyapplications.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1
          - description: Day-2 operation for generating server dumps
            displayName: OpenLibertyDump
            kind: OpenLibertyDump
            name: openlibertydumps.apps.openliberty.io
            version: v1beta2
          - description: Day-2 operation for generating server performance data
            displayName: OpenLibertyPerformanceData
            kind: OpenLibertyPerformanceData
            name: openlibertyperformancedata.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1
          - description: Day-2 operation for gathering server traces
            displayName: OpenLibertyTrace
            kind: OpenLibertyTrace
            name: openlibertytraces.apps.openliberty.io
            version: v1beta2
        description: |
          Open Liberty is a lightweight open framework for building fast and efficient cloud-native Java microservices. Open Liberty is the most flexible server runtime available to Java developers in this solar system. It is optimized for modern cloud technologies and practices, making it an ideal choice for container and Kubernetes-based deployments.

          Open Liberty Operator allows you to deploy and manage containerized Liberty applications securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform Day-2 operations such as gathering traces and memory dumps easily using the operator.
          ## Documentation
          See our [**documentation**](https://ibm.biz/olo-docs) for more information. If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/olo-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Application Lifecycle
          You can deploy your Liberty application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Management of Lightweight Third-Party Authentication (LTPA)
          LTPA provides a method of SSO configuration to authenticate users to access applications. The Liberty operator can generate and manage the LTPA key for applications. A single LTPA key can be used per each namespace and shared with participating microservices and applications to allow users to authenticate only once.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Integration with Semeru Cloud Compiler
          Use a single toggle to stand up a compatible Semeru Cloud Compiler instance to handle compilation requests from the multiple instances of the same application. Liberty Operator automatically sets up encrypted communication between the compiler and the application. It also handles application upgrade scenarios gracefully to ensure compiler and application instances remain compatible.
          ### Exposing metrics to Prometheus
          Expose the Liberty application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the OpenLibertyApplication's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Single Sign-On (SSO)
          Liberty runtime provides capabilities to delegate authentication to external providers. Your application users can log in using their existing social media credentials from providers such as Google, Facebook, LinkedIn, Twitter, GitHub, and any OpenID Connect (OIDC) or OAuth 2.0 clients. Open Liberty Operator allows you to easily configure and manage the single sign-on information for your applications.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the OpenLibertyApplication's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your applications can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless applications using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the Liberty server.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode. You can also easily configure and use single storage for serviceability-related day-2 operations, such as gathering server traces and dumps.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Open Liberty
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Open Liberty
        - MicroProfile
        - Jakarta EE
        - Java
        - Application Server
        - Application Runtime
        - Runtime
        - Spring
        - Application
        - Open Source
        - OpenLiberty
        - OLO
        links:
        - name: Documentation
          url: https://ibm.biz/olo-docs
        - name: Open Liberty
          url: https://openliberty.io
        - name: Open Liberty Container Images
          url: https://ibm.biz/ol-images
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - registry.connect.redhat.com/ibm/open-liberty-operator-bundle@sha256:c0336e5079f255ab6cd38e4e2e61d0cb4c6a35a976528fb44409bfde57e70dde
        - icr.io/appcafe/open-liberty-operator@sha256:c3b031fe14054f65d6612cc505b7681336873dc32ce1091451d8c7799034bc83
        version: 1.5.0
      entries:
      - name: open-liberty-operator.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: open-liberty-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Quantum
      provider-url: ""
    name: openshift-qiskit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-qiskit-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "qiskit.ibm.com/v1alpha1",
                "kind": "QiskitPlayground",
                "metadata": {
                  "name": "qiskitplayground-sample"
                },
                "spec": {
                  "image": "qiskit/jupyter:0.1",
                  "imagePullPolicy": "Always",
                  "resources": {
                    "requests": {
                      "memory": "2G",
                      "cpu": "1"
                    },
                    "limits": {
                      "memory": "3G",
                      "cpu": "1.5"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: docker.io/qiskit/operator:0.1
          createdAt: 2021-004-23 22:25:53
          description: |-
            Launches a developement environment Jupyter notebok development environment for
            developing quantum applications using Qiskit.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
          support: https://github.com/qiskit-community/openshift-quantum-operators/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: QiskitPlayground is the Schema for the qiskitplaygrounds
              API
            displayName: Qiskit Playground
            kind: QiskitPlayground
            name: qiskitplaygrounds.qiskit.ibm.com
            version: v1alpha1
        description: |-
          ### Introduction
          Launches a developement Jupiter lab environment with Qiskit.

          QiskitPlayground is a Jupyter notebook with the pre installed qiskit libraries for implementing quantum application development.
          It comes with pre installed python packages for visualising results. QiskitPlayground can run quantum
          applications on simulator or on real quantum machines using IBM Q account.

          ### Documentation
          For detailed installation guide [follow](https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator)

          ### Contributing
          You can contribute by
            * Raising any [issues](https://github.com/qiskit-community/openshift-quantum-operators/issues) you find using ibm-quantum-operator community operator
            * Fixing issues by opening [Pull Requests](https://github.com/qiskit-community/openshift-quantum-operators/pulls)
            * Talking about openshift-qiskit-operator

          ### License
          kubernetes-qiskit-operator is licensed under the [Apache 2.0 license](https://github.com/qiskit-community/openshift-quantum-operators/blob/master/LICENSE)

          #### Note
          Any api could be changed any time without any notice. That said, your feedback is very important and appreciated to make this project more stable and useful.
        displayName: QiskitPlayground
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Qiskit
        - Open Source
        - Quantum
        links:
        - name: GitHub
          url: https://github.com/qiskit-community/openshift-quantum-operators/tree/master/operators-examples/kubernetes-qiskit-operator
        maintainers:
        - email: blublinsky@ibm.com
          name: boris
        maturity: alpha
        provider:
          name: IBM Quantum
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/openshift-qiskit-operator@sha256:633ce4190b02034d70b52d7e624fc9d417ef997607f9710a36e73bbc0963fe4d
        - docker.io/qiskit/operator:0.1
        version: 0.2.1
      entries:
      - name: openshift-qiskit-operator.v0.2.1
        version: 0.2.1
      - name: openshift-qiskit-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: openshift-qiskit-operator
    provider:
      name: IBM Quantum
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jason Skrzypek
      provider-url: https://github.com/shpwrck
    name: openshift-node-upgrade-mutex-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-node-upgrade-mutex-operator.v0.2.66
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrule-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrule",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrule-sample"
                },
                "spec": {
                  "jobName": "example",
                  "jobNamespace": "node-upgrade-mutex-operator",
                  "jobSpec": {
                    "selector": {},
                    "template": {
                      "metadata": {
                        "name": "mutex"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              "20"
                            ],
                            "image": "ubi9/toolbox",
                            "name": "mutex"
                          }
                        ],
                        "restartPolicy": "Never"
                      }
                    }
                  },
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexRunConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutexrunconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutexrunconfig",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutexrunconfig-sample"
                },
                "spec": {
                  "canary": {
                    "name": "canary-example"
                  },
                  "mutexRules": [
                    {
                      "name": "mutexrule-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ],
                  "mutexTargets": [
                    {
                      "name": "mutextarget-sample",
                      "namespace": "node-upgrade-mutex-operator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "openshift-optional.skrzypek.io/v1alpha1",
                "kind": "MutexTarget",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openshift-node-upgrade-mutex-operator",
                    "app.kubernetes.io/instance": "mutextarget-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mutextarget",
                    "app.kubernetes.io/part-of": "openshift-node-upgrade-mutex-operator"
                  },
                  "name": "mutextarget-sample"
                },
                "spec": {
                  "machineConfigPools": [
                    "worker"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/jskrzypek/openshift-node-upgrade-mutex-operator
          createdAt: "2025-05-30T17:49:00Z"
          description: |
            This software injects required operations into the worker node upgrade process.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: node-upgrade-mutex-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          provider: Jason Skrzypek
          repository: https://github.com/shpwrck/openshift-upgrade-rollout-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A node mutex rule definition in the form of a kubernetes
              job spec
            displayName: Mutex Rule
            kind: MutexRule
            name: mutexrules.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: Lists of Targets and Rules to apply to them, as well as an
              upgrade facilitating Machine Config Pool
            displayName: Mutex Run Config
            kind: MutexRunConfig
            name: mutexrunconfigs.openshift-optional.skrzypek.io
            version: v1alpha1
          - description: A list of Machine Config Pools to enforce upgrade procedure.
            displayName: Mutex Target
            kind: MutexTarget
            name: mutextargets.openshift-optional.skrzypek.io
            version: v1alpha1
        description: |
          ## Introduction

          This operator injects required operations in the form of a Kubernetes job into the worker node upgrade process using MachineConfigPools. In short, when provided a list of operations (`MutexRules`) and a list of MachineConfigPools (`MutexTargets`) , the OpenShift Node Upgrade Mutex Operator will execute each operation for a given target, enable updates for the given target by modifying it's MachineConfigPool membership, and wait until the node has been upgraded before proceeding onto subsequent operations and targets.

          ## Dependencies

          ### Release Version Label

          In order to confirm the successful upgrade of each target node, the NFD Operator is used to generate a version feature on every worker node. In order to proceed through given targets, each node must have this label
          `'machineconfiguration.openshift.io/release-image-version'` with a key providing the semantic version of this node (e.g. `'4.16.4'`)

          There are several ways to accomplish this, but an example MachineConfig has been provided [here](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator/blob/main/machineconfig.yaml).

          ### Canary Machine Config Pool

          In addition to node labels, a MachineConfigPool that will be used to accomodate targets subject to upgrades must be provided matching the following criteria:
          * selects only worker nodes
          * is not paused

          Applying the example below will create a compatible canary pool:

          ```yaml
          ---
          apiVersion: machineconfiguration.openshift.io/v1
          kind: MachineConfigPool
          metadata:
            name: canary
          spec:
            machineConfigSelector:
              matchLabels:
                machineconfiguration.openshift.io/role: worker
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ''
              matchExpressions:
                - key: 'isCanary'
                  operator: Exists
            paused: false
          ```

          ## Usage / Custom Resources Definitions

          #### MutexRule:

          A resource storing a single rule referring to a template for a Kubernetes job to be created.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRule
          metadata:
            name: k8s-example
          spec:
            jobName: "example" # Jobs will be run with this prefix name
            jobNamespace: "node-upgrade-mutex-operator" # Jobs will be run in this namespace
            jobSpec: # Accepts full 'batch/v1' job spec
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            cleanUpJobSpec: # Accepts full 'batch/v1' job spec (Optional)
              selector: {}
              template:
                metadata:
                  name: mutex
                spec:
                  containers:
                  - command:
                    - sleep
                    args:
                    - "30"
                    image: ubi9/toolbox
                    name: mutex
                  restartPolicy: Never
            type: kubernetes
          ```

          #### MutexTarget:

          A resource storing a set of MachineConfigPools to pause and incrementally upgrade.
          This resource **does not** modify existing resources.

          ```yaml
          ---
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexTarget
          metadata:
            name: worker-example
          spec:
            machineConfigPools:
              - worker # Name of existing machineconfigpool
          ```

          #### MutexRunConfig:

          A resource that references existing MutexRules and MutexTargets as well as an existing MachineConfigPool that will allow members to be upgraded. Every rule defined in the `mutexRules` list will be added as a mandatory task to every node that is a member of the each MachineConfigPool defined in the `mutexTargets` list.
          This resource **does** modify existing resources.

          ```yaml
          apiVersion: openshift-optional.skrzypek.io/v1alpha1
          kind: MutexRunConfig
          metadata:
            name: example
          spec:
            skipTLSVerify: false # Defaults to false if omitted
            canary:
              name: canary # Name of unpaused machineconfigpool to modify
            mutexRules: # List of existing mutexRules
              - name: k8s-example
                namespace: node-upgrade-mutex-operator
            mutexTargets: # List of existing mutexTargets
              - name: worker-example
                namespace: node-upgrade-mutex-operator
          ```

          ## Procedure

          #### MutexRule

          1. Upon creation/modification, the schema of each K8s rule is validated.
          2. Status is updated to indicate valid or invalid configuration.

          #### MutexTargets

          1. Upon creation/modification, the existence of each target is validated.
          2. Status is updated to indicate valid or invalid configuration.
          3. The nodes assigned to each target are added to the status field `nodes`.

          #### MutexRunConfigs

          1. Upon creation/modification, the existence as well as validation status of each rule, target, and canary pool is confirmed.
          2. If valid, the existence of an upgrade is queried. An upgrade is available if the following is true:
              * The `status.history` of the `cluster` ClusterVersion equals `Completed`
              * The `release-image-version` of any worker node is less than the `spec.desiredUpdate.version` of the `cluster` ClusterVersion.
          3. If an update exists, each target MachineConfigPool has `spec.paused` set to `true`.
          4. If each pool is paused, each node is updated to include a label for each rule with the following format:
              * `mutexrule.openshift-optional.skrzypek.io/{{MutexRule_Name}}.{{MutexRule_Namespace}}`
          5. If labels have been applied, the provided canary pool is modified to exclude nodes with labels provided by step (4).
          6. Assuming steps (2-5) have completed successfully, then the process for running jobs and updating nodes begins. The process per node is as follows:
              * Launch each K8S job sequentially
              * Remove the current node's rule labels
              * Wait until the `release-image-version` of the node matches the `spec.desiredUpdate.version` of the `cluster` ClusterVersion
              * Launch each K8S cleanup job sequentially
          7. If the update succeeds, all machine config pools are unpaused and all match expressions are removed.

          ## Permissions Granted

          In addition to the permissions required to operate the included CRDs, the following access is granted to the operator's service account.

          | Resource | Permissions |
          | :---: | :---: |
          | Jobs | Create & Read |
          | Nodes | Read & Update |
          | ClusterVersion | Read |
          | MachineConfigPools | Read & Update |

          ## Notes

          * This operator was built using the operator-sdk's ansible plugin and therefore has all of the inherited capabilities and restrictions.
          * This operator has undergone minimal testing and should not be run in production environments without acknowledging the risks.
          * This operator is currently alpha and subject to change based on requirements.

          ## Additional Information
          For additional information refer to the [GitHub Repository](https://github.com/shpwrck/openshift-node-upgrade-mutex-operator)
        displayName: OpenShift Node Upgrade Mutex Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenShift Optional
        - Upgrade
        - Mutex
        links:
        - name: Openshift Node Upgrade Mutex Operator
          url: https://github.com/shpwrck/openshift-node-upgrade-mutex-operator
        maintainers:
        - email: jskrzype@redhat.com
          name: Jason Skrzypek
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Jason Skrzypek
          url: https://github.com/shpwrck
        relatedImages:
        - quay.io/jskrzypek/openshift-node-upgrade-mutex-operator@sha256:d4c28e8b256c00429d9999595c2386f1a25de597ee8e6c33871bb7b22b84c4af
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/community-operator-pipeline-prod/openshift-node-upgrade-mutex-operator@sha256:5ac79c9193934440464a595434fe4241e038505d32650843afcad5f5bc2ce8a7
        version: 0.2.66
      entries:
      - name: openshift-node-upgrade-mutex-operator.v0.2.66
        version: 0.2.66
      - name: openshift-node-upgrade-mutex-operator.v0.2.65
        version: 0.2.65
      - name: openshift-node-upgrade-mutex-operator.v0.2.60
        version: 0.2.60
      - name: openshift-node-upgrade-mutex-operator.v0.2.59
        version: 0.2.59
      - name: openshift-node-upgrade-mutex-operator.v0.2.53
        version: 0.2.53
      - name: openshift-node-upgrade-mutex-operator.v0.2.38
        version: 0.2.38
      - name: openshift-node-upgrade-mutex-operator.v0.2.34
        version: 0.2.34
      - name: openshift-node-upgrade-mutex-operator.v0.2.15
        version: 0.2.15
      - name: openshift-node-upgrade-mutex-operator.v0.2.14
        version: 0.2.14
      - name: openshift-node-upgrade-mutex-operator.v0.1.22
        version: 0.1.22
      name: alpha
    defaultChannel: alpha
    packageName: openshift-node-upgrade-mutex-operator
    provider:
      name: Jason Skrzypek
      url: https://github.com/shpwrck
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicast-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: alpha
    - currentCSV: apicast-community-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIcast",
                "metadata": {
                  "name": "example-apicast"
                },
                "spec": {
                  "adminPortalCredentialsRef": {
                    "name": "mysecretname"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/apicast-operator:v0.8.0
          createdAt: "2019-10-27T22:40:00Z"
          description: APIcast is an API gateway built on top of NGINX. It is part
            of the Red Hat 3scale API Management Platform
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/apicast-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIcast is the Schema for the apicasts API.
            displayName: APIcast
            kind: APIcast
            name: apicasts.apps.3scale.net
            version: v1alpha1
        description: |
          The APIcast Operator creates and maintains the Red Hat 3scale API Gateway in several deployment configurations.

          APIcast is an API gateway built on top of [NGINX](https://www.nginx.com/). It is part of the [Red Hat 3scale API Management Platform](https://www.redhat.com/en/technologies/jboss-middleware/3scale).

          ### Supported Features
          * **Installer** A way to install an APIcast gateway instance on Kubernetes
          * **Upgrade** Upgrade from previously installed APIcast gateway instance
          * **Reconcilliation** Tunable CRD parameters after the APIcast gateway is installed

          ### Upgrading your installation
          The APIcast Operator understands how to run and upgrade between a set of APIcast versions.
          See [the upgrade guide](https://github.com/3scale/apicast-operator/blob/v0.8.0/doc/operator-user-guide.md#upgrading-APIcast) for more information.

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/apicast-operator/tree/v0.8.0).

          ### Getting help
          If you encounter any issues while using operator, you can create an issue on our [website](https://github.com/3scale/apicast-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using APIcast Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/apicast-operator/pulls)
          * Improving [documentation](https://github.com/3scale/apicast-operator)
          * Talking about APIcast Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/apicast-operator/issues).

          ### License
          APIcast Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/apicast-operator/blob/master/LICENSE)
        displayName: APIcast
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        - api
        - APIcast
        - apicast
        - gateway
        - api-gateway
        - openresty
        links:
        - name: GitHub
          url: https://github.com/3scale/apicast-operator
        - name: Documentation
          url: https://github.com/3scale/apicast-operator/tree/v0.8.0
        maintainers:
        - email: eastizle+apicastoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/apicast-operator:v0.8.0
        - quay.io/community-operator-pipeline-prod/apicast-community-operator@sha256:b0e697be3b4af0554de09306016b5852a4dadacd1962ab9417f6e682491519cd
        version: 0.8.0
      entries:
      - name: apicast-community-operator.v0.8.0
        version: 0.8.0
      - name: apicast-community-operator.v0.7.1
        version: 0.7.1
      - name: apicast-community-operator.v0.6.0
        version: 0.6.0
      - name: apicast-community-operator.v0.5.1
        version: 0.5.1
      name: stable
    defaultChannel: stable
    packageName: apicast-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tang-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tang-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
          createdAt: 2023-11-23T12:49:40
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a77d39ded03271f2deb8b438abd4328a1846ee0d4886cce14d869ba61f84d600
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:f43c96035f43040b3644e37119a3b35fc433f366d4b636459d1b9b02a05b5d3d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:97cade2c1ee468261aec5400728c8d44de387b459134aec7a4c3b5ec5a335d2c
        version: 1.0.0
      entries:
      - name: tang-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: tang-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "daemons.redhat.com/v1alpha1",
                "kind": "TangServer",
                "metadata": {
                  "finalizers": [
                    "finalizer.daemons.tangserver.redhat.com"
                  ],
                  "name": "tangserver",
                  "namespace": "nbde"
                },
                "spec": {
                  "healthScript": "/usr/bin/tangd-health-check",
                  "image": "registry.redhat.io/rhel9/tang",
                  "keyPath": "/var/db/tang",
                  "persistentVolumeClaim": "tangserver-pvc",
                  "replicas": 1,
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
          createdAt: 2024-12-04T19:01:27
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TangServer is the Schema for the tangservers API
            displayName: Tang Server
            kind: TangServer
            name: tangservers.daemons.redhat.com
            version: v1alpha1
        description: NBDE Tang Server operator allows Tang Server deployment on OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nbde
        - tang
        links:
        - name: Tang Operator
          url: https://github.com/latchset/tang-operator
        maintainers:
        - email: sarroutb@redhat.com
          name: Sergio Arroutbi Braojos
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/nbde-tang-server/tang-operator-bundle@sha256:a6f3a3a8aa52de4441ef3249a76dcff1b6b6defba417a9fdbba37eb458adee04
        - registry.redhat.io/nbde-tang-server/tang-rhel9-operator@sha256:1cb2f794689817282ac570cacb106d75a2751a1926d89d3b83b619e02d869e5e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1172e150fff22c5eeab572f26961f3f53fbf896ee76d08c7503cfe2777c55458
        version: 1.0.3
      entries:
      - name: tang-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: alpha
    packageName: tang-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-xtradb-cluster-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:6208f64c7fe75f39ab7511e2b480c493d2ecc8d1e76a1ac498a93a6a24e8beb5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        version: 1.18.0
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0
        version: 1.18.0
      - name: percona-xtradb-cluster-operator-certified.v1.17.0
        version: 1.17.0
      name: stable
    - currentCSV: percona-xtradb-cluster-operator-certified.v1.18.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBCluster",
                "metadata": {
                  "name": "cluster1",
                  "finalizers": [
                    "percona.com/delete-pxc-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.18.0",
                  "tls": {
                    "enabled": true
                  },
                  "initContainer" : {
                      "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49"
                    },
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 4 * * *"
                  },
                  "pxc": {
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3",
                    "autoRecovery": true,
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "6G"
                          }
                        }
                      }
                    },
                    "gracePeriod": 600
                  },
                  "haproxy": {
                    "enabled": true,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "proxysql": {
                    "enabled": false,
                    "size": 3,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5",
                    "resources": {
                      "requests": {
                        "memory": "1G",
                        "cpu": "600m"
                      }
                    },
                    "affinity": {
                      "antiAffinityTopologyKey": "kubernetes.io/hostname"
                    },
                    "volumeSpec": {
                      "persistentVolumeClaim": {
                        "resources": {
                          "requests": {
                            "storage": "2G"
                          }
                        }
                      }
                    },
                    "podDisruptionBudget": {
                      "maxUnavailable": 1
                    },
                    "gracePeriod": 30
                  },
                  "logcollector": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12",
                    "resources": {
                      "requests": {
                        "memory": "100M",
                        "cpu": "200m"
                      }
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e",
                    "serverHost": "monitoring-service",
                    "resources": {
                      "requests": {
                        "memory": "150M",
                        "cpu": "300m"
                      }
                    }
                  },
                  "backup": {
                    "image": "registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e",
                    "pitr": {
                      "enabled": false,
                      "storageName": "STORAGE-NAME-HERE",
                      "timeBetweenUploads": 60,
                      "timeoutSeconds": 60
                    },
                    "storages": {
                      "s3-us-west": {
                        "type": "s3",
                        "verifyTLS": true,
                        "s3": {
                          "bucket": "S3-BACKUP-BUCKET-NAME-HERE",
                          "credentialsSecret": "my-cluster-name-backup-s3",
                          "region": "us-west-2"
                        }
                      },
                      "azure-blob": {
                        "type": "azure",
                        "azure": {
                          "credentialsSecret": "azure-secret",
                          "container": "test"
                        }
                      },
                      "fs-pvc": {
                        "type": "filesystem",
                        "volume": {
                          "persistentVolumeClaim": {
                            "accessModes": [
                              "ReadWriteOnce"
                            ],
                            "resources": {
                              "requests": {
                                "storage": "6G"
                              }
                            }
                          }
                        }
                      }
                    },
                    "schedule": [
                      {
                        "name": "daily-backup",
                        "schedule": "0 0 * * *",
                        "retention": {
                          "type": "count",
                          "count": 5,
                          "deleteFromStorage": true
                        },
                        "storageName": "fs-pvc"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "storageName": "fs-pvc"
                }
              },
              {
                "apiVersion": "pxc.percona.com/v1",
                "kind": "PerconaXtraDBClusterRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "pxcCluster": "cluster1",
                  "backupName": "backup1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
          createdAt: "2025-04-08T07:32:10.3Z"
          description: Percona Operator for MySQL based on Percona XtraDB Cluster
            manages the lifecycle of Percona XtraDB cluster instances.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-xtradb-cluster-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona XtraDB Cluster
            displayName: PerconaXtraDBCluster
            kind: PerconaXtraDBCluster
            name: perconaxtradbclusters.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Backup
            displayName: PerconaXtraDBClusterBackup
            kind: PerconaXtraDBClusterBackup
            name: perconaxtradbclusterbackups.pxc.percona.com
            version: v1
          - description: Instance of a Percona XtraDB Cluster Restore
            displayName: PerconaXtraDBClusterRestore
            kind: PerconaXtraDBClusterRestore
            name: perconaxtradbclusterrestores.pxc.percona.com
            version: v1
        description: |2

          ## Percona is Cloud Native

          Percona Operator for MySQL based on Percona XtraDB Cluster is an open-source drop in replacement for MySQL Enterprise with synchronous replication running on Kubernetes. It automates the deployment and management of the members in your Percona XtraDB Cluster environment. It can be used to instantiate a new Percona XtraDB Cluster, or to scale an existing environment.

          Consult the [documentation](https://percona.github.io/percona-xtradb-cluster-operator/) on the Percona Operator for MySQL based on Percona XtraDB Cluster for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster**  change the `size` parameter to [add or remove members](https://percona.github.io/percona-xtradb-cluster-operator/install/scaling) of the cluster. Three is the minimum recommended size for a functioning cluster.

          * **Manage Your Users**  [add, remove, or change](https://percona.github.io/percona-xtradb-cluster-operator/configure/users) the privileges of database users

          * **Automate Your Backups**  [configure cluster backups](https://percona.github.io/percona-xtradb-cluster-operator/configure/operator) to run on a scheduled basis. Backups can be stored on a persistent volume or S3-compatible storage. Leverage [Point-in-time recovery](https://www.percona.com/doc/kubernetes-operator-for-pxc/backups.html#storing-binary-logs-for-point-in-time-recovery) to reduce RPO/RTO.
          * **Proxy integration**  choose HAProxy or ProxySQL as a proxy in front of the Percona XtraDB Cluster. Proxies are deployed and configured automatically with the Operator.

          ### Common Configurations

          * **Set Resource Limits** - set limitation on requests to CPU and memory resources.

          * **Customize Storage** - set the desired Storage Class and Access Mode for your database cluster data.

          * **Control Scheduling** - define how your PXC Pods are scheduled onto the K8S cluster with tolerations, pod disruption budgets, node selector and affinity settings.
          * Automatic synchronization of MySQL users with ProxySQL
          * HAProxy Support
          * Fully automated minor version updates (Smart Update)
          * Update Reader members before Writer member at cluster upgrades
          * Support MySQL versions 5.7, 8.0 and 8.4 (technical preview)
          ### Before You Start

          Add the PXC user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-secrets
          type: Opaque
          data:
            root: cm9vdF9wYXNzd29yZA==
            xtrabackup: YmFja3VwX3Bhc3N3b3Jk
            monitor: bW9uaXRvcg==
            proxyadmin: YWRtaW5fcGFzc3dvcmQ=
            pmmserverkey: c3VwYXxefHBheno=
            operator: b3BlcmF0b3JhZG1pbg==
            replication: cmVwbF9wYXNzd29yZAo=
          ``` ###  Release Highlights
          * **PMM3 support**
          The Operator is natively integrated with PMM 3, enabling you to monitor the health and performance of your Percona Distribution for MySQL deployment and at the same time enjoy enhanced performance, new features, and improved security that PMM 3 provides.
          * **Improved monitoring for clusters in multi-region or multi-namespace deployments in PMM**
          Now you can define a custom name for your clusters deployed in different data centers. This name helps Percona Management and Monitoring (PMM) Server to correctly recognize clusters as connected and monitor them as one deployment. Similarly, PMM Server identifies clusters deployed with the same names in different namespaces as separate ones and correctly displays performance metrics for you on dashboards.
          * **More resilient database restores without matching user Secrets**
          You no longer need matching user Secrets between your backup and your target cluster to perform a restore. The Operator now has a post-restore step that changes user passwords in the restored database to the ones from the local Secret. Also, it creates missing system users and adds missing grants.
          * **Improved backup retention for streamlined management of scheduled backups in cloud storage**
          A new backup retention configuration gives you more control over how backups are managed in storage and retained in Kubernetes. With the deleteFromStorage flag, you can disable automatic deletion from AWS S3 or Azure Blob storage and instead rely on native cloud lifecycle policies. This makes backup cleanup more efficient and better aligned with flexible storage strategies.
        displayName: Percona Operator for MySQL based on Percona XtraDB Cluster
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - percona
        - database
        - pxc
        - galera
        - sql
        - operator
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-xtradb-cluster-operator/
        - name: Github
          url: https://github.com/percona/percona-xtradb-cluster-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ad865bf20782fecedb247247cdd24dc4707eafe986f5ca623157bab7d6a7a57e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ec3985eb1556103f2c9c5b4dfcf6871487af061480a58636914d2f6a6cce0b12
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator@sha256:8c38ea962d8bd397168b05f5e400cd89e17724b635f1b5ee536f5ac7250eeb49
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-bundle@sha256:11d64a921a98b7ec1a2126a03e446a055a5ae1705dc2005bb8b73cbbcf7ecf99
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:5d3e213b88036a88759d221a4b8b8e87b1e06a91e82c59f1913997b21900bce2
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:d9c1993de5b20661445488f991bc06c98e93099034c4c17901610d4a2797f82f
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:2444c2ea0e9e28ef3743a118ace4dbdf055ded87765ea8b7a54ebcf0213f7b4d
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:8ca29f400d9ac5cabfb6881a9c9142d4d42e170babfbdcd07f11b09ab2296d17
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:c9d08d4b43f0fce5dda5a5ca84b23c9e64495de7a8c8714a78d97ab60b482f0e
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:f13e44bf74869459ba43236fd36f6b60f9ad5241bb7f7e3c4a98cb06d84ec174
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:ff8020a618bebcc20ca748e93d9b70fa1bacf951cd8e6ad526f9dd07efd2f1d5
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:4ab5bddfaf75df4c140e9be3b28a256afa1fe5bf6227cb1dd176d506b4a2613b
        - registry.connect.redhat.com/percona/percona-xtradb-cluster-operator-containers@sha256:749799c8e1562047209bfd3b0c60bb2ad34c38b37490461307d31c745809d6f3
        version: 1.18.0-cw
      entries:
      - name: percona-xtradb-cluster-operator-certified.v1.18.0-cw
        version: 1.18.0-cw
      - name: percona-xtradb-cluster-operator-certified.v1.17.0-cw
        version: 1.17.0-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-xtradb-cluster-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Commvault
      provider-url: ""
    name: hedvig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hedvig-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\":\"hedvig.io/v1\",\n    \"kind\":\"HedvigDeploy\",\n
            \   \"metadata\": {\n      \"name\": \"testdeploy\",\n      \"namespace\":
            \"hedvig\"\n    },\n    \"spec\": {\n      \"cluster\": {\n        \"name\":
            \"hedvigcl\",\n        \"seeds\": [\n          \"hedvigcl1.hedvig.io\",\n
            \         \"hedvigcl2.hedvig.io\",\n          \"hedvigcl3.hedvig.io\"\n
            \       ]\n      },\n      \"k8s\": {\n        \"name\": \"k8scl\"\n      },\n
            \     \"proxy\": {\n        \"repository\": \"hedvig/hedvig-proxy\",\n
            \       \"tag\": \"v-4.3.1\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"controller\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"node\": {\n        \"repository\": \"hedvig/hedvig-csi-volume-driver\",\n
            \       \"tag\": \"v-2.0\",\n        \"imagePullPolicy\": \"IfNotPresent\"\n
            \     },\n      \"sidecars\": [\n        {\n          \"name\": \"csi-provisioner\",\n
            \         \"repository\": \"quay.io/k8scsi/csi-provisioner\",\n          \"tag\":
            \"v1.6.0\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n        },\n
            \       {\n          \"name\": \"csi-snapshotter\",\n          \"repository\":
            \"quay.io/k8scsi/csi-snapshotter\",\n          \"tag\": \"v2.1.0\",\n
            \         \"imagePullPolicy\": \"IfNotPresent\"\n        },              \n
            \       {\n          \"name\": \"csi-resizer\",\n          \"repository\":
            \"quay.io/k8scsi/csi-resizer\",\n          \"tag\": \"v0.5.0\",\n          \"imagePullPolicy\":
            \"IfNotPresent\"\n        },              \n        {\n          \"name\":
            \"csi-node-driver-registrar\",\n          \"repository\": \"quay.io/k8scsi/csi-node-driver-registrar\",\n
            \         \"tag\": \"v1.0-canary\",\n          \"imagePullPolicy\": \"IfNotPresent\"\n
            \       }\n      ]\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: hedvig/hedvig-operator:2.0.0-ubi
          createdAt: ""
          description: Hedvig Operator is the official operator to deploy and manage
            Hedvig components in container orchestrators such as Kubernetes and Openshift.
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hedvig Deploy
            displayName: HedvigDeploy
            kind: HedvigDeploy
            name: hedvigdeploys.hedvig.io
            version: v1
        description: "The Hedvig Distributed Storage Platform is a software-defined,
          scale-out storage solution. It distributes data simultaneously across multiple
          locations – from on-premises data centers to the cloud – and scales capacity
          on-demand by leveraging the storage of commodity x86/ARM servers. The distributed
          write maximizes availability and protects data from hardware failures ranging
          from a single disk to an entire site, improving disaster recovery planning.\n\nHedvig
          provides a CSI Driver for accelerating your journey into the container ecosystem.
          CSI Driver enables containerized applications to maintain state by dynamically
          provisioning and consuming Hedvig virtual disks as persistent volumes. \n###
          \n\nHedvig Operator deploys the Hedvig Storage Proxy and the Hedvig CSI
          Driver to manage the life cycle of persistent storage in Kubernetes and
          Openshift.\n\n**Supported Container Orchestrators:**\n* Kubernetes 1.16+\n*
          Red Hat OpenShift 4.6+\n\n**Supported Hedvig Releases:**\n* Hedvig 4.0+\nPlease
          refer to the [documentation](https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_Operator_User_Guide.pdf)
          for steps to install and use the operator.\n\n\n\n\n\n"
        displayName: Hedvig Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - CSI
        - Hedvig
        - Cloud
        - Block
        - NFS
        - iSCSI
        - Snapshot
        - Volume
        links:
        - name: Documentation
          url: https://documentation.commvault.com/commvault/hedvig/article
        - name: Hedvig CSI User Guide
          url: https://documentation.commvault.com/commvault/hedvig/others/pdf/Hedvig_CSI_User_Guide.pdf
        maintainers:
        - email: ashenoy@commvault.com
          name: Abhijith Shenoy
        maturity: stable
        provider:
          name: Commvault
        relatedImages:
        - hedvig/hedvig-operator:2.0.0-ubi
        - quay.io/openshift-community-operators/hedvig-operator@sha256:81d956c12057f18bf920044a4e39c27c99b2c23afdccc0ff9468f0484e1961d7
        version: 1.0.2
      entries:
      - name: hedvig-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: hedvig-operator
    provider:
      name: Commvault
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kiali-ossm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kiali-operator.v2.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
          createdAt: "2024-12-12T10:44:15Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.1.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:976a31ca8cf17ac0887b8a586edfcc29eb30175802461b3114758af525e7b800
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-operator-bundle@sha256:f40e7955ac649c383134ab6e77bdd521b75b21ecd9b476436405d0674c7fc855
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-ossmc-rhel9@sha256:99643f294c3f92a6fc82f321cab7c756e4879fefb48fa824f1bf49ff7182658d
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9-operator@sha256:0612dfa52f2bf39a1d87e640409313585fedc7cfcff6d00e77bb14809d27d646
        - registry.redhat.io/openshift-service-mesh-tech-preview/kiali-rhel9@sha256:06fde6c696a33d17b0d2f7620ec46422c82277fd17a3fe6d7319de05c732aace
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:e37d76dd7c7a2db79ff1e96afef19b07c3beaa8ebf0c7bd85b808908b8f0ce77
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:27a6aebffa3d049ee48050126e6107db2e07c49e9ce9297175facb6368e1e792
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:66e4d933a9d69158bf87dd45835183bf2308b77c48908ab64496c5d22c816293
        version: 2.1.2
      entries:
      - name: kiali-operator.v2.1.2
        version: 2.1.2
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: candidate
    - currentCSV: kiali-operator.v2.11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "Kiali",
                "metadata": {
                  "name": "kiali"
                },
                "spec": {
                  "version": "default",
                  "istio_namespace": "istio-system",
                  "deployment": {
                    "logger": {
                      "log_level": "info"
                    },
                    "view_only_mode": false
                  }
                }
              },
              {
                "apiVersion": "kiali.io/v1alpha1",
                "kind": "OSSMConsole",
                "metadata": {
                  "name": "ossmconsole"
                },
                "spec": {
                  "version": "default",
                  "kiali": {
                    "serviceName": "",
                    "serviceNamespace": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
          createdAt: "2025-09-10T00:28:49Z"
          description: This productized operator provides Kiali and OSSMC. Kiali is
            the Istio observability and management Console. OSSMC is the OpenShift
            Service Mesh Console plugin, powered by Kiali.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/kiali/kiali
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A configuration file for a Kiali installation.
            displayName: Kiali
            kind: Kiali
            name: kialis.kiali.io
            version: v1alpha1
          - description: A configuration file for a OpenShift Service Mesh Console
              installation.
            displayName: OpenShift Service Mesh Console
            kind: OSSMConsole
            name: ossmconsoles.kiali.io
            version: v1alpha1
        description: |-
          ## About the managed application

          A Microservice Architecture breaks up the monolith into many smaller pieces
          that are composed together. Patterns to secure the communication between
          services like fault tolerance (via timeout, retry, circuit breaking, etc.)
          have come up as well as distributed tracing to be able to see where calls
          are going.

          A service mesh can now provide these services on a platform level and frees
          the application writers from those tasks. Routing decisions are done at the
          mesh level.

          Kiali works with OpenShift Service Mesh to visualize the service
          mesh topology, to provide visibility into features like circuit breakers,
          request rates and more. It offers insights about the mesh components at
          different levels, from abstract Applications to Services and Workloads.

          See [https://www.kiali.io](https://www.kiali.io) to read more.

          ### Accessing the UI

          By default, the Kiali operator exposes the Kiali UI as an OpenShift Route.

          If on OpenShift, you can create an OSSMConsole CR to have the operator
          install the OpenShift ServiceMesh Console plugin to the OpenShift Console
          thus providing an interface directly integrated with the OpenShift Console.

          ## About this Operator

          ### Kiali Custom Resource Configuration Settings

          For quick descriptions of all the settings you can configure in the Kiali
          Custom Resource (CR), see
          [the kiali.io docs](https://kiali.io/docs/configuration/kialis.kiali.io/).

          ## Prerequisites for enabling this Operator

          Kiali is a companion tool for OpenShift Service Mesh. So before you install Kiali, you must have
          already installed Service Mesh.
        displayName: Kiali Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - service-mesh
        - observability
        - monitoring
        - istio
        links:
        - name: Installation Guide
          url: https://www.kiali.io/docs/installation
        - name: Features
          url: https://www.kiali.io/docs/features
        - name: Documentation Home
          url: https://www.kiali.io
        - name: Blogs and Articles
          url: https://medium.com/kialiproject
        - name: Server and UI Source Code
          url: https://github.com/kiali/kiali
        - name: Operator Source Code
          url: https://github.com/kiali/kiali-operator
        - name: OpenShift Service Mesh Console Code
          url: https://github.com/kiali/openshift-servicemesh-plugin
        maintainers:
        - email: kiali-dev@googlegroups.com
          name: Kiali Developers Google Group
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:9a0d4862a25c2999a3f85b9a7a0332dab4ec1744548191d7a00c838b06ba26d2
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel8@sha256:81db94df331799f4087596378914b712d85d117dde3df2d4902079a82d4a9c55
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:7487fbb4b776dfca28ad09311ba857e07ad98bc91ca5ee6fcaaa5b7783d04fbd
        - registry.redhat.io/openshift-service-mesh/kiali-ossmc-rhel9@sha256:fe25be563d50e5ad4e5475f153e03d4b24c7db094d7e824c51ebaae79418b1f7
        - registry.redhat.io/openshift-service-mesh/kiali-rhel8@sha256:d729ef8d74c2f6e02d8c68ca761b1fb3c7c3706f1290b684468f0acb0ef313e0
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9-operator@sha256:2970861bd8e65e36d5080dc4d56340d08a52dbb1217be352998d135eb854732f
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:85eeb972e46f26ff30e04991e76ec3050d98b334e7b7f201a094f8d50b06854e
        - registry.redhat.io/openshift-service-mesh/kiali-rhel9@sha256:a00ec00e466e080592dca5c1cc153698b1412a54e1979647f74281f79b5691cf
        - registry.redhat.io/openshift-service-mesh/kiali-operator-bundle@sha256:d04033c2cd2d236f6319a3e35880b89c8f0c60d53cd32f6f097d07d6d27e5f2c
        version: 2.11.3
      entries:
      - name: kiali-operator.v2.11.3
        version: 2.11.3
      - name: kiali-operator.v2.11.2
        version: 2.11.2
      - name: kiali-operator.v2.11.1
        version: 2.11.1
      - name: kiali-operator.v2.4.7
        version: 2.4.7
      - name: kiali-operator.v2.4.6
        version: 2.4.6
      - name: kiali-operator.v2.4.5
        version: 2.4.5
      - name: kiali-operator.v2.4.4
        version: 2.4.4
      - name: kiali-operator.v2.4.3
        version: 2.4.3
      - name: kiali-operator.v1.89.10
        version: 1.89.10
      - name: kiali-operator.v1.89.9
        version: 1.89.9
      - name: kiali-operator.v1.89.8
        version: 1.89.8
      - name: kiali-operator.v1.89.7
        version: 1.89.7
      - name: kiali-operator.v1.89.6
        version: 1.89.6
      - name: kiali-operator.v1.89.1
        version: 1.89.1
      - name: kiali-operator.v1.73.10
        version: 1.73.10
      - name: kiali-operator.v1.73.8
        version: 1.73.8
      - name: kiali-operator.v1.73.7
        version: 1.73.7
      - name: kiali-operator.v1.73.4
        version: 1.73.4
      - name: kiali-operator.v1.65.11
        version: 1.65.11
      - name: kiali-operator.v1.65.10
        version: 1.65.10
      - name: kiali-operator.v1.65.9
        version: 1.65.9
      - name: kiali-operator.v1.65.8
        version: 1.65.8
      - name: kiali-operator.v1.65.7
        version: 1.65.7
      - name: kiali-operator.v1.65.6
        version: 1.65.6
      - name: kiali-operator.v1.57.7
        version: 1.57.7
      name: stable
    defaultChannel: stable
    packageName: kiali-ossm
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-ati-updates
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-ati-updates.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ati.loadcore.keysight.com/v1alpha1",
                "kind": "Updates",
                "metadata": {
                  "name": "updates-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "installWait": {
                    "enabled": true,
                    "image": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:40:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Updates
            name: updates.ati.loadcore.keysight.com
            version: v1alpha1
        description: Keysight ATI Updates Operator
        displayName: Keysight ATI Updates Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Ati Updates
          url: https://keysight-ati-updates.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-ati-updates-bundle@sha256:3d06a58b2a277a67339c22f3fe10c1daac1c8eabaac8540b01177f6f5688784c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-ati-updates@sha256:abda76eb50d77e5a95191c9dbb21eefe64af5b794511ceafdfda007c15449c6e
        version: 0.1.21
      entries:
      - name: keysight-ati-updates.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-ati-updates
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatchlogs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatchlogs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudwatchlogs.services.k8s.aws/v1alpha1",
                "kind": "LogGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
          createdAt: "2025-09-19T20:51:16Z"
          description: AWS CloudWatch Logs controller is a service controller for
            managing CloudWatch Logs resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LogGroup represents the state of an AWS cloudwatchlogs LogGroup
              resource.
            displayName: LogGroup
            kind: LogGroup
            name: loggroups.cloudwatchlogs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch Logs resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch Logs**

          You can use Amazon CloudWatch Logs to monitor, store, and access your log files from Amazon Elastic Compute Cloud (Amazon EC2) instances, AWS CloudTrail, Route 53, and other sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch Logs
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatchlogs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Logs Developer Resources
          url: https://aws.amazon.com/cloudwatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatchlogs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatchlogs-controller@sha256:94145cae2b8cef1410acbb4f1c1eb571e71408341716af1742f27ddf4dc350b6
        version: 1.1.1
      entries:
      - name: ack-cloudwatchlogs-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudwatchlogs-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatchlogs-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudwatchlogs-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatchlogs-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatchlogs-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatchlogs-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatchlogs-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatchlogs-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatchlogs-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatchlogs-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatchlogs-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatchlogs-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatchlogs-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatchlogs-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatchlogs-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatchlogs-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatchlogs-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatchlogs-controller.v0.0.7
        version: 0.0.7
      - name: ack-cloudwatchlogs-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudwatchlogs-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatchlogs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle Corporation
      provider-url: https://github.com/oracle/coherence-operator
    name: coherence-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coherence-operator.v3.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "Coherence",
                "metadata": {
                  "name": "coherence-sample"
                },
                "spec": {
                  "replicas": 3
                }
              },
              {
                "apiVersion": "coherence.oracle.com/v1",
                "kind": "CoherenceJob",
                "metadata": {
                  "name": "coherence-job-sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "false"
          containerImage: container-registry.oracle.com/middleware/coherence-operator:3.5.4
          createdAt: "2025-08-28T16:15:09Z"
          description: The Oracle Coherence Kubernetes Operator enables easy management
            of Coherence clusters in a Kubernetes environment.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/arch.amd64: supported
          operatorframework.io/arch.arm64: supported
          operatorframework.io/os.linux: supported
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/oracle/coherence-operator
          support: Oracle Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coherence is the top level schema for the Coherence API and
              custom resource definition (CRD).
            displayName: Coherence
            kind: Coherence
            name: coherence.coherence.oracle.com
            version: v1
          - description: |-
              CoherenceJob is the top level schema for the CoherenceJob API and custom resource definition (CRD)
              for configuring Coherence Job workloads.
            displayName: Coherence Job
            kind: CoherenceJob
            name: coherencejob.coherence.oracle.com
            version: v1
        description: |
          The Oracle Coherence Kubernetes Operator enables easy management of Coherence clusters in a Kubernetes environment.

          ### Installation Guide & Configuration

          For the complete installation guide and all configuration options please refer to
          [Coherence Operator Documentation](https://oracle.github.io/coherence-operator/docs/latest/#/about/01_overview).
        displayName: Coherence Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Coherence
        - Oracle
        - In Memory Data
        - Caching
        links:
        - name: Coherence Kubernetes Operator
          url: https://oracle.github.io/coherence-operator/docs/latest
        - name: Oracle Coherence
          url: https://coherence.community
        maintainers:
        - email: jonathan.knight@oracle.com
          name: Jonathan Knight
        maturity: mature
        minKubeVersion: 1.26.0
        provider:
          name: Oracle Corporation
          url: https://github.com/oracle/coherence-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/coherence-operator:3.5.4
        - container-registry.oracle.com/middleware/coherence-operator:3.5.4
        version: 3.5.4
      entries:
      - name: coherence-operator.v3.5.4
        version: 3.5.4
      - name: coherence-operator.v3.5.2
        version: 3.5.2
      - name: coherence-operator.v3.5.0
        version: 3.5.0
      - name: coherence-operator.v3.4.3
        version: 3.4.3
      name: stable
    defaultChannel: stable
    packageName: coherence-operator
    provider:
      name: Oracle Corporation
      url: https://github.com/oracle/coherence-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Google
      provider-url: https://cloud.google.com/alloydb/omni
    name: alloydb-omni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloydb-omni-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "backupPlanRef": "backupplan1",
                  "dbclusterRef": "dbcluster-sample",
                  "manual": true,
                  "physicalBackupSpec": {
                    "backupType": "full"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "BackupPlan",
                "metadata": {
                  "name": "backupplan1"
                },
                "spec": {
                  "backupRetainDays": 14,
                  "backupSchedules": {
                    "full": "0 0 * * 0",
                    "incremental": "0 21 * * *"
                  },
                  "dbclusterRef": "dbcluster-sample",
                  "paused": false
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "clone1"
                },
                "spec": {
                  "clonedDBClusterConfig": {
                    "dbclusterName": "new-dbcluster-sample"
                  },
                  "pointInTime": "2024-02-23T19:59:43Z",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "dbcluster-sample"
                },
                "spec": {
                  "databaseVersion": "16.3.0",
                  "primarySpec": {
                    "adminUser": {
                      "passwordRef": {
                        "name": "db-pw-dbcluster-sample"
                      }
                    },
                    "resources": {
                      "cpu": 1,
                      "disks": [
                        {
                          "name": "DataDisk",
                          "size": "10Gi"
                        }
                      ],
                      "memory": "5Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "dbcluster-sample-rp-1"
                },
                "spec": {
                  "dbcParent": {
                    "name": "dbcluster-sample"
                  },
                  "instanceType": "ReadPool",
                  "nodeCount": 2,
                  "resources": {
                    "cpu": 2,
                    "disks": [
                      {
                        "name": "DataDisk",
                        "size": "15Gi"
                      }
                    ],
                    "memory": "6Gi"
                  },
                  "schedulingconfig": {
                    "nodeaffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "preference": {
                            "matchExpressions": [
                              {
                                "key": "another-node-label-key",
                                "operator": "In",
                                "values": [
                                  "another-node-label-value"
                                ]
                              }
                            ]
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "app",
                                  "operator": "In",
                                  "values": [
                                    "store"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "kubernetes.io/hostname"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "podAntiAffinity": {
                      "preferredDuringSchedulingIgnoredDuringExecution": [
                        {
                          "podAffinityTerm": {
                            "labelSelector": {
                              "matchExpressions": [
                                {
                                  "key": "security",
                                  "operator": "In",
                                  "values": [
                                    "S1"
                                  ]
                                }
                              ]
                            },
                            "topologyKey": "topology.kubernetes.io/zone"
                          },
                          "weight": 1
                        }
                      ]
                    },
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane",
                        "operator": "Exists"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Failover",
                "metadata": {
                  "name": "failover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "PgBouncer",
                "metadata": {
                  "name": "mypgbouncer"
                },
                "spec": {
                  "allowSuperUserAccess": true,
                  "dbclusterRef": "dbcluster-sample",
                  "parameters": {
                    "default_pool_size": "15",
                    "ignore_startup_parameters": "extra_float_digits",
                    "max_client_conn": "800",
                    "max_db_connections": "160",
                    "pool_mode": "transaction"
                  },
                  "podSpec": {
                    "image": "gcr.io/alloydb-omni/operator/g-pgbouncer:1.5.0",
                    "resources": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "replicaCount": 1,
                  "serviceOptions": {
                    "type": "ClusterIP"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-downstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "downstream": {
                    "control": "setup",
                    "host": "10.10.10.10",
                    "password": {
                      "name": "ha-rep-pw-dbcluster-sample"
                    },
                    "port": 5432,
                    "replicationSlotName": "dbcluster_sample_replication_upstream_sample",
                    "username": "alloydbreplica"
                  }
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Replication",
                "metadata": {
                  "name": "replication-upstream-sample"
                },
                "spec": {
                  "dbcluster": {
                    "name": "dbcluster-sample"
                  },
                  "upstream": {}
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backup": "backup1",
                  "sourceDBCluster": "dbcluster-sample"
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Sidecar",
                "metadata": {
                  "name": "sidecar-sample"
                },
                "spec": {
                  "sidecars": [
                    {
                      "args": [
                        "-c",
                        "while [ true ]\ndo\ndate\nset -x\nls -lh /logs/diagnostic\nset +x\ndone\n"
                      ],
                      "command": [
                        "/bin/sh"
                      ],
                      "image": "busybox",
                      "name": "sidecar-sample",
                      "volumeMounts": [
                        {
                          "mountPath": "/logs",
                          "name": "obsdisk"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "alloydbomni.dbadmin.goog/v1",
                "kind": "Switchover",
                "metadata": {
                  "name": "switchover-sample"
                },
                "spec": {
                  "dbclusterRef": "dbcluster-sample",
                  "newPrimary": "aaaa-dbcluster-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
          createdAt: "2025-07-25T04:12:04Z"
          operatorframework.io/suggested-namespace: alloydb-omni-system
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/internal-objects: '["backuprepositories.alloydbomni.internal.dbadmin.goog",
            "failovers.alloydbomni.internal.dbadmin.goog", "instancebackupplans.alloydbomni.internal.dbadmin.goog","instancebackups.alloydbomni.internal.dbadmin.goog","instancerestores.alloydbomni.internal.dbadmin.goog","instances.alloydbomni.internal.dbadmin.goog","instanceswitchovers.alloydbomni.internal.dbadmin.goog","instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog","lrojobs.alloydbomni.internal.dbadmin.goog","replicationconfigs.alloydbomni.internal.dbadmin.goog","sidecars.alloydbomni.internal.dbadmin.goog","createstandbyjobs.alloydbomni.internal.dbadmin.goog","deletestandbyjobs.alloydbomni.internal.dbadmin.goog"]'
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BackupPlan is the Schema for the backupplans API.
            displayName: Backup Plan
            kind: BackupPlan
            name: backupplans.alloydbomni.dbadmin.goog
            version: v1
          - description: Backup is the Schema for the backups API.
            displayName: Backup
            kind: Backup
            name: backups.alloydbomni.dbadmin.goog
            version: v1
          - description: CreateStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Create Standby Job
            kind: CreateStandbyJob
            name: createstandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: DBCluster is the Schema for the DBClusters API.
            displayName: DB Cluster
            kind: DBCluster
            name: dbclusters.alloydbomni.dbadmin.goog
            version: v1
          - description: DBInstance is the Schema for the DBInstances API.
            displayName: DB Instance
            kind: DBInstance
            name: dbinstances.alloydbomni.dbadmin.goog
            version: v1
          - description: DeleteStandbyJob is an internal workflow tracking object.
              Users should not directly interact with this.
            displayName: (Internal) Delete Standby Job
            kind: DeleteStandbyJob
            name: deletestandbyjobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover represents the parameters and status of a single
              failover operation.
            displayName: Failover
            kind: Failover
            name: failovers.alloydbomni.dbadmin.goog
            version: v1
          - description: PgBouncer is the Schema for the PgBouncer API.
            displayName: PgBouncer
            kind: PgBouncer
            name: pgbouncers.alloydbomni.dbadmin.goog
            version: v1
          - description: Replication is the Schema for the Replications API.
            displayName: Replication
            kind: Replication
            name: replications.alloydbomni.dbadmin.goog
            version: v1
          - description: Restore is the Schema for the Restores API.
            displayName: Restore
            kind: Restore
            name: restores.alloydbomni.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.dbadmin.goog
            version: v1
          - description: Switchover represents the parameters and status of a single
              switchover.
            displayName: Switchover
            kind: Switchover
            name: switchovers.alloydbomni.dbadmin.goog
            version: v1
          - description: UserDefinedAuthentication is the Schema for the UserDefinedAuthentications
              API.
            displayName: User Defined Authentication
            kind: UserDefinedAuthentication
            name: userdefinedauthentications.alloydbomni.dbadmin.goog
            version: v1
          - description: BackupRepository is the Schema for the backuprepositories
              API.
            displayName: (Internal) Backup Repository
            kind: BackupRepository
            name: backuprepositories.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Failover is the Schema for the failover API.
            displayName: (Internal) Failover
            kind: Failover
            name: failovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackupPlan is the Schema for the InstanceBackupPlan
              API.
            displayName: (Internal) Instance Backup Plan
            kind: InstanceBackupPlan
            name: instancebackupplans.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceBackup is the Schema for the InstanceBackup API.
            displayName: (Internal) Instance Backup
            kind: InstanceBackup
            name: instancebackups.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceRestore is the Schema for the InstanceRestore API.
            displayName: (Internal) Instance Restore
            kind: InstanceRestore
            name: instancerestores.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Instance is the Schema for the Instance API.
            displayName: (Internal) Instance
            kind: Instance
            name: instances.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceSwitchover is the Schema for the InstanceSwitchover
              API.
            displayName: (Internal) Instance Switchover
            kind: InstanceSwitchover
            name: instanceswitchovers.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: InstanceUserDefinedAuthentication is the Schema for the InstanceUserDefinedAuthentication
              API.
            displayName: (Internal) Instance User Defined Authentication
            kind: InstanceUserDefinedAuthentication
            name: instanceuserdefinedauthentications.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: LROJob is the Schema for the LROJob API.
            displayName: (Internal) LRO Job
            kind: LROJob
            name: lrojobs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: ReplicationConfig is the Schema for the ReplicationConfig
              API.
            displayName: (Internal) Replication Config
            kind: ReplicationConfig
            name: replicationconfigs.alloydbomni.internal.dbadmin.goog
            version: v1
          - description: Sidecar is the Schema for the Sidecar API.
            displayName: (Internal) Sidecar
            kind: Sidecar
            name: sidecars.alloydbomni.internal.dbadmin.goog
            version: v1
        description: |-
          AlloyDB Omni is a database that lets you deploy a streamlined edition of AlloyDB for PostgreSQL in your own computing environment. While being 100% compatible with PostgreSQL applications and tools, AlloyDB Omni delivers higher performance for both OLTP and OLAP workloads and makes administration easy with a number of autopilots. AlloyDB AI provides powerful tools to build generative AI applications with the ability to create, store, and index vector embeddings right in your database.

          ## Post-installation

          After installing the operator, you need to manually create the cert-manager resources in your cluster as this is a strict dependency for our operator. For this, follow these steps:

          ```
          kubectl create ns ${NAMESPACE:?}
          ```

          ```
          kubectl apply -f - <<EOF
          apiVersion: cert-manager.io/v1
          kind: ClusterIssuer
          metadata:
            name: alloydbomni-selfsigned-cluster-issuer
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: fleet-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: fleet-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - fleet-webhook-service.alloydb-omni-system.svc
            - fleet-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: fleet-selfsigned-issuer
            secretName: fleet-webhook-server-cert
          ---
          apiVersion: cert-manager.io/v1
          kind: Issuer
          metadata:
            name: local-selfsigned-issuer
            namespace: ${NAMESPACE:?}
          spec:
            selfSigned: {}
          ---
          apiVersion: cert-manager.io/v1
          kind: Certificate
          metadata:
            name: local-serving-cert
            namespace: ${NAMESPACE:?}
          spec:
            dnsNames:
            - local-webhook-service.alloydb-omni-system.svc
            - local-webhook-service.alloydb-omni-system.svc.cluster.local
            issuerRef:
              kind: Issuer
              name: local-selfsigned-issuer
            secretName: local-webhook-server-cert
          EOF
          ```

          Replace the following:
          * `NAMESPACE` : the namespace where you have installed your operator. Suggested namespace: `alloydb-omni-system`.

          ## Creating a database cluster

          After you install the AlloyDB Omni operator on your Kubernetes cluster, you can create an AlloyDB Omni database cluster. For each database cluster you create, you must first create a password secret using the following format:

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: db-pw-${DB_CLUSTER_NAME:s?}
            namespace: ${DB_CLUSTER_NAMESPACE:?}
          type: Opaque
          data:
            ${DB_CLUSTER_NAME:?}: "ENCODED_PASSWORD"
          ```

          Replace the following:
          * `DB_CLUSTER_NAME`: the name of this database cluster. For example, `my-db-cluster`.
          * `DB_CLUSTER_NAMESPACE` (Optional): the namespace where you want to create the database cluster. For example, `my-db-cluster-namespace`.
          * `ENCODED_PASSWORD`: the database login password for the default postgres user role, encoded as a base64 string. For example, `Q2hhbmdlTWUxMjM=` for password `ChangeMe123`.

          Follow [these][database-cluster] steps for creating a database cluster.

          ## Connect to AlloyDB Omni running on Kubernetes
          Follow [these][connect] steps to connect to AlloyDB Omni.

          [database-cluster]: https://cloud.google.com/alloydb/omni/docs/deploy-kubernetes#create
          [connect]: https://cloud.google.com/alloydb/omni/docs/run-connect#connect-kubernetes
        displayName: AlloyDB Omni Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - sql
        - postgres
        - vector
        - ai
        - analytics
        - postgresql
        - alloydb
        links:
        - name: Documentation
          url: https://cloud.google.com/alloydb/omni/docs
        - name: Release Notes
          url: https://cloud.google.com/alloydb/docs/release-notes
        maintainers:
        - email: alloydb-omni-ext@google.com
          name: Google Cloud
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Google
          url: https://cloud.google.com/alloydb/omni
        relatedImages:
        - gcr.io/alloydb-omni/operator/fleet-operator:1.5.0
        - gcr.io/alloydb-omni/operator/local-operator:1.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/community-operator-pipeline-prod/alloydb-omni-operator@sha256:08fafb811f709780c0ccb3c22e3c6417cb14a70a9305e46a7b50457647a82102
        version: 1.5.0
      entries:
      - name: alloydb-omni-operator.v1.5.0
        version: 1.5.0
      - name: alloydb-omni-operator.v1.4.1
        version: 1.4.1
      - name: alloydb-omni-operator.v1.4.0
        version: 1.4.0
      - name: alloydb-omni-operator.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: alloydb-omni-operator
    provider:
      name: Google
      url: https://cloud.google.com/alloydb/omni
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: lifecycle-agent
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: lifecycle-agent.v4.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "ImageBasedUpgrade",
                "metadata": {
                  "name": "upgrade"
                },
                "spec": {
                  "autoRollbackOnFailure": {},
                  "extraManifests": [
                    {
                      "name": "sno-extramanifests",
                      "namespace": "openshift-lifecycle-agent"
                    }
                  ],
                  "oadpContent": [
                    {
                      "name": "oadp-cm-sno-backup",
                      "namespace": "openshift-adp"
                    }
                  ],
                  "seedImageRef": {
                    "image": "quay.io/xyz",
                    "version": "4.16.0"
                  },
                  "stage": "Idle"
                }
              },
              {
                "apiVersion": "lca.openshift.io/v1",
                "kind": "SeedGenerator",
                "metadata": {
                  "name": "seedimage"
                },
                "spec": {
                  "seedImage": "quay.io/xyz"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
          description: The Lifecycle Agent for OpenShift provides local lifecycle
            management services for Single Node Openshift (SNO) clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-lifecycle-agent
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-lifecycle-agent",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: |-
            [
              "OpenShift Container Platform",
              "OpenShift Platform Plus"
            ]
          operators.operatorframework.io/builder: operator-sdk-v1.28.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/lifecycle-agent
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ImageBasedUpgrade is the Schema for the ImageBasedUpgrades
              API
            displayName: Image-based Cluster Upgrade
            kind: ImageBasedUpgrade
            name: imagebasedupgrades.lca.openshift.io
            version: v1
          - description: SeedGenerator is the Schema for the seedgenerators API
            displayName: Seed Generator
            kind: SeedGenerator
            name: seedgenerators.lca.openshift.io
            version: v1
        description: "# Lifecycle Agent for OpenShift\nThe Lifecycle Agent for OpenShift
          provides local lifecycle management services \nfor Single Node Openshift
          (SNO) clusters.\n\n## Where to find more information\nYou can find additional
          guidance in the [agent repository](https://github.com/openshift-kni/lifecycle-agent).\n"
        displayName: Lifecycle Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - upgrade
        links:
        - name: Lifecycle Agent
          url: https://lifecycle-agent.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/recert-rhel9@sha256:95e2bf53a32b53943350d63d09595a13419435f8dd91c1019c5708a128cc65a7
        - registry.redhat.io/openshift4/lifecycle-agent-operator-bundle@sha256:1d67bf321a84edd246a4348458369c674b6791492f519388a9008a513e8d836d
        - registry.redhat.io/openshift4/lifecycle-agent-rhel9-operator@sha256:7e8c393bf5a6a1c728f1850867e31025fe9e3a379a5bfd8c01df159bd6de058c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:683e74056df40f38004b2145b8a037dd43b56376061915f37ccb50b5ed19b404
        version: 4.18.0
      entries:
      - name: lifecycle-agent.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: lifecycle-agent
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc
      provider-url: ""
    name: openshift-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openshift-gitops-operator.v1.10.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.8.19\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:b5c8a6c7438a45125fc7fd970946f715ab7193b7446d625fae80dde9d1f3f914
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:b0dc693854319ada341e2e3872b34d7db9823b470ed369edfde88d3ae73a40e1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5d5061dbc2ec5082cf14b6c600fb5400b83cf91d7ccebfa80680a238d275db
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a74e0a1fe3589ebe9f226a2bad20e9a5d5601a0e62c5f3290902f466f55b7160
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:ed530db9b4a29fb591ea208cd0c6b3b83b49537ff5a0a4f501b54ce667f9b303
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:ece2fee4227ad1becadca47028b78ceaf8717a98ec84e3f1b5188a135c4b093c
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:4c842f0ea8f0e9b53eb16d94488fc2f4298bd8e4be2f48713be9157da24549e6
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:c779ad87a7f71d851c2af818abf16e5151c43d3fd54e64eac2fc7117c00753f5
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:0905ba3f5cf87780a1e95b6a00ea7cd340d4ed1b5054e4b2c82406270baeaefa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d35dff5d59e748ef0404e7e76b0237ea96b2bb19b86fa04a3a34df7a379277b2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:2bec31b0c5407ffcecfd8b860d8afb1629eaab31a250f60191f94a824e0a52f4
        version: 1.10.6
      entries:
      - name: openshift-gitops-operator.v1.10.6
        version: 1.10.6
      - name: openshift-gitops-operator.v1.10.5
        version: 1.10.5
      - name: openshift-gitops-operator.v1.10.4
        version: 1.10.4
      - name: openshift-gitops-operator.v1.10.3
        version: 1.10.3
      - name: openshift-gitops-operator.v1.10.2
        version: 1.10.2
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.10
    - currentCSV: openshift-gitops-operator.v1.11.7-0.1724840231.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.11.7
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.9.21\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:959de6a95ef8978f66fdbe1033bcaa0ddeff9a376023e5d1d2698c5acb417305
        - registry.redhat.io/rhel8/redis-6@sha256:c4a98010f7dd5faec2047bb9803061704233a8efc930be2a02386ac16d062d86
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:42f809f32f1d43186b849e3cc78063d3ebb2b54c55b2a1b77908af09a5ce044b
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:eade50b8321c1b2e0a8ac4dcb34596d43dea0a137b8b97b6eb68c6d469c8b200
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c94c2167dc957856682a104b9a6187d3d321dec71ae794acc0dc37794d2cc373
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:0b926fc9eae775425faa1ab9709a36fdac15afe4bfe523720e17ecf5df45bb28
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d6ad2ee900bf99a3672049cc8ce9c94870d41176501ded1abbc1b229b9553401
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:31bdff225bd825d02ce0f468b1edc77541ce85b7b9111f14a583c418c984232f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:4d22a7a6cc2f0442889512267ae3d9356f05ff0dfdbbe137c3d8ebebc24818c8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:0df37cf0a0e48efae02a0711c6da61c6165e7856db97097e98d56c8b6a137bb9
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:7ea96000768b3cf7711150531c4c1c7edff2c45a2f29fe10bd9811ab6b281909
        version: 1.11.7+0.1724840231.p
      entries:
      - name: openshift-gitops-operator.v1.11.7-0.1724840231.p
        version: 1.11.7+0.1724840231.p
      - name: openshift-gitops-operator.v1.11.7
        version: 1.11.7
      - name: openshift-gitops-operator.v1.11.6
        version: 1.11.6
      - name: openshift-gitops-operator.v1.11.5
        version: 1.11.5
      - name: openshift-gitops-operator.v1.11.4
        version: 1.11.4
      - name: openshift-gitops-operator.v1.11.3
        version: 1.11.3
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.11
    - currentCSV: openshift-gitops-operator.v1.12.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.10.18\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbf3038daf7c797123c511f9a609c3485cb1a5bcae53c614fad17e0a65fa7ee3
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a1d4e26ea755181818401697f18260b9e876aa441be14830f9142081ee22a24d
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:7d82c719b1cef64607d30e8aaa5ab293955771c5b1dbd430e0befc988a43356d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:1cf485a889c7586cb4944f000b53c0e97d08f97ba14e580bec57dad7cb13712e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:b5736d1106db29fcdf42328ca20b96382d4339f90b97a07120ea73ee174ea1ad
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:80198f65f7e3f9022a34e0342e9423807b855dcb7bda810172eebc9a27afb5c2
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:33bbb88343285a0510670384b87d0226dd6de8aa12c78b68921362c4d6f5a6aa
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9e0fee30842a16d7b8a39d336123c80c5271cb2ceb50759fc7ee51a4ed09ceb3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:48dd1254fd3a887aef2eae9239f68637eae18514a50f72ea504414d83dc727b8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:73918f3b3c507fc64006b7a500ecfa9d58151d694852f696c41a4412edffc78e
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        version: 1.12.6
      entries:
      - name: openshift-gitops-operator.v1.12.6
        version: 1.12.6
      - name: openshift-gitops-operator.v1.12.5-0.1727371523.p
        version: 1.12.5+0.1727371523.p
      - name: openshift-gitops-operator.v1.12.5
        version: 1.12.5
      - name: openshift-gitops-operator.v1.12.4
        version: 1.12.4
      - name: openshift-gitops-operator.v1.12.3
        version: 1.12.3
      - name: openshift-gitops-operator.v1.12.2
        version: 1.12.2
      - name: openshift-gitops-operator.v1.12.1
        version: 1.12.1
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.12
    - currentCSV: openshift-gitops-operator.v1.13.3-0.1741683398.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          olm.substitutesFor: openshift-gitops-operator.v1.13.3-0.1738142397.p
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.11.12\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3f2804bc2af437ff1f06c214b48a1a780dfcfab995650d901c03270057eb9f3c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:06a6637c4776f4285bcb295d327ee60b1aed61b7adc5f7ab52898d79281ced70
        - registry.redhat.io/rhel8/redis-6@sha256:a848cc4edb4405846c673187dd68ce66c328dd1ad42f193f5737288c7971399d
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b982df02ea7c45a7603d83fe82cac12e1faac303a1b72af6a727512bb97441df
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d4d5724f67859d24c32f0f5735f5af3e325f77db97f58bc7f17ed4b831b7e534
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:a366229e6ee26ebe2d7f765b2a2da6057a29e783835655b5d97545ac563b541a
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a7c3e31ce414e6c6fecf1652343b22c52502902676bf4d7a75bf42bdabfca06
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:e87515d016565bb69cb2c2ac40432e0127dd2cb20a490ac82019836032ba15e5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c02dc2c79ee24ce5450ea4e19493ac2a1997eeb5bfd359291642c529826d0cdd
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7c6bb4a525964d9cdf1c64c646174f91273501dea93224cab5382f81533c6642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:d41034bcf5c28bcb8afef08678793d0cc27a4032e2811a0c496c3bd0f10ecc17
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:f04d565eb5162fd5a9c78cd8443af6a3b184a660277537f1d6d575caa2e85441
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:bde11fa83b7e708677b08ea995041f480b88edfb93e4a62c10380d101c0a349c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a90baa81d2272e681a16a91deacdc1c212b00142b0d9736aff14f772f68be11
        version: 1.13.3+0.1741683398.p
      entries:
      - name: openshift-gitops-operator.v1.13.3-0.1741683398.p
        version: 1.13.3+0.1741683398.p
      - name: openshift-gitops-operator.v1.13.3-0.1738142397.p
        version: 1.13.3+0.1738142397.p
      - name: openshift-gitops-operator.v1.13.3
        version: 1.13.3
      - name: openshift-gitops-operator.v1.13.2
        version: 1.13.2
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.13
    - currentCSV: openshift-gitops-operator.v1.14.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
          createdAt: "2025-05-22 12:40:45"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.12.13-1\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_gitops/
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:a9f30ed08abcab0122556a5485fbcf6e0fc03860f19954dcebdbf5e754876a45
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:bcf6b3136a5e71803cb518c8763c165bb46ee37ef7bf55503911d7fcb36b3a8f
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:773592dfb9964970f444871a8812e8b74db3f205faaa0bf148defa1a2b87b649
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:97e32825f59c33cca7d8dba3580095205c20435ccd15b91cf0aff37db9841e0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:42a79aad641e8607fc339e4ce7f9cf1d258be01f863dbc31a94d2c27edb645bd
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:1112160243f227a5d393eeee165bd4fed0d3f42646a9280cbe9af60ef728aeb2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c369df980874aae270ca7bc6455f1ca45467d75d25fe80d7bcb06a7826bdeaa3
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8096f824bd96e3148cf4b6936766c888d1fff0af9e9d0be440ae4feb6adc03f2
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:7cf924af77d9c56e70710b99e080ed8aee162f05ec472f32085a76ef85cf9879
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:2c45cfc01f9ec7577afc09cadc4c1a6f470836f6a0a7c24d66b791913f89bb05
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:b7d3c882fc25fbbea28951afb197d4d47ace253864889339396e20f0c6897a7f
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:7358787959af7e7e589edfc043a97d80ebab87536bde365d33be7a600071bf0e
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:cd4965d91c163160bcf43a02c54ca60b8704281c95de904f904df942bfd83bcd
        version: 1.14.4
      entries:
      - name: openshift-gitops-operator.v1.14.4
        version: 1.14.4
      - name: openshift-gitops-operator.v1.14.3-0.1746016855.p
        version: 1.14.3+0.1746016855.p
      - name: openshift-gitops-operator.v1.14.3-0.1742836272.p
        version: 1.14.3+0.1742836272.p
      - name: openshift-gitops-operator.v1.14.3
        version: 1.14.3
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.14
    - currentCSV: openshift-gitops-operator.v1.15.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
          createdAt: "2024-11-25T03:57:37Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.13.9\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:a9e2360f00c5c9840a100a6748bd905b8f34169140a33fb0ca4af9cf6db71642
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ad49a3d14f843bde0d2365093d00ffa6d3b9d59a75c81e0c56710ad37b53bd92
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:0c7f72546d0d918bfd21aff7157b983ff7c832b269f5efb84d2ed723a09bf612
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:911733d78ae6f36f642b4d63f044cd102b7eee170508eaf518ac7acf0f339698
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:5bda26e2c16beebf66dd8f3c3c55d48e2f00db870f2d9d4d44570d9d3b8bddb0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:0b1771f697a700263c5b827dbde56a5c9bab610e9d7df55ef93885a4e54cc0e9
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:d41454824632475f0e88aa6e0031ec643bd37e1d575df36cd3ee63daf1403281
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:41717c27915251bda8afeef651b609b8a9ed410e0fa09c4cf8fa7e91d6a94292
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:6cec76d7ea70d77923cf4285177e375d34e96e3aecbe77b68cd5ff769e6e540b
        version: 1.15.4
      entries:
      - name: openshift-gitops-operator.v1.15.4
        version: 1.15.4
      - name: openshift-gitops-operator.v1.15.3
        version: 1.15.3
      - name: openshift-gitops-operator.v1.15.2
        version: 1.15.2
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.15
    - currentCSV: openshift-gitops-operator.v1.16.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
          createdAt: "2025-03-07T13:13:07Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v2.14.16\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:ceff1cde6d043daae599b578eec9759c6d49cd5cf770460b9d075fa879a16e81
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:58284a12174d19c2d04723fa1f551d96b490872e0fa2ad94af1b3e52dc6ad955
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:cab809fd1288d31c93f4f5c4facb7efaff9f844e21a859d1e4557831575a13c8
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a5e2a12ec5dea607a6651462d8e7920cc73f3272dba4041c397a14f2c73d7532
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:56653df97e4498ecd67b9645d75e44f65c625594e312dbcf0f09937b57d76010
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:03d0d9093422396330aa8611adf77abdc9be759f3d2e23850e4abe778bd6002b
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:2598a3339bc920b91018d3f7acbe78ccbee684b5182c874e48296744d2373f9d
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8441322a0fce02df407573675b5b5d92d56de97c8aec72541b33d2dadc4050d5
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:ead539d7cf54992cf1bfeae6a73a453059a7dffef44505991ff33319c8d40c50
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c37aa141387c847eac825363dddda3eb59203d3694120dba439b00cf0c66f320
        version: 1.16.3
      entries:
      - name: openshift-gitops-operator.v1.16.3
        version: 1.16.3
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.16
    - currentCSV: openshift-gitops-operator.v1.17.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
          createdAt: "2025-07-30T13:03:16Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.0.14\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:5c9ea426cd60e7b8d1d8e4fe763909200612434c65596855334054e26cbfe3d0
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:4d945dd1a6fc0e8b6cfc16f4359128f1427b094e76efd4ff31325a1839246107
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:cc45e212e7eb4e7e8daac6ea74071d5d98260de0b1f6d1cea0e7876111ccea8c
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:be93be0341649215fb001725aac0c3c5925343adbf1ad00c979b8c3d489512d6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:d73b4f5b355f556208751281923bb8b43edffc061a673c04d53b88f0e50ecb26
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:2dd72760f736ab13e87d6e507e654c390d2d6af3154c6adb85610b5f58085b9c
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:3fc4941fa4a6f2dbe9dd82622a0d3fb794ecd5a9dee286a38b60a7f6a947d685
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:2f5f997bce924445de735ae0508dca1a7bba561bc4acdacf659928488233cb8a
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:c0a513ed54b631629efafe6e94b775d5bd0a943c44fbe96f02a0d5bb1f989099
        - registry.redhat.io/rhel9/redis-7@sha256:2fca0decc49230122f044afb2e7cd8f64921a00141c8c22c2f1402f3564f87f8
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:1b86eb4e071b1c526931b6d18c3d8f51522293699059c514488602c320271049
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:a2ae66772d90e734bb3385a59852e6016325a21b916a2580a33cf0da7cb0b527
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:70e5e3900d0c89999c69be1d58adf84c1b3664aacbffa9c5e0efc8ae1b6ef9c1
        version: 1.17.1
      entries:
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      name: gitops-1.17
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      name: gitops-1.18
    - currentCSV: openshift-gitops-operator.v1.6.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.4.28\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b22190de3162d3e03481cae8904ce563912756eea7ac7e7feec0acbbdae5e6f6
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0cc4fa704e581cd5581c816668183f602fae6ef9ab3c04ce0d99d561287ef140
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:104cc7564fcc3261faa3fbf9c95720b4434e15103b876c080ed012b244ef1dae
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:dbcad68b70ab7918e944099aef4177c9bc3378e9ca49499dbb94a175927d3209
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:06f1150114d4c0bd236d2a5d4716c9c3e95a9733058297790389d6ba2a1bf264
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:e9e2f950cda8153cff112e43872b98e4977d3cbdf59713b8de92defef6d34b6d
        version: 1.6.6
      entries:
      - name: openshift-gitops-operator.v1.6.6
        version: 1.6.6
      - name: openshift-gitops-operator.v1.6.5
        version: 1.6.5
      - name: openshift-gitops-operator.v1.6.4
        version: 1.6.4
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.6
    - currentCSV: openshift-gitops-operator.v1.7.4-0.1690486082.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
          createdAt: "2021-09-15T22:31:21Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.substitutesFor: openshift-gitops-operator.v1.7.4
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD v2.5.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficient privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces. \n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc.
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:53598a6effeb90e4f1b005b2521beffd2fa2b0c52d0e7f2347ee2abd2577cab3
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:e0967a4058fe5a9bfe7e998a3347529c99536cf3a44de9e40f9632035ff5a2fe
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:c55aadf45397083bec4e42d6c0514ed79169a07fd3d9d75671904df2c0520e52
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:26050fb95978e90e8d00e895032045e77cb236008a0e1417459e1aeca8a48c69
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:5c11be8ec9db54331893f2f298aa27006e77b0eaded2d88d0531c77da91a421f
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:c2f39be75173b4cc57498c5845b14c464bf9af92b49716df9661c8135ac2cf7a
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f1371d2012796e9aaca105112190ff89ec3d0be95968e05b19db101480f67975
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:45789c49d9334e0ed062641313714c101bd0117dc5633316b1ec63e7a6c76c15
        version: 1.7.4+0.1690486082.p
      entries:
      - name: openshift-gitops-operator.v1.7.4-0.1690486082.p
        version: 1.7.4+0.1690486082.p
      - name: openshift-gitops-operator.v1.7.4
        version: 1.7.4
      - name: openshift-gitops-operator.v1.7.3
        version: 1.7.3
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.7
    - currentCSV: openshift-gitops-operator.v1.8.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.6.153\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:8cc59901689858e06f5c2942f8c975449c17454fa8364da6153d9b5a3538a985
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:b2d58f4dc20eb7e17e09af0e8c5de55fa04f9f4f697e10c6be81a7a6a08381b2
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:f77594bc053be144b33ff9603ee9675c7e82946ec0dbfb04d8f942c8d73155da
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:0d13c03048b4435d9f5045a6fb7bd4b8808461b26aeca60c0047fa5937e1d45d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:16d9184345e6d5a19f326a7b52f76d8a59b96763f233c50546ff017a625e5ff5
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:d5829e880db4b82a50a4962d61ea148522a93644174931b256d7ad866eadcf40
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:8aa68fddef88921aeeede24a849d3dabbceff01dc03185342c21e6ee58033eaf
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f32df77b3c23d6a833ff887386285ca7a5bd436641f0ec1ffc934c8247fa315f
        version: 1.8.6
      entries:
      - name: openshift-gitops-operator.v1.8.6
        version: 1.8.6
      - name: openshift-gitops-operator.v1.8.5
        version: 1.8.5
      - name: openshift-gitops-operator.v1.8.4-0.1693214052.p
        version: 1.8.4+0.1693214052.p
      - name: openshift-gitops-operator.v1.8.4
        version: 1.8.4
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.8
    - currentCSV: openshift-gitops-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "argo-rollout"
                },
                "spec": null
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          olm.skipRange: ""
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n*
          Automated GitOps bootstrapping using Tekton and Argo CD with [GitOps Application
          Manager CLI](https://github.com/redhat-developer/kam)\n\n## Components\n*
          Argo CD 2.7.16\n* GitOps Application Manager CLI ([download](https://github.com/redhat-developer/kam/releases))\n\n##
          How to Install \nAfter installing the OpenShift GitOps operator, an instance
          \ of Argo CD is installed in the `openshift-gitops` namespace which has
          sufficent privileges for managing cluster configurations. You can create
          additional Argo CD instances using the `ArgoCD` custom resource within the
          desired namespaces.\n```yaml\napiVersion: argoproj.io/v1alpha1\nkind: ArgoCD\nmetadata:\n
          \ name: argocd\nspec:\n  server:\n    route:\n      enabled: true\n```\n\nOpenShift
          GitOps is a layered product on top of OpenShift that enables teams to adopt
          GitOps principles for managing cluster configurations and automating secure
          and repeatable application delivery across hybrid multi-cluster Kubernetes
          environments. OpenShift GitOps is built around Argo CD as the core upstream
          project and assists customers to establish an end-to-end application delivery
          workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        links:
        - name: Release Notes
          url: https://docs.openshift.com/container-platform/4.10/cicd/gitops/gitops-release-notes.html
        - name: Day 1 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day1
        - name: Day 2 Operations
          url: https://github.com/redhat-developer/kam/tree/master/docs/journey/day2
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:136d3a936759e112d3ea2e78f7c02bd882b0fd3011f4c452f45c32c43822772d
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:43bb961d847477902e66a92edd1539322cac8d5f74706b048a458ad037a0c7a3
        - registry.redhat.io/rhel8/redis-6@sha256:edbd40185ed8c20ee61ebdf9f2e1e1d7594598fceff963b4dee3201472d6deda
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:3873d1f18b5643e1de2601419410270342650a21428bb26a4cbee20c0c63f98b
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:11a6084e11aa32511a34bd6cfb4f4995aec690c1f5ed61861e99f6e42d62afb4
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:a568c8c647e2e2d6abdf12138eeef8f8bdde1fdcbf786e9ca8c8f93bc79acaf9
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:86012145d4b8798dc12f6f4f07a9316a193188aaac865e324b63f4e60e8ec584
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:27efa0ab1d9c69047a924bc1dcd2b56108f5f390a6da56afb6ab7a37cb4578e2
        - registry.redhat.io/openshift-gitops-1/kam-delivery-rhel8@sha256:80f92842fe4cba008a873ad2ab6059a3a24e76e16671d8ece034dbbe666996bc
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:edf7ce748b703e195220b7bd7b42fa2caa4cdfd96840445e096036a0d85f1ff2
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ec9f60018694dcc5d431ba47d5536b761b71cb3f66684978fe6bb74c157679ac
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:3d17786e2f4063154b9e3a40b96bf8ec0a14b86395cea5fe7ca18c5b43764478
        version: 1.9.4
      entries:
      - name: openshift-gitops-operator.v1.9.4
        version: 1.9.4
      - name: openshift-gitops-operator.v1.9.3
        version: 1.9.3
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: gitops-1.9
    - currentCSV: openshift-gitops-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Rollout",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "RolloutManager",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "rbac": {
                    "defaultPolicy": "",
                    "policy": "g, system:cluster-admins, role:admin\n",
                    "scopes": "[groups]"
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "resourceExclusions": "- apiGroups:\n  - tekton.dev\n  clusters:\n  - '*'\n  kinds:\n  - TaskRun\n  - PipelineRun        \n",
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "openShiftOAuth": true,
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "pipelines.openshift.io/v1alpha1",
                "kind": "GitopsService",
                "metadata": {
                  "name": "gitopsservice-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          console.openshift.io/plugins: '["gitops-plugin"]'
          containerImage: registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
          createdAt: "2025-08-21T01:20:45Z"
          description: Enables teams to adopt GitOps principles for managing cluster
            configurations and application delivery across hybrid multi-cluster Kubernetes
            environments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.18.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-gitops-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["gitopsservices.pipelines.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-developer/gitops-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An AnalysisRun is an instantiation of an AnalysisTemplate.
              AnalysisRuns are like Jobs in that they eventually complete.
            displayName: AnalysisRun
            kind: AnalysisRun
            name: analysisruns.argoproj.io
            version: v1alpha1
          - description: An AnalysisTemplate is a template spec which defines how
              to perform a canary analysis, such as the metrics, its frequency, and
              the values which are considered successful or failed.
            displayName: AnalysisTemplate
            kind: AnalysisTemplate
            name: analysistemplates.argoproj.io
            version: v1alpha1
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: ApplicationSet is the representation of an ApplicationSet
              controller deployment.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: Argo CD is the representation of an Argo CD deployment.
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: A ClusterAnalysisTemplate is like an AnalysisTemplate,but
              it is not limited to its namespace. It can be used by any Rollout throughout
              the cluster.
            displayName: ClusterAnalysisTemplate
            kind: ClusterAnalysisTemplate
            name: clusteranalysistemplates.argoproj.io
            version: v1alpha1
          - description: An Experiment is limited run of one or more ReplicaSets for
              the purposes of analysis. Experiments typically run for a pre-determined
              duration, but can also run indefinitely until stopped.
            displayName: Experiment
            kind: Experiment
            name: experiments.argoproj.io
            version: v1alpha1
          - description: GitopsService is the Schema for the gitopsservices API
            displayName: Gitops Service
            kind: GitopsService
            name: gitopsservices.pipelines.openshift.io
            version: v1alpha1
          - kind: NamespaceManagement
            name: namespacemanagements.argoproj.io
            version: v1beta1
          - description: NotificationsConfigurations contains the notification template
              used to generate notifications that notify users about important changes
              in the application state.
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
          - description: A controller for managing Argo Rollouts
            displayName: RolloutManager
            kind: RolloutManager
            name: rolloutmanagers.argoproj.io
            version: v1alpha1
          - kind: Rollout
            name: rollouts.argoproj.io
            version: v1alpha1
        description: "Red Hat OpenShift GitOps is a declarative continuous delivery
          platform based on [Argo CD](https://argoproj.github.io/argo-cd/). It enables
          teams to adopt GitOps principles for managing cluster configurations and
          automating secure and repeatable application delivery across hybrid multi-cluster
          Kubernetes environments. Following GitOps and infrastructure as code principles,
          you can store the configuration of clusters and applications in Git repositories
          and use Git workflows to roll them out to the target clusters.\n\n## Features\n*
          Automated install and upgrades of Argo CD\n* Manual and automated configuration
          sync from Git repositories to target OpenShift and Kubernetes clusters\n*
          Support for the Helm and Kustomize templating tools\n* Configuration drift
          detection and visualization on live clusters\n* Audit trails of rollouts
          to the clusters\n* Monitoring and logging integration with OpenShift\n##Components\n*
          Argo CD v3.1.6\n\n## How to Install \nAfter installing the OpenShift GitOps
          operator, an instance  of Argo CD is installed in the `openshift-gitops`
          namespace which has sufficient privileges for managing cluster configurations.
          You can create additional Argo CD instances using the `ArgoCD` custom resource
          within the desired namespaces.\n```yaml\napiVersion: argoproj.io/v1beta1\nkind:
          ArgoCD\nmetadata:\n  name: argocd\nspec:\n  server:\n    route:\n      enabled:
          true\n```\n\nOpenShift GitOps is a layered product on top of OpenShift that
          enables teams to adopt GitOps principles for managing cluster configurations
          and automating secure and repeatable application delivery across hybrid
          multi-cluster Kubernetes environments. OpenShift GitOps is built around
          Argo CD as the core upstream project and assists customers to establish
          an end-to-end application delivery workflow on GitOps principles.\n"
        displayName: Red Hat OpenShift GitOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - devtools
        - gitops
        - pipelines
        maintainers:
        - email: team-gitops@redhat.com
          name: OpenShift GitOps Team
        maturity: GA
        provider:
          name: Red Hat Inc
        relatedImages:
        - registry.redhat.io/openshift-gitops-1/argo-rollouts-rhel8@sha256:c2283997d47298d49378e8200b149ea613c7735317ef12e74b7915a590f41ce2
        - registry.redhat.io/openshift-gitops-1/dex-rhel8@sha256:adca9c55b80d899d500bf0a7234cf61ae30ffed1b2c927ddea8d9fd07c9acb6a
        - registry.redhat.io/openshift4/ose-haproxy-router@sha256:7f6ceae56c60fd432425c36f93a9e9bbbb96ef9422333ad6ea41ff3a2eaf8374
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/openshift-gitops-1/argocd-extensions-rhel8@sha256:1eef9ecd58cf8391f29360ef74dd8aa5dcb48e1582d1500e445706d8302c951e
        - registry.redhat.io/openshift-gitops-1/gitops-operator-bundle@sha256:8546401f9a96b0c397da8ab26306cca412e6fa6d11471c739177a5f533ba6f11
        - registry.redhat.io/openshift-gitops-1/argocd-agent-rhel8@sha256:08efd3e7a79c2945693c99ed31ec912223becff7238b1e28d1b1da5d63f93019
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/openshift-gitops-1/console-plugin-rhel8@sha256:640e3298c7cd2faa84971b0d861446c4892ab8602108af05b8b5b049a4de624f
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8-operator@sha256:f6cddda1dc00211d1896bcf0f331bd7fa3962d632ae8b9fe66ddd31cd8a361b0
        - registry.redhat.io/openshift-gitops-1/gitops-rhel8@sha256:6284873bcf1593c210003e0292ac5583f03f03e80d937098d808d48fa0d01cf6
        - registry.redhat.io/rhel9/redis-7@sha256:b4f58ad83ab8719a2e102f38ee526546cd1d56428901afc3add45c9a0e2a2f12
        - registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:8a0544c14823492165550d83a6d8ba79dd632b46144d3fdcb543793726111d76
        - registry.redhat.io/openshift-gitops-1/must-gather-rhel8@sha256:5a09fcba3d9e6fc462719e81a46af6022cf7f70a05bcb387ed959a49f6f37dbc
        version: 1.18.0
      entries:
      - name: openshift-gitops-operator.v1.18.0
        version: 1.18.0
      - name: openshift-gitops-operator.v1.17.1
        version: 1.17.1
      - name: openshift-gitops-operator.v1.17.0
        version: 1.17.0
      - name: openshift-gitops-operator.v1.16.2
        version: 1.16.2
      - name: openshift-gitops-operator.v1.16.1
        version: 1.16.1
      - name: openshift-gitops-operator.v1.16.0-0.1746014725.p
        version: 1.16.0+0.1746014725.p
      - name: openshift-gitops-operator.v1.16.0
        version: 1.16.0
      - name: openshift-gitops-operator.v1.15.1
        version: 1.15.1
      - name: openshift-gitops-operator.v1.15.0
        version: 1.15.0
      - name: openshift-gitops-operator.v1.15.0-0.1738074324.p
        version: 1.15.0+0.1738074324.p
      - name: openshift-gitops-operator.v1.14.2
        version: 1.14.2
      - name: openshift-gitops-operator.v1.14.2-0.1738140086.p
        version: 1.14.2+0.1738140086.p
      - name: openshift-gitops-operator.v1.14.1
        version: 1.14.1
      - name: openshift-gitops-operator.v1.14.0
        version: 1.14.0
      - name: openshift-gitops-operator.v1.13.1
        version: 1.13.1
      - name: openshift-gitops-operator.v1.13.0
        version: 1.13.0
      - name: openshift-gitops-operator.v1.12.0
        version: 1.12.0
      - name: openshift-gitops-operator.v1.11.2
        version: 1.11.2
      - name: openshift-gitops-operator.v1.11.1
        version: 1.11.1
      - name: openshift-gitops-operator.v1.11.0
        version: 1.11.0
      - name: openshift-gitops-operator.v1.10.1
        version: 1.10.1
      - name: openshift-gitops-operator.v1.10.0
        version: 1.10.0
      - name: openshift-gitops-operator.v1.9.2
        version: 1.9.2
      - name: openshift-gitops-operator.v1.9.1
        version: 1.9.1
      - name: openshift-gitops-operator.v1.9.0
        version: 1.9.0
      - name: openshift-gitops-operator.v1.8.3
        version: 1.8.3
      - name: openshift-gitops-operator.v1.8.2
        version: 1.8.2
      - name: openshift-gitops-operator.v1.8.1
        version: 1.8.1
      - name: openshift-gitops-operator.v1.8.0
        version: 1.8.0
      - name: openshift-gitops-operator.v1.7.2
        version: 1.7.2
      - name: openshift-gitops-operator.v1.7.1
        version: 1.7.1
      - name: openshift-gitops-operator.v1.7.0
        version: 1.7.0
      - name: openshift-gitops-operator.v1.6.3
        version: 1.6.3
      - name: openshift-gitops-operator.v1.6.2
        version: 1.6.2
      - name: openshift-gitops-operator.v1.6.1-0.1666124167.p
        version: 1.6.1+0.1666124167.p
      - name: openshift-gitops-operator.v1.6.1
        version: 1.6.1
      - name: openshift-gitops-operator.v1.6.0
        version: 1.6.0
      - name: openshift-gitops-operator.v1.5.4
        version: 1.5.4
      - name: openshift-gitops-operator.v1.5.3
        version: 1.5.3
      - name: openshift-gitops-operator.v1.5.2
        version: 1.5.2
      name: latest
    defaultChannel: latest
    packageName: openshift-gitops-operator
    provider:
      name: Red Hat Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: volume-expander-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: volume-expander-operator.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage,Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
          createdAt: "2022-05-05T13:25:53Z"
          description: This operator provides automation expanding pod's volumes when
            they are about to run out of space.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: volume-expander-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/volume-expander-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The purpose of the volume-expander-operator is to expand volumes when they are running out of space.
          This is achieved by using the [volume expansion feature](https://kubernetes.io/docs/concepts/storage/persistent-volumes/#expanding-persistent-volumes-claims).

          The operator periodically checks the `kubelet_volume_stats_used_bytes` and `kubelet_volume_stats_capacity_bytes published` by the kubelets to decide when to expand a volume.
          Notice that these metrics are generated only when a volume is mounted to a pod. Also the kubelet takes a minute or two to start generating accurate values for these metrics. The operator accounts for that.

          This operator works based on the following annotations to PersistentVolumeClaim resources:

          | Annotation | Default &nbsp;  &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;  | Description  |
          | :--- | :--- | :--- |
          | `volume-expander-operator.redhat-cop.io/autoexpand`  | N/A  | if set to "true" enables the volume-expander-operator to watch on this PVC  |
          | `volume-expander-operator.redhat-cop.io/polling-frequency`  | `"30s"` | How frequently to poll the volume metrics. Express this value as a valid golang [Duration](https://golang.org/pkg/time/#ParseDuration)  |
          | `volume-expander-operator.redhat-cop.io/expand-threshold-percent` | `"80"` | the percentage of used storage after which the volume will be expanded. This must be a positive integer. |
          | `volume-expander-operator.redhat-cop.io/expand-by-percent` | `"25"` | the percentage by which the volume will be expanded, relative to the current size. This must be an integer between 0 and 100 |
          | `volume-expander-operator.redhat-cop.io/expand-up-to` | MaxInt64 | the upper bound for this volume to be expanded to. The default value is the largest quantity representable and is intended to be interpreted as infinite. If the default is used it is recommend to ensure the namespace has a quota on the used storage class. |

          Note that not all of the storage driver implementations support volume expansion. It is a responsibility of the user/platform administrator to ensure that storage class and the persistent volume claim meet all the requirements needed for the volume expansion feature to work properly.

          This operator was tested with [OCS](https://www.redhat.com/en/technologies/cloud-computing/openshift-container-storage), but should work with any other storage driver that supports volume expansion.
        displayName: Volume Expander Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - volume
        - storage
        - csi
        - expansion
        links:
        - name: Volume Expander Operator
          url: https://github.com/redhat-cop/volume-expander-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/volume-expander-operator@sha256:b775df99f891e2cee9c2169e9c948a8d61d65b4cc90f36752b46a56e2d4be602
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/volume-expander-operator@sha256:e4059a66a9af7c786c8c651abb8a57a55a0e528342e3a59f8c264b40f87d6e25
        version: 0.3.6
      entries:
      - name: volume-expander-operator.v0.3.6
        version: 0.3.6
      - name: volume-expander-operator.v0.3.5
        version: 0.3.5
      - name: volume-expander-operator.v0.3.4
        version: 0.3.4
      - name: volume-expander-operator.v0.3.3
        version: 0.3.3
      - name: volume-expander-operator.v0.3.2
        version: 0.3.2
      - name: volume-expander-operator.v0.3.1
        version: 0.3.1
      - name: volume-expander-operator.v0.3.0
        version: 0.3.0
      - name: volume-expander-operator.v0.2.0
        version: 0.2.0
      - name: volume-expander-operator.v0.1.8
        version: 0.1.8
      - name: volume-expander-operator.v0.1.7
        version: 0.1.7
      - name: volume-expander-operator.v0.1.6
        version: 0.1.6
      - name: volume-expander-operator.v0.1.5
        version: 0.1.5
      - name: volume-expander-operator.v0.1.4
        version: 0.1.4
      - name: volume-expander-operator.v0.1.3
        version: 0.1.3
      - name: volume-expander-operator.v0.1.2
        version: 0.1.2
      - name: volume-expander-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: volume-expander-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: APIMatic.io
      provider-url: https://www.apimatic.io/
    name: apimatic-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apimatic-kubernetes-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicodegen.apimatic.io/v1beta1",
                "kind": "APIMatic",
                "metadata": {
                  "name": "apimatic-sample"
                },
                "spec": {
                  "licensespec": {
                    "licenseSourceName": "apimaticlicense",
                    "licenseSourceType": "ConfigMap"
                  },
                  "podspec": {
                    "apimaticContainerSpec": {
                      "image": "apimaticio/apimatic-codegen",
                      "imagePullSecret": "apimaticimagesecret"
                    }
                  },
                  "replicas": 3,
                  "servicespec": {
                    "apimaticserviceport": {
                      "nodePort": 32000,
                      "port": 8070
                    },
                    "servicetype": "NodePort"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
          createdAt: "2021-08-24 12:00:00"
          description: Generate client SDKs and interactive Documentation for your
            APIs in minutes
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apimatic/apimatic-kubernetes-operator
          support: APIMatic.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMatic is the Schema for the apimatics API
            displayName: APIMatic App
            kind: APIMatic
            name: apimatics.apicodegen.apimatic.io
            version: v1beta1
        description: "##About APIMatic \n\n APIMatic is a Developer Experience   platform
          for Web APIs that empowers users to create multi-language SDKs with API
          reference docs, dynamic code samples, assisted authentication and more -
          all auto-generated to provide a comprehensive developer experience beyond
          any API documentation tool.\n\n## Table of contents\n\n* [Introduction](#introduction)\n*
          [Features](#features)\n* [Running the Sample](#running-the-sample)\n  *
          [Prerequisites](#prerequisites)\n  * [Steps for Direct Deployment](#steps-for-direct-deployment)\n
          \ * [Steps for OLM Deployment](#steps-for-olm-deployment)\n* [Technical
          Support](#technical-support)\n* [Copyrights](#copyrights)\n\n## Introduction\n\nAPIMatic
          Operator simplifies the configuration and lifecycle management of the APIMatic
          code and docs generation solution on different Kubernetes distributions
          and OpenShift. The Operator encapsulates key operational knowledge on how
          to configure and upgrade the APIMatic CodeGen application, making it easy
          to get up and running.\n\n\nMore information about the underlying APIMatic
          CodeGen API that is exposed\nby this operator can be found [here](https://apimatic-core-v3-docs.netlify.app/#/http/getting-started/overview-apimatic-core).\n\n##
          Features\n\nAPIMatic Operator provides the following features:\n- Deploys
          the APIMatic CodeGen Web API service within the Kubernetes or OpenShift
          cluster.\n- Exposes the APIMatic CodeGen API external to the cluster, using
          Service type as [NodePort](https://kubernetes.io/docs/concepts/services-networking/service/#nodeport),
          [LoadBalancer](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          or [ExternalName](https://kubernetes.io/docs/concepts/services-networking/service/#externalname).\n-
          For exposing the service through an ingress resource, create an Ingress
          resource in the namespace of your APIMatic CR and set owned APIMatic service
          created by the operator as a backed service. More information can be found
          [here](https://kubernetes.io/docs/concepts/services-networking/ingress/).\n-
          Manual horizontal scaling of pods.\n  ```sh\n  kubectl scale apm apimatic-sample--replicas=2\n
          \ ```\n\n## Running the Sample \n\n### Prerequisites\n\nPlease contact APIMatic
          at [support@apimatic.io](mailto:support@apimatic.io) to register with the
          APIMatic CodeGen Docker Hub registry and acquire a valid license to run
          the APIMatic CodeGen API.\n\nFurther prerequisites for running the sample
          include:\n\n- [go](https://golang.org/) v1.16.*\n- [git](https://git-scm.com/)\n-
          [make](https://www.gnu.org/software/make/)\n- [Operator SDK](https://sdk.operatorframework.io/docs/overview/)\n-
          A running Kubernetes cluster with [kubectl](https://kubernetes.io/docs/reference/kubectl/overview/)
          on client. For testing purposes, you can use [Minikube](https://minikube.sigs.k8s.io/docs/)
          or [kind](https://kind.sigs.k8s.io/)\n- For checking the service created
          by the APIMatic operator on-prem, you can use [MetalLB](https://metallb.org/)\n\n###
          Steps for Direct Deployment\n\nTo run the sample for checking the APIMatic
          operator:\n\n- Clone the APIMatic repository into your working directory
          using the following command:\n  ```sh\n  git clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n
          \ ```\n- Run `make deploy` to set up the APIMatic operator resources. This
          will deploy the '`apimatic-system`' namespace as well as the CRD and the
          RBAC manifests.\n\n- Create a secret named '`apimaticimagesecret`' to allow
          pulling the APIMatic CodeGen image using the Docker Hub username registered
          with APIMatic. If you have not done so, please contact support@apimatic.io
          for the steps required.\n  ```sh\n  kubectl create secret docker-registry
          apimaticimagesecret --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named '`apimaticlicense`' that will serve as the volume storing
          the APIMatic license. An example of this is given below which will create
          the ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n\n- Once done, you can remove the APIMatic resources using the
          following command:\n  ```sh\n  make undeploy\n  ```\n\n### Steps for OLM
          Deployment\n\nThe following steps can be used to utilize [Operator LifeCycle
          Manager (OLM)](https://olm.operatorframework.io/docs/) to deploy the operator
          and run the sample. The steps are as follows:\n\n- If not already done so,
          clone the APIMatic repository into your working directory:\n  ```sh\n  git
          clone https://github.com/apimatic/apimatic-kubernetes-operator.git\n  ```
          \n- [Install OLM in your Kubernetes cluster](https://olm.operatorframework.io/docs/getting-started/#installing-olm-in-your-cluster).\n\n-
          Run the following script to install the resources required by OLM to deploy
          the APIMatic operator in the Kubernetes cluster within the `apimatic-system`
          namespace. Information about the different resources required can be found
          using the steps given [here](https://olm.operatorframework.io/docs/tasks/).\n
          \ ```sh\n  kubectl apply -f olm/manifests.yaml\n  ```\n\n- This should spin
          up the `ClusterServiceVersion` of the operator in the `apimatic-system`
          namespace, following which the operator pod will spin up. To ensure the
          that the operator installed successfully, check for the `ClusterServiceVersion`
          and the operator deployment in the namespace it was installed in.\n  ```sh\n
          \ kubectl get csv -n apimatic-system\n  \n  kubectl get deployment -n apimatic-system\n
          \ ```\n\n  - Create a secret named `apimaticimagesecret` to allow pulling
          the APIMatic CodeGen image using the Docker Hub username registered with
          APIMatic. If you have not done so, please contact support@apimatic.io for
          the steps required.\n  ```sh\n  kubectl create secret docker-registry apimaticimagesecret
          --docker-server=https://index.docker.io/v1/ --docker-username=<your-name>
          --docker-password=<your-pword> --docker-email=<your-email>\n  ```\n- Create
          a [configmap](https://kubernetes.io/docs/concepts/configuration/configmap/)
          resource named `apimaticlicense` that will serve as the volume storing the
          APIMatic license. An example of this is given below which will create the
          ConfigMap using the *License.lic* file located at */usr/local/apimatic/license/*\n
          \ ```sh\n  kubectl create configmap apimaticlicense --from-file /usr/local/apimatic/license/License.lic\n
          \ ```\n- This will deploy a ConfigMap resource with the following definition:\n
          \ ```sh\n  apiVersion: v1\n  data:\n    License.lic: \\\"<License file contents
          here>\\\"\n  kind: ConfigMap\n  metadata:\n    name: apimaticlicense\n    namespace:
          default\n  ```\n- Now run the sample using the following command:\n  ```sh\n
          \ kubectl apply -f config/samples/apicodegen_v1beta1_apimatic.yaml\n  ```\n\n-
          You will now see a new [Deployment](https://kubernetes.io/docs/concepts/workloads/controllers/deployment/)
          with replica count of 3 and [Service](https://kubernetes.io/docs/concepts/services-networking/service/)
          of type `NodePort` created, both named ***apimatic-sample***. Accessing
          http://localhost:32000 from your browser (or using curl from within the
          Minikube or Kind cluster, if using that) should now show the APIMatic Titan
          page.\n- You can now use the exposed APIMatic CodeGen API to generate API
          SDKs and Docs using [curl](https://curl.se/), [Postman](https://www.postman.com/)
          or your own custom Web application that consumes the APIMatic CodeGen API
          service.\n- Once done, you can remove the APIMatic operator resources using
          the follow script:\n  ```sh\n  kubectl delete -f olm/manifests.yaml\n  ```\n\n##
          Technical Support\n\n- To request additional features in the future, or
          if you notice any discrepancy regarding this document, please drop an email
          to [support@apimatic.io](mailto:support@apimatic.io).\n\n### Copyrights\n\n&copy;
          2021 APIMatic.io\n"
        displayName: APIMatic Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Rest APIs
        - API SDK Generator
        - API Docs Generator
        - API Developer Portal
        - APIMatic
        - Developer Tools
        links:
        - name: Overview
          url: https://www.apimatic.io/
        - name: Developer Experience Portal
          url: https://www.apimatic.io/developer-experience-portal
        - name: Code-Gen as a Service
          url: https://www.apimatic.io/continuous-code-generation
        - name: API Transformer
          url: https://www.apimatic.io/transformer
        - name: Contact
          url: https://www.apimatic.io/contact
        - name: Documentation
          url: https://docs.apimatic.io/
        - name: Blogs
          url: https://www.apimatic.io/blog/
        maintainers:
        - email: support@apimatic.io
          name: Support
        maturity: alpha
        provider:
          name: APIMatic.io
          url: https://www.apimatic.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/apimatic-kubernetes-operator@sha256:a5f68205291caa9bb7903180afc974fe86252ba87e671978c8b49e0206976d06
        - docker.io/apimaticio/apimatic-kubernetes-operator:v0.0.1
        version: 0.0.1
      entries:
      - name: apimatic-kubernetes-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: apimatic-kubernetes-operator
    provider:
      name: APIMatic.io
      url: https://www.apimatic.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MySQL
      provider-url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
    name: ndb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndb-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "mysql.oracle.com/v1","kind": "NdbCluster","metadata":
            {"name": "example-ndb"},"spec": {"redundancyLevel": 2,"dataNode": {"nodeCount":
            2},"mysqlNode": {"nodeCount":2}}}]'
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
          createdAt: "2024-12-18T15:05:46.000Z"
          description: The MySQL NDB Operator is a Kubernetes operator for managing
            a MySQL NDB Cluster setup inside a Kubernetes Cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Ndb Cluster
            displayName: NdbCluster
            kind: NdbCluster
            name: ndbclusters.mysql.oracle.com
            version: v1
        description: |
          ## NdbCluster Custom Resource
          The NdbCluster custom resource is defined with custom fields that map to the configuration of a MySQL Cluster. To deploy a MySQL Cluster inside the K8s Cluster, an NdbCluster object with the desired configuration of the MySQL Cluster has to be created in the K8s Cluster. The NDB Operator watches for any changes made to any NdbCluster objects and will receive an event when the NdbCluster object is created in the K8s Cluster. It will then start deploying the MySQL Cluster, with the specified configuration, inside the K8s Cluster using native Kubernetes workloads and resources. Once the MySQL Cluster is ready, the NdbCluster resource can be further modified to make changes to the MySQL Cluster configuration. The Operator will receive this event as well and will update the configuration of the already running MySQL Cluster.

          The documentation at [NdbCluster-CRD.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md) explains the various custom fields of the NdbCluster CRD and the effect they have on the MySQL Cluster configuration. The [examples](https://github.com/mysql/mysql-ndb-operator/tree/main/docs/examples) folder has a few examples of the NdbCluster resource objects which can be used to deploy MySQL Cluster inside the K8s Cluster.
          ## Deploy a simple MySQL Cluster

          The [examples/example-ndb.yaml](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/examples/example-ndb.yaml) file has a NdbCluster object example-ndb that defines a simple MySQL Cluster with 2 data nodes and 2 MySQL Servers. To create this object in the default namespace of the K8s Cluster, run this command from the root of the repo :
          ```
          kubectl apply -f docs/examples/example-ndb.yaml
          ```

          Various NdbCluster status fields report back the actual status of the MySQL Cluster running inside the K8s Cluster. Wait for the MySQL Cluster nodes to be started and become ready using the [UpToDate](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbclusterconditiontypestring-alias) Condition defined in the NdbCluster object status :
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```
          The MySQL Cluster is ready for transactions once the NdbCluster is UpToDate, i.e. the UpToDate condition becomes true.

          At any point, a brief status of an NdbCluster object can be viewed using kubectl get ndb command.
          ```
          kubectl get ndb example-ndb
          ```

          The output of the command gives a short summary the MySQL Cluster nodes controlled by the respective NdbCluster object.
          ```
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 3m    True
          ```

          To list all the pods created by the NDB Operator, run :
          ```
          kubectl get pods -l mysql.oracle.com/v1=example-ndb
          ```

          For example-ndb NdbCluster object, two management nodes (*-mgmd-* pods), two multi-threaded data nodes (*-ndbmtd-* pods) and two MySQL Servers (*-mysqld-*) would have been created by the NDB operator :
          ```
          NAME                   READY   STATUS    RESTARTS   AGE
          example-ndb-mgmd-0     1/1     Running   0          3m53s
          example-ndb-mgmd-1     1/1     Running   0          2m5s
          example-ndb-mysqld-0   1/1     Running   0          78s
          example-ndb-mysqld-1   1/1     Running   0          78s
          example-ndb-ndbmtd-0   1/1     Running   0          3m53s
          example-ndb-ndbmtd-1   1/1     Running   0          3m53s
          ```
          If running inside a openshift cluster, pods from statefulset "example-ndb-mysqld" will not start up. By default openshift runs all the pod with "restricted" scc, but the statefulset "example-ndb-mysqld" required higher privilege. Use the below command to check the exact scc required
          ```
          oc get pods example-ndb-mysqld-0 -o yaml | oc adm policy scc-subject-review -f -
          ```
          Create a user account with the scc and set the user to the statefulset "example-ndb-mysqld". Below commands creates an user with "anyuid" and assign it to the statefulset.
          ```
          oc create sa any-uid-user
          oc adm policy add-scc-to-user anyuid -z any-uid-user
          oc set sa sts example-ndb-mysqld any-uid-user
          ```
          ## Connect to the MySQL Cluster

          The NDB Operator, by default, creates few Services to expose the services offered by the MySQL Cluster nodes within the Kubernetes Cluster.
          ```
          kubectl get services -l mysql.oracle.com/v1=example-ndb
          ```

          The above command will generate an output similar to :
          ```
          NAME                 TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S) AGE
          example-ndb-mgmd     ClusterIP   10.100.207.255   <none>        1186/TCP 5m
          example-ndb-mysqld   ClusterIP   10.103.182.157   <none>        3306/TCP 5m
          example-ndb-ndbmtd   ClusterIP   None             <none>        1186/TCP 5m
          ```

          The NDB Operator will also create a root user with a random password on all the MySQL Servers. The root password will be stored in a K8s Secret which can be retrieved from the NdbCluster object status :
          ```
          kubectl get ndb example-ndb -o jsonpath={.status.generatedRootPasswordSecretName}
          ```

          To extract and decode the password from the Secret :
          ```
          base64 -d <<< \
          $(kubectl get secret example-ndb-mysqld-root-password \
          -o jsonpath={.data.password})
          ```

          One can also specify the password and host to be used with the root account via the NdbCluster spec. See the [CRD documentation](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/NdbCluster-CRD.md#ndbmysqldspec) for more details on this.
          ## Access MySQL Cluster from inside K8s

          To connect to the Management and MySQL Servers from within the Kubernetes Cluster, you can use the example-ndb-mgmd and example-ndb-mysqld services as the MySQL Cluster connectstring and MySQL host respectively. You can also use the mgmd and mysqld pods' hostnames to directly connect to that particular pod. The mysql, ndb_mgm clients and any other ndb tools will work as expected.

          A demonstration is available at [connect-from-inside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-inside-k8s-demo.md).
          ## Access MySQL Cluster from outside K8s

          By default, the Management and MySQL services created by the NDB Operator are of type [ClusterIP](https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) and are accessible only from within the K8s Cluster. To expose them outside the K8s Cluster, the services have to be upgraded into a LoadBalancer type. This can be done by setting the spec.managementNode.enableLoadBalancer and spec.mysqlNode.enableLoadBalancer fields of the NdbCluster resource to true. These options can be enabled when the NdbCluster resource object is created or via an update to the object when the MySQL Cluster is already running.

          Once enabled, the previously ClusterIP type example-ndb-mgmd and example-ndb-mysqld services will be upgraded to LoadBalancer type, and they will be available at the external IP assigned to them by the cloud provider.

          Another way to access these services without enabling the LoadBalancers support is to use the kubectl port-forward command.

          In both ways, only the mysql and ndb_mgm clients work. Any NDB tool which uses the NDBAPI to connect to the MySQL data nodes will not work as expected from outside the K8s Cluster.

          A demonstration is available at [connect-from-outside-k8s-demo.md](https://github.com/mysql/mysql-ndb-operator/blob/main/docs/connect-from-outside-k8s-demo.md).
          ## Updating MySQL Cluster Configuration

          Once a MySQL Cluster has been deployed by the NDB Operator inside the K8s Cluster, its configuration can be updated by updating the spec of the NdbCluster resource object that represents it. The NDB Operator will pick up any changes made to the NdbCluster resource object and will apply the updated configuration to the respective MySQL Cluster. The NDB Operator takes care of updating the MySQL Cluster config file and restarting all the management/data nodes if required. When the update is being handled by the NdbOperator, it will set the UpToDate condition of the NdbCluster resource object to false and will not accept any further updates until it completes the current one.

          The kubectl get ndb example-ndb command will report the UpToDate status of an NdbCluster. It can be used to check if the current update has been completed.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE   UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:1/2    Ready:2/2 10m   False
          ```

          The UpToDate condition can also be used to wait for the operator to complete the update.
          ```
          kubectl wait --for=condition=UpToDate ndb example-ndb --timeout=10m
          ```

          Once the update is complete, the UpToDate condition will be set back to true by the operator.
          ```
          $ kubectl get ndb example-ndb
          NAME          REPLICA   MANAGEMENT NODES   DATA NODES   MYSQL SERVERS AGE      UP-TO-DATE
          example-ndb   2         Ready:2/2          Ready:2/2    Ready:2/2 10m50s   True
          ```

          ## Delete a MySQL Cluster

          To stop and remove the MySQL Cluster running inside the K8s Cluster, delete the NdbCluster resource object.
          ```
          kubectl delete -f docs/examples/example-ndb.yaml
          ```

          (or)
          ```
          kubectl delete ndb example-ndb
          ```

          This will delete all the data, the pods running the MySQL Cluster nodes and also delete all other associated K8s resources created by the NDB Operator.
        displayName: Ndb Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mysql
        - ndb
        - in-memory
        links:
        - name: Document
          url: https://dev.mysql.com/doc/ndb-operator/en/
        maintainers:
        - email: paulo.p.silva@oracle.com
          name: paulofasilva
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: MySQL
          url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
        relatedImages:
        - container-registry.oracle.com/mysql/community-ndb-operator:9.1.0-1.6.0
        - quay.io/community-operator-pipeline-prod/ndb-operator@sha256:5f33c2a8181ea8e91844afd1f4042affbd9eb878266b6b2bcc6906f0829ed00b
        version: 1.6.0
      entries:
      - name: ndb-operator.v1.6.0
        version: 1.6.0
      name: alpha
    defaultChannel: alpha
    packageName: ndb-operator
    provider:
      name: MySQL
      url: https://artifacthub.io/packages/helm/mysql-ndb-operator/ndb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KEDA Community
      provider-url: ""
    name: keda
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keda.v2.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "KedaController",
                "metadata": {
                  "name": "keda",
                  "namespace": "keda"
                },
                "spec": {
                  "watchNamespace": "",
                  "operator": {
                    "logLevel": "info",
                    "logEncoder": "console"
                  },
                  "metricsServer": {
                    "logLevel": "0"
                  },
                  "serviceAccount": {}
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledJob",
                "metadata": {
                  "name": "scaledjob-sample"
                },
                "spec": {
                  "failedJobsHistoryLimit": 5,
                  "jobTargetRef": {
                    "activeDeadlineSeconds": 600,
                    "backoffLimit": 6,
                    "completions": 1,
                    "parallelism": 1,
                    "template": null
                  },
                  "maxReplicaCount": 100,
                  "pollingInterval": 30,
                  "successfulJobsHistoryLimit": 5,
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ScaledObject",
                "metadata": {
                  "name": "example-scaledobject"
                },
                "spec": {
                  "cooldownPeriod": 300,
                  "maxReplicaCount": 100,
                  "minReplicaCount": 0,
                  "pollingInterval": 30,
                  "scaleTargetRef": {
                    "name": "example-deployment"
                  },
                  "triggers": [
                    {
                      "metadata": {
                        "property": "examle-property"
                      },
                      "type": "example-trigger"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "TriggerAuthentication",
                "metadata": {
                  "name": "example-triggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              },
              {
                "apiVersion": "keda.sh/v1alpha1",
                "kind": "ClusterTriggerAuthentication",
                "metadata": {
                  "name": "example-clustertriggerauthentication"
                },
                "spec": {
                  "secretTargetRef": [
                    {
                      "key": "example-role-key",
                      "name": "example-secret-name",
                      "parameter": "exmaple-secret-parameter"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Cloud Provider
          certified: "false"
          containerImage: ghcr.io/kedacore/keda-olm-operator:2.7.1
          createdAt: "2022-05-11T00:00:00.000Z"
          description: '[DEPRECATED] Use Custom Metrics Autoscaler Operator instead'
          operatorframework.io/suggested-namespace: keda
          operators.operatorframework.io/builder: operator-sdk-v1.1.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/kedacore/keda-olm-operator
          support: KEDA Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              Represents an installation of a particular version of KEDA Controller.
            displayName: KedaController
            kind: KedaController
            name: kedacontrollers.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your jobs and what the triggers
            displayName: ScaledJob
            kind: ScaledJob
            name: scaledjobs.keda.sh
            version: v1alpha1
          - description: Defines how KEDA should scale your application and what the
              triggers are
            displayName: ScaledObject
            kind: ScaledObject
            name: scaledobjects.keda.sh
            version: v1alpha1
          - description: Describes authentication parameters
            displayName: TriggerAuthentication
            kind: TriggerAuthentication
            name: triggerauthentications.keda.sh
            version: v1alpha1
          - description: Describes cluster wide authentication parameters
            displayName: ClusterTriggerAuthentication
            kind: ClusterTriggerAuthentication
            name: clustertriggerauthentications.keda.sh
            version: v1alpha1
        description: "## Deprecated\n This KEDA community operator is deprecated,
          please use `Custom Metrics Autoscaler` Operator instead.\n## About the managed
          application\nKEDA is a Kubernetes-based event driven autoscaler.  KEDA can
          monitor event sources like Kafka, RabbitMQ, or cloud event sources and feed
          the metrics from those sources into the Kubernetes horizontal pod autoscaler.
          \ With KEDA, you can have event driven and serverless scale of deployments
          within any Kubernetes cluster.\n## About this Operator\nThe KEDA Operator
          deploys and manages installation of KEDA Controller in the cluster. Install
          this operator and follow installation instructions on how to install KEDA
          in you cluster.\n\n## Prerequisites for enabling this Operator\nPlease note
          that you can not run both KEDA 1.x and 2.x on the same Kubernetes cluster.
          You need to uninstall KEDA 1.x first, in order to install and use KEDA 2.x.\n\nDon't
          forget to uninstall KEDA v1 CRDs as well, to ensure that, please run:\n\n```
          oc delete crd scaledobjects.keda.k8s.io ```\n\n``` oc delete crd triggerauthentications.keda.k8s.io
          \ ```\n## How to install KEDA in the cluster\nThe installation of KEDA is
          triggered by the creation of `KedaController` resource, please refer to
          [KedaController Example](https://github.com/kedacore/keda-olm-operator#kedacontroller-spec)
          for more deatils on available options.\n\nOnly resource named `keda` in
          namespace `keda` will trigger the installation, reconfiguration or removal
          of the KEDA Controller resource.\n\nThere could be only one KEDA Controller
          in the cluster. \n"
        displayName: '[DEPRECATED] KEDA'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keda
        - scaling
        - event
        links:
        - name: Homepage
          url: https://keda.sh
        - name: Source Code
          url: https://github.com/kedacore/keda
        - name: Slack Channel
          url: https://kubernetes.slack.com/messages/CKZJ36A5D
        maintainers:
        - email: cncf-keda-maintainers@lists.cncf.io
          name: KEDA Community
        maturity: deprecated
        minKubeVersion: 1.17.0
        provider:
          name: KEDA Community
        relatedImages:
        - quay.io/openshift-community-operators/keda@sha256:50816f675c76d36dbc62846579150477018ef87df325861751f3fc02a70c8781
        - ghcr.io/kedacore/keda-olm-operator:2.7.1
        version: 2.7.1
      entries:
      - name: keda.v2.7.1
        version: 2.7.1
      - name: keda.v2.6.1
        version: 2.6.1
      - name: keda.v2.6.0
        version: 2.6.0
      - name: keda.v2.5.0
        version: 2.5.0
      - name: keda.v2.4.0
        version: 2.4.0
      - name: keda.v2.3.0
        version: 2.3.0
      - name: keda.v2.2.0
        version: 2.2.0
      - name: keda.v2.1.0
        version: 2.1.0
      - name: keda.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keda
    provider:
      name: KEDA Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        - quay.io/community-operator-pipeline-prod/dynatrace-operator@sha256:c475f9cc0f6caa2bb316ff42a35a40b1d4bf8ef7c8a3384d9c78eb205bda7076
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.3
        version: 1.6.3
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      - name: dynatrace-operator.v0.6.0
        version: 0.6.0
      - name: dynatrace-operator.v0.5.1
        version: 0.5.1
      - name: dynatrace-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kaoto.io
      provider-url: https://kaoto.io
    name: kaoto-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kaoto-operator.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "designer.kaoto.io/v1alpha1",
                "kind": "Kaoto",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "kaoto-operator",
                    "app.kubernetes.io/instance": "kaoto-demo",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kaoto",
                    "app.kubernetes.io/part-of": "kaoto-operator"
                  },
                  "name": "kaoto-demo"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
          createdAt: "2025-09-03T12:48:58Z"
          description: Integration Editor for Apache Camel
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/KaotoIO/kaoto-operator
          support: contact@kaoto.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kaoto is the Schema for the kaotoes API.
            displayName: Kaoto
            kind: Kaoto
            name: kaotoes.designer.kaoto.io
            version: v1alpha1
        description: Kaoto is a low code and no code integration designer to edit
          and deploy based on Apache Camel . It is designed to be extendable and flexible
          and adaptable to different usecases.
        displayName: Kaoto Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - kaoto
        - camel-k
        - designer
        links:
        - name: Kaoto Operator
          url: https://kaoto-operator.domain
        maintainers:
        - email: info@kaoto.io
          name: kaoto
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: kaoto.io
          url: https://kaoto.io
        relatedImages:
        - quay.io/kaotoio/kaoto-operator@sha256:590d4958f148954ea1edb9825fb7fb4935d041ddcdb4a99302da14a605380071
        - quay.io/community-operator-pipeline-prod/kaoto-operator@sha256:4b6cc43eba8efbe17ffec746431d6cf6b045886ff865673a08a0820abecebdcd
        version: 2.7.0
      entries:
      - name: kaoto-operator.v2.7.0
        version: 2.7.0
      - name: kaoto-operator.v2.6.0
        version: 2.6.0
      - name: kaoto-operator.v2.5.0
        version: 2.5.0
      - name: kaoto-operator.v2.4.0
        version: 2.4.0
      - name: kaoto-operator.v2.0.1
        version: 2.0.1
      - name: kaoto-operator.v2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: kaoto-operator
    provider:
      name: kaoto.io
      url: https://kaoto.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <=2.4.0
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d4883d7c622683b3319b5e6b3a7edfbf2594c18060131a8bf64504805f875522
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:a3033b50ae434b97de4de1ca03013dabcfac580d810a6ee8c0fe05b6a9301952
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:5ea2dd102074798d1ee005dc03fbbd24f4324df831f279355dc2dfc262c9d880
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator-rhmp.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/aerospike-kubernetes-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - quay.io/aerospike/aerospike-kubernetes-operator@sha256:e6d6bd7c58e036b3728e52475a9b764ca8c46b83cbc4751cac66294495bf2b31
        - registry.connect.redhat.com/aerospike/aerospike-kubernetes-operator-bundle-rhmp@sha256:e4dec1f39f8e24043cf16703efcac2cc3b953d3f4c37b3956fce0c65ef829764
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator-rhmp.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator-rhmp.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator-rhmp.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator-rhmp.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator-rhmp.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator-rhmp.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator-rhmp.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator-rhmp.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator-rhmp.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator-rhmp.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator-rhmp.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator-rhmp
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-data-reporter-operator
      app.kubernetes.io/managed-by: ibm-data-reporter-operator
      app.kubernetes.io/name: ibm-data-reporter-operator
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-data-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-data-reporter-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "DataReporterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-data-reporter-operator",
                    "app.kubernetes.io/instance": "datareporterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datareporterconfig",
                    "app.kubernetes.io/part-of": "ibm-data-reporter-operator"
                  },
                  "name": "datareporterconfig"
                },
                "spec": {
                  "userConfig": [
                    {
                      "metadata": {
                        "ameta1": "ametadata1",
                        "bmeta1": "bmetadata1",
                        "cmeta1": "cmetadata1",
                        "dmeta1": "dmetadata1"
                      },
                      "userName": "system:serviceaccount:redhat-marketplace:ibm-data-reporter-operator-api"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
          createdAt: ""
          description: The IBM Data Reporter Operator accepts events and transforms
            them into reports submitted to the Data Service of the IBM Metrics Operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReporterConfig is the Schema for the datareporterconfigs
              API
            displayName: Data Reporter Config
            kind: DataReporterConfig
            name: datareporterconfigs.marketplace.redhat.com
            version: v1alpha1
          required:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
        description: "# IBM Data Reporter Operator\n\n# Introduction\n\nThe IBM Data
          Reporter Operator accepts events and transforms them into reports submitted
          to the IBM Metrics Operator Data Service.\n\n# Details\n\nThe IBM Data Reporter
          Operator deploys a service that provides an endpoint to which callers can
          send raw json event data. The event data is transformed into a report and
          is sent to the IBM Metrics Operator's Data Service. The IBM Metrics Operator
          Data Service periodically uploads the reports to IBM Software Central. It
          can also filter, transform and forward events to alternate endpoints.\n\n##
          Prerequisites\n\n- OpenShift Container Platform, major version 4 with any
          available supported minor version\n- Install IBM Metrics Operator\n  - IBM
          Data Reporter Operator prerequisties the IBM Metrics Operator's Data Service\n
          \ - Register the Cluster by creating a `redhat-marketplace-pull-secret`,
          as per the instructions\n  - Once `rhm-data-service` has started IBM Data
          Reporter Operator will accept requests\n\n## SecurityContextConstraints
          Requirements\n\n- The operator runs under Red Hat restricted SCC\n\n## Resources
          Required\n\n- The operator requires at least 85 Mi memory.\n\n## Limitations\n\n-
          Only runs on amd64, s390x and ppc64le architectures\n\n## Installing\n\n-
          A user with the cluster administrator role.\n- Install this operator in
          the same namespace as the IBM Metrics Operator\n  - default namespace: `ibm-software-central`\n\n##
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n# Configuration\n\nOptional:\n- Configure the
          Custom Resource Definition `DataReporterConfig` named `datareporterconfig`
          as per the following sample\n  - Configuration that can not be reconciled
          successfully will be reported on the DataReporterConfig Status\n  - jsonPath
          expressions are handled by: https://github.com/ohler55/ojg\n    - jsonPath
          comparison: https://cburgmer.github.io/json-path-comparison/\n  - Transformer
          type currently supported is `kazaam`\n    - https://github.com/willie68/kazaam\n\nSample
          DataReporterConfig:\n```YAML\napiVersion: marketplace.redhat.com/v1alpha1\nkind:
          DataReporterConfig\nmetadata:\n  name: datareporterconfig\nspec:\n  # true:
          report generated for 1 event and provide return code from data-service,
          confirming delivery\n  # false: events will be accumulated before building
          report, and immediately return 200, delivery unfconfirmed\n  confirmDelivery:
          false\n  dataFilters:   # dataFilters matches the first dataFilter selector,
          applies transforms and forwards to data-service and destinations\n  - altDestinations:
          \  # List of desintations to transform and foward events to\n    - authorization:
          \  # optional endpoint if an authorization token must first be requested
          \n        authDestHeader: Authorization   # http header key to be appended
          to destination request\n        authDestHeaderPrefix: 'Bearer '   # optional
          prefix for constructing destination request http header value\n        bodyData:
          \  # body data to send to authorization request, often an apiKey\n          secretKeyRef:\n
          \           name: auth-body-data-secret\n            key: bodydata\n        header:
          \  # secret map to use for authorization request headers\n          secret:\n
          \           name: auth-header-map-secret\n        tokenExpr: $.token   #
          optionally extract the returned body data using a jsonPath expression\n
          \       url: https://127.0.0.1:8443/api/2.0/accounts/my_account_id/apikeys/token
          \  # the url to make an authorization request to\n      header:   # secret
          map to use for destination request headers\n        secret:\n          name:
          dest-header-map-secret\n      transformer:   # The transformer to apply
          to event data for this destination\n        configMapKeyRef:\n          key:
          kazaam.json\n          name: kazaam-configmap\n        type: kazaam\n      url:
          https://127.0.0.1:8443/metering/api/1.0/usage   # the url to send the event
          to\n      urlSuffixExpr: $.properties.productId   # optional jsonPath expression
          on the event to extract and set a suffix to the url path\n    manifestType:
          dataReporter   # override manifestType of report sent to data-service\n
          \   selector:   # Matches against event data and a user. Empty selector
          matches all.\n      matchExpressions:   # jsonPath expression must return
          a result for all expressions to match (AND)\n      - $[?($.event == \"Account
          Contractual Usage\")]\n      - $.properties.productId\n      - $[?($.properties.source
          != null)]\n      - $[?($.properties.unit == \"AppPoints\")]\n      - $[?($.properties.quantity
          >= 0)]\n      - $[?($.timestamp != null)]\n      matchUsers:   # must match
          one of these users (OR). Omitting users matches any user\n      - system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n
          \   transformer:   # The transformer to apply to the event for data-service\n
          \     configMapKeyRef:\n        key: kazaam.json\n        name: kazaam-configmap\n
          \     type: kazaam\n  tlsConfig:   # TLS configuration for requests outbound
          to destinations\n    caCerts:\n    - name: tls-ca-certificates\n      key:
          ca.crt\n    certificates:\n    - clientCert:\n        secretKeyRef:\n          name:
          tls-ca-certificates\n          key: tls.crt\n      clientKey:\n        secretKeyRef:\n
          \         name: tls-ca-certificates\n          key: tls.key\n    cipherSuites:
          \n    - TLS_AES_128_GCM_SHA256\n    - TLS_AES_256_GCM_SHA384\n    - TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\n    - TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\n
          \   - TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\n    insecureSkipVerify: false\n
          \   minVersion: VersionTLS12\n  userConfig:   # Optional metadata to apply
          to report sent to data-service for a specific user\n  - metadata:\n      ameta1:
          ametadata1\n      bmeta1: bmetadata1\n      cmeta1: cmetadata1\n      dmeta1:
          dmetadata1\n    userName: system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api\n```\n\nSample
          Header Secrets:\n```YAML\napiVersion: v1\nstringData:\n  accept: | \n    */*\nkind:
          Secret\nmetadata:\n  name: dest-header-map-secret\ntype: Opaque\n---\napiVersion:
          v1\nstringData:\n  accept: application/json\n  Content-Type: application/json\nkind:
          Secret\nmetadata:\n  name: auth-header-map-secret\ntype: Opaque\n```\n\nSample
          Authentication Body Secret:\n```YAML\napiVersion: v1\nstringData:\n  bodydata:
          |\n    {\"apikey\": \"myapikey\"}\nkind: Secret\nmetadata:\n  name: auth-body-data-secret\ntype:
          Opaque\n```\n\nSample TLS Secret:\n```YAML\napiVersion: v1\ndata:\n  ca.crt:
          LS0t...\n  tls.key: LS0t...\n  tls.crt: LS0t...\nkind: Secret\nmetadata:\n
          \ name: tls-ca-certificates\n```\n\nSample Kazaam ConfigMap:\n```YAML\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: kazaam-configmap\ndata:\n  kazaam.json:
          |\n    [\n        {\n            \"operation\": \"timestamp\",\n            \"spec\":
          {\n                \"timestamp\": {\n                \"inputFormat\": \"2006-01-02T15:04:05.999999+00:00\",\n
          \               \"outputFormat\": \"$unixmilli\"\n                }\n            }\n
          \       },\n        {\n            \"operation\": \"shift\",\n            \"spec\":
          {\n                \"instances[0].instanceId\": \"properties.source\",\n
          \               \"instances[0].startTime\": \"timestamp\",\n                \"instances[0].endTime\":
          \"timestamp\",\n                \"instances[0].metricUsage[0].metricId\":
          \"properties.unit\",\n                \"instances[0].metricUsage[0].quantity\":
          \"properties.quantity\"\n            }\n        },\n        {\n            \"operation\":
          \"default\",\n            \"spec\": {\n                \"meteringModel\":
          \"point-in-time\",\n                \"meteringPlan\": \"contract\"\n            }\n
          \       }\n    ]\n```\n\n### API Service User Configuration\n\n- The ClusterRole
          for api access is `clusterrole/ibm-data-reporter-operator-api`\n- The default
          ServiceAccount provided as an api user is `system:serviceaccount:ibm-software-central:ibm-data-reporter-operator-api`\n
          \ - The default ClusterRoleBinding for this user is `clusterrolebinding/ibm-data-reporter-operator-api`\n\nOptional:\n\n-
          To create an additional ServiceAccount and ClusterRoleBinding for api access\n\n```SHELL\nNAMESPACE=$(oc
          config view --minify -o jsonpath='{..namespace}') && \\\noc create serviceaccount
          my-api-service-account && \\\noc create clusterrolebinding ibm-data-reporter-operator-api
          --clusterrole=ibm-data-reporter-operator-api --serviceaccount=${NAMESPACE}:my-api-service-account\n```\n\n-
          Update datareporterconfig to attach metadata to reports associated with
          this user\n\n## API Service Usage\n\n- Get Token & Host\n\n```SHELL\noc
          project ibm-software-central\nexport DRTOKEN=$(oc create token ibm-data-reporter-operator-api
          --namespace ibm-software-central --duration 1h)\nexport DRHOST=$(oc get
          route ibm-data-reporter --template='{{ .spec.host }}')\n```\n\n- Get the
          Status\n\n```SHELL\ncurl -k -H \"Authorization: Bearer ${DRTOKEN}\" https://${DRHOST}/v1/status
          \n```\n\n- Post an Event\n\n```SHELL\ncurl -k -H \"Authorization: Bearer
          ${DRTOKEN}\" -X POST -d '{\"event\":\"myevent\"}' https://${DRHOST}/v1/event\n```\n\n##
          Storage\n\n- The operator temporarily stores events in pod memory, and writes
          to the IBM Metrics Operator data-service, which requires a PersistentVolume\n\n##
          License\n\nCopyright IBM Corporation 2023. All Rights Reserved.\n\nLicensed
          under the Apache License, Version 2.0 (the \"License\");\nyou may not use
          this file except in compliance with the License.\nYou may obtain a copy
          of the License at\n\n    http://www.apache.org/licenses/LICENSE-2.0\n\nUnless
          required by applicable law or agreed to in writing, software\ndistributed
          under the License is distributed on an \"AS IS\" BASIS,\nWITHOUT WARRANTIES
          OR CONDITIONS OF ANY KIND, either express or implied.\nSee the License for
          the specific language governing permissions and\nlimitations under the License.\n"
        displayName: IBM Data Reporter Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-data-reporter-operator@sha256:b6a286f79dfbad9af018138774843e4f7f0012d659090676bd535992f63a508a
        - registry.connect.redhat.com/rh-marketplace/ibm-data-reporter-operator-bundle@sha256:7fe9e82ce279206bd53b27fd81391a39bc0caea8f492f7aea03282bb69694781
        version: 2.20.0
      entries:
      - name: ibm-data-reporter-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-data-reporter-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
          createdAt: "2025-04-16T06:54:48Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.4.2.5,990005"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:ed8230e90f0ab0586a743d5e760a7469345b1136971aad17ef675c1beccf1982
        - registry.connect.redhat.com/ca/uma-operator@sha256:8ff4aab6d0043ccd2542a86b7e9c24d60670a3c376021ef2d02fc1a289130275
        version: 2025.4.5-5
      entries:
      - name: uma-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
          createdAt: "2025-07-24T15:07:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.5.2.12,990012""
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:08995782514e496c24d888352e9d1e6b788dd87adb761e0c2e745be2975ffcaf
        - registry.connect.redhat.com/ca/uma-operator@sha256:90b4b21d93bce41f2530699cec01f03e5daa73bfbf4f30cbcc8fbd6e75e97da6
        version: 2025.5.4-12
      entries:
      - name: uma-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-operator.v2025.7.4-23
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
          createdAt: "2025-08-25T09:00:19Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.7.1.23,990023"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:7b324e897fbcc881ececa529822d7b738f82cd4c3adf4cc54b66080b888f5d9c
        - registry.connect.redhat.com/ca/uma-operator@sha256:35f74109aa1916524c1fb483972caa31cb84381c3cff7c3b26fd9b676f9d748c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.4-23
      entries:
      - name: uma-operator.v2025.7.4-23
        version: 2025.7.4-23
      name: "2025.7"
    - currentCSV: uma-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
          createdAt: "2025-08-15T03:19:46Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "25.8.1.9,990009"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:e2a329fc74ed5f17e395367126c7e96197178e8f886d475b657a7910a4a14152
        - registry.connect.redhat.com/ca/uma-operator@sha256:adea186632253865406e95a0062abbb312dee23c92ff108f5914e325dc0b8833
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        version: 2025.8.2-9
      entries:
      - name: uma-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
          createdAt: "2025-07-30T08:06:11Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                   microservicesAgent:
                     enabled: false
                     enableKernelVersionCheck: true
                     # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                     # If set to false, application metrics are reported along with pod infra metrics.
                     createAgentPerApplication: true
                     filterType: whitelist
                     tracer:
                       enableHttpTracer: true
                       enableHttpsTracer: false
                       enableMySQLTracer: true
                       enableMongoTracer: true
                       enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   memory: "300Mi"
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3250Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                     nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
                mixedMode: "false"
                buildTag: "99.99.0.serp_F152813,19"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:c892787785c8c5e0de1692709a9e3f54eef36f1f00c53f1b9feabb613769a48c
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:0218d68f97e89b48bbdd9c7cd750d4a666cf33af1cb8a76507c067a332a7b6bd
        version: 2025.7.2-19
      entries:
      - name: uma-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
          createdAt: "2025-09-11T11:15:33Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.8,990008"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:2bc56e149c14920662b549a5367f98daf9afad0df29b4746b195f6afc0c9c597
        - registry.connect.redhat.com/ca/uma-operator@sha256:dc15dfdc8b82ee159b52036cc9d057e6525b08c478ed759939cf5c3fc7daeeec
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgent",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "false",
                    "serverCert": "",
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "agentSidecarEnabled": true,
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "nginx": {
                          "enabled": true
                        },
                        "nodejs": {
                          "enabled": false,
                          "filterNpts": "node",
                          "whitelistedProcesses": "add:[\"^node\\\\s+.*pm2-runtime\\\\s+.*\\\\.(yaml|yml|json|js).*$\"]\n"
                        },
                        "php": {
                          "enabled": true
                        },
                        "probe": {
                          "enabled": true,
                          "manageWebhookCaBundle": false,
                          "tlsEnabled": true,
                          "webhookEnabled": false
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "microservicesAgent": {
                        "createAgentPerApplication": true,
                        "enableKernelVersionCheck": true,
                        "enabled": false,
                        "filterType": "whitelist",
                        "tracer": {
                          "enableAppFlow": true,
                          "enableHttpTracer": true,
                          "enableHttpsTracer": false,
                          "enableMongoTracer": true,
                          "enableMySQLTracer": true
                        }
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "gcLoggingEnabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "800Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true,
                          "version2": {
                            "addLabelKeyInMetric": true,
                            "agentNaming": "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}",
                            "allowCustomVertex": false,
                            "attributesAlias": [],
                            "autoDiscovery": "global",
                            "clusterinfoPollInterval": "15s",
                            "dataQueueSize": 300,
                            "enabled": false,
                            "excludeLabels": [],
                            "excludeMetricByLabels": [],
                            "excludeMetrics": [],
                            "includeLabels": [
                              ".*"
                            ],
                            "includeMetricByLabels": [
                              ".*"
                            ],
                            "includeMetrics": [
                              ".*"
                            ],
                            "labelNameClamp": "200",
                            "labelPriorityOrder": [],
                            "labelValueClamp": "200",
                            "metricLengthClamp": "1000",
                            "metricsAlias": [],
                            "metricsPrefix": [
                              {
                                ".*": "{type}|{k8s_pod_name}"
                              }
                            ],
                            "mode": "k8s",
                            "namespaceExcludeList": [],
                            "parallelProcessingRoutines": 50,
                            "scrapeInterval": "15s",
                            "scrapeTimeout": "10s",
                            "scrapperVertexQueueSize": 300,
                            "supportabilityMetrics": true
                          }
                        },
                        "limits": {
                          "cpu": "1",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "httpCollector": {
                      "enabled": true,
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ssl": {
                          "enabled": false
                        }
                      },
                      "memoryGrid": {
                        "enabled": ""
                      },
                      "replicas": 1,
                      "server": {
                        "host": "0.0.0.0",
                        "limits": {
                          "cpu": "2",
                          "memory": "4096Mi"
                        },
                        "port": 8085,
                        "requests": {
                          "cpu": "500m",
                          "memory": "500Mi"
                        }
                      }
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "iibia": {
                      "aceConfig": [
                        {
                          "mqChannel": "",
                          "mqHost": "127.0.0.1",
                          "mqPort": "1414",
                          "mqQmdgrName": "",
                          "profileName": "default",
                          "restPort": "7600"
                        }
                      ],
                      "debug": false,
                      "enabled": false,
                      "iibpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "iibusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "mqpassword": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "mqusername": {
                        "value": "",
                        "valueFrom": {
                          "secretKeyRef": {
                            "key": "",
                            "name": ""
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "brokerenabled": false,
                        "keystoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "keystorePassword": {
                          "value": ""
                        },
                        "mqcipherspec": "",
                        "qmenabled": false,
                        "truststoreKey": {
                          "value": "",
                          "valueFrom": {
                            "secretKeyRef": {
                              "key": "",
                              "name": ""
                            }
                          }
                        },
                        "truststorePassword": {
                          "value": ""
                        }
                      },
                      "statistics": {
                        "enabled": true,
                        "nodes": "jvm"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "prefix": "k8s",
                  "role": "common",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
          createdAt: "2025-09-24T12:22:41Z"
          description: The Universal Monitoring Agent for Openshift (UMA for Openshift)
            acts as single deployment that automatically discovers and monitors cloud
            and container infrastructures and containerized application processes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent
            kind: UniversalMonitoringAgent
            name: universalmonitoringagents.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent** for OpenShift acts as a single deployment that automatically discovers and monitors cloud and container infrastructures and containerized application processes. UMA for OpenShift discovers and monitors all the containers on a host. This capability allows to monitor both the Infrastructure ( like Node, Pods, Containers, Cluster Services etc ) and applications running on this infrastructure & correlate them together which can help the triager to isolate if the the problem is related to the application or infrastructure.

          For more information:
          https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgent
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true

             role: common

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   probe:
                     enabled: true
                     webhookEnabled: false
                     tlsEnabled: true
                     manageWebhookCaBundle: false
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     # The agentMode to enable to CSA Java Agent in UMA, it is lightweight agent than default java agent
                     # csa - Containerized Service Architecture
                     # default - Existing default java agent
                     #agentMode: csa
                     # The agentSidecarEnabled to enable to use sidecar approach for java-agent,
                     # It auto creates new deployment revision with sidecar, and roll-out updated revision.
                     agentSidecarEnabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                   nginx:
                     enabled: true
                   # This feature for enabling auto attach node.js probe into node.js application deployment
                   # nodejs.enabled - to switch on/off node.js application discovery and attach
                   # filterNpts - Filters supported node process tools node and pm2 delimited by comma. Example "node,pm2"
                   # whitelistedProcesses - to attach known Node.js processes using regex, use prefix with 'add:' to append to default list, otherwise it replaces with current ones.
                   nodejs:
                     enabled: false
                     filterNpts: "node"
                     whitelistedProcesses: |
                       add:["^node\\s+.*pm2-runtime\\s+.*\\.(yaml|yml|json|js).*$"]
                  # This feature for enabling auto attach php probe into PHP application deployment
                  # php.enabled - to switch on/off PHP application discovery and attach
                   php:
                     enabled: true
                 microservicesAgent:
                   enabled: false
                   enableKernelVersionCheck: true
                   # When createAgentPerApplication is set to true UMA creates a new agent per pod as 'clustername|namespace_deployment|podname'.
                   # If set to false, application metrics are reported along with pod infra metrics.
                   createAgentPerApplication: true
                   filterType: whitelist
                   tracer:
                     enableHttpTracer: true
                     enableHttpsTracer: false
                     enableMySQLTracer: true
                     enableMongoTracer: true
                     enableAppFlow: true
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               httpCollector:
                 enabled: true
                 replicas: 1
                 server:
                   host: 0.0.0.0
                   port: 8085
                   # The below properties define the resources for 'apmia-http-collector' deployment.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "4096Mi"
                   requests:
                     cpu: "500m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "500Mi"
                 memoryGrid:
                   enabled:
                 # Enable this ingress resource creation if there is Ingress Controller deployed
                 # It requires session affinity using HTTP Headers by LB Ingress Controller
                 ingress:
                   enabled: false
                   host:
                   ssl:
                     enabled: false
               container:
                 prometheus:
                   exporter:
                     enabled: true
                     version2:
                       enabled: false
                       autoDiscovery: "global"
                       mode: "k8s"
                       clusterinfoPollInterval: "15s"
                       parallelProcessingRoutines: 50
                       dataQueueSize: 300
                       scrapperVertexQueueSize: 300
                       allowCustomVertex: false
                       supportabilityMetrics: true
                       agentNaming: "{k8s_cluster_name}|{k8s_namespace_name}|{k8s_owner_name}"
                         # example:
                         # metricsPrefix:
                         #   - "go_.*": "{type}|{k8s_pod_name}|go_stats"
                         #   - ".*": "{type}|{k8s_pod_name}"
                       metricsPrefix:
                           - ".*": "{type}|{k8s_pod_name}"
                       addLabelKeyInMetric: true
                       namespaceExcludeList: [ ]
                       includeLabels: [ ".*" ]
                       includeMetricByLabels: [ ".*" ]
                       includeMetrics: [ ".*" ]
                         # example:
                         # metricsAlias:
                         #   - "abc*":'test'
                         #   - "go*":"internal"
                       metricsAlias: []
                       attributesAlias: []
                         # example
                         # labelPriorityOrder:
                         #   - "redis_exporter_build_info":"version,golang_version,commit_sha"
                       labelPriorityOrder: []
                       excludeLabels: [ ]
                       excludeMetrics: [ ]
                       excludeMetricByLabels: [ ]
                       labelValueClamp: "200"
                       labelNameClamp: "200"
                       metricLengthClamp: "1000"
                       scrapeTimeout: "10s"
                       scrapeInterval: "15s"
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "1"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               iibia:
                 enabled: false
                 statistics:
                   enabled: true
                   nodes: "jvm"
                 debug: false
                 ssl:
                   brokerenabled: false
                   qmenabled: false
                   mqcipherspec: ""
                   truststoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   truststorePassword:
                     value: ""
                   keystoreKey:
                     value: ""
                     valueFrom:
                       secretKeyRef:
                         name: ""
                         key: ""
                   keystorePassword:
                     value: ""
                 # The below properties define the properties for different ibm ace port, subscriber properties
                 aceConfig:
                   - profileName: "default"
                     restPort: "7600"
                     mqHost: "127.0.0.1"
                     mqPort: "1414"
                     mqQmdgrName: ""
                     mqChannel: ""
                 iibusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 iibpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqusername:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 mqpassword:
                   value: ""
                   valueFrom:
                     secretKeyRef:
                       name: ""
                       key: ""
                 # The below properties define the resources for 'apm-iib-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate.
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "false"
               buildTag: "25.9.1.12,990012"
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}
                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false

          ```
        displayName: DX APM Universal Monitoring Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Install-UMA-for-OpenShift.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:5c62c26f5d6eea41219b5eeda619b8cfeb19e417da2808fd14f53fae402b87cd
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator-bundle@sha256:fd45df2c1531ef2871c1dcdcc52e2cb5c7771d07b57d0604b3370d228cd41262
        version: 2025.9.1-12
      entries:
      - name: uma-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eks-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eks-controller.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "FargateProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eks.services.k8s.aws/v1alpha1",
                "kind": "Nodegroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
          createdAt: "2025-09-19T22:01:50Z"
          description: AWS EKS controller is a service controller for managing EKS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessEntry represents the state of an AWS eks AccessEntry
              resource.
            displayName: AccessEntry
            kind: AccessEntry
            name: accessentries.eks.services.k8s.aws
            version: v1alpha1
          - description: Addon represents the state of an AWS eks Addon resource.
            displayName: Addon
            kind: Addon
            name: addons.eks.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS eks Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.eks.services.k8s.aws
            version: v1alpha1
          - description: FargateProfile represents the state of an AWS eks FargateProfile
              resource.
            displayName: FargateProfile
            kind: FargateProfile
            name: fargateprofiles.eks.services.k8s.aws
            version: v1alpha1
          - description: IdentityProviderConfig represents the state of an AWS eks
              IdentityProviderConfig resource.
            displayName: IdentityProviderConfig
            kind: IdentityProviderConfig
            name: identityproviderconfigs.eks.services.k8s.aws
            version: v1alpha1
          - description: Nodegroup represents the state of an AWS eks Nodegroup resource.
            displayName: Nodegroup
            kind: Nodegroup
            name: nodegroups.eks.services.k8s.aws
            version: v1alpha1
          - description: PodIdentityAssociation represents the state of an AWS eks
              PodIdentityAssociation resource.
            displayName: PodIdentityAssociation
            kind: PodIdentityAssociation
            name: podidentityassociations.eks.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Kubernetes Service (EKS) resources in AWS from within your Kubernetes cluster.

          **About Amazon EKS**

          Amazon Elastic Kubernetes Service (Amazon EKS) is a managed service that you can use to run Kubernetes on AWS without needing to install, operate, and maintain your own Kubernetes control plane or nodes. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eks
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EKS Developer Resources
          url: https://aws.amazon.com/eks/getting-started/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eks-controller:1.9.2
        - quay.io/community-operator-pipeline-prod/ack-eks-controller@sha256:120566a02e487aaa8823323423fccfaa377135037e2e1cb9f3ff14313cf3bfeb
        version: 1.9.2
      entries:
      - name: ack-eks-controller.v1.9.2
        version: 1.9.2
      - name: ack-eks-controller.v1.9.1
        version: 1.9.1
      - name: ack-eks-controller.v1.9.0
        version: 1.9.0
      - name: ack-eks-controller.v1.8.0
        version: 1.8.0
      - name: ack-eks-controller.v1.7.6
        version: 1.7.6
      - name: ack-eks-controller.v1.7.5
        version: 1.7.5
      - name: ack-eks-controller.v1.7.4
        version: 1.7.4
      - name: ack-eks-controller.v1.7.3
        version: 1.7.3
      - name: ack-eks-controller.v1.7.2
        version: 1.7.2
      - name: ack-eks-controller.v1.6.4
        version: 1.6.4
      - name: ack-eks-controller.v1.6.3
        version: 1.6.3
      - name: ack-eks-controller.v1.6.2
        version: 1.6.2
      - name: ack-eks-controller.v1.6.1
        version: 1.6.1
      - name: ack-eks-controller.v1.6.0
        version: 1.6.0
      - name: ack-eks-controller.v1.5.5
        version: 1.5.5
      - name: ack-eks-controller.v1.5.4
        version: 1.5.4
      - name: ack-eks-controller.v1.5.3
        version: 1.5.3
      - name: ack-eks-controller.v1.5.2
        version: 1.5.2
      - name: ack-eks-controller.v1.5.1
        version: 1.5.1
      - name: ack-eks-controller.v1.5.0
        version: 1.5.0
      - name: ack-eks-controller.v1.4.7
        version: 1.4.7
      - name: ack-eks-controller.v1.4.6
        version: 1.4.6
      - name: ack-eks-controller.v1.4.5
        version: 1.4.5
      - name: ack-eks-controller.v1.4.4
        version: 1.4.4
      - name: ack-eks-controller.v1.4.3
        version: 1.4.3
      - name: ack-eks-controller.v1.4.2
        version: 1.4.2
      - name: ack-eks-controller.v1.4.1
        version: 1.4.1
      - name: ack-eks-controller.v1.4.0
        version: 1.4.0
      - name: ack-eks-controller.v1.3.5
        version: 1.3.5
      - name: ack-eks-controller.v1.3.3
        version: 1.3.3
      - name: ack-eks-controller.v1.3.1
        version: 1.3.1
      - name: ack-eks-controller.v1.3.0
        version: 1.3.0
      - name: ack-eks-controller.v1.2.3
        version: 1.2.3
      - name: ack-eks-controller.v1.2.2
        version: 1.2.2
      - name: ack-eks-controller.v1.2.1
        version: 1.2.1
      - name: ack-eks-controller.v1.2.0
        version: 1.2.0
      - name: ack-eks-controller.v1.1.3
        version: 1.1.3
      - name: ack-eks-controller.v1.1.2
        version: 1.1.2
      - name: ack-eks-controller.v1.1.1
        version: 1.1.1
      - name: ack-eks-controller.v1.1.0
        version: 1.1.0
      - name: ack-eks-controller.v1.0.5
        version: 1.0.5
      - name: ack-eks-controller.v1.0.4
        version: 1.0.4
      - name: ack-eks-controller.v1.0.3
        version: 1.0.3
      - name: ack-eks-controller.v1.0.2
        version: 1.0.2
      - name: ack-eks-controller.v1.0.1
        version: 1.0.1
      - name: ack-eks-controller.v1.0.0
        version: 1.0.0
      - name: ack-eks-controller.v0.1.9
        version: 0.1.9
      - name: ack-eks-controller.v0.1.8
        version: 0.1.8
      - name: ack-eks-controller.v0.1.7
        version: 0.1.7
      - name: ack-eks-controller.v0.1.6
        version: 0.1.6
      - name: ack-eks-controller.v0.1.5
        version: 0.1.5
      - name: ack-eks-controller.v0.1.4
        version: 0.1.4
      - name: ack-eks-controller.v0.1.3
        version: 0.1.3
      - name: ack-eks-controller.v0.1.2
        version: 0.1.2
      - name: ack-eks-controller.v0.1.1
        version: 0.1.1
      - name: ack-eks-controller.v0.1.0
        version: 0.1.0
      - name: ack-eks-controller.v0.0.15
        version: 0.0.15
      - name: ack-eks-controller.v0.0.14
        version: 0.0.14
      - name: ack-eks-controller.v0.0.13
        version: 0.0.13
      - name: ack-eks-controller.v0.0.12
        version: 0.0.12
      - name: ack-eks-controller.v0.0.11
        version: 0.0.11
      - name: ack-eks-controller.v0.0.10
        version: 0.0.10
      - name: ack-eks-controller.v0.0.9
        version: 0.0.9
      - name: ack-eks-controller.v0.0.8
        version: 0.0.8
      - name: ack-eks-controller.v0.0.7
        version: 0.0.7
      - name: ack-eks-controller.v0.0.6
        version: 0.0.6
      - name: ack-eks-controller.v0.0.5
        version: 0.0.5
      - name: ack-eks-controller.v0.0.4
        version: 0.0.4
      - name: ack-eks-controller.v0.0.3
        version: 0.0.3
      - name: ack-eks-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-eks-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EPAM Delivery Platform
      provider-url: https://epam.github.io/edp-install/
    name: sonar-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonar-operator.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "Sonar",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonar-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonar",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonar-sample"
                },
                "spec": {
                  "defaultPermissionTemplate": "edp-default",
                  "secret": "sonar-secret",
                  "settings": [
                    {
                      "key": "sonar.global.exclusions",
                      "values": [
                        "**/*.jsx",
                        "**/*.tsx"
                      ]
                    },
                    {
                      "fieldValues": {
                        "beginBlockRegexp": ".*",
                        "endBlockRegexp": ".*"
                      },
                      "key": "sonar.issue.ignore.block"
                    },
                    {
                      "key": "sonar.dbcleaner.hoursBeforeKeepingOnlyOneSnapshotByDay",
                      "value": "20"
                    }
                  ],
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonargroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonargroup",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonargroup-sample"
                },
                "spec": {
                  "description": "Developers",
                  "name": "sonar-developers",
                  "permissions": [
                    "admin"
                  ],
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarPermissionTemplate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarpermissiontemplate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarpermissiontemplate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarpermissiontemplate-sample"
                },
                "spec": {
                  "default": true,
                  "description": "EDP permission templates",
                  "groupsPermissions": {
                    "sonar-developers": [
                      "scan"
                    ],
                    "sonar-users": [
                      "issueadmin"
                    ]
                  },
                  "name": "edp-default",
                  "projectKeyPattern": ".+",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityGate",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualitygate-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualitygate",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualitygate-sample"
                },
                "spec": {
                  "conditions": {
                    "new_duplicated_lines_density": {
                      "error": "1",
                      "op": "GT"
                    }
                  },
                  "default": true,
                  "name": "sonarqualitygate-sample",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarQualityProfile",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonarqualityprofile-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonarqualityprofile",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonarqualityprofile-sample"
                },
                "spec": {
                  "default": true,
                  "language": "go",
                  "name": "sonarqualityprofile-sample",
                  "rules": {
                    "go:S1151": {
                      "params": "max=\"6\"",
                      "severity": "MAJOR"
                    }
                  },
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              },
              {
                "apiVersion": "edp.epam.com/v1alpha1",
                "kind": "SonarUser",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "sonar-operator",
                    "app.kubernetes.io/instance": "sonaruser-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sonaruser",
                    "app.kubernetes.io/part-of": "sonar-operator"
                  },
                  "name": "sonaruser-sample"
                },
                "spec": {
                  "email": "sonaruser@mail.com",
                  "groups": [
                    "sonar-administrators"
                  ],
                  "login": "sonaruser",
                  "name": "sonar user",
                  "permissions": [
                    "scan"
                  ],
                  "secret": "sonaruser-secret",
                  "sonarRef": {
                    "name": "sonar"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Developer Tools
          containerImage: docker.io/epamedp/sonar-operator:3.1.0
          createdAt: "2023-09-08T12:53:57Z"
          description: An Operator for managing Sonar entities.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-sonar-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonarGroup is the Schema for the sonar group API.
            displayName: Sonar Group
            kind: SonarGroup
            name: sonargroups.edp.epam.com
            version: v1alpha1
          - description: SonarPermissionTemplate is the Schema for the sonar permission
              template API.
            displayName: Sonar Permission Template
            kind: SonarPermissionTemplate
            name: sonarpermissiontemplates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityGate is the Schema for the sonarqualitygates
              API
            displayName: Sonar Quality Gate
            kind: SonarQualityGate
            name: sonarqualitygates.edp.epam.com
            version: v1alpha1
          - description: SonarQualityProfile is the Schema for the sonarqualityprofiles
              API
            displayName: Sonar Quality Profile
            kind: SonarQualityProfile
            name: sonarqualityprofiles.edp.epam.com
            version: v1alpha1
          - description: Sonar is the Schema for the sonars API.
            displayName: Sonar
            kind: Sonar
            name: sonars.edp.epam.com
            version: v1alpha1
          - description: SonarUser is the Schema for the sonarusers API
            displayName: Sonar User
            kind: SonarUser
            name: sonarusers.edp.epam.com
            version: v1alpha1
        description: |
          Sonar Operator is an operator that is responsible for establishing
          a connection to provided SonarQube server, reconciling Sonar entities (users,
          groups, permission templates, etc) according to the created CRs.
        displayName: EDP Sonar Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - analysis
        - automation
        - code
        - continuous
        - devops
        - edp
        - inspection
        - metrics
        - metrics
        - monitoring
        - operator
        - quality
        - security
        - software
        - static
        - technical
        - vulnerabilities
        links:
        - name: Sonar Operator
          url: https://github.com/epam/edp-sonar-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/sonar-operator:3.1.0
        - quay.io/openshift-community-operators/sonar-operator@sha256:5e7a87f52955162f090c1d7fc69f6d0917444b553355a11c061ab60f289a17d7
        version: 3.1.0
      entries:
      - name: sonar-operator.v3.1.0
        version: 3.1.0
      name: alpha
    defaultChannel: alpha
    packageName: sonar-operator
    provider:
      name: EPAM Delivery Platform
      url: https://epam.github.io/edp-install/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-documentdb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-documentdb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "documentdb.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
          createdAt: "2025-09-19T22:13:37Z"
          description: AWS DocumentDB controller is a service controller for managing
            DocumentDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBCluster represents the state of an AWS documentdb DBCluster
              resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS documentdb DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.documentdb.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS documentdb DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.documentdb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon DocumentDB resources in AWS from within your Kubernetes cluster.

          **About Amazon DocumentDB**

          Amazon DocumentDB (with MongoDB compatibility) is a fast, reliable, and fully managed database service. Amazon DocumentDB makes it easy to set up, operate, and scale MongoDB-compatible databases in the cloud. With Amazon DocumentDB, you can run the same application code and use the same drivers and tools that you use with MongoDB.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon DocumentDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - documentdb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon DocumentDB Developer Resources
          url: https://aws.amazon.com/documentdb/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: documentdb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/documentdb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-documentdb-controller@sha256:8a991c76f281ec30258b6317bd623894b78a4f39d9790a1cbb32e6f1b7ff63f5
        version: 1.1.1
      entries:
      - name: ack-documentdb-controller.v1.1.1
        version: 1.1.1
      - name: ack-documentdb-controller.v1.1.0
        version: 1.1.0
      - name: ack-documentdb-controller.v1.0.14
        version: 1.0.14
      - name: ack-documentdb-controller.v1.0.13
        version: 1.0.13
      - name: ack-documentdb-controller.v1.0.12
        version: 1.0.12
      - name: ack-documentdb-controller.v1.0.11
        version: 1.0.11
      - name: ack-documentdb-controller.v1.0.10
        version: 1.0.10
      - name: ack-documentdb-controller.v1.0.9
        version: 1.0.9
      - name: ack-documentdb-controller.v1.0.7
        version: 1.0.7
      - name: ack-documentdb-controller.v1.0.6
        version: 1.0.6
      - name: ack-documentdb-controller.v1.0.5
        version: 1.0.5
      - name: ack-documentdb-controller.v1.0.2
        version: 1.0.2
      - name: ack-documentdb-controller.v1.0.1
        version: 1.0.1
      - name: ack-documentdb-controller.v1.0.0
        version: 1.0.0
      - name: ack-documentdb-controller.v0.0.9
        version: 0.0.9
      - name: ack-documentdb-controller.v0.0.8
        version: 0.0.8
      - name: ack-documentdb-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-documentdb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: local-storage-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: local-storage-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "local.storage.openshift.io/v1",
                "kind": "LocalVolume",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassDevices": [
                    {
                      "devicePaths": [
                        "/dev/disk/by-id/ata-crucial"
                      ],
                      "fsType": "ext4",
                      "storageClassName": "foobar",
                      "volumeMode": "Filesystem"
                    }
                  ]
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeSet",
                "metadata": {
                  "name": "example-localvolumeset"
                },
                "spec": {
                  "deviceInclusionSpec": {
                    "deviceMechanicalProperties": [
                      "Rotational",
                      "NonRotational"
                    ],
                    "deviceTypes": [
                      "disk"
                    ],
                    "maxSize": "100G",
                    "minSize": "10G"
                  },
                  "maxDeviceCount": 10,
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  },
                  "storageClassName": "example-storageclass",
                  "volumeMode": "Block"
                }
              },
              {
                "apiVersion": "local.storage.openshift.io/v1alpha1",
                "kind": "LocalVolumeDiscovery",
                "metadata": {
                  "name": "auto-discover-devices"
                },
                "spec": {
                  "nodeSelector": {
                    "nodeSelectorTerms": [
                      {
                        "matchExpressions": [
                          {
                            "key": "kubernetes.io/hostname",
                            "operator": "In",
                            "values": [
                              "worker-0",
                              "worker-1"
                            ]
                          }
                        ]
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
          createdAt: "2019-08-14T00:00:00Z"
          description: Configure and use local storage volumes.
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-local-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["localvolumediscoveryresults.local.storage.openshift.io"]'
          repository: https://github.com/openshift/local-storage-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manage local storage volumes for OpenShift
            displayName: Local Volume
            kind: LocalVolume
            name: localvolumes.local.storage.openshift.io
            version: v1
          - description: A Local Volume set allows you to filter a set of storage
              volumes, group them and create a dedicated storage class to consume
              storage from the set of volumes.
            displayName: Local Volume Set
            kind: LocalVolumeSet
            name: localvolumesets.local.storage.openshift.io
            version: v1alpha1
          - description: Discover list of potentially usable disks on the chosen set
              of nodes
            displayName: Local Volume Discovery
            kind: LocalVolumeDiscovery
            name: localvolumediscoveries.local.storage.openshift.io
            version: v1alpha1
          - description: Disc inventory of available disks from selected nodes
            displayName: Local Volume Discovery Result
            kind: LocalVolumeDiscoveryResult
            name: localvolumediscoveryresults.local.storage.openshift.io
            version: v1alpha1
        description: |
          Operator that configures local storage volumes for use in Kubernetes and OpenShift. OpenShift 4.2 and above are the only supported OpenShift versions.
        displayName: Local Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - local storage
        links:
        - name: Documentation
          url: https://github.com/openshift/local-storage-operator/tree/master/docs
        - name: Source Repository
          url: https://github.com/openshift/local-storage-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-local-storage-diskmaker-rhel9@sha256:57af9e32c31dad769bda449ebe35e09ea135e8a7d06cddb42be4d1a09005f518
        - registry.redhat.io/openshift4/ose-local-storage-mustgather-rhel9@sha256:ab3cb65b88c34fd58524a2a51cd510c611924e5860834965bad0b4418a8b6dd3
        - registry.redhat.io/openshift4/ose-local-storage-rhel9-operator@sha256:19140e94c1cb7e756d48a3e9e650ddec47d1081eb4abd64475c01abc1ac85ea1
        - registry.redhat.io/openshift4/ose-local-storage-operator-bundle@sha256:d303fc63f5b849f073e34a7791ea92cdc3cff269cd1b593e27480ae96cac1a15
        version: 4.18.0-202509100344
      entries:
      - name: local-storage-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: local-storage-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: local-storage-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: local-storage-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: local-storage-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: local-storage-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: local-storage-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: local-storage-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: local-storage-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: local-storage-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: local-storage-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: local-storage-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: local-storage-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: local-storage-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: local-storage-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: local-storage-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: local-storage-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtc-operator.v1.7.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
          createdAt: "2024-09-04T17:53:42Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.7.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-velero-rhel8@sha256:07cafb83ba90378c1fa60338cf951a9d7145ec245710637e90fdb508875dcc45
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:d8232366bb3abee6454fcbff6842ec4deaa60cde762f3868eae6218c888cbfe7
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:b95a14eda5ae107ba3747995f1ea74a6422f9e5c20d685505e954ca7630b810d
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:1d86075469312e84f3d16379efcd6130ec9e8b23a1343e3945f071d82beff6e3
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:dc10544e1bb24883f38373a0445944745282e5a69ed1744f428f98871a7a39c9
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:31b3658dd79c18111c21e99aea52cc198f8e95b458a1a4f8d2bfc4d21e9bdacb
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:1b8008a16a0f53c09ecc5a219964194689a1aeb01acee8313aa49a9c26259429
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:a539064234088a4ee13b8a9666ca9507208fe38189ee9b8b673a38a58aa75599
        - registry.redhat.io/rhmtc/openshift-migration-velero-restic-restore-helper-rhel8@sha256:b137e1345038c9e566767f24cce60c03ec278f5a2efc28fa08e0bd8bfa016486
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-aws-rhel8@sha256:c001dc96a1478a58f30461ab018e40bede13a78f3cdce869a70dcb8b9bba22b1
        - registry.redhat.io/rhmtc/openshift-velero-plugin-rhel8@sha256:92a3162c705537147005a065cdddbb410e1863d80c08d8b0f27b7ff5588500cd
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:8fbfffb0bc3adfb4d72961dad381e1daba414bf4a7084f03b5208f6be4191ceb
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-gcp-rhel8@sha256:035855c3d4b32a29232846d73dc398e81f941fd90a03e33241ca06f6e80b2204
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:f86e96edb0418811df328debdee93b4a51cf7f651134ee35e9c50b9dbfc29fd8
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-microsoft-azure-rhel8@sha256:084f1e54ef0d7f457d5a7cfc57b567ad8d85237c12f707c82f7a30a54e3652ba
        version: 1.7.17
      entries:
      - name: mtc-operator.v1.7.17
        version: 1.7.17
      - name: mtc-operator.v1.7.16
        version: 1.7.16
      - name: mtc-operator.v1.7.15
        version: 1.7.15
      - name: mtc-operator.v1.7.13
        version: 1.7.13
      - name: mtc-operator.v1.7.12
        version: 1.7.12
      - name: mtc-operator.v1.7.11
        version: 1.7.11
      - name: mtc-operator.v1.7.10
        version: 1.7.10
      - name: mtc-operator.v1.7.9
        version: 1.7.9
      - name: mtc-operator.v1.7.8
        version: 1.7.8
      - name: mtc-operator.v1.7.7
        version: 1.7.7
      - name: mtc-operator.v1.7.6
        version: 1.7.6
      - name: mtc-operator.v1.7.5
        version: 1.7.5
      - name: mtc-operator.v1.7.4
        version: 1.7.4
      - name: mtc-operator.v1.7.3
        version: 1.7.3
      - name: mtc-operator.v1.7.2
        version: 1.7.2
      - name: mtc-operator.v1.7.1
        version: 1.7.1
      - name: mtc-operator.v1.7.0
        version: 1.7.0
      name: release-v1.7
    - currentCSV: mtc-operator.v1.8.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigrationController",
                  "metadata": {
                    "name": "migration-controller",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "azure_resource_group": "",
                    "cluster_name": "host",
                    "migration_velero": true,
                    "migration_controller": true,
                    "migration_ui": true,
                    "migration_log_reader": true,
                    "olm_managed": true,
                    "restic_timeout": "1h",
                    "mig_pv_limit": "100",
                    "mig_pod_limit": "100",
                    "mig_namespace_limit": "10"
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageMigration",
                  "metadata": {
                    "name": "directimagemigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectImageStreamMigration",
                  "metadata": {
                    "name": "directimagestreammigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigAnalytic",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {
                    "plan": {
                      "name": "foo",
                      "namespace": "openshift-migration"
                    },
                    "analyzeK8SResources": true,
                    "analyzeImageCount": true,
                    "analyzePVCapacity": true
                  }
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigCluster",
                  "metadata": {
                    "name": "host",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "DirectVolumeMigration",
                  "metadata": {
                    "name": "direct",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                      "kind": "DirectVolumeMigrationProgress",
                      "metadata": {
                        "name": "direct-progress",
                        "namespace": "openshift-migration"
                      },
                      "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigPlan",
                  "metadata": {
                    "name": "migplan",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigMigration",
                  "metadata": {
                    "name": "migmigration",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "migration.openshift.io/v1alpha1",
                  "kind": "MigStorage",
                  "metadata": {
                    "name": "migstorage",
                    "namespace": "openshift-migration"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
          createdAt: "2025-08-22T15:27:29Z"
          description: Facilitates migration of container workloads from OpenShift
            3.x to OpenShift 4.x
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.5.0 <1.8.10'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-migration
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/mig-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: a request to migrate PVCs from a set of namespaces directly
              to a target
            displayName: DirectVolumeMigration
            kind: DirectVolumeMigration
            name: directvolumemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate watch the progress of rsync pod migrating
              a PVC to a target
            displayName: DirectVolumeMigrationProgress
            kind: DirectVolumeMigrationProgress
            name: directvolumemigrationprogresses.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a set of namespaces
              directly to a target
            displayName: DirectImageMigration
            kind: DirectImageMigration
            name: directimagemigrations.migration.openshift.io
            version: v1alpha1
          - description: a request to migrate local images from a single imagestream
              directly to a target
            displayName: DirectImageStreamMigration
            kind: DirectImageStreamMigration
            name: directimagestreammigrations.migration.openshift.io
            version: v1alpha1
          - description: A request to Analyze a MigPlan and its namespaces
            displayName: MigAnalytic
            kind: MigAnalytic
            name: miganalytics.migration.openshift.io
            version: v1alpha1
          - description: A cluster defined for migration
            displayName: MigCluster
            kind: MigCluster
            name: migclusters.migration.openshift.io
            version: v1alpha1
          - description: A migration process
            displayName: MigMigration
            kind: MigMigration
            name: migmigrations.migration.openshift.io
            version: v1alpha1
          - description: A migration plan
            displayName: MigPlan
            kind: MigPlan
            name: migplans.migration.openshift.io
            version: v1alpha1
          - description: A migration contoller deployment
            displayName: MigrationController
            kind: MigrationController
            name: migrationcontrollers.migration.openshift.io
            version: v1alpha1
          - description: A migration hook
            displayName: MigHook
            kind: MigHook
            name: mighooks.migration.openshift.io
            version: v1alpha1
          - description: A migration storage location
            displayName: MigStorage
            kind: MigStorage
            name: migstorages.migration.openshift.io
            version: v1alpha1
        description: |-
          The Migration Toolkit for Containers Operator enables installation of the application migration tool components.

          After you have installed the Operator, you can install components by creating a MigrationController CR.

          By default, the Operator installs the following components on an OpenShift Container Platform 4 target cluster:

          * velero and restic, to back up data on the source cluster and restore it on the target cluster
          * mig-controller, to coordinate the backup and restore processes
          * mig-ui, the web console for managing migrations

          If you experience any issues or have feature requests, please file an [issue](https://github.com/konveyor/mig-operator/issues)

          For more information, see the [OpenShift Container Platform Migration Guide](https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html).
        displayName: Migration Toolkit for Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - crane
        - mtc
        links:
        - name: Migrations
          url: https://github.com/konveyor/mig-controller
        - name: Documentation
          url: https://docs.openshift.com/container-platform/latest/migration_toolkit_for_containers/about-mtc.html
        - name: Crane Operator
          url: https://github.com/konveyor/mig-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhmtc/openshift-migration-must-gather-rhel8@sha256:f9492269b75608057fbc5edd53603807f1f80cd8581dea12032c88e04cf8f4cb
        - registry.redhat.io/rhmtc/openshift-migration-rsync-transfer-rhel8@sha256:442e7c7f2cc075bf468403ace8846c96ad2fe43eb7f6562d566caa24e4f86873
        - registry.redhat.io/rhmtc/openshift-migration-ui-rhel8@sha256:86e747e0330b0156379d273d0c29c13a65dafbf4f536656723e2bc746872286c
        - registry.redhat.io/rhmtc/openshift-migration-hook-runner-rhel8@sha256:adaa4f1afefb4d5853713821cb6e7de0d9f5a640f0243f20b41a44be84707caf
        - registry.redhat.io/rhmtc/openshift-migration-log-reader-rhel8@sha256:4f76bcf3df3bb7073b8aabd3da072cacfec05ecf4bfb47e23340074756f7fc30
        - registry.redhat.io/rhmtc/openshift-migration-openvpn-rhel8@sha256:d21b99d8edefa7a1bab30c85969ec7d7560d06be308bd787e698e409d54cbc19
        - registry.redhat.io/rhmtc/openshift-migration-operator-bundle@sha256:9f5479ac795eb50fa5ea597828b2d50ddaa4878faa1a07b3de0e611cb4af5fc6
        - registry.redhat.io/rhmtc/openshift-migration-velero-plugin-for-mtc-rhel8@sha256:5f46ba04146438f2abd6905468d2ecbeb19374a83771d364904ac2ee6afdbefa
        - registry.redhat.io/rhmtc/openshift-migration-registry-rhel8@sha256:accef2a0dd69b254bd120232e5dc86a56aa8f25bcfdba3f9bab7e4127c157762
        - registry.redhat.io/rhmtc/openshift-migration-rhel8-operator@sha256:8a44fe2aea3e842dcccb1e20df4818f4ca573fe13f51427ae8b603d0000a9411
        - registry.redhat.io/rhmtc/openshift-migration-controller-rhel8@sha256:c776387c662e21aa61378d0136b2495826e1c157548f9ca9c95d044af38ee169
        version: 1.8.10
      entries:
      - name: mtc-operator.v1.8.10
        version: 1.8.10
      - name: mtc-operator.v1.8.9
        version: 1.8.9
      - name: mtc-operator.v1.8.8
        version: 1.8.8
      - name: mtc-operator.v1.8.7
        version: 1.8.7
      - name: mtc-operator.v1.8.6
        version: 1.8.6
      - name: mtc-operator.v1.8.5
        version: 1.8.5
      - name: mtc-operator.v1.8.4
        version: 1.8.4
      - name: mtc-operator.v1.8.3
        version: 1.8.3
      - name: mtc-operator.v1.8.0
        version: 1.8.0
      name: release-v1.8
    defaultChannel: release-v1.8
    packageName: mtc-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odigos
      provider-url: https://odigos.io
    name: odigos-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odigos-operator.v1.0.212
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.odigos.io/v1alpha1",
                "kind": "Odigos",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "odigos-operator"
                  },
                  "name": "odigos-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing
          containerImage: registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
          createdAt: "2025-07-18T13:23:10Z"
          description: Odigos enables automatic distributed tracing with OpenTelemetry
            and eBPF.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Odigos Enterprise subscription
            (for enterprise features)
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Odigos
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odigos is the Schema for the odigos API
            displayName: Odigos
            kind: Odigos
            name: odigos.operator.odigos.io
            version: v1alpha1
        description: |-
          The Odigos Operator provides options for installing and configuring Odigos.

          Upon creation of an Odigos resource, the Odigos Operator installs the following Odigos components:

          - Instrumentor
          - Odiglet
          - Autoscaler
          - Scheduler
          - Frontend UI

          With Odigos installed, follow the Odigos docs at docs.odigos.io for instructions on instrumenting applications and configuring destinations.
        displayName: Odigos Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - OpenTelemetry
        - eBPF
        - tracing
        - observability
        - distributed tracing
        - otel
        - monitoring
        links:
        - name: Odigos
          url: https://odigos.io
        - name: Odigos Documentation
          url: https://docs.odigos.io
        - name: Odigos on Github
          url: https://github.com/odigos-io/odigos
        maintainers:
        - email: mike@odigos.io
          name: Mike Dame
        maturity: alpha
        minKubeVersion: 1.20.15
        provider:
          name: Odigos
          url: https://odigos.io
        relatedImages:
        - registry.connect.redhat.com/odigos/odigos-instrumentor-ubi9@sha256:16712373c8e927b7208e43e9fe6e21180289975dbf3a81df60578e8550cf7150
        - registry.connect.redhat.com/odigos/odigos-scheduler-ubi9@sha256:9e613c9b22d2fbfa3d21cc24f3b1e2dbc2e7287790e1f4286f1817c92565c865
        - registry.connect.redhat.com/odigos/odigos-ui-ubi9@sha256:b2a27cdfc94f1968a06ae18de1977127bb7d4e7d553e875d09a212693302f9c6
        - registry.connect.redhat.com/odigos/operator-bundle@sha256:bc8a47d5f8ffe3a2fe48819f08c82525677131cfb4f3c0facd82470c0f10acee
        - registry.connect.redhat.com/odigos/odigos-autoscaler-ubi9@sha256:692193851b40f72ca79f7c82903136686d721c7fed77ef599037269ac6e55fdb
        - registry.connect.redhat.com/odigos/odigos-enterprise-instrumentor-ubi9@sha256:647e526cdbe38023bb221a802b180f0b95cb25734a0b39fa11cbbe22a07482a3
        - registry.connect.redhat.com/odigos/odigos-enterprise-odiglet-ubi9@sha256:aec9b66461375994dc9aa92e5b2cc7d4151cf0332150726840cec049a7cbe7d9
        - registry.connect.redhat.com/odigos/odigos-collector-ubi9@sha256:b67138d3ffd9e721602c5bf83f0c03ae1dcee230eca5e15335dcb9abced281ba
        - registry.connect.redhat.com/odigos/odigos-certified-operator-ubi9@sha256:c221eb2be22c2fcabcee728429bbca1255e0e31642c572e4dc40096d1d35e0d1
        - registry.connect.redhat.com/odigos/odigos-odiglet-ubi9@sha256:770193901b4e5abcec59833f56f1530b43d4ec9514f3b5a5cb0bbb4356e06c67
        version: 1.0.212
      entries:
      - name: odigos-operator.v1.0.212
        version: 1.0.212
      name: alpha
    defaultChannel: alpha
    packageName: odigos-operator
    provider:
      name: Odigos
      url: https://odigos.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nimbusmosaic-dusim
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nimbusmosaic-dusim.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Nimbusmosaicdusim",
                "metadata": {
                  "name": "nimbusmosaicdusim-sample"
                },
                "spec": {
                  "development": false,
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "microservices": {
                    "jvm": "-Xmx300M -Xms64M -Xss256k"
                  },
                  "service": {
                    "type": "ClusterIP"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:56Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Nimbusmosaicdusim
            name: nimbusmosaicdusims.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Nimbus Mosaic DUSim Operator
        displayName: Keysight Nimbus Mosaic DUSim Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nimbusmosaic Dusim
          url: https://keysight-nimbusmosaic-dusim.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-nimbusmosaic-dusim@sha256:655dd6595a9c8e82947bed24991ac52f8fa55be99ca410e7d9f0026c51ef4520
        - registry.connect.redhat.com/loadcore/keysight-nimbusmosaic-dusim-bundle@sha256:97d5f906ef83272a8d7a2a249f38f8b9b186f8e9b64729a049a7eac0dbda8b1a
        version: 0.1.21
      entries:
      - name: keysight-nimbusmosaic-dusim.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nimbusmosaic-dusim
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://redhat.com
    name: pelorus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pelorus-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.pelorus.dora-metrics.io/v1alpha1",
                "kind": "Pelorus",
                "metadata": {
                  "name": "pelorus-sample"
                },
                "spec": {
                  "exporters": {
                    "global": {},
                    "instances": [
                      {
                        "app_name": "deploytime-exporter",
                        "exporter_type": "deploytime"
                      },
                      {
                        "app_name": "failuretime-exporter",
                        "enabled": false,
                        "exporter_type": "failure"
                      },
                      {
                        "app_name": "committime-exporter",
                        "exporter_type": "committime"
                      },
                      {
                        "app_name": "webhook-exporter",
                        "exporter_type": "webhook"
                      }
                    ]
                  },
                  "federate_openshift_monitoring": {
                    "enabled": false
                  },
                  "openshift_prometheus_basic_auth_pass": "changeme",
                  "openshift_prometheus_htpasswd_auth": "internal:{SHA}+pvrmeQCmtWmYVOZ57uuITVghrM=",
                  "prometheus_retention": "1y",
                  "prometheus_retention_size": "1GB",
                  "prometheus_storage": false,
                  "prometheus_storage_pvc_capacity": "2Gi",
                  "prometheus_storage_pvc_storageclass": "gp2"
                }
              }
            ]
          capabilities: Basic Install
          categories: |
            Modernization & Migration,Developer Tools,Monitoring,Integration & Delivery
          containerImage: quay.io/pelorus/pelorus-operator:0.0.9
          createdAt: "2024-05-27T20:04:01Z"
          description: |
            Tool that helps IT organizations measure their impact on the overall performance of their organization
          operatorframework.io/suggested-namespace: pelorus
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dora-metrics/pelorus/
          support: Pelorus Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Pelorus is the Schema for pelorus API to adopt instance to
              the requested workflow
            displayName: Pelorus
            kind: Pelorus
            name: peloruses.charts.pelorus.dora-metrics.io
            version: v1alpha1
        description: |
          Pelorus is a tool that helps IT organizations measure their impact on the overall performance of their organization. It does this by gathering metrics about team and organizational behaviors over time in some key areas of IT that have been shown to impact the value they deliver to the organization as a whole. Some of the key outcomes Pelorus can focus on are:

          - Software Delivery Performance
          - Product Quality and Sustainability
          - Customer experience

          For more background on the project you can read [@trevorquinn](https://github.com/trevorquinn)'s blog post on [Metrics Driven Transformation](https://www.openshift.com/blog/exploring-a-metrics-driven-approach-to-transformation).

          ## Software Delivery Performance as an outcome

          Currently, Pelorus functionality can capture proven metrics that measure Software Delivery Performance -- a significant outcome that IT organizations aim to deliver.

          Pelorus is a Grafana dashboard that can easily be deployed to an OpenShift cluster, and provides an organizational-level view of the [four critical measures of software delivery performance](https://blog.openshift.com/exploring-a-metrics-driven-approach-to-transformation/).

          ## Software Delivery Metrics Dashboard

          A short video describing each of metrics provided by Pelorus is available [here](https://www.youtube.com/watch?v=7-iB_KhUaQg).

          ## Demo

          [YouTube Video](https://www.youtube.com/watch?v=VPCOIfDcgso) with the Pelorus in action recorded during online Konveyor Community event.

          ## Documentation

          Pelorus documentation is available at [pelorus.readthedocs.io](https://pelorus.readthedocs.io/en/latest/).

          ## Contributing to Pelorus

          If you are interested in contributing to the Pelorus project, please review our Contribution guide which can be found in the [contribution guide](https://github.com/dora-metrics/pelorus/blob/master/CONTRIBUTING.md)

          ## Statement of Support

          Our support policy can be found in the [Upstream Support statement](https://github.com/dora-metrics/pelorus/blob/master/docs/UpstreamSupport.md)

          ## Code of Conduct
          Refer to dora-metrics's Code of Conduct [here](https://github.com/dora-metrics/pelorus/blob/master/CODE_OF_CONDUCT.md).

          ## License

          This repository is licensed under the terms of [Apache-2.0 License](https://raw.githubusercontent.com/dora-metrics/pelorus/master/LICENSE).
        displayName: Pelorus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pelorus-operator
        - dora
        - dora metrics
        - pelorus
        - metrics
        - transformation
        - devops
        links:
        - name: Pelorus Docs
          url: https://pelorus.readthedocs.io/en/latest/
        - name: Pelorus Getting Started
          url: https://pelorus.readthedocs.io/en/latest/GettingStarted/Overview/
        - name: Pelorus GIT repository
          url: https://github.com/dora-metrics/pelorus/
        maintainers:
        - email: kgranger@redhat.com
          name: Kevin Granger
        - email: whayutin@redhat.com
          name: Wesley Hayutin
        - email: mpryc@redhat.com
          name: Michal Pryc
        - email: msouzaol@redhat.com
          name: Mateus Souza Oliveira
        - email: esauer@redhat.com
          name: Eric Sauer
        maturity: alpha
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - quay.io/pelorus/pelorus-operator:0.0.9
        - quay.io/community-operator-pipeline-prod/pelorus-operator@sha256:f54c42a30675897a971cac278d5d8ea24edf3121570264c07cde04f11cc4e7a8
        version: 0.0.9
      entries:
      - name: pelorus-operator.v0.0.9
        version: 0.0.9
      - name: pelorus-operator.v0.0.8
        version: 0.0.8
      - name: pelorus-operator.v0.0.7
        version: 0.0.7
      - name: pelorus-operator.v0.0.6
        version: 0.0.6
      - name: pelorus-operator.v0.0.5
        version: 0.0.5
      - name: pelorus-operator.v0.0.4
        version: 0.0.4
      - name: pelorus-operator.v0.0.3
        version: 0.0.3
      - name: pelorus-operator.v0.0.2
        version: 0.0.2
      - name: pelorus-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: pelorus-operator
    provider:
      name: Red Hat
      url: https://redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cli-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cli-manager-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "CliManager",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-cli-manager-operator"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2024-06-04"
          description: "The CLI Manager is a comprehensive tool designed to simplify
            the management of \nOpenShift CLI plugins within the OpenShift environment.
            \nModeled after the popular krew plugin manager, it offers seamless integration,
            \neasy installation, and efficient handling of a wide array of plugins,
            \nenhancing your OpenShift command-line experience.\n"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-cli-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/cli-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CLI Manager
            kind: CliManager
            name: climanagers.operator.openshift.io
            version: v1
          - displayName: CLI Plugin
            kind: Plugin
            name: plugins.config.openshift.io
            version: v1alpha1
        description: "The CLI Manager is a comprehensive tool designed to simplify
          the management of \nOpenShift CLI plugins within the OpenShift environment.
          \nModeled after the popular krew plugin manager, it offers seamless integration,
          \neasy installation, and efficient handling of a wide array of plugins,
          \nenhancing your OpenShift command-line experience.\n"
        displayName: CLI Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cli-manager-operator
        - cli-manager
        links:
        - name: Source Code
          url: https://github.com/openshift/cli-manager
        - name: Source Code
          url: https://github.com/openshift/cli-manager-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/cli-manager/cli-manager-rhel9-operator@sha256:be0fc230fef111cff36a8b846bbac9aaa98aae94ad9200416e841aa487121de4
        - registry.redhat.io/cli-manager/cli-manager-rhel9@sha256:a6b7857c1acd5173c770e6ba1d03947eb13804d2512f7f7cd6bb96e5e5cdaa17
        - registry.redhat.io/cli-manager/cli-manager-operator-bundle@sha256:74c88f1cfef4cc31b794fc5ed8dcc10d3b0e6869c228d808f1d98819e69554b6
        version: 0.1.1
      entries:
      - name: cli-manager-operator.v0.1.1
        version: 0.1.1
      name: tech-preview
    defaultChannel: tech-preview
    packageName: cli-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: advanced-cluster-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: advanced-cluster-management.v2.12.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-21T21:16:17Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.12.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.12/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d239df44dafac3bde3eb7ae8c80cb342fa4bb4eabe22f27d2c92d283324e6e10
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:8ccbbc058d324429aef0c31c99de8c7b3a5abc02ec6f7a14cc60eef63a8f5c69
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:7050f71c3a038fec16db8d0eee69d7d5a4294c043c13e9cf69c3cb2e28897f4c
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:3ac71b363aef8db127b88e0a63ee616d0d06342142f83522e34c9cac4dd77b2e
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:b782190b3a285003d319321be09a25ee1376c951cb0cd146cbb634c9b1ca97b3
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:3f286e3c24fe73b4b42e6cf4b6769f91e333014450e1ab135b1e5b4a271cbaec
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:3d3260ff206c5ee249443b1f8da49562584f7ee1abd284691b42f1728b6ed5dc
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:70ddf96707d398d5444bb7bbde6c684275cff62b8a83c7cf6143ff529daf0dee
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:1e5c12ed0de090198e097cc5deb133ad075fba224f239b8b892e6e1672ec9883
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:7375991ce3b9e64cb7542dc176414f0974047667d08beac184e2adadd9f83532
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:794ef15591f217be629d6c680c56e84962923993f6dd77514ca5d2bdbd96217b
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:ec65a5744945ca20f804d94999a4002efab9ddad360a58f9498800bffa9fa0f4
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:7c548dd869f3458fb70a536708306cb615b9166c7e866cc1215e82f183b39529
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:58f08a93cd35bd2ac19f94a1779205c81a2fc3a552d1e77b7c659fd51c9dd34f
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:bd24cfabb81b464a31c3d8c21fa6371e46b83094d5c750acc33346203fe15e84
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:6a7c54cb7e6c3cc506ffb8a4f3ec7443ef80798e2c917e20b5f06d39bde02bc4
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3a00f55dee152e0345e2616ac14f652dcb4f37f3b9dfca6f049ac5522874c6e
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:c4c3bc4f45ae194b884d89319625f2a0970dada823877a35feecd8c733cb2c8d
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:a0e4c432590e65818e7ebaab127d483d68fdc5ccae9a13820665a8aef122fde7
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:3faa2318f159d264a9194fc01239d3a2098c184b365b153e3c56e555d5d4992e
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7968e96b48b2ed63e7657793cf6cc2564055809df19cb43b7e284ac03f2cc26c
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cdaf7be69a5d66cfa4777feb8266421dafd55353d6703e7e591bb9fb23017abc
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:08dd3f6805fa02478671927982a2c77bdec0a5b7e7fd1c9faf20c8af2bb7b428
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:b85f3de5231f2f1bc3b4e9c95917f562f39e09dc9b7db35690024163e3a5afe5
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:c46657831915ed5f42f4a55e725bc9d17e0671191f4509e610468fda160710c0
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:fbfade039739179995828b2ce4f4411b1315ac9a0e4a98e96a115012676a859d
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:c402a4ac61e11dc9e45fdc7bc7fcb5bf5ca3475a66d782aea60d9f6524e51a24
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:a5c65dee1f7e4177050244109c10e384899e01d1924b41a540373aa7daf4f274
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:f1c371fef66e96199c4a96e21192b47bca423b282d6f99e927644840aaedfd80
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:fff4fb0924f866b305ad0ecc5d2682da6cb20de15448a2a8c640df182719b4e2
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:c2259b90decf24f8fb1cb058d4751c75fd1558e3c9fd1079857ad9c47ca9e8d2
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:1419e274b4e2e4c37985208c3b1b9ecd0cdef493427ace3c71ff35ed8fcc394b
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:6f2c4f1abc4388e76414544e65f08238c96fc4d6fbf71fa79330d9dca117d221
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:875e4bf42ad192cfb954e38c8474a529aa938855d69dc0d83c4ee4a4777c1028
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:93df0586f9281fa2879eed0fdf2fb115be355d572cc24dc5c5c9a92137044877
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:c10a72bf8c0d76b9b4d9d7c0bfca4d83a3f7dd7f6a37b4cc24ee7591638a2482
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:b83e630e77bd011c374d73d801f10a3230f6798ede10d7f39bf2e8ff6437cf2c
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:dca845ad05b02619b81f231eb57dd826720c221d22282fe7422ceebe57191371
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:9c297b4c1f9ba17908594f1843ea6de6c5a5e0f3500258c23fbb8b5684249c9e
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:1ccd17b482df844ec42936993ff162f665b5a473ad344270ff2425cd63d428fc
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:28730c8f2344e51b0b9ca603ac63b60885b39be86582943bd1cc200779eee0c5
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:22e262c0353c6f0d5d1b916074836c53437e22697f67c85a43e6d717daf651fa
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:7b64803ddd3b7886ad3ec3de943a38cec5aeb9604732ab7a5ea5061748bf1411
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:5156129064dbe4e9989b41018b1bb250b2eb3f02e0c0afd75a874da9b54ad3c2
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:187879345174e515863bc1b29a961a0a0d9ad44006a6e12de715b4508dec38a8
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:8e62fa4decf01a3beae7080f19469f33822f5010a688ada5db8d2d23b008e498
        version: 2.12.4
      entries:
      - name: advanced-cluster-management.v2.12.4
        version: 2.12.4
      - name: advanced-cluster-management.v2.12.3
        version: 2.12.3
      - name: advanced-cluster-management.v2.12.2
        version: 2.12.2
      - name: advanced-cluster-management.v2.12.1
        version: 2.12.1
      - name: advanced-cluster-management.v2.12.0
        version: 2.12.0
      name: release-2.12
    - currentCSV: advanced-cluster-management.v2.13.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-09-08T18:37:07Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <2.13.4'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n#### Upgrading from release 2.4\nThe
          special considerations described above also apply when upgrading from Red
          Hat Advanced Cluster Management for Kubernetes release 2.4.\n- If you are
          using a filtered mirror catalog, update your catalog-building procedures
          to ensure that the `multicluster-engine` package is included in your mirror
          catalog.\n- Before changing the update channel to trigger the update, edit
          your existing `MultiClusterHub` resource to add the `mce-subscription-spec`
          annotation as shown in the example above so that the upgraded version of
          the Red Hat Advanced Cluster Management for Kubernetes knows the correct
          catalog source to use to find the `multicluster-engine` package.\n### Where
          to find more installation information\nYou can find additional installation
          guidance in the [install documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.13/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:1393744757874e79d1fb058cb5d3fe7dcbbcd2f187bdcdcdf34e67227c61fbb9
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:f74fb6b4d42f1d7431bfc9d2424f953732fed06d2abd8d1fc27b878b4d1322d5
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:588678bea97237d87cd2e55a0e8000e73461cfc421cc91d554a2f1e352b68d14
        - registry.redhat.io/rhacm2/acm-flightctl-ocp-ui-rhel9@sha256:89863ce0eb5fb73272bda8762d74461552f8502fbe262b16cbde01064cc533b3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:a0218b9d5f1803f60d9b295d1504a870780895cffe8f1c5b3ce4c3553da5e4f9
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:ba51f7d077858a0abe32a254970390b50cd7d75433b5661e17f887c48c56a30f
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:5aec572fc18c8ed533620553aba2c676453297ad2440336a5228a5a4f50f5782
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:72c6896a5a4dbb7717d76d9eeade96130d97909eee07e8d3cc17af0658b306ab
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:04bcddae590e2debf35090055dd52ae4bcaac4b369727266be53ac64670203d0
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:296a31e1aad9f6288bdf80fb3191642e3dd404e9e8fcad198016bedbd3abe9b4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:b1620cebc62240f93b12e6937d15760440be119c421c127214974bdfa9d08ebd
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:7f742f024c787932932b8a4100a5329e70aad39537e0c20c2a4e245fc6245e25
        - registry.redhat.io/rhacm2/acm-flightctl-ui-rhel9@sha256:86800d67ec8ed63e5408947ba6a441e5afc1050b72a86487b9357674ed495a7e
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:8b152051ef2c4c077379e72aa510c0bd613b219254cad66fab0cd3ec0d1d4bf1
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:7b15c6b492e5cb2725b9b345aa3bfae4108e2815bc233ac063b81a67c6550fb6
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:65e1f0eb4e6c797c6dc6518674b28370db3fdaa6d7e17f4233c07207224dcb48
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:4d068813e3eb704ea7b2d342d5a2ce8fa2ac71793b1b37634c18704ef0a5e8d7
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:49e99f41cbea3faefc3ae93385a10b544ce91df0d6639cfee65760be6ae0df00
        - registry.redhat.io/rhacm2/console-rhel9@sha256:5526edf70b784bddfc2b7e1c154a44201c346fc20117635500bb54d325921c25
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:2da264ca59278f63990e579939371d45afec3bc7d9922a608520e7647a0377ac
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:fc16203de469dd1ba16e428a9d7efbb16a5f511d866ef6fcc253530de1f032a6
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:b3535378919524df259b85587ca83f5620433914418c9941374951c36e0530ae
        - registry.redhat.io/rhacm2/acm-flightctl-worker-rhel9@sha256:57f69874b2df77c185374ddcea530af6783410a0f35528ee1a44cc2dee6f7fa2
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:1b7e6e6a4ae36209f98b1539c37c5975a82363b33fe5ac1e3a320e58b99be239
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:3028d011e7b88dd9597821128275f606fdba639e5b83e4e6a8783fa2616cba6f
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:a2ab651af1c97e54d2c14de78ff554c0f25e8d000af93a74fc7b7ed913f4e338
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:3cbb9d271c2f23d430a623c83606bbc8a80379c24d13e610886ad51fd21ce1e4
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:5fc35a944895cf7ea6afa5929c45304aaa9f17397b4f1e42cb8c799d9aead9d5
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:dc271885425ab74e0554b531a2078133ce02006d0bb9688f3b54a850cfefd8f8
        - registry.redhat.io/openshift4/ose-cli-artifacts-rhel9@sha256:67ff9f19edac8d623ce9080a8c05cde3da73afd68e7352e4ee5b22a9ccb596bc
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:e23115b10a92c323b4cdaf0412919c273dbca33f19910120498bd49cb5810ff5
        - registry.redhat.io/rhel9/redis-7@sha256:848f4298a9465dafb7ce9790e991bd8a11de2558e3a6685e1d7c4a6e0fc5f371
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:0272215ab43b5688295d94cce0de4490a2fbabc1dd09ed41645be7acef1e3c9b
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:ae1f175ffbd3698f2b3036bf9b62b5c9b24729cfb8248345b8bdd1525ccc5bb4
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:f1db9b10481319dc142ca6046e4e3141addb4c0afab482eeaf80643fad9ad039
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:6f95308a508499d2d995653d3fd2def94ac5170b4c3f391d32a1e0594250c33a
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:ccc9c0524149820d701c39355af89287ad749832053ea55d1585e77347fa0ec0
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d29ceda561f36faa581d76d8b889f7e01d253d33c88709078855ae26cada4917
        - registry.redhat.io/rhacm2/acm-flightctl-api-rhel9@sha256:8302b536c896290533f3f50084cb9047cb55af3a4404887b70704d23862619fa
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:3490daccc203fade26e7df348c74a907ab0ffc8901f4965606c8227849087c44
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:669e9e2178476bc7a4688f3ebf4c20402608bb4b10e4df6863ec821c7e572299
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/rhacm2/acm-flightctl-periodic-rhel9@sha256:2c9b3c3d6cd7e435da337447580dfbfde8c4932308405eeae5d97ed5ec57fbfc
        - registry.redhat.io/rhel8/postgresql-13@sha256:826a1a75f4186f36217b5c1f1aa270838c4b55b1021d33a4876566bfc3b8e629
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:242d754d5fe1f2a07091ae93da284b8912e0f4b4c973200cf6d1241e6abb88a9
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:ab1e4f8685896e2016a0b4c4479378058314f365e4f09ef42cba67f3649bb306
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:81cc71faaf20a986d233fee799a7d2b954cd3f43875a2689f76c53920ccf4df0
        - registry.redhat.io/rhel8/postgresql-16@sha256:89dfeaefc7e9e974e47a3365e22f354314f8a5e906e93a0152afe4e42f6a82d0
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:ab0192a51ee37b7508f3e4db66ffa53b904dcb1f76d9aa1ac0e80447d8ff9e6a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:ce342242c50c103fc5e4b0126551231977b35bd477eb627fb8f5d2a6510ddaa0
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:67939d6e4342372820d7af6777f0ffd123e86e4e8186132c3d4892cb717a5e7c
        - registry.redhat.io/rhacm2/memcached-rhel9@sha256:9ef9f1e2010d77211126482c03e8c72dfb7458940dc95ac952ee8dd1dedd2991
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:84109be1814719c343b8b4676981cdbcea0356f012b6c854221c271bdb518fea
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:6ed6f62f5c05a24938335287b785b80c253424e1ee20d83311cd87c6a2090895
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:0040791500337f88af7d922895487549915faa3b69dacba1a23540e7d6f2e580
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:0d8202cd10170bb0a138d52a1d71e630ff74df466541897ba86ed324a75ea1d8
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:96bdb2e25ae0f381de0c94a09a3cd11f3c0bdc1d8cfb4fd5ced2cd7b24fbe2c2
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:f9ba60c2a415a863c6bd6030fb4663913368a9c271bb40b8acfdc294dcf3a7eb
        version: 2.13.4
      entries:
      - name: advanced-cluster-management.v2.13.4
        version: 2.13.4
      - name: advanced-cluster-management.v2.13.3
        version: 2.13.3
      - name: advanced-cluster-management.v2.13.2
        version: 2.13.2
      - name: advanced-cluster-management.v2.13.1
        version: 2.13.1
      - name: advanced-cluster-management.v2.13.0
        version: 2.13.0
      name: release-2.13
    - currentCSV: advanced-cluster-management.v2.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          createdAt: "2025-07-25T02:10:33Z"
          description: Advanced provisioning and management of OpenShift and Kubernetes
            clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.13.0 <2.14.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"open-cluster-management"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: open-cluster-management
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          repository: https://github.com/open-cluster-management/multiclusterhub-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: "Red Hat Advanced Cluster Management for Kubernetes provides
          the multicluster hub, a central management console for managing multiple
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. You can use the hub to create Red Hat OpenShift Container Platform
          clusters on selected providers, or import existing Kubernetes-based clusters.
          After the clusters are managed, you can set compliance requirements to ensure
          that the clusters maintain the specified security requirements. You can
          also deploy business applications across your clusters.\n\n## How to install\nUse
          of this Red Hat product requires a licensing and subscription agreement.\n\nInstall
          the Red Hat Advanced Cluster Management for Kubernetes operator by following
          instructions presented when you click the `Install` button. After installing
          the operator, create an instance of the `MultiClusterHub` resource to install
          the hub.\n\nNote that if you will be using the hub to manage non-OpenShift
          4.x clusters, you will need to create a Kubernetes `Secret` resource containing
          your OpenShift pull secret and specify this `Secret` in the `MultiClusterHub`
          resource, as described in the install documentation.\n### Special considerations
          for disconnected network environments\nIf you are installing Red Hat Advanced
          Cluster Management for Kubernetes in a disconnected network environment,
          there are some special install and upgrade requirements.\n#### Include multicluster
          engine for Kubernetes in mirrored catalog\nInstalling an OLM operator in
          a disconnected network environment involves the use of a local mirrored
          operator catalog and image registry. Red Hat Advanced Cluster Management
          for Kubernetes requires (and automatically installs) the multicluster engine
          for Kuberenetes operator so this operator package is needed in your local
          mirrored operator catalog in addition to the one for Red Hat Advanced Cluster
          Management for Kubernetes.\n\nif you are using a full mirror of the Red
          Hat operators catalog, no special action is needed when creating the mirror
          catalog.\n\nIf you are using a filtered (rather than full) mirror of the
          OLM catalog, be sure to include both the `advanced-cluster-management` **and**
          `multicluster-engine` packages in your mirrored catalog when creating it.\n####
          Add annotation to MultiClusterHub resource\nIf you are installing or upgrading
          to a Red Hat Advanced Cluster Management for Kubernetes release prior to
          release 2.7, an `mce-subscription-spec` annotation is required in the instance
          of the `MultiClusterHub` resource you create to set up your hub. This annotation
          identifies the catalog source for your local mirrored operator catalog and
          is required when installing releases prior to release 2.7 regardless of
          whether your local mirrored operator catalog is a full mirror or a filtered
          one.\n\nThe `mce-subscription-spec` annotation is not usually needed, and
          thus can usually be omitted, when installing or upgrading to release 2.7
          or a later release. Starting with release 2.7, the Red Hat Advanced Cluster
          Management for Kubernetes operator is able to automatically determine the
          catalog source to use in typical scenarios.\n\nThe following is an example
          of the `mce-subscription-spec` annotation:\n\n```\n apiVersion: operator.open-cluster-management.io/v1\n
          kind: MultiClusterHub\n metadata:\n \tannotations:\n \t\tinstaller.open-cluster-management.io/mce-subscription-spec:
          '{\"source\": \"my-catalog-source\"}'\n spec: {}\n\n```\n\nReplace `my-catalog-source`
          with the name of your catalog source.\n\nIf you are creating the `MultiClusterHub`
          resource using the OperatorHub UI, switch to the YAML view for the resource
          in order to insert the annotation.\n### Where to find more installation
          information\nYou can find additional installation guidance in the [install
          documentation](https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/html/install/installing). "
        displayName: Advanced Cluster Management for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - acm
        - advanced-cluster-management
        - install
        - multiclusterhub
        - rhacm
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_advanced_cluster_management_for_kubernetes/2.14/
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ee742e7baa8bcbbfccf79abcf15d22347d14f280eafaf22c42cef71051b10fba
        - registry.redhat.io/rhacm2/kube-rbac-proxy-rhel9@sha256:0ab21d4189669901fe085c7f3a4b4c9778c853f8abd706bbfffcb59942fbc8e6
        - registry.redhat.io/rhacm2/prometheus-alertmanager-rhel9@sha256:143f49d2ce1f58282e276396722ba37165cd1bf2051d537bb4e6208786e5aed6
        - registry.redhat.io/rhacm2/insights-metrics-rhel9@sha256:d9fa60b5b1d6b6c94c5d8b610c9a7b979202de8b46a344e0a65d06f6ef21a7ef
        - registry.redhat.io/rhem/flightctl-api-rhel9@sha256:83c6c7819e04aa58d961518367ce6ad3159a28300ca097967f630044d181be34
        - registry.redhat.io/rhem/flightctl-ui-ocp-rhel9@sha256:973eb3ca3da32c87ba308928a21fbd25514b2f6a6662300d4928479cb554cff3
        - registry.redhat.io/rhacm2/multicluster-observability-rhel9-operator@sha256:0ea80a1039d217e899ea012b77e99a7c5832e5b7e292dca473759c449bf2d1ea
        - registry.redhat.io/rhacm2/acm-must-gather-rhel9@sha256:556d737195a4878877c650a93619af1528e52694f2a8e47e13463a603d2f4900
        - registry.redhat.io/rhel9/redis-7@sha256:ff31822cba556e40b6b00720839cc56f474fdb9257fb39b7604fafc11761d379
        - registry.redhat.io/rhacm2/insights-client-rhel9@sha256:53deacec64238607afda29c27463d72ac3b582c9670e2e77a9c7e48851214795
        - registry.redhat.io/rhacm2/acm-prometheus-config-reloader-rhel9@sha256:44e305ca15e7c9b50bd7fa8bb391f5ee66906aa26ecc5f612d029aaee5bb4cb8
        - registry.redhat.io/rhacm2/observatorium-rhel9-operator@sha256:c70242a037e36690d8dd1030254dde778917356b4f1f61db2cecb03b6c2d7380
        - registry.redhat.io/rhacm2/acm-grafana-rhel9@sha256:ce5ff191023b5e14c339b2a03aa1be05a12d024efa5302652065597e630996b6
        - registry.redhat.io/rhacm2/acm-search-indexer-rhel9@sha256:c6ff6cae5bc6447d0ed24e458fc5d5fc709b2106fe93b143f9546bd962ea7f87
        - registry.redhat.io/rhacm2/thanos-receive-controller-rhel9@sha256:f3e34b200e9480363c56da1cd88b2de37fc6a490691e4596a132217f8ea286d8
        - registry.redhat.io/rhacm2/kube-state-metrics-rhel9@sha256:0cecc3d87afcd0d63bdeae3ecae37f8a38534fbdc8bad5c548e0d4f4fb0156a4
        - registry.redhat.io/rhacm2/rbac-query-proxy-rhel9@sha256:935baf67627d5e76c0766d5d2a9e36b398a12535f097749a987dd0c5de788570
        - registry.redhat.io/rhacm2/acm-search-v2-rhel9@sha256:40b3e063435eea6e526a57d6bd1ae489ca703e81eeca82dda1c2e6aedec1262c
        - registry.redhat.io/rhel9/memcached@sha256:955138db9a7a23ee31198aabebd0811438fee11bb0b624c38ddb90efe52bd07c
        - registry.redhat.io/rhacm2/acm-volsync-addon-controller-rhel9@sha256:e063c54c7182b70d78d24b5026f174436885383980c9b1e3e7664fa2adfad883
        - registry.redhat.io/rhacm2/acm-governance-policy-framework-addon-rhel9@sha256:c329cfd004c58a7fcc40f4fbb986399dc2fe08d7922d2b4f210664f58751b1a6
        - registry.redhat.io/rhacm2/governance-policy-propagator-rhel9@sha256:d65aef370a7ed01c0c7cc0ad7f0800359db7863db4ec8845de5b450992f26933
        - registry.redhat.io/rhacm2/cert-policy-controller-rhel9@sha256:69f0b9a7fb494572adf2795be865d6d6ba4b67db0eed0e9d61f5dcb5e9bf6efe
        - registry.redhat.io/rhacm2/metrics-collector-rhel9@sha256:46c5835a4bfc0e4767d0959bc47e939c195be197e9715aae17788d4d0e799d56
        - registry.redhat.io/rhacm2/observatorium-rhel9@sha256:cee29bdb03927dc1866c267232d391ea5ee762b23c6b04a777e7d61d14a61bba
        - registry.redhat.io/rhem/flightctl-worker-rhel9@sha256:1b1f6f85af0910f1acd029f54a608b25d3a2977c3d8b323266391543324305b4
        - registry.redhat.io/rhacm2/acm-search-v2-api-rhel9@sha256:d56b997a9084d56c63f539147052e64ba30258afe91a9ff832f3673cbccbe467
        - registry.redhat.io/rhacm2/console-rhel9@sha256:d4c42a203110050a3834be8d5a0eee384ad7e9fbb991da589d75acd15055b8b1
        - registry.redhat.io/rhacm2/acm-prometheus-rhel9@sha256:ad534f7b0efea45b9f4a0a026c76a2a030cb608acdb835dfea3852332acfcae1
        - registry.redhat.io/rhacm2/volsync-rhel9@sha256:90d399736ca81bb244415efaf05738beff806912a6ac7444a7d4839089b7a081
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:a48758b69b9a7451b7e0caea465a181f854b41f1e74ce86c15216ee1bb31fcb9
        - registry.redhat.io/rhacm2/acm-operator-bundle@sha256:437a786a301ea720cdb5651dff878a9f51470bab46ba28daae4928c7074c9157
        - registry.redhat.io/rhacm2/node-exporter-rhel9@sha256:b489229190f9a3da66bd154e65490c3a35d1999d2db6707759f06eadb02b968d
        - registry.redhat.io/rhacm2/multiclusterhub-rhel9@sha256:d550013d385552d6242b56c5b60f15b3ebb488c2033f808de54bcc639abc2f30
        - registry.redhat.io/rhacm2/klusterlet-addon-controller-rhel9@sha256:ad7f30c61bb11c0d32686e75b01f3b286a1fb06edd8d2ef6a9e2d924fba2b565
        - registry.redhat.io/rhem/flightctl-ui-rhel9@sha256:940c0564ced75b7912d3ffc923a1f6bab9aed0f4a52962f7f894d08937b7d37e
        - registry.redhat.io/rhacm2/acm-cluster-permission-rhel9@sha256:0e1eba23e0e8454f2dec240704e7ccc922351d58880f72514d3b2bc30c9e995e
        - registry.redhat.io/rhacm2/memcached-exporter-rhel9@sha256:8e57af77cc5dee3aaa83f6ae6e62e232c5fc4fbe4dca8eb755226351963058e3
        - registry.redhat.io/rhacm2/config-policy-controller-rhel9@sha256:b7b7581e7743148d19283250d3300d292b96c48ecae96f6e6202a7ca8204e2ab
        - registry.redhat.io/rhacm2/thanos-rhel9@sha256:0cc23b5063c0d3264ed345dad8453901e4db18cc9bed1656f2799f72de75bd53
        - registry.redhat.io/rhacm2/submariner-addon-rhel9@sha256:2b1b1c707e0b5217c66a96fb82070de47fc0342747081d4be0049c4542944c8a
        - registry.redhat.io/rhacm2/acm-multicluster-observability-addon-rhel9@sha256:8f4f59e237f366f08367cabbea3eb51e766ad52db0d35c5efc56182c13fc5c09
        - registry.redhat.io/rhem/flightctl-periodic-rhel9@sha256:ee59f307e468e2b7b38f7c040898d587d441f8fb3d59cb022ae23e4717739e3f
        - registry.redhat.io/rhacm2/search-collector-rhel9@sha256:f9a9e10fde9065d650a7dc8eba83349e5db0a21727027464cf144724d0de8cfa
        - registry.redhat.io/rhacm2/endpoint-monitoring-rhel9-operator@sha256:7ee0aea84bf00b7b42027fa3e970c1f4b279e345a698b4050c9f413eb0383bb1
        - registry.redhat.io/rhacm2/grafana-dashboard-loader-rhel9@sha256:8253a9225f30b0f0a405e2c3293924acbfbf81f3d9c50bc4431fe4052a9fe908
        - registry.redhat.io/rhacm2/multicluster-operators-application-rhel9@sha256:0252cf95019717f03c7ac72fb034b5bd9ad7aef2da0a92dd21c56cf47f03b812
        - registry.redhat.io/rhacm2/prometheus-rhel9@sha256:cb5346afe1c7b58eb758cdced94f2c1161c1621033b52f29757efe3a50320a65
        - registry.redhat.io/rhem/flightctl-cli-artifacts-rhel9@sha256:3491ec932d90112e89f3a4cebe5bdd9ed6a7811f116df43bab8bb5f827129361
        - registry.redhat.io/openshift4/ose-configmap-reloader-rhel9@sha256:c2ef90b4704251dc9f7ef5288f5bad7aeb4cd2b3706ec531c60f6f4346073a77
        - registry.redhat.io/rhacm2/multicluster-operators-subscription-rhel9@sha256:8151a89ebfebf2e6d336849fdd329b123b7f486938808ee80bf6bd8d44f6985c
        - registry.redhat.io/rhacm2/multicloud-integrations-rhel9@sha256:95085fc33a0150a39b6ef4e616fd23f0bddd9dcad0e75c4a89ed27113626e6eb
        - registry.redhat.io/rhel8/postgresql-16@sha256:d4abbd9492347e0aa7e1b0ff967e74d4b6a4edd6fd20d896f2f2ae03c623181c
        - registry.redhat.io/rhacm2/acm-governance-policy-addon-controller-rhel9@sha256:9c79b1153fc927e7582c5346ea0408e03f794eb983597f02a5e816ae961ba86b
        - registry.redhat.io/rhacm2/cluster-backup-rhel9-operator@sha256:55ef2c163b0c0deb3402e39acf37ff6bb05e2a5d268dcae5f8a400128d67c584
        - registry.redhat.io/rhacm2/acm-cli-rhel9@sha256:364c7416a9642dfb37f05e2eac4427918959b12b159ac90a9ca310c584d6434a
        - registry.redhat.io/rhacm2/acm-siteconfig-rhel9@sha256:bcb5949ebc31bbf6b45dfd34aae9c80c82a08531e8a3d9f6eef3354f942a5a5d
        - registry.redhat.io/rhacm2/multicluster-operators-channel-rhel9@sha256:8e6ee423acfafe28c002911e832d47dffc00a5227325bd6a5b1ab359c7eee5f4
        version: 2.14.0
      entries:
      - name: advanced-cluster-management.v2.14.0
        version: 2.14.0
      name: release-2.14
    defaultChannel: release-2.14
    packageName: advanced-cluster-management
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sriov-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-network-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovIBNetwork",
                "metadata": {
                  "name": "example-sriovibnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "linkState": "auto",
                  "networkNamespace": "default",
                  "resourceName": "mellanoxnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetwork",
                "metadata": {
                  "name": "example-sriovnetwork",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"host-local\",\n  \"subnet\": \"10.56.217.0/24\",\n  \"rangeStart\": \"10.56.217.171\",\n  \"rangeEnd\": \"10.56.217.181\",\n  \"routes\": [{\n    \"dst\": \"0.0.0.0/0\"\n  }],\n  \"gateway\": \"10.56.217.1\"\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "intelnics",
                  "spoofChk": "on",
                  "trust": "off",
                  "vlan": 0
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodePolicy",
                "metadata": {
                  "name": "policy-1",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "deviceType": "vfio-pci",
                  "mtu": 9000,
                  "nicSelector": {
                    "pfNames": [
                      "eth1"
                    ],
                    "rootDevices": [
                      "0000:01:00.1"
                    ],
                    "vendor": "8086"
                  },
                  "nodeSelector": {
                    "feature.node.kubernetes.io/network-sriov.capable": "true"
                  },
                  "numVfs": 6,
                  "priority": 99,
                  "resourceName": "intelnics"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkNodeState",
                "metadata": {
                  "name": "minikube",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "interfaces": []
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovNetworkPoolConfig",
                "metadata": {
                  "name": "sriovnetworkpoolconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "sriovnetwork.openshift.io/v1",
                "kind": "SriovOperatorConfig",
                "metadata": {
                  "name": "default",
                  "namespace": "openshift-sriov-network-operator"
                },
                "spec": {
                  "enableInjector": true,
                  "enableOperatorWebhook": true,
                  "logLevel": 2
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
          createdAt: "2024-11-21T23:48:37Z"
          description: An operator for configuring SR-IOV components and initializing
            SRIOV network devices in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "sriovnetwork.openshift.io/v1",
              "kind": "SriovOperatorConfig",
              "metadata": {
                "name": "default"
              },
              "spec": {
                "configurationMode": "daemon",
                "disableDrain": false,
                "enableInjector": true,
                "enableOperatorWebhook": true,
                "logLevel": 2
              }
            }
          operatorframework.io/suggested-namespace: openshift-sriov-network-operator
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-sriov-network-operator",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/internal-objects: '["sriovnetworknodestates.sriovnetwork.openshift.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/k8snetworkplumbingwg/sriov-network-operator
          support: Red Hat
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OVSNetwork is the Schema for the ovsnetworks API
            displayName: OVSNetwork
            kind: OVSNetwork
            name: ovsnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovIBNetwork is the Schema for the sriovibnetworks API
            displayName: Sriov IBNetwork
            kind: SriovIBNetwork
            name: sriovibnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodePolicy is the Schema for the sriovnetworknodepolicies
              API
            displayName: Sriov Network Node Policy
            kind: SriovNetworkNodePolicy
            name: sriovnetworknodepolicies.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetworkNodeState is the Schema for the sriovnetworknodestates
              API
            displayName: Sriov Network Node State
            kind: SriovNetworkNodeState
            name: sriovnetworknodestates.sriovnetwork.openshift.io
            version: v1
          - kind: SriovNetworkPoolConfig
            name: sriovnetworkpoolconfigs.sriovnetwork.openshift.io
            version: v1
          - description: SriovNetwork is the Schema for the sriovnetworks API
            displayName: Sriov Network
            kind: SriovNetwork
            name: sriovnetworks.sriovnetwork.openshift.io
            version: v1
          - description: SriovOperatorConfig is the Schema for the sriovoperatorconfigs
              API
            displayName: Sriov Operator Config
            kind: SriovOperatorConfig
            name: sriovoperatorconfigs.sriovnetwork.openshift.io
            version: v1
        description: |
          # SR-IOV Network Operator for Openshift

          ## Introduction
          The sriov-network-operator is generally responsible for configuring the sriov components in a openshift cluster.

          ### Supported Features
          * Initialize the SR-IOV NICs on nodes.
          * provision SR-IOV device plugin on selected node.
          * provision SR-IOV CNI plugin on selected nodes.
          * manage configuration of SR-IOV device plugin.
          * generate net-att-def CRs for SR-IOV CNI plugin.

          This operator has to run in namespace 'openshift-sriov-network-operator'. An Operator Group is also required to install this operator:

          ```
          $ oc create namespace openshift-sriov-network-operator

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: sriov-network-operators
            namespace: openshift-sriov-network-operator
          spec:
            targetNamespaces:
            - openshift-sriov-network-operator
          EOF
          ```
        displayName: SR-IOV Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - network
        - sr-iov
        - sriov
        links:
        - name: Source Code
          url: https://github.com/k8snetworkplumbingwg/sriov-network-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-sriov-network-operator-bundle@sha256:4c8271d0bf6d94061fe8df8110b041cb0e1a840fb6263315f552166fc5053e73
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-sriov-network-metrics-exporter-rhel9@sha256:2e71d4125a35f6e7bd2a04cdd8cb6aae058d745eda372ae3d3fd37783778f486
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin-rhel9@sha256:57a57f4f3a0adb8b7234683f2968352988ec9a414824f4f444cca096825a5df5
        - registry.redhat.io/openshift4/ose-sriov-network-rhel9-operator@sha256:07d0b5269c9788c5b953cc2fab28f0102e8b02de6142c95d7cc7836b18fc24c4
        - registry.redhat.io/openshift4/ose-sriov-network-webhook-rhel9@sha256:45835aba95af72aa8e73b630d1d2b4c1079217531f025f4715a2eb8e8856205f
        - registry.redhat.io/openshift4/sriov-cni-rhel9@sha256:8b412541a797aaff1b49aa229e310957a8d7da03a3d38e6bbfd65f2ca36c2bdd
        - registry.redhat.io/openshift4/ose-sriov-dp-admission-controller-rhel9@sha256:6eb244a301ae5395b69b1bfcd0baac37c094f4c38a1a330c0a70fb81e71f6157
        - registry.redhat.io/openshift4/ose-sriov-network-config-daemon-rhel9@sha256:3b87d49b200ac4fdeac97f0e97a46001e91c8f4680e2975e90bd94ec37864403
        - registry.redhat.io/openshift4/ose-sriov-infiniband-cni-rhel9@sha256:0f18db6a5bbc952bee8f3e4ac9e93e807d1e85cd132affe94b1f4005c1e128ce
        - registry.redhat.io/openshift4/ose-sriov-rdma-cni-rhel9@sha256:65864bc235689ff38b63d2a7b2c72874ee155241a493fa9429fbe87d5100cfcf
        version: 4.18.0-202509100344
      entries:
      - name: sriov-network-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: sriov-network-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: sriov-network-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: sriov-network-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: sriov-network-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: sriov-network-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: sriov-network-operator.v4.18.0-202505260733
        version: 4.18.0-202505260733
      - name: sriov-network-operator.v4.18.0-202505121333
        version: 4.18.0-202505121333
      - name: sriov-network-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: sriov-network-operator.v4.18.0-202504221235
        version: 4.18.0-202504221235
      - name: sriov-network-operator.v4.18.0-202504160002
        version: 4.18.0-202504160002
      - name: sriov-network-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: sriov-network-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: sriov-network-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: sriov-network-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: sriov-network-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: sriov-network-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: sriov-network-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: operator
      app.kubernetes.io/managed-by: kustomize
      app.kubernetes.io/name: zabbix-operator-certified
      app.kubernetes.io/part-of: zabbix-operator-certified
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zabbix LLC
      provider-url: ""
    name: zabbix-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zabbix-operator-certified-rhmp.v6.0.40
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent",
                "metadata": {
                  "labels": {
                    "app": "agent",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "log_remote_commands": true,
                  "max_lines_per_second": 20,
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixAgent2",
                "metadata": {
                  "labels": {
                    "app": "agent2",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-agent2"
                },
                "spec": {
                  "active_allow": true,
                  "activeservers": "",
                  "allow_key": "",
                  "allow_privileged": true,
                  "buffer_send": 5,
                  "buffer_size": 100,
                  "debug_level": 3,
                  "deny_key": "system.run[*]",
                  "enable_persistent_buffer": false,
                  "enable_status_port": false,
                  "host_interface": "",
                  "host_interface_item": "",
                  "hostname": "",
                  "hostname_item": "system.hostname",
                  "load_module": "",
                  "metadata": "",
                  "metadata_item": "",
                  "passive_allow": true,
                  "passive_servers": "",
                  "persistent_buffer_claim": "",
                  "persistent_buffer_period": "1h",
                  "refresh_active_checks": 120,
                  "server_host": "zabbix-server",
                  "start_agents": 3,
                  "timeout": 3,
                  "tls_accept": "unencrypted",
                  "tls_ca_file_name": "",
                  "tls_cert_file_name": "",
                  "tls_cipher_all": "",
                  "tls_cipher_all_13": "",
                  "tls_cipher_cert": "",
                  "tls_cipher_cert_13": "",
                  "tls_cipher_psk": "",
                  "tls_cipher_psk_13": "",
                  "tls_connect": "unencrypted",
                  "tls_crl_file_name": "",
                  "tls_key_file_name": "",
                  "tls_psk_file_name": "",
                  "tls_psk_identity": "",
                  "tls_server_cert_issuer": "",
                  "tls_server_cert_subject": "",
                  "unsafe_user_parameters": false,
                  "use_host_node_name": true
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixFull",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-full"
                },
                "spec": {
                  "builtin_report_service": true,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysql_volumeclaim": "mysql-full-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxyMysql",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-mysql"
                },
                "spec": {
                  "db_server_port": 3306,
                  "enable_snmp_traps": false,
                  "internal_db": true,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix_proxy",
                  "proxy": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "zabbix_mysql_volumeclaim": "mysql-proxy-zabbix",
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixProxySqlite",
                "metadata": {
                  "labels": {
                    "app": "proxy",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-proxy-sqlite"
                },
                "spec": {
                  "db_data_volumeclaim": "zabbix-proxy-sqlite",
                  "enable_snmp_traps": false,
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "proxy": {
                    "cache_size": "8M",
                    "config_frequency": 3600,
                    "data_sender_frequency": 1,
                    "debug_level": 3,
                    "enable_remote_commands": false,
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "hostname": "",
                    "hostname_item": "system.hostname",
                    "housekeeping_frequency": 1,
                    "log_remote_commands": true,
                    "log_slow_queries": 0,
                    "node_name_as_db_name": true,
                    "proxy_heartbeat_frequency": 60,
                    "proxy_local_buffer": 0,
                    "proxy_mode": 0,
                    "proxy_offline_buffer": 1,
                    "server_host": "zabbix-server",
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_history_pollers": 1,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "timeout": 4,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_connect": "unencrypted",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "tls_psk_file_name": "",
                    "tls_psk_identity": "",
                    "tls_server_cert_issuer": "",
                    "tls_server_cert_subject": "",
                    "trapper_timeout": 300,
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10
                  },
                  "proxy_size": 1,
                  "snmp_traps_volumeclaim": "zabbix-snmptraps",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  }
                }
              },
              {
                "apiVersion": "kubernetes.zabbix.com/v1alpha1",
                "kind": "ZabbixServer",
                "metadata": {
                  "labels": {
                    "app": "server",
                    "vendor": "zabbix"
                  },
                  "name": "zabbix-server"
                },
                "spec": {
                  "builtin_report_service": true,
                  "db_server_host": "mysql-server",
                  "db_server_port": 3306,
                  "enable_report_service": false,
                  "enable_snmp_traps": false,
                  "history_storage_url": "",
                  "java_gateway": {
                    "debug_level": "info",
                    "start_pollers": 5,
                    "timeout": 3
                  },
                  "java_gateway_size": 1,
                  "mysql_database": "zabbix",
                  "server": {
                    "allow_unsupported_db_versions": false,
                    "cache_size": "32M",
                    "cache_update_frequency": 60,
                    "db_tls_cipher": "",
                    "db_tls_cipher_13": "",
                    "db_tls_connect": "",
                    "debug_level": 3,
                    "export_file_size": "",
                    "history_cache_size": "16M",
                    "history_index_cache_size": "4M",
                    "history_storage_date_index": true,
                    "history_storage_types": "",
                    "housekeeping_frequency": 1,
                    "load_module": "",
                    "log_slow_queries": 0,
                    "max_housekeeper_delete": 5000,
                    "proxy_config_frequency": 3600,
                    "proxy_data_frequency": 1,
                    "service_manager_sync_frequency": 60,
                    "start_alerters": 3,
                    "start_db_syncers": 4,
                    "start_discoverers": 1,
                    "start_escalators": 1,
                    "start_history_pollers": 5,
                    "start_http_pollers": 1,
                    "start_ipmi_pollers": 0,
                    "start_java_pollers": 5,
                    "start_lld_processors": 2,
                    "start_odbc_pollers": 1,
                    "start_pingers": 1,
                    "start_pollers": 5,
                    "start_pollers_unreachable": 1,
                    "start_preprocessors": 3,
                    "start_proxy_pollers": 0,
                    "start_report_writers": 1,
                    "start_timers": 1,
                    "start_trappers": 5,
                    "start_vmware_collectors": 0,
                    "stats_allowed_ip": "",
                    "timeout": 4,
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_cipher_all": "",
                    "tls_cipher_all_13": "",
                    "tls_cipher_cert": "",
                    "tls_cipher_cert_13": "",
                    "tls_cipher_psk": "",
                    "tls_cipher_psk_13": "",
                    "tls_crl_file_name": "",
                    "tls_key_file_name": "",
                    "trapper_timeout": 300,
                    "trend_cache_size": "4M",
                    "trend_function_cache_size": "4M",
                    "unavailable_delay": 60,
                    "unreachable_delay": 15,
                    "unreachable_period": 45,
                    "value_cache_size": "8M",
                    "vault_db_path": "",
                    "vmware_cache_size": "8M",
                    "vmware_frequency": 60,
                    "vmware_perf_frequency": 60,
                    "vmware_timeout": 10,
                    "web_service_url": ""
                  },
                  "snmp_traps_volumeclaim": "",
                  "snmptraps": {
                    "snmp_trap_date_format": "+%Y%m%d.%H%M%S",
                    "snmp_trap_format": "\\n"
                  },
                  "vault_token": "",
                  "vault_url": "",
                  "web": {
                    "db_cipher_list": "",
                    "db_double_ieee754": true,
                    "db_encryption": false,
                    "db_verify_host": false,
                    "deny_gui_access": false,
                    "enable_web_access_log": true,
                    "gui_access_ip_range": "",
                    "gui_warning_msg": "Zabbix is under maintenance.",
                    "history_storage_types": "",
                    "max_execution_time": 300,
                    "max_input_time": 300,
                    "memory_limit": "128M",
                    "post_max_size": "16M",
                    "server_name": "Kubernetes installation",
                    "session_name": "zbx_sessionid",
                    "sso_settings": "",
                    "timezone": "Europe/Riga",
                    "upload_max_filesize": "2M",
                    "vault_db_path": ""
                  },
                  "web_enable_route": true,
                  "web_service": {
                    "allowed_ip": "0.0.0.0/0",
                    "debug_level": 3,
                    "ignore_url_cert_errors": false,
                    "timeout": 3,
                    "tls_accept": "unencrypted",
                    "tls_ca_file_name": "",
                    "tls_cert_file_name": "",
                    "tls_key_file_name": ""
                  },
                  "web_size": 2,
                  "zabbix_mysqlsecret": "zabbix-full-secrets"
                }
              }
            ]
          ansible.operator-sdk/reconcile-period: 5s
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          certified: "true"
          containerImage: registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
          createdAt: "2025-05-26T07:41:22Z"
          description: Zabbix operator with multiple deployment variants and different
            components
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/zabbix-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <6.0.40
          operatorframework.io/suggested-namespace: zabbix
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: ""
          support: Zabbix
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Zabbix agent 2 is a new generation of Zabbix agent and may
              be used in place of Zabbix agent
            displayName: Zabbix agent 2
            kind: ZabbixAgent2
            name: zabbixagent2s.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix agent is deployed on a monitoring nodes to actively
              monitor local resources and applications
            displayName: Zabbix agent
            kind: ZabbixAgent
            name: zabbixagents.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database server, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Full
            kind: ZabbixFull
            name: zabbixfulls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with MySQL database server
            displayName: Zabbix proxy (MySQL)
            kind: ZabbixProxyMysql
            name: zabbixproxymysqls.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix proxy with SQLite database
            displayName: Zabbix proxy (SQLite3)
            kind: ZabbixProxySqlite
            name: zabbixproxysqlites.kubernetes.zabbix.com
            version: v1alpha1
          - description: Zabbix server with MySQL database support, Nginx web-server
              and Zabbix Java Gateway
            displayName: Zabbix Server
            kind: ZabbixServer
            name: zabbixservers.kubernetes.zabbix.com
            version: v1alpha1
        description: "## About this Operator\n\nZabbix helps you to real-time monitoring
          of millions of metrics collected from tens of thousands of servers, virtual
          machines and network devices.\nThe Zabbix Operator allows users to easily
          deploy, manage, and maintain Zabbix deployments on OpenShift. By installing
          this integration you will be able to deploy Zabbix server / proxies and
          other components with a single command.\n## Supported Features\n* **Zabbix
          Server** - Simple Zabbix installation with included Zabbix server, Zabbix
          web-interface and Zabbix Java Gateway with MySQL database support. The feature
          does not provide MySQL service and requires an external MySQL database.
          \n* **Zabbix Server (Full)** - Zabbix installation with included Zabbix
          server, Zabbix web-interface, Zabbix Java Gateway and MySQL server instance.\n*
          **Zabbix proxy (SQLite3)** - Very simple way to gain power of Zabbix proxy.
          The feature has  SQLite3 support for Zabbix proxies and allow to specify
          amount of proxies. \n* **Zabbix proxy (MySQL)** - Another option of Zabbix
          proxy. The option support and deliver MySQL database.\n* **Zabbix agent**
          - Zabbix agent can be deployed on each available node for stability and
          performance monitoring on remote nodes. It allows to gather metrics with
          full automation!* \n## Prerequisites\nAll deployment options are require
          additional information during deployment. Please, check the following instructions
          and provide required configuration:\n* **Zabbix Server** - MySQL database
          host information and MySQL database credentials in specially formatted *Secret*.
          Additionally it is possible to specify SSL certificates for HTTPS support
          in *Secret*.\n* **Zabbix Server (Full)** - MySQL database credentials in
          specially formatted *Secret*. MySQL database volume name information. Additionally
          it is possible to specify SSL certificates for HTTPS support in *Secret*.\n*
          **Zabbix proxy (SQLite3)** - Zabbix server host information only.\n* **Zabbix
          proxy (MySQL)** - MySQL database credentials in specially formatted *Secrets*
          and Zabbix server host information.\n* **Zabbix agent** - Zabbix server
          host information only for outgoing and incoming connections. Zabbix agent
          uses \"privileged\" mode to monitor node resources! For example, running
          processes.* \n### MySQL credentials\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-full-secrets\ndata:\n  mysql_root_password: emFiYml4X3N1cGVyX3Jvb3Q=
          [1]\n  mysql_zabbix_password: emFiYml4X3VzZXJfcGFzc3dvcmQ= [2]\n  mysql_zabbix_username:
          emFiYml4 [3]\ntype: Opaque\n```\nUsing MySQL root password (*mysqlrootpassword*
          [1]) Zabbix server / proxy will try to create MySQL database schema with
          grant permissions to *mysqlzabbixusername* [2] and *mysqlzabbixpassword*
          [3].\n### SSL certificates for HTTPS\n```\nkind: Secret\napiVersion: v1\nmetadata:\n
          \ name: zabbix-web-sslsecret\ndata:\n  ssl.crt: >-\n   < ssl.crt data>\n
          \ ssl.key: >-\n\t < ssl.key data >\n  dhparam.pem: >-\n   <  dhparam.pem
          data >\n```\nFiles *ssl.crt*, *ssl.key* and *dhparam.perm* are required
          for Zabbix web-interface for SSL support.\n"
        displayName: Zabbix Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - zabbix
        - appliance
        - agent
        - metrics
        links:
        - name: Zabbix
          url: https://www.zabbix.com
        - name: Zabbix Official Documentation
          url: https://www.zabbix.com/documentation/6.0/en/manual/quickstart
        - name: Zabbix Operator Installation
          url: https://www.zabbix.com/documentation/6.0/en/manual/installation/containers/openshift
        - name: Downloads
          url: https://www.zabbix.com/download
        maintainers:
        - email: alexey.pustovalov@zabbix.com
          name: Alexey Pustovalov
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Zabbix LLC
        relatedImages:
        - registry.connect.redhat.com/zabbix/zabbix-proxy-mysql-60@sha256:c2ae74d97c083ff348dc9c55932a33da4779cd9efdd9d7c5a79acfffae5b9ed0
        - registry.connect.redhat.com/zabbix/zabbix-server-mysql-60@sha256:637484faec6e9a55049c13dbd1f100c03d757c3a21749388c0605040f02d05ce
        - registry.connect.redhat.com/zabbix/zabbix-web-mysql-60@sha256:611d2670e1bf95b81226976c0a1f64434b2a7766f196dadbd646a7d8319f4cb1
        - registry.connect.redhat.com/zabbix/zabbix-agent2-60@sha256:424003d52f3cb4360adf2b3619d8b9f8de9196c630ee7e4ceb5d177c1c5a36e4
        - registry.connect.redhat.com/zabbix/zabbix-web-service-60@sha256:4a333094e110713470a897b6406c5e340e962caf32a429dddf19865041fcc448
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified-bundle-marketplace@sha256:4145f7a04d7dbdda226b75af4e4cabca63a803f290b1581ea4ae0dc223eca7df
        - registry.redhat.io/rhel9/mysql-80@sha256:c4b5246d0c0d089a766ea07768c22d9d6be8049ac1379bd345db99390ffe2a36
        - quay.io/zabbix/zabbix-operator@sha256:43d26488a19967bdb2a8aafde5f43820be95011811da9cc5ec623c9029801675
        - registry.connect.redhat.com/zabbix/zabbix-java-gateway-60@sha256:a4feb725c12c4ac90eb9244c6fafdb6708679f2e0e7f6aa683c765127306412e
        - registry.connect.redhat.com/zabbix/zabbix-proxy-sqlite-60@sha256:62e62ebc8c1bc13fb610f7a83412e65d852f74b3c2c212baa75ea27fc7fa03eb
        - registry.connect.redhat.com/zabbix/zabbixoperator-certified@sha256:e548b555e656012082ce188383990e9d65eaca484329695c7547671bb3c05d28
        - registry.connect.redhat.com/zabbix/zabbix-agent-60@sha256:2895b5e45785aaf528d4a6f78c8af0a0d494dbe9362403fdd470673d911d5004
        - registry.connect.redhat.com/zabbix/zabbix-snmptraps-60@sha256:e61e6b1c551dec8488cce0998f94d83d6d10b5d89a465d7119b088313032cb81
        version: 6.0.40
      entries:
      - name: zabbix-operator-certified-rhmp.v6.0.40
        version: 6.0.40
      - name: zabbix-operator-certified-rhmp.v6.0.38
        version: 6.0.38
      - name: zabbix-operator-certified-rhmp.v6.0.36
        version: 6.0.36
      - name: zabbix-operator-certified-rhmp.v6.0.33
        version: 6.0.33
      - name: zabbix-operator-certified-rhmp.v6.0.29
        version: 6.0.29
      - name: zabbix-operator-certified-rhmp.v6.0.27
        version: 6.0.27
      - name: zabbix-operator-certified-rhmp.v6.0.26
        version: 6.0.26
      - name: zabbix-operator-certified-rhmp.v6.0.25
        version: 6.0.25
      - name: zabbix-operator-certified-rhmp.v6.0.17
        version: 6.0.17
      - name: zabbix-operator-certified-rhmp.v6.0.14
        version: 6.0.14
      - name: zabbix-operator-certified-rhmp.v6.0.13
        version: 6.0.13
      name: lts
    defaultChannel: lts
    packageName: zabbix-operator-certified-rhmp
    provider:
      name: Zabbix LLC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-appsec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-appsec.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Appsec",
                "metadata": {
                  "name": "appsec-sample"
                },
                "spec": {
                  "appsecDataModel": {
                    "port": 8080
                  },
                  "appsecResourceService": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "remoteRepository": "docker.io",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:39:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Appsec
            name: appsecs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP APPSEC Operator
        displayName: Keysight WAP APPSEC Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Appsec
          url: https://keysight-wap-appsec.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-appsec@sha256:e39ff225c26b750f86553693dda47d0dcc9834a4ab0975aea154361ee3527314
        - registry.connect.redhat.com/loadcore/keysight-wap-appsec-bundle@sha256:e5049593a52fb26c124de7005b45919f26fa0430293f18283a20c8cde981b025
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-appsec.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-appsec
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MuleSoft
      provider-url: anypoint.mulesoft.com
    name: runtime-fabric-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: alpha
    - currentCSV: runtime-fabric-operator.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rtf.operators.mulesoft.com/v1",
                "kind": "RuntimeFabric",
                "metadata": {
                  "name": "runtime-fabric"
                },
                "spec": {
                  "activationData": null,
                  "agent": {
                    "replicas": 1,
                    "app": {
                      "resources": {
                        "limits": {
                          "cpu": "1000m",
                          "memory": "500Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      }
                    },
                    "rtfd": {
                      "resources": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "150m",
                          "memory": "200Mi"
                        }
                      },
                      "watch": {
                        "enabled": true,
                        "components": {
                          "deployments": {
                            "enabled": true
                          },
                          "ingress": {
                            "enabled": true
                          },
                          "secrets": {
                            "enabled": true
                          },
                          "persistencegateways": {
                            "enabled": true
                          },
                          "nodes": {
                            "enabled": true
                          },
                          "jobs": {
                            "enabled": true
                          },
                          "routes": {
                            "enabled": true
                          }
                        }
                      }
                    }
                  },
                  "resource-cache": {
                    "replicas": 1,
                    "nginx": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "250Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    },
                    "fetcher": {
                      "resources": {
                        "limits": {
                          "cpu": "50m",
                          "memory": "100Mi"
                        },
                        "requests": {
                          "cpu": "50m",
                          "memory": "50Mi"
                        }
                      }
                    }
                  },
                  "mule-clusterip-service": {
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "global": {
                    "cluster": {
                      "flavor": "byok",
                      "type": "openshift"
                    },
                    "containerLogPaths": [
                      "/var/lib/docker/containers",
                      "/var/log/containers",
                      "/var/log/pods"
                    ],
                    "core": {
                      "affinity": {},
                      "tolerations": [],
                      "topologySpreadConstraints": [
                        {
                          "maxSkew": 1,
                          "topologyKey": "topology.kubernetes.io/zone",
                          "whenUnsatisfiable": "ScheduleAnyway"
                        }
                      ]
                    },
                    "worker": {
                      "affinity": {},
                      "tolerations": []
                    },
                    "deployment": {
                      "nodeSelector": ""
                    },
                    "image": {
                      "provider": "aws",
                      "pullSecretName": "rtf-pull-secret",
                      "rtfRegistry": "rtf-runtime-registry.kprod.msap.io",
                      "repositoryNamespace": "mulesoft"
                    },
                    "service": {
                      "clusterIp": null
                    },
                    "authorizedNamespaces": false,
                    "fipsEnabled": false,
                    "crds": {
                      "install": true,
                      "keep": true,
                      "flex": {
                        "enabled": false
                      }
                    },
                    "nodeWatcherEnabled": true,
                    "nodeReconciliation": true,
                    "deploymentRateLimitPerSecond": 1
                  },
                  "muleLicense": null,
                  "proxy": {
                    "http_no_proxy": null,
                    "http_proxy": null,
                    "monitoring_proxy": null
                  },
                  "customLog4jEnabled": false
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
          createdAt: "2025-07-31T21:12:41Z"
          description: Runtime Fabric Operator Application from Mulesoft
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mulesoft/rtf-ocp-operator
          support: MuleSoft
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Runtime Fabric Operator from Mulesoft
              **Activation data**
              Used to install Runtime Fabric on a Kubernetes service. Copy this value from the Runtime Fabric setup page in Runtime Manager.
              **Registry URL**
              The registry from which to retrieve the images needed to install and run Runtime Fabric. If you're using a local registry, specify the URL here.
                * Default registry URLs:
                    - US control plane: rtf-runtime-registry.kprod.msap.io
                    - EU control plane: rtf-runtime-registry.kprod-eu.msap.io
              **Registry pull secret**
              Used to retrieve images from the registry. Change this value only if you did so during the operator installation.
              **Mule license key**
              Your company’s Mule license key. The key is not required during install, but it is required to deploy applications to Runtime Fabric. Must be Base64 encoded.
            displayName: Runtime Fabric
            kind: RuntimeFabric
            name: runtimefabrics.rtf.operators.mulesoft.com
            version: v1
        description: |-
          Anypoint Runtime Fabric is a container service that you install in your existing Red Hat OpenShift cluster that enables you to run Mule applications, composite APIs, and API gateways across any environment with centralized management via Anypoint Platform.
          Capabilities of Runtime Fabric include:
          * Isolation between applications by running a separate Mule runtime server per application.
          * Ability to run multiple versions of Mule runtime server on the same set of resources.
          * Scaling applications across multiple replicas.
          * Automated application fail-over.
          * Application management with Anypoint Runtime Manager.
          Runtime Fabric supports the following Red Hat OpenShift deployment options:
          * Red Hat OpenShift Service on AWS
          * Microsoft Azure Red Hat OpenShift
          * Red Hat OpenShift Dedicated
          * Red Hat OpenShift on IBM Cloud
          * Self-managed Red Hat OpenShift editions (Performance Plus, OCP, Kubernetes engine)
          Runtime Fabric supports Red Hat OpenShift versions 4.16 to 4.19.
        displayName: Runtime Fabric Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - agent
        - rtf
        - runtime
        - fabric
        - mulesoft
        links:
        - name: Runtime Fabric Operator
          url: https://catalog.redhat.com/software/container-stacks/detail/62901512fefb1b137ba23946
        - name: Documentation
          url: https://docs.mulesoft.com/release-notes/runtime-fabric/runtime-fabric
        maintainers:
        - email: support@mulesoft.com
          name: Mulesoft
        maturity: stable
        provider:
          name: MuleSoft
          url: anypoint.mulesoft.com
        relatedImages:
        - quay.io/rtf/runtime-fabric-operator@sha256:e8a4a6053fe36d111270ce281c5da456b161ad82f95c94cad035e9b932d9a2e5
        - registry.connect.redhat.com/rtf/runtime-fabric-operator@sha256:7a11451b49caafbcc3b0312ac91232529ee3689c139badcd050e349e87409efe
        - quay.io/rtf/kube-rbac-proxy@sha256:a2523c532c0c3d51a5396a901d7ded23e402a9a1492c783aae27af6d0c1d2ec5
        version: 2.11.0
      entries:
      - name: runtime-fabric-operator.v2.11.0
        version: 2.11.0
      - name: runtime-fabric-operator.v2.10.34
        version: 2.10.34
      - name: runtime-fabric-operator.v2.10.17
        version: 2.10.17
      - name: runtime-fabric-operator.v2.10.0
        version: 2.10.0
      - name: runtime-fabric-operator.v2.9.2
        version: 2.9.2
      name: stable
    defaultChannel: stable
    packageName: runtime-fabric-operator
    provider:
      name: MuleSoft
      url: anypoint.mulesoft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: vertical-pod-autoscaler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vertical-pod-autoscaler.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "autoscaling.k8s.io/v1",
                "kind": "VerticalPodAutoscaler",
                "metadata": {
                  "name": "myapp-vpa"
                },
                "spec": {
                  "resourcePolicy": {
                    "containerPolicies": [
                      {
                        "containerName": "*",
                        "controlledResources": [
                          "cpu",
                          "memory"
                        ],
                        "maxAllowed": {
                          "cpu": 1,
                          "memory": "500Mi"
                        },
                        "minAllowed": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      }
                    ]
                  },
                  "targetRef": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "myapp-deployment"
                  }
                }
              },
              {
                "apiVersion": "autoscaling.openshift.io/v1",
                "kind": "VerticalPodAutoscalerController",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "minReplicas": 2,
                  "podMinCPUMillicores": 25,
                  "podMinMemoryMb": 250,
                  "safetyMarginFraction": 0.15
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
          createdAt: "2024-12-05T23:41:17Z"
          description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
            Pod Autoscaler (VPA) can be configured to monitor a workload's resource
            utilization, and then adjust its CPU and memory limits by updating the
            pod (future) or restarting the pod with updated limits.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.5.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-vertical-pod-autoscaler
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/vertical-pod-autoscaler-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents VPA checkpoints used by the VPA controllers to
              manage workloads
            displayName: VPA Checkpoint
            kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscalerCheckpoint
            name: verticalpodautoscalercheckpoints.autoscaling.k8s.io
            version: v1beta2
          - description: Represents an instance of the set of VPA controllers
            displayName: VPA Controller
            kind: VerticalPodAutoscalerController
            name: verticalpodautoscalercontrollers.autoscaling.openshift.io
            version: v1
          - description: An application whose resources are managed by the VPA
            displayName: VerticalPodAutoscaler
            kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1
          - kind: VerticalPodAutoscaler
            name: verticalpodautoscalers.autoscaling.k8s.io
            version: v1beta2
        description: An operator to run the OpenShift Vertical Pod Autoscaler. Vertical
          Pod Autoscaler (VPA) can be configured to monitor a workload's resource
          utilization, and then adjust its CPU and memory limits by updating the pod
          (future) or restarting the pod with updated limits.
        displayName: Vertical Pod Autoscaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - verticalpodautoscaler
        - vpa
        - autoscaling
        - scaling
        links:
        - name: Vertical Pod Autoscaler Documentation
          url: https://docs.openshift.com/container-platform/latest/nodes/pods/nodes-pods-vertical-autoscaler.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-operator-bundle@sha256:f5df5d878ef0a0b494c9bafe99c35f3152d30f6c426be789fcd636dd13eff6e6
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9-operator@sha256:689b5155cba91852f7b9f3acd85f906ba62ab7be549d738dd7eee06bfd10b966
        - registry.redhat.io/openshift4/ose-vertical-pod-autoscaler-rhel9@sha256:5da084c4dc81251a707db1bd9347c13b39a1e01660120aa2e6a47f9d1d981fe9
        version: 4.18.0-202509100344
      entries:
      - name: vertical-pod-autoscaler.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: vertical-pod-autoscaler.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: vertical-pod-autoscaler.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: vertical-pod-autoscaler.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: vertical-pod-autoscaler.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: vertical-pod-autoscaler.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: vertical-pod-autoscaler.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: vertical-pod-autoscaler.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: vertical-pod-autoscaler.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: vertical-pod-autoscaler.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: vertical-pod-autoscaler.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: vertical-pod-autoscaler.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: vertical-pod-autoscaler.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: vertical-pod-autoscaler.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: vertical-pod-autoscaler.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: vertical-pod-autoscaler
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: enmasse
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-online.1.7.4-0.1657725339.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "StandardInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "1Gi",
                      "storage": "5Gi"
                    },
                    "addressFullPolicy": "FAIL"
                  },
                  "router": {
                    "linkCapacity": 50,
                    "resources": {
                      "memory": "512Mi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "BrokeredInfraConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "broker": {
                    "resources": {
                      "memory": "4Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressPlan",
                "metadata": {
                  "name": "standard-small-queue"
                },
                "spec": {
                  "addressType": "queue",
                  "shortDescription": "Small Queue",
                  "resources": {
                    "router": 0.01,
                    "broker": 0.1
                  }
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta2",
                "kind": "AddressSpacePlan",
                "metadata": {
                  "name": "standard-small"
                },
                "spec": {
                  "addressSpaceType": "standard",
                  "infraConfigRef": "default",
                  "shortDescription": "Small Address Space Plan",
                  "resourceLimits": {
                    "router": 1.0,
                    "broker": 2.0,
                    "aggregate": 2.0
                  },
                  "addressPlans": [
                    "standard-small-queue"
                  ]
                }
              },
              {
                "apiVersion": "admin.enmasse.io/v1beta1",
                "kind": "AuthenticationService",
                "metadata": {
                  "name": "standard-authservice"
                },
                "spec": {
                  "type": "standard",
                  "standard": {
                    "storage": {
                      "claimName": "standard-authservice",
                      "deleteClaim": true,
                      "size": "1Gi",
                      "type": "persistent-claim"
                    }
                  }
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "AddressSpace",
                "metadata": {
                  "name": "myspace"
                },
                "spec": {
                  "plan": "standard-small",
                  "type": "standard"
                }
              },
              {
                "apiVersion": "enmasse.io/v1beta1",
                "kind": "Address",
                "metadata": {
                  "name": "myspace.myqueue"
                },
                "spec": {
                  "address": "myqueue",
                  "plan": "standard-small-queue",
                  "type": "queue"
                }
              },
              {
                "apiVersion": "user.enmasse.io/v1beta1",
                "kind": "MessagingUser",
                "metadata": {
                  "name": "myspace.user"
                },
                "spec": {
                  "authentication": {
                    "password": "ZW5tYXNzZQ==",
                    "type": "password"
                  },
                  "authorization": [
                    {
                      "addresses": [
                        "myqueue"
                      ],
                      "operations": [
                        "send",
                        "recv"
                      ]
                    }
                  ],
                  "username": "user"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
          createdAt: "2019-02-19T00:00:00Z"
          description: Red Hat Integration - AMQ Online provides messaging as a managed
            service on OpenShift
          olm.substitutesFor: amq-online.1.7.4
          operators.openshift.io/capability: '["license-required"]'
          operators.openshift.io/valid-licenses: '["Red Hat AMQ", "Red Hat Integration"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ"]'
          repository: ""
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A messaging user that can connect to an Address Space
            displayName: Messaging User
            kind: MessagingUser
            name: messagingusers.user.enmasse.io
            version: v1beta1
          - description: A resource representing the available schema of plans and
              authentication services
            displayName: AddressSpaceSchema
            kind: AddressSpaceSchema
            name: addressspaceschemas.enmasse.io
            version: v1beta1
          - description: A group of messaging addresses that can be accessed via the
              same endpoint
            displayName: Address Space
            kind: AddressSpace
            name: addressspaces.enmasse.io
            version: v1beta1
          - description: A messaging address that can be used to send/receive messages
              to/from
            displayName: Address
            kind: Address
            name: addresses.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the standard address
              space type
            displayName: Standard Infra Config
            kind: StandardInfraConfig
            name: standardinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Infrastructure configuration template for the brokered address
              space type
            displayName: Brokered Infra Config
            kind: BrokeredInfraConfig
            name: brokeredinfraconfigs.admin.enmasse.io
            version: v1beta1
          - description: Plan describing the resource usage of a given address type
            displayName: Address Plan
            kind: AddressPlan
            name: addressplans.admin.enmasse.io
            version: v1beta2
          - description: Plan describing the capabilities and resource limits of a
              given address space type
            displayName: Address Space Plan
            kind: AddressSpacePlan
            name: addressspaceplans.admin.enmasse.io
            version: v1beta2
          - description: Authentication service configuration available to address
              spaces.
            displayName: Authentication Service
            kind: AuthenticationService
            name: authenticationservices.admin.enmasse.io
            version: v1beta1
          - description: Console Service Singleton for deploying global console.
            displayName: Console Service
            kind: ConsoleService
            name: consoleservices.admin.enmasse.io
            version: v1beta1
        description: "**Red Hat Integration - AMQ Online** provides messaging as a
          managed service on OpenShift.\nWith Red Hat Integration - AMQ Online, administrators
          can configure a cloud-native,\nmulti-tenant messaging service either in
          the cloud or on premise.\nDevelopers can provision messaging using the Red
          Hat Integration - AMQ Online Console.\nMultiple development teams can provision
          the brokers and queues from the\nconsole, without requiring each team to
          install, configure, deploy,\nmaintain, or patch any software. \n\n**The
          core capabilities include:**\n\n  * **Built-in authentication and authorization**
          - Use the built-in authentication service or\n    plug in your own authentication
          service for authentication and\n    authorization of messaging clients.\n\n
          \ * **Self-service messaging for applications** - The service administrator
          deploys\n    and manages the messaging infrastructure, while applications
          can request\n    messaging resources regardless of the messaging infrastructure.\n\n
          \ * **Support for a wide variety of messaging patterns** - Choose between\n
          \   JMS-style messaging with strict guarantees, or messaging that supports\n
          \   a larger number of connections and higher throughput.\n\n## Post-installation
          tasks\n\nTo fully use Red Hat Integration - AMQ Online, you need to create
          a few\ninfrastructure components after the installation, including:\n\n
          \ * An authentication service\n  * Infrastructure configuration for the
          standard and brokered address space\n  * Address and address space plans\n
          \ * (Optional) Create RBAC roles to allow users to discover available plans\n
          \ * (Optional) Create RBAC roles to allow users to self-manage addresses
          and\n    address spaces.\n\nFor a complete overview of how to configure
          Red Hat Integration - AMQ Online, see\n[Configuring Red Hat Integration
          - AMQ Online](https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/installing_and_managing_amq_online_on_openshift/configuring-messaging).\n\n###
          Quickstart infrastructure configuration\n\nIf you simply want to get started
          quickly, you can import the following\nYAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          StandardInfraConfig\nmetadata:\n  name: default\n---\napiVersion: admin.enmasse.io/v1beta2\nkind:
          AddressPlan\nmetadata:\n  name: standard-small-queue\nspec:\n  addressType:
          queue\n  resources:\n    router: 0.01\n    broker: 0.1\n---\napiVersion:
          admin.enmasse.io/v1beta2\nkind: AddressSpacePlan\nmetadata:\n  name: standard-small\nspec:\n
          \ addressSpaceType: standard\n  infraConfigRef: default\n  addressPlans:\n
          \ - standard-small-queue\n  resourceLimits:\n    router: 2.0\n    broker:
          3.0\n    aggregate: 4.0\n---\napiVersion: admin.enmasse.io/v1beta1\nkind:
          AuthenticationService\nmetadata:\n  name: standard-authservice\nspec:\n
          \ type: standard\n  standard:\n    storage:\n      claimName: standard-authservice\n
          \     deleteClaim: true\n      size: 1Gi\n      type: persistent-claim\n~~~\n\n###
          Create RBAC roles to allow users to discover available plans\n\nFor users
          to discover the available plans, cluster-wide roles to read the available\nschema
          can be created.  Import the following YAML by saving the content to a file
          and apply it by running the \n`oc apply -f <file>` command. You can also
          split the content (at the `---` marker)\nand import the single YAML document
          using the Web UI: \n\n~~~yaml\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRole\nmetadata:\n  name: enmasse.io:schema\n  labels:\n    app: enmasse\nrules:\n
          \ - apiGroups: [ \"enmasse.io\" ]\n    resources: [ \"addressspaceschemas\"
          ]\n    verbs: [ \"get\", \"list\", \"watch\" ]\n---\napiVersion: rbac.authorization.k8s.io/v1\nkind:
          ClusterRoleBinding\nmetadata:\n  name: \"enmasse.io:schema\"\n  labels:\n
          \   app: enmasse\nroleRef:\n  apiGroup: rbac.authorization.k8s.io\n  kind:
          ClusterRole\n  name: enmasse.io:schema\nsubjects:\n- kind: Group\n  name:
          system:authenticated\n~~~\n\n### Creating infrastructure using the Web UI\n\nYou
          must create a new instance of each of the following custom resources. You
          can\nuse the example data directly, which is provided when using the\nWeb
          UI:\n\n  * *Authentication Service* – Create an authentication service.\n
          \ * *Brokered Infra Config* – Create the broker infrastructure configuration.\n
          \ * *Standard Infra Config* – Create the standard infrastructure\n    configuration.\n\nYou
          must also create at least one address space plan and one address plan.\n\n*Note*:
          The name of the address space plan and address plan is required\nlater when
          creating new addresses. Some examples use specific plan\nnames, which might
          not be available in your environment. You can\ncreate those plans, or edit
          the examples to use different plan names.\n\n## Configuration for messaging
          tenants\n\nWhile service administrators perform the infrastructure configuration,
          the following\nresources are for the actual users of the system, the messaging
          tenants.\n\nYou need to create those resources to satisfy your particular
          use case.\n\n  * *Address space* – A container for addresses\n  * *Address*
          – A messaging address (address, topic, queue, and so on)\n  * *Messaging
          user* – Manages access to an address\n"
        displayName: Red Hat Integration - AMQ Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amq
        - messaging
        - amqp
        - iot
        - mqtt
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq/7.7/html/using_amq_online_on_openshift/index
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-broker@sha256:1ec0437bbd82722148fdbb02dc55682b114c311cd6f521a8bc621a8c1cb4b32b
        - registry.redhat.io/redhat-sso-7/sso73-openshift@sha256:18d15980f28f0d0c6b8582131a1f062d84463577512875b5391a612c3d7d1096
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6d489a7897bd831ec273034f1ff0e223427a40afd01b9835ebb020a8bf9ddc66
        - registry.redhat.io/amq7/amq-online-1-agent@sha256:f98ed8ee97ac59d8fdd40ec8b2c063a5afd360a3e183057c186e7a77a32801ee
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator-metadata@sha256:d6d0e71f5451e8255edfff87ef28ba7743fe7057efd53b25bbbd01bf9e1d0307
        - registry.redhat.io/amq7/amq-online-1-address-space-controller@sha256:9acc555c8e869aec7d72f53b60320e1a96cc82d748638e0d50970837229fdbee
        - registry.redhat.io/amq7/amq-online-1-auth-plugin@sha256:68926f434b2f6ce4d9698938bc3c53d514c21c3f76c73b6b4cdf7fcf51f7cb0d
        - registry.redhat.io/amq7/amq-online-1-console-server-rhel7@sha256:1dcb3d0cc7bc6e84f4f96962ff88849511b7a7686a1f5dbd9300c4c2c8c05e2f
        - registry.redhat.io/amq7/amq-online-1-topic-forwarder@sha256:de33a701339d91a4e9275f7cf65f917ff75368a59b8ab36dd66c90bc306875aa
        - registry.redhat.io/amq7/amq-online-1-controller-manager-rhel7-operator@sha256:2aaf87224a96aa56dc595e96d8e7ec8d138aaf20240afa25a1e805b405941355
        - registry.redhat.io/amq7/amq-online-1-console-init@sha256:0970804909b82fdc3d4e7d5d88280e41157d743107de8bfa2d1baa1367543176
        - registry.redhat.io/amq7/amq-interconnect@sha256:5ebd4c4a0ed0db2725ae466550c915531b32933c420344cc2083c616da3f326f
        - registry.redhat.io/openshift3/oauth-proxy@sha256:59b20f23c03d0012ec5c95c4725892ee64277c22dedc9d0a90a7b6b0bc80fc72
        - registry.redhat.io/amq7/amq-online-1-none-auth-service@sha256:a013fb1806c002b408a7753ff943cee24fbcdf8a657eff1e9582c42b5a7b0844
        - registry.redhat.io/amq7/amq-online-1-standard-controller@sha256:070349af30cec52fa8efc185fc940077a64bb40e3e36310d7cc6e39ce0971436
        - registry.redhat.io/amq7/amq-online-1-broker-plugin@sha256:2466a336c0d78b196635de3fad636aeff2d9265839f7271b8ef94f3ba97d9bc0
        version: 1.7.4+0.1657725339.p
      entries:
      - name: amq-online.1.7.4
        version: 1.7.4
      - name: amq-online.1.7.4-0.1657725339.p
        version: 1.7.4+0.1657725339.p
      - name: amq-online.1.7.3
        version: 1.7.3
      - name: amq-online.1.7.2
        version: 1.7.2
      - name: amq-online.1.7.1
        version: 1.7.1
      - name: amq-online.1.7.1-0.1628610187.p
        version: 1.7.1+0.1628610187.p
      - name: amq-online.1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: amq-online
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/lws-operator
    name: leader-worker-set
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: leader-worker-set.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "LeaderWorkerSetOperator",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed",
                  "operatorLogLevel": "Normal"
                }
              }
            ]
          capabilities: Basic Install
          categories: Openshift Optional
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-04-14T08:12:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lws-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          repository: https://github.com/openshift/lws-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: LeaderWorkerSetOperator
            name: leaderworkersetoperators.operator.openshift.io
            version: v1
        description: "Leader Worker Set Operator is based on the [LWS](https://lws.sigs.k8s.io/docs/)
          open source project.\nLeader Worker Set: An API for deploying a group of
          pods as a unit of replication. \nIt aims to address common deployment patterns
          of AI/ML inference workloads, \nespecially multi-host inference workloads
          where the LLM will be sharded and run across multiple devices on multiple
          nodes.\n"
        displayName: Leader Worker Set Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - leader-worker-set
        links:
        - name: Leader Worker Set
          url: https://lws.sigs.k8s.io/docs/overview/
        - name: Operand Source Code
          url: https://github.com/openshift/kubernetes-sigs-lws
        - name: Operator Source Code
          url: https://github.com/openshift/lws-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/lws-operator
        relatedImages:
        - registry.redhat.io/leader-worker-set/lws-rhel9@sha256:affb303b1173c273231bb50fef07310b0e220d2f08bfc0aa5912d0825e3e0d4f
        - registry.redhat.io/leader-worker-set/lws-operator-bundle@sha256:f641cf3b2d6659e26eb0ae49c67a3416152a2fbc2f26f490889a2bd169b9ea06
        - registry.redhat.io/leader-worker-set/lws-rhel9-operator@sha256:c202bfa15626262ff22682b64ac57539d28dd35f5960c490f5afea75cef34309
        version: 1.0.0
      entries:
      - name: leader-worker-set.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: leader-worker-set
    provider:
      name: Red Hat
      url: https://github.com/openshift/lws-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzo-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v1",
                "kind": "Anzo",
                "metadata": {
                  "name": "agent01"
                },
                "spec": {
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-agent01",
                      "template": {
                        "metadata": {
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                },
                                "requests": {
                                  "cpu": "4000m",
                                  "memory": "12Gi"
                                }
                              }
                            }
                          ],
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoAgent"
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
          createdAt: "2021-08-02"
          description: kubernetes operator for Anzo
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzo-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Anzo is the Schema for the anzos API
            displayName: Anzo Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Operator provides the way to install and configure an anzo agent setup on Red Hat K8S environment.
          Currently, this is supported only through when deployed as an Anzo Agent as part of an Anzo Unstructured deployment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Anzo Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v2.0.0/README_openshift.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:72a54c9457e311001b25c1b414b7001c6d7f9a34b89efb1b4198084586386f31
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-operator@sha256:563e964f5771b9bfd860ea993b6e0936a54504908f5b2d1dff6f99ade74bb7a0
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo@sha256:d1c2d54ca205d669c5ea1bb6aae105afe2385cc5695cecd78423ffa1d98929f8
        version: 2.0.0
      entries:
      - name: anzo-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzo-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: QuestLabs
      provider-url: https://www.questlabs.in
    name: alloy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: alloy.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.alloys.com/v1alpha1",
                "kind": "Alloy",
                "metadata": {
                  "name": "alloy"
                },
                "spec": {
                  "acmeEmail": "",
                  "dnsFQDN": "",
                  "sslTermination": "true"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Alloy
            name: alloys.charts.alloys.com
            version: v1alpha1
        description: |-
          Alloy is platform to build trusted data in OSDU, available on-premise or in the cloud. It is equipped with built-in data quality rules derived from PPDM and OSDU standards that control the ingestion of data.

              • Applicable for conventional and new energy streams
              • Highly Scalable

              Benefits
              • Helps to build robust applications/solutions on stable data foundation using full data lifecycle APIs & built-in data/business rules
              • Help data governance team to monitor/oversea data across organisation and intervene when required
              • Substantially low implementation time for any new data entity or new domain (solar, wind etc.) because of modular rules packages
              • Capability to deploy workflows for complex corporate specific data quality and enrichment requirements
              • Spend less time & resources to deploy fit-for-purpose data
              • Schema based checks out of the box for conventional and new energy streams.
              • Sustainable and future proof


              Prerequisites
              1. Please create alloy project before installing alloy operator

          Getting Help

          Once Alloy is successfully installed you can also browse documentation at https://your.domain.com/alloy/docs/introduction
          Please visit https://www.questlabs.in/alloy for more details and if you have any queries or suggestions, mail us at info@questlabs.in
        displayName: Alloy
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - alloy
        links:
        - name: alloy
          url: https://www.questlabs.in
        maintainers:
        - email: info@questlabs.in
          name: QuestLabs
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: QuestLabs
          url: https://www.questlabs.in
        relatedImages:
        - alloy.azurecr.io/alloy@sha256:bf29f69889d408507cf1a9a940beb80d95ba9faff1da2d70a4aea82cc235d96d
        - registry.connect.redhat.com/questlabs/alloy-operator@sha256:fe4e83fc4a50e1aa04e6a9ba2a57acc737077b86681c1d7c597646fad337f221
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d291e0d38bd95eb2731a10bd5ff06bfa07c6621528d3c2bbe888397d87b3feaa
        version: 1.0.4
      entries:
      - name: alloy.v1.0.4
        version: 1.0.4
      - name: alloy.v1.0.3
        version: 1.0.3
      - name: alloy.v1.0.2
        version: 1.0.2
      - name: alloy.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: alloy
    provider:
      name: QuestLabs
      url: https://www.questlabs.in
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kove
      provider-url: https://kove.com
    name: kove-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kove-operator.v3.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveMemoryConfig",
                "metadata": {
                  "labels": {
                    "KoveMemoryConfig": ""
                  },
                  "name": "example-memory-config"
                },
                "spec": {
                  "container": "",
                  "namespace": "",
                  "size": ""
                }
              },
              {
                "apiVersion": "containers.kove.com/v1",
                "kind": "KoveSystem",
                "metadata": {
                  "labels": {
                    "KoveSystem": ""
                  },
                  "name": "kove-system"
                },
                "spec": {
                  "autogrowInitialIncrement": "1 GiB",
                  "autogrowMaxIncrement": "1 TiB",
                  "l4MaxMemoryPercent": 75,
                  "managementConsoleAddress": "",
                  "memoryPlusLocalMemoryMaxSize": "0 MiB"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
          createdAt: "2025-09-25T16:31:36Z"
          description: Configure containers with Kove® pooled memory
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kove-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Kove
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configure containers with Kove® pooled memory
            displayName: Kove Memory Configuration
            kind: KoveMemoryConfig
            name: kovememoryconfigs.containers.kove.com
            version: v1
          - description: Configure cluster with Kove:SDM™
            displayName: Kove System Configuration
            kind: KoveSystem
            name: kovesystems.containers.kove.com
            version: v1
        description: |
          ## About the Operator
          The Kove:SDM™ operator integrates Kove's software-defined memory
          solution with Kubernetes and OpenShift clusters. The operator enables
          containers to transparently provision any amount of memory, enabling total
          container memory capacity to exceed physical hardware limitations.

          Use the Kove:SDM™ operator to:

          * Create custom resources to automatically allocate memory to specific
            containers
          * Increase container flexibility with dynamic memory capacity that
            auto-grows as needed
          * Scale up container density by removing memory as an orchestration
            bottleneck

          ## Using the Operator
          Using the operator is as simple as creating a custom resource which
          specifies an initial Kove memory allocation size, a namespace, and a
          container name. Containers in the specified namespace with a matching
          container name will automatically be provided with a Kove memory allocation
          which will auto-grow as needed.

          ## About Kove
          Kove makes memory a provisionable resource, just like storage, CPUs, and
          networking. Kove:SDM™ provides a transparent, linearly scalable
          memory resource accessible to any application without code changes, from
          bare-metal apps to virtual machines and containers.
        displayName: Kove:SDM™ Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kove
        - Pooled Memory
        - Memory Pool
        - Memory Target
        - External Memory
        - Memory
        links:
        - name: Kove
          url: https://kove.com
        maintainers:
        - email: contact-kove@kove.com
          name: Kove
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Kove
          url: https://kove.com
        relatedImages:
        - quay.io/kove/kove-client-drivers@sha256:1c1786e1bbf16609218154126b4340b2f7c4b71cbd9045589402f75c032ea38d
        - quay.io/kove/kove-client-drivers@sha256:716ccda2d8f082a954f2042fd189fec257bd2dcb2c50e497bad42b9a9e5f0e3d
        - quay.io/kove/kove-operator@sha256:0fb5fa2e6702d5e5bf692c25f190f65d306fc72cdceebf0814ff64adcb5f6994
        - registry.access.redhat.com/ubi8/pause@sha256:eacbaaadc932633330e799eda24e9b731b5a4ba7ed823ee70791d21029724a9a
        - registry.connect.redhat.com/kove/kove-operator@sha256:04bf970760c4dfe910465b0ef6a18e9c17fa70f6388131f12ae0aa0cd646f663
        - quay.io/kove/kove-xmem@sha256:41bfc536a278153b0a4460d2836b56762b56cf50b94d17e123a63be066c28fe7
        - quay.io/kove/kove-compatibility-detection@sha256:0303830a1f785b5e4ce6824db7c50f8feee2678fd5e87ded983686ab3f60b232
        - quay.io/kove/kove-container-hook@sha256:66ee33d59ced63406d9b40f3a3226fab7f52071f8bed9aa1ba7d859837038528
        - quay.io/kove/kove-agent@sha256:9d026b925e1d414a531530a30ed8b5f10b421ee12c7c13949ea0ee9907d36b75
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fa0acefda43e309e7f9f31f5247c933d0b868ff1573ece2f38e0549d7e2ec942
        version: 3.0.2
      entries:
      - name: kove-operator.v3.0.2
        version: 3.0.2
      - name: kove-operator.v3.0.1
        version: 3.0.1
      - name: kove-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: kove-operator
    provider:
      name: Kove
      url: https://kove.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: logging-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: logging-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Elasticsearch","metadata":{"name":"elasticsearch"},"spec":{"esClusterName":"prod","esVersion":"7.17.0","esSecurity":{"autoGeneratePassword":true,"tlsEnabled":true}}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Fluentd","metadata":{"name":"fluentd"},"spec":{"esCluster":{"host":"elasticsearch-master"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"},"indexNameStrategy":"namespace_name"}},{"apiVersion":"logging.logging.opstreelabs.in/v1beta1","kind":"Kibana","metadata":{"name":"kibana"},"spec":{"replicas":1,"esCluster":{"host":"https://elasticsearch-master:9200","esVersion":"7.16.0","clusterName":"elasticsearch"},"esSecurity":{"tlsEnabled":true,"existingSecret":"elasticsearch-password"}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  Monitoring, Big Data
          certified: "false"
          containerImage: quay.io/opstree/logging-operator:v0.4.0
          createdAt: 04-07-2022
          description: A golang-based operator to create and manage EFK (Elasticsearch,
            Fluentd, and Kibana) stack on Kubernetes.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          repository: https://github.com/ot-container-kit/logging-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elasticsearch
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Life Cycle
            displayName: IndexLifeCycle
            kind: IndexLifeCycle
            name: indexlifecycles.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Index Template
            displayName: IndexTemplate
            kind: IndexTemplate
            name: indextemplates.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Fluentd
            displayName: Fluentd
            kind: Fluentd
            name: fluentds.logging.logging.opstreelabs.in
            version: v1beta1
          - description: Kibana
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.logging.opstreelabs.in
            version: v1beta1
        description: |2+

          Logging Operator is an operator created in Golang to set up and manage EFK(Elasticsearch, Fluentd, and Kibana) cluster inside Kubernetes and Openshift environment. This operator is capable of setting up each individual component of EFK cluster separately.
          ### Supported Features
          - Setup of elasticsearch cluster with different nodes type:- master, data, ingestion, and client.
          - Customizable configuration for elasticsearch, fluentd stack.
          - Setup of fluentd as light-weight log-shipper as Daemonset.
          - Kibana will be set up as visualization tool for elastic stack.
          - Seamless upgrades of elasticsearch, fluentd, and kibana.
          - Security best practices support for complete stack such as TLS, elastic security.
          - Kubernetes resources objects support like:- resources, securityContext, affinity, tolerations, etc.
          - Elasticsearch plugins and keystore seamless support with operator
          ### Documentation
          **[Documentation](https://ot-logging-operator.netlify.app/)**
          ### Getting Help
          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-
          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation
          ### License
          MongoDB Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/logging-operator/blob/master/LICENSE)

        displayName: Logging Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Operator
        - Elasticsearch
        - Fluentd
        - Kibana
        - Opstree
        - Logging
        - Monitoring
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-logging-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep
        maturity: beta
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/logging-operator@sha256:d3ce3322f7f25c7b8bfabdd6e9fd32b833239fc39592240e0ccff3a13757fb17
        - quay.io/opstree/logging-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: logging-operator.v0.4.0
        version: 0.4.0
      - name: logging-operator.v0.3.0
        version: 0.3.0
      name: beta
    defaultChannel: beta
    packageName: logging-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-cert-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1
    - currentCSV: cert-manager-operator.v1.15.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
          createdAt: 2025-03-13T18:26:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.0 <1.15.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/infrastructure-features: '["proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.15.5](https://github.com/cert-manager/cert-manager/tree/v1.15.2), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:20efff60a0caf5eafb38986fd21611697b5bc534c2e789da233983a9739938ed
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:6af3ee8b2a5a87042fb7158bda8d6cf2e6324d1e265974acf77214d4cd0ea0d3
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9573d74bd2b926ec94af76f813e6358f14c5b2f4e0eedab7c1ff1070b7279a5c
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:f252d6f89a7cdcc520c3af461ca57d0dfabf736b9c38277d77477c4c880e182c
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:f328263e2d29e34ede65e4501f0447b2d9f84e9445a365c2fa2fbb253939e274
        version: 1.15.1
      entries:
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.15
    - currentCSV: cert-manager-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
          createdAt: 2025-07-10T08:52:43
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.15.1 <1.16.1'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.16.5](https://github.com/cert-manager/cert-manager/tree/v1.16.5), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:831e4b1e5aa6eed6fbb22f779ddbcb9c73e4bf6cc079d0c6652737a6a44e6320
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:33308217d2226c2555eb18020710d16175295a8e349263a5435e4b158bbc735f
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:97d03647567ddb6f04edea39244bc0814e8f7e557251060d9b241a5fabe04dfc
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:d30e755a266bb035d50e2b257e7a26c1fe6a5ccb66125576db99d537aa03ec27
        version: 1.16.1
      entries:
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.16
    - currentCSV: cert-manager-operator.v1.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "authorizationURL": "https://acme-staging-v02.api.letsencrypt.org/acme/authz-v3/XXXXX",
                  "dnsName": "sample.dns.name",
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "key": "XXX",
                  "solver": {
                    "dns01": {
                      "route53": {
                        "accessKeyID": "XXX",
                        "hostedZoneID": "XXX",
                        "region": "us-east-1",
                        "secretAccessKeySecretRef": {
                          "key": "awsSecretAccessKey",
                          "name": "aws-secret"
                        }
                      }
                    },
                    "selector": {
                      "dnsNames": [
                        "sample.dns.name"
                      ]
                    }
                  },
                  "token": "XXX",
                  "type": "DNS-01",
                  "url": "https://acme-staging-v02.api.letsencrypt.org/acme/chall-v3/XXXXXX/XXXXX",
                  "wildcard": false
                }
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "selfsigned-ca",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "selfsigned-ca.dns.name",
                  "isCA": true,
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "selfsigned-issuer"
                  },
                  "privateKey": {
                    "algorithm": "ECDSA",
                    "size": 256
                  },
                  "secretName": "ca-root-secret"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "tls-cert",
                  "namespace": "default"
                },
                "spec": {
                  "commonName": "sample.dns.name",
                  "dnsNames": [
                    "sample.dns.name"
                  ],
                  "isCA": false,
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "secretName": "tls-cert"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "annotations": {
                    "cert-manager.io/certificate-name": "tls-cert",
                    "cert-manager.io/certificate-revision": "1",
                    "cert-manager.io/private-key-secret-name": "tls-cert-sample"
                  },
                  "name": "tls-cert-sample",
                  "namespace": "default"
                },
                "spec": {
                  "groups": [
                    "system:serviceaccounts",
                    "system:serviceaccounts:cert-manager",
                    "system:authenticated"
                  ],
                  "issuerRef": {
                    "kind": "Issuer",
                    "name": "letsencrypt-staging"
                  },
                  "request": "XXX",
                  "username": "system:serviceaccount:cert-manager:cert-manager"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "selfsigned-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "ca-issuer",
                  "namespace": "default"
                },
                "spec": {
                  "ca": {
                    "secretName": "ca-root-secret"
                  }
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "letsencrypt-staging",
                  "namespace": "default"
                },
                "spec": {
                  "acme": {
                    "email": "aos-ci-cd@redhat.com",
                    "privateKeySecretRef": {
                      "name": "letsencrypt-staging"
                    },
                    "server": "https://acme-staging-v02.api.letsencrypt.org/directory",
                    "solvers": [
                      {
                        "dns01": {
                          "route53": {
                            "accessKeyID": "ACCESS_KEY_ID",
                            "hostedZoneID": "HOSTED_ZONE_ID",
                            "region": "AWS_REGION",
                            "secretAccessKeySecretRef": {
                              "key": "access-key",
                              "name": "sample-aws-secret"
                            }
                          }
                        },
                        "selector": {
                          "dnsNames": [
                            "sample.dns.name"
                          ]
                        }
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "CertManager",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logLevel": "Normal",
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "IstioCSR",
                "metadata": {
                  "annotations": {
                    "kubernetes.io/description": "Creating this resource requires the istio-csr tech-preview feature to be enabled, which otherwise has no effect. Please refer to the cert-manager documentation for more information on enabling the istio-csr feature. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process"
                  },
                  "name": "default",
                  "namespace": "istio-csr"
                },
                "spec": {
                  "istioCSRConfig": {
                    "certManager": {
                      "issuerRef": {
                        "group": "cert-manager.io",
                        "kind": "Issuer",
                        "name": "istio-csr-issuer"
                      }
                    },
                    "istio": {
                      "namespace": "istio-system"
                    },
                    "istiodTLSConfig": {
                      "trustDomain": "cluster.local"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
          createdAt: 2025-08-06T08:21:40
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=1.16.1 <1.17.0'
          operator.openshift.io/uninstall-message: The cert-manager Operator for Red
            Hat OpenShift will be removed from cert-manager-operator namespace. If
            your Operator configured any off-cluster resources, these will continue
            to run and require manual cleanup. All operands created by the operator
            will need to be manually cleaned up. Please refer to https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/cert-manager-operator-uninstall.html
            for additional steps.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-manager-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/cert-manager-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: CertManager is the Schema for the certmanagers API
            displayName: CertManager
            kind: CertManager
            name: certmanagers.operator.openshift.io
            version: v1alpha1
          - kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - kind: IstioCSR
            name: istiocsrs.operator.openshift.io
            version: v1alpha1
          - kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          The cert-manager Operator for Red Hat OpenShift provides seamless support for [cert-manager v1.17.4](https://github.com/cert-manager/cert-manager/tree/v1.17.4), which automates certificate management.
          For more information, see the [cert-manager Operator for Red Hat OpenShift documentation](https://docs.openshift.com/container-platform/latest/security/cert_manager_operator/index.html).
        displayName: cert-manager Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cert-manager-operator
        - cert
        - certificates
        - security
        - TLS
        links:
        - name: Documentation
          url: https://github.com/openshift/cert-manager-operator/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:4d5e238300ce6f427a1045d51d6b37a4e5c5633985208ebb44f91e7dd53897d9
        - registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:4f7c045819c39e176a6090efdaba6ec736edf772d88fc87dd1c6fb33d3b5b26b
        - registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:96d51e3a64bf30cbd92836c7cbd82f06edca16eef78ab1432757d34c16628659
        - registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:9ea2c29a384b964cef14f853278821df3cd30320f25afab8823897192f67fc7e
        - registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:cc20a2ab116597b080d303196825d7f5c81c6f30268f7866fb2911706efea210
        version: 1.17.0
      entries:
      - name: cert-manager-operator.v1.17.0
        version: 1.17.0
      - name: cert-manager-operator.v1.16.1
        version: 1.16.1
      - name: cert-manager-operator.v1.16.0
        version: 1.16.0
      - name: cert-manager-operator.v1.15.1
        version: 1.15.1
      - name: cert-manager-operator.v1.15.0
        version: 1.15.0
      name: stable-v1.17
    defaultChannel: stable-v1
    packageName: openshift-cert-manager-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: https://www.turbonomic.com/
    name: t8c-tsc-client-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
          createdAt: "2025-08-07T07:15:28Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2-candidate-20250806145822000'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:f3497acc3216d820014cce7332e9d0683f4ca6438782ace02210192f787ca097
        - icr.io/cpopen/t8c-client-operator@sha256:8db7396c956e7a04aee92e1e90491cd3eb533cebe5f5616ca43ff16d9f3ed6e2
        version: 1.4.2-candidate-20250806145822000
      entries:
      - name: t8c-tsc-client-certified.v1.4.2-candidate-20250806145822000
        version: 1.4.2-candidate-20250806145822000
      name: candidate
    - currentCSV: t8c-tsc-client-certified.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "TurbonomicClient",
                "metadata": {
                  "name": "turbonomicclient-release"
                },
                "spec": {
                  "global": {
                    "version": "8.16.1"
                  },
                  "probes": {
                    "actionScript": {
                      "enabled": false
                    },
                    "actionStreamKafka": {
                      "enabled": false
                    },
                    "appDynamics": {
                      "enabled": false
                    },
                    "compellent": {
                      "enabled": false
                    },
                    "datadog": {
                      "enabled": false
                    },
                    "dynatrace": {
                      "enabled": false
                    },
                    "flexera": {
                      "enabled": false
                    },
                    "hds": {
                      "enabled": false
                    },
                    "hpe3par": {
                      "enabled": false
                    },
                    "hyperFlex": {
                      "enabled": false
                    },
                    "hyperV": {
                      "enabled": false
                    },
                    "ibmStorageFlashSystem": {
                      "enabled": false
                    },
                    "instana": {
                      "enabled": false
                    },
                    "jvm": {
                      "enabled": false
                    },
                    "mssql": {
                      "enabled": false
                    },
                    "mysql": {
                      "enabled": false
                    },
                    "netApp": {
                      "enabled": false
                    },
                    "newRelic": {
                      "enabled": false
                    },
                    "nutanix": {
                      "enabled": false
                    },
                    "oneView": {
                      "enabled": false
                    },
                    "oracle": {
                      "enabled": false
                    },
                    "powerVM": {
                      "enabled": false
                    },
                    "pure": {
                      "enabled": false
                    },
                    "scaleIO": {
                      "enabled": false
                    },
                    "serviceNow": {
                      "enabled": false
                    },
                    "snmp": {
                      "enabled": false
                    },
                    "terraform": {
                      "enabled": false
                    },
                    "tomcat": {
                      "enabled": false
                    },
                    "ucs": {
                      "enabled": false
                    },
                    "vCenter": {
                      "enabled": false
                    },
                    "vmax": {
                      "enabled": false
                    },
                    "vmm": {
                      "enabled": false
                    },
                    "vplex": {
                      "enabled": false
                    },
                    "webLogic": {
                      "enabled": false
                    },
                    "webSphere": {
                      "enabled": false
                    },
                    "wmi": {
                      "enabled": false
                    },
                    "xtremIO": {
                      "enabled": false
                    }
                  }
                }
              },
              {
                "apiVersion": "clients.turbonomic.ibm.com/v1alpha1",
                "kind": "VersionManager",
                "metadata": {
                  "name": "versionmanager-release"
                },
                "spec": {
                  "url": "http://remote-nginx-tunnel:9080/cluster-manager/clusterConfiguration"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
          createdAt: "2025-08-12T23:15:08Z"
          description: Turbonomic Secure Connect for Hybrid environment management
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.4.2'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TurbonomicClient is the Schema for the turbonomicclients
              API
            displayName: Turbonomic Client
            kind: TurbonomicClient
            name: turbonomicclients.clients.turbonomic.ibm.com
            version: v1alpha1
          - description: VersionManager is the Schema for the versionmanagers API
            displayName: Version Manager
            kind: VersionManager
            name: versionmanagers.clients.turbonomic.ibm.com
            version: v1alpha1
        description: "### Turbonomic Secure Connect for Hybrid environment management\nTurbonomic
          Workload Automation for Multicloud simultaneously optimizes performance,
          compliance, and cost in real-time. Workloads are precisely resourced, automatically,
          to perform while satisfying business constraints:\n* Continuous placement
          of workload across multiple clouds both on-prem and public clouds providers.\n*
          Continuous scaling for applications and the underlying infrastructure.\n*
          Capability to connect both public cloud and on-prem targets on single saas
          instance\nIt assures application performance by giving workloads the resources
          they need when they need them.\n\n### How do I start?\nTurbonomic documentation
          can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm.\nTo
          setup Turbonomic Secure Connect client, follow the instructions below:\n1.
          Click Install on the landing page\n2. Upon moving to Install Operator page,
          do not modify the default selection for any of the configuration\n3. Select
          the Installed namespace from the drop down or add a new namespace and click
          on create project to create the entered namespace.\n4. Click Install on
          the Install Operator page.\nAt this point operator is installed.\n\nOn the
          Details tab, under Provided APIs: \n1. Click Create instance for the Turbonomic
          Client API. Enable/Disable the probes based on the requirement, and click
          Create.\n2. Click Create instance for the Version Manager API. Do not modify
          the JSON/YAML file, click on Create.\n\n### How does it work?\nTurbonomic
          uses a public APIs already exposed by application and infrastructure instrumentation
          to discover and monitor your environment.\nTurbonomic determines the right
          actions that drive continuous health, including continuous placement and
          continuous scaling for applications and the underlying cluster.\nTurbonomic
          leverages the built-on orchestration provided by the application and infrastructure
          deployment tools and automates the execution of these actions to continiously
          meet the respective service level objective of each application service."
        displayName: Turbonomic Secure Connect Client
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - turbonomic
        - turbo secure connect
        links:
        - name: T8c Tsc Client Certified
          url: https://t8c-tsc-certified.domain
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        relatedImages:
        - icr.io/cpopen/t8c-client-operator@sha256:626e024327388410f7a77bd9b8e5a5df29993fcafdb0176cd0bea08b05713c61
        - registry.connect.redhat.com/turbonomic/t8c-tsc-client-operator-bundle@sha256:23f2f66e9f6d79d05e2be5a10c394834f093f4677ab06d8935ed33980408e8d3
        version: 1.4.2
      entries:
      - name: t8c-tsc-client-certified.v1.4.2
        version: 1.4.2
      name: stable
    defaultChannel: stable
    packageName: t8c-tsc-client-certified
    provider:
      name: Turbonomic, Inc.
      url: https://www.turbonomic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-storage-odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-storage-odf-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.7.1'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:24eae9d1502ce98bb024d3eba154ddec07fdac0d78ecad0c3a64b55f5b3600f7
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:c6d6c3e0408b851cdc9589cc0ee642d77397cdfaa40f50cf42557c5aaf9c0690
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:2dd8608645b13f81330466e25664b132d46af480a3ab2b0dd48fb1feba38870a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        version: 1.7.1
      entries:
      - name: ibm-storage-odf-operator.v1.7.1
        version: 1.7.1
      name: stable-v1.7
    - currentCSV: ibm-storage-odf-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.ibm.com/v1alpha1",
                "kind": "FlashSystemCluster",
                "metadata": {
                  "name": "flashsystemcluster-sample",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "defaultPool": {
                    "fsType": "ext4",
                    "poolName": "Pool0",
                    "spaceEfficiency": "thick",
                    "storageclassName": "odf-flashsystemcluster-sample",
                    "volumeNamePrefix": "odf"
                  },
                  "insecureSkipVerify": true,
                  "name": "flashsystem-xxx",
                  "secret": {
                    "name": "fs-secrets-example",
                    "namespace": "openshift-storage"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          console.openshift.io/plugins: '["ibm-storage-odf-plugin"]'
          containerImage: registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
          createdAt: ""
          description: IBM Storage ODF operator provides basic storage capabilities
            and extended management functions through OpenShift data foundation framework
            for applications.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1 <1.8.0'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlashSystemCluster is the Schema for the FlashSystemClusters
              API
            displayName: Flash System Cluster
            kind: FlashSystemCluster
            name: flashsystemclusters.odf.ibm.com
            version: v1alpha1
        description: IBM Storage ODF operator provides basic storage capabilities
          and extended management functions through OpenShift data foundation framework
          for applications.
        displayName: IBM Storage ODF operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ODF
        - storage
        - block storage
        - flashsystem
        - ibm
        links:
        - name: IBM Storage ODF Operator
          url: https://github.com/IBM/ibm-storage-odf-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator-bundle@sha256:ff314ab983699455c32ab0ef27bb552d00825974a2d7088df2000f09afd8a2e2
        - registry.connect.redhat.com/ibm/ibm-storage-odf-operator@sha256:7432340ac156e620240eea73a19321a871fa89cc513c340a91d97d3f3fd4ded2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e8eb296cc9b3702378212d46276015ddef581dccb008af99fda697dbaf87bfbc
        - quay.io/ibmodffs/ibm-storage-odf-plugin@sha256:3344c7e987674fb98173bbc6ae1b189613e3ccefb468f1af9866734ddbf0408f
        version: 1.8.0
      entries:
      - name: ibm-storage-odf-operator.v1.8.0
        version: 1.8.0
      name: stable-v1.8
    defaultChannel: stable-v1.8
    packageName: ibm-storage-odf-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-dr-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-dr-manager.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "HostName_1": "100.200.30.41",
                    "HostName_2": "100.200.30.43",
                    "abc": "HostName_2",
                    "www": "HostName_1",
                    "xyz": "HostName_3"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:5f10cfc1bc8571f44a99d88755adade0743106219673bbdd2e4d3a04ee714778
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:a98f9cc51f2c7baeec6dbb4933a42ae8ae20ff1a42c9d48a37174cf5211bbcf8
        version: 8.0.220
      entries:
      - name: infoscale-dr-manager.v8.0.220
        version: 8.0.220
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-dr-manager.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          createdAt: "2024-06-17T10:14:13Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:9f90cd7ab8c0ab5da6eb12f7f0dbd177df62f07c38fb759a97989899cfcbe75d
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:c46cbf78631fc790c77b6c13c1d1d4b5ab35f49164f7c2e36c7ec59220cbdce2
        version: 8.0.330
      entries:
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-dr-manager.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "cloudVendor": "Local",
                  "currentPrimary": "Clus1",
                  "force": false,
                  "lbEnabled": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "disable",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteCloudVendor": "Local",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteLbEnabled": false,
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 5,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "maximumMetadataCopies": 5,
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
          createdAt: "2024-02-22T14:14:42Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721648-168743109)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/168721626-168743109-0/v168721654-168743109)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membership
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membership
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-dr-operator-bundle@sha256:15d6d2d189c8acd42867a8ed6d843b8be819a07e71b42aab24ea5fdb1fd7bfec
        - registry.connect.redhat.com/arctera/infoscale-dr-operator@sha256:7dff20a0aff34e1e66a97c9a39cda883642dd1e94b872c8fc2f62efc7a4db1bb
        version: 8.0.410
      entries:
      - name: infoscale-dr-manager.v8.0.410
        version: 8.0.410
      - name: infoscale-dr-manager.v8.0.409
        version: 8.0.409
      - name: infoscale-dr-manager.v8.0.400
        version: 8.0.400
      - name: infoscale-dr-manager.v8.0.330
        version: 8.0.330
      - name: infoscale-dr-manager.v8.0.320
        version: 8.0.320
      - name: infoscale-dr-manager.v8.0.310
        version: 8.0.310
      - name: infoscale-dr-manager.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-dr-manager.v8.0.100
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DNS",
                "metadata": {
                  "name": "test-dns"
                },
                "spec": {
                  "domain": "vcslnxdns.com",
                  "resRecord": {
                    "abc": "r7515-054-vm9",
                    "r7515-054-vm10": "10.221.85.83",
                    "r7515-054-vm8": "10.221.85.81",
                    "r7515-054-vm9": "10.221.85.82",
                    "www": "r7515-054-vm8",
                    "xyz": "r7515-054-vm10"
                  },
                  "stealthMasters": [
                    "1.2.3.4"
                  ],
                  "tsigKeyFile": "/Kvcslnxdns.com.+157+10641.private"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DataReplication",
                "metadata": {
                  "name": "test-datareplication"
                },
                "spec": {
                  "currentPrimary": "Clus1",
                  "force": false,
                  "localHostAddress": "10.0.0.1",
                  "localNIC": "eth0",
                  "localNetMask": "255.255.240.0",
                  "remoteClusterDetails": [
                    {
                      "bandwidthLimit": "N/A",
                      "clusterName": "Clus2",
                      "latencyProtection": "fail",
                      "logProtection": "autodcm",
                      "networkTransportProtocol": "TCP",
                      "remoteHostAddress": "10.0.0.2",
                      "remoteNIC": "eth0",
                      "remoteNetMask": "255.255.240.0",
                      "replicationState": "start",
                      "replicationType": "async"
                    }
                  ],
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "DisasterRecoveryPlan",
                "metadata": {
                  "name": "test-disaster-recovery-plan"
                },
                "spec": {
                  "clusterFailOverPolicy": "Manual",
                  "dataReplicationPointer": "test-datareplication",
                  "force": false,
                  "preferredClusterList": [
                    "Clus1",
                    "Clus2"
                  ],
                  "primaryCluster": "Clus1",
                  "selector": {
                    "namespace": "mysql"
                  }
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "GlobalClusterMembership",
                "metadata": {
                  "name": "global-cluster-membership"
                },
                "spec": {
                  "backupClusterScopeCRD": true,
                  "counterMissTolerance": 5,
                  "datarepRefreshStatusFrequency": 10,
                  "globalClusterOperation": "none",
                  "globalMemberClusters": [
                    {
                      "clusterID": "Clus1",
                      "drControllerAddress": "10.0.10.1",
                      "drControllerPort": "8080"
                    },
                    {
                      "clusterID": "Clus2",
                      "drControllerAddress": "10.0.10.2",
                      "drControllerPort": "9090"
                    }
                  ],
                  "localClusterName": "Clus1",
                  "metadataBackupInterval": 15
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ DR Manager manages the Disaster Recovery operations
            of InfoScale™ for container environment
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataReplication is the Schema for the storagereplications
              API
            displayName: Data Replication
            kind: DataReplication
            name: datareplications.infoscale.veritas.com
            version: v1
          - description: DisasterRecoveryPlan is the Schema for the disasterrecoveryplans
              API
            displayName: Disaster Recovery Plan
            kind: DisasterRecoveryPlan
            name: disasterrecoveryplans.infoscale.veritas.com
            version: v1
          - description: DNS is the Schema for the DNSs API
            displayName: DNS
            kind: DNS
            name: dns.infoscale.veritas.com
            version: v1
          - description: GlobalClusterMembership is the Schema for the globalclustermemberships
              API
            displayName: Global Cluster Membership
            kind: GlobalClusterMembership
            name: globalclustermemberships.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ DR Manager

          InfoScale™ DR Manager manages the Disaster Recovery operations of InfoScale™ for container environment

          ## Overview

          - Disaster Recovery(DR) is provided to applications hosted in container ecosystems. Native container HA capabilities provide high availability to application components within a cluster. However, DR functionality provides disaster recovery in the event of a entire cluster failure and application components can be restored on another peer cluster in membership.
          - You can form a logical notion called 'Global Cluster' comprising clusters that can be used to migrate DR-enabled objects. DR-enabled objects migrate to peer cluster in case of a disaster like entire cluster going down, loss of connectivity with a particular cluster, user-initiated planned migration across cluster.

          ---

          ## Pre-requisites

          **1.** InfoScale must be installed and InfoScale pods must be configured on the clusters. [[More Details](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v155481520-155483981)]

          **2.** Create storage class with name, which must always be **`infoscale-dr-csi-sc`.**

          ```
          apiVersion: storage.k8s.io/v1
          kind: StorageClass
          metadata:
            name: infoscale-dr-csi-sc
          parameters:
            fstype: vxfs
            layout: mirror
          allowVolumeExpansion: true
          provisioner: org.veritas.infoscale
          reclaimPolicy: Delete

          ```

          **3.** Create persistent volume claim with name, which must always be **`infoscale-dr-meta-bkp-pvc`.**

          ```
          apiVersion: v1
          kind: PersistentVolumeClaim
          metadata:
            name: infoscale-dr-meta-bkp-pvc
          spec:
            accessModes:
            - ReadWriteOnce
            resources:
              requests:
                storage: 4Gi
            storageClassName: infoscale-dr-csi-sc
            volumeMode: Filesystem

          ```

          - **Note:** User must use InfoScale™ CSI provisioned storage.

          - [For more details, please refer to 'Installing InfoScale DR Manager on openshift' section from official documentation](https://www.veritas.com/content/support/en_US/doc/155483965-155483981-0/v151984431-155483981)

          ## InfoScale™ DR Manager custom resources

          - **Note:** InfoScale™ DR Manager custom resources must be created only in the following order

          **1. Global Cluster Membership:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: GlobalClusterMembership
          metadata:
            name: global-cluster-membership
          spec:
            # Local cluster name in the global membership
            localClusterName: Clus1

            globalMemberClusters:
                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus1

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.1"

                # Port used for DR controller
                drControllerPort: "8080"

                # Cluster ID of member of global cluster membershrip
              - clusterID: Clus2

                # Address Used For Communicating With Peer Cluster's DR Controller
                drControllerAddress: "10.0.10.2"

                # Port used for DR controller
                drControllerPort: "9090"

          ```

          **2. Data Replication:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DataReplication
          metadata:
            name: test-datareplication
          spec:
            # Primary cluster details

            # Virtual IP address to configure VVR
            localHostAddress: 10.0.0.1

            # Corresponding netmask to configure VVR
            localNetMask: 255.255.240.0

            # Corresponding network interface to configure VVR (If NIC name is identical for all nodes)
            localNIC: eth0

            # Namespace and optional labels for which you want to configure data replication
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # Current primary cluster name - Name of the cluster you want to back up
            currentPrimary: Clus1

            # Secondary cluster details
            remoteClusterDetails:
                # ID of the Cluster to be used for a backup
              - clusterName: Clus2

                # Virtual IP address for VVR configuration of this cluster
                remoteHostAddress: 10.0.0.2

                # Corresponsding Netmask of this cluster
                remoteNetMask: 255.255.240.0

                # Corresponding Network interface of this cluster
                remoteNIC: eth0

                # (optional) replication type can be sync or async.
                # default value will be async if not specified.
                replicationType: async

          ```
          **3. Disaster Recovery Plan:**

          ```
          apiVersion: infoscale.veritas.com/v1
          kind: DisasterRecoveryPlan
          metadata:
            name: test-disaster-recovery-plan
          spec:
            # Name of cluster that should be treated as primary for this DR plan
            primaryCluster: Clus1

            # (optional) Set force to True if peer cluster(s) is not reachable and local cluster needs to perform takeover
            force: false

            # List of member cluster(s) where this DRPlan can failover.
            # Sequence of MemberCluster specified in this list denotes relative preference of member cluster(s)
            # Must be subset of Global Cluster Membership
            preferredClusterList: ["Clus1", "Clus2"]

            # Kind of corrective action in case of disaster
            # default value will be "Manual" if not specified
            clusterFailOverPolicy: Manual

            # Specify namespace and optional labels to decide what all needs to be part of the disaster recovery plan
            selector:
              namespace: mysql
              #labels:
              #  app: db

            # (optional) Pointer to manage data replication
            dataReplicationPointer: test-datareplication

          ```
        displayName: InfoScale™ DR Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software
        - Defined Storage
        - Clustered storage
        - CFS
        - DR
        - Disaster Recovery
        links:
        - name: Infoscale DR Manager
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com/
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager-bundle@sha256:71ff93b137b6569f8fdefbd446cf12cd95477e4294bdea98fb5318e5d7d1cdd6
        - registry.connect.redhat.com/veritas-technologies/infoscale-dr-manager@sha256:eb999187a1f9a9ce5566089e24adf598b1c229d31c5e57ddcfe4439ba25e97dc
        version: 8.0.100
      entries:
      - name: infoscale-dr-manager.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-dr-manager
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:7a04dfd687dfe1646bd59a39c9e88781ffb815085bbd15f6305f6eac5f59bc49
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:d1a9246def81537e3417a6d48d1e818459583db658b54a87ca5af6911c8a3f53
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:e88676c0346c245b315f1a3bf7f85b10515bae3620c786fedda0533b39d2a400
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:66bd3eba0b68c317fe128c6713488cb6374f9647095768d3f7eed8120bf02e77
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-bundle@sha256:29a44046aa2cf273798b5a58ce3c674e51d5e923b7b6798322a8c365e79ad28a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-2
      entries:
      - name: couchbase-operator.v2.8.1-2
        version: 2.8.1-2
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.8.0
          createdAt: 2021-12-08T10-00-00Z
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino-operator:v0.8.0
        - quay.io/openshift-community-operators/authorino-operator@sha256:8fe34f94f87c567401cc2bbbbcc3a3e719364b80a2b17851496f86c85af5c09a
        version: 0.8.0
      entries:
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: authorino-operator.v0.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/authorino-operator:v0.16.0
          createdAt: "2024-11-25T13:22:24Z"
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: The operator to manage instances of Authorino
        displayName: Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: Authorino Operator
          url: https://github.com/Kuadrant/authorino-operator
        - name: Authorino
          url: https://github.com/Kuadrant/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/kuadrant/authorino:v0.20.0
        - quay.io/community-operator-pipeline-prod/authorino-operator@sha256:92ad93f27c34f1406fc05a248d04709b0109adf37f2f9424f48fc6a73d9d359e
        - quay.io/kuadrant/authorino-operator:v0.16.0
        version: 0.16.0
      entries:
      - name: authorino-operator.v0.16.0
        version: 0.16.0
      - name: authorino-operator.v0.15.1
        version: 0.15.1
      - name: authorino-operator.v0.15.0
        version: 0.15.0
      - name: authorino-operator.v0.14.0
        version: 0.14.0
      - name: authorino-operator.v0.13.1
        version: 0.13.1
      - name: authorino-operator.v0.13.0
        version: 0.13.0
      - name: authorino-operator.v0.12.0
        version: 0.12.0
      - name: authorino-operator.v0.11.1
        version: 0.11.1
      - name: authorino-operator.v0.10.0
        version: 0.10.0
      - name: authorino-operator.v0.9.0
        version: 0.9.0
      - name: authorino-operator.v0.8.0
        version: 0.8.0
      - name: authorino-operator.v0.7.0
        version: 0.7.0
      - name: authorino-operator.v0.6.0
        version: 0.6.0
      - name: authorino-operator.v0.5.0
        version: 0.5.0
      - name: authorino-operator.v0.4.1
        version: 0.4.1
      - name: authorino-operator.v0.4.0
        version: 0.4.0
      - name: authorino-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle@sha256:2a28b64f7321df23928c8226b3eb3161cf520d38cc57d875b3cbfc828c164edc
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
          createdAt: "2024-07-29T15:35:56Z"
          description: '[DEPRECATED] Use the certified Hazelcast Operator instead'
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          ## Deprecated
          Use the certified Hazelcast Operator instead.
          # Hazelcast Platform Operator #

          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: '[DEPRECATED] Hazelcast Platform Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:99005afda5c71ceeac200928dfed4ea35480f871053101e3aeaa44005e3fda14
        - quay.io/community-operator-pipeline-prod/hazelcast-platform-operator@sha256:1209876fc9103eb2d69c34e97e4e2e0b384a9a778ae7e6a53222b8552fbf2f26
        version: 5.13.0
      entries:
      - name: hazelcast-platform-operator.v5.13.0
        version: 5.13.0
      - name: hazelcast-platform-operator.v5.12.0
        version: 5.12.0
      - name: hazelcast-platform-operator.v5.11.0
        version: 5.11.0
      - name: hazelcast-platform-operator.v5.10.0
        version: 5.10.0
      - name: hazelcast-platform-operator.v5.9.0
        version: 5.9.0
      - name: hazelcast-platform-operator.v5.8.0
        version: 5.8.0
      - name: hazelcast-platform-operator.v5.7.0
        version: 5.7.0
      - name: hazelcast-platform-operator.v5.6.0
        version: 5.6.0
      - name: hazelcast-platform-operator.v5.5.0
        version: 5.5.0
      - name: hazelcast-platform-operator.v5.4.0
        version: 5.4.0
      - name: hazelcast-platform-operator.v5.3.0
        version: 5.3.0
      - name: hazelcast-platform-operator.v5.2.0
        version: 5.2.0
      - name: hazelcast-platform-operator.v5.1.0
        version: 5.1.0
      - name: hazelcast-platform-operator.v5.0.0
        version: 5.0.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kms-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kms-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "kms.services.k8s.aws/v1alpha1",
                "kind": "Key",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
          createdAt: "2025-09-19T22:02:02Z"
          description: AWS KMS controller is a service controller for managing KMS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS kms Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.kms.services.k8s.aws
            version: v1alpha1
          - description: Grant represents the state of an AWS kms Grant resource.
            displayName: Grant
            kind: Grant
            name: grants.kms.services.k8s.aws
            version: v1alpha1
          - description: Key represents the state of an AWS kms Key resource.
            displayName: Key
            kind: Key
            name: keys.kms.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Key Management Service (KMS) resources in AWS from within your Kubernetes cluster.

          **About Amazon KMS**

          AWS Key Management Service (AWS KMS) is a managed service that makes it easy for you to create and control the cryptographic keys that are used to protect your data. AWS KMS uses hardware security modules (HSM) to protect and validate your AWS KMS keys under the [FIPS 140-2 Cryptographic Module Validation Program](https://csrc.nist.gov/projects/cryptographic-module-validation-program/Certificate/3139), except in the China (Beijing) and China (Ningxia) Regions.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon KMS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kms
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon KMS Developer Resources
          url: https://aws.amazon.com/kms/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kms maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-kms-controller@sha256:a2c60484a6556949318aa1d730c601b5bb05b1795b945a607209ef3434cce72f
        - public.ecr.aws/aws-controllers-k8s/kms-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-kms-controller.v1.1.1
        version: 1.1.1
      - name: ack-kms-controller.v1.1.0
        version: 1.1.0
      - name: ack-kms-controller.v1.0.31
        version: 1.0.31
      - name: ack-kms-controller.v1.0.30
        version: 1.0.30
      - name: ack-kms-controller.v1.0.29
        version: 1.0.29
      - name: ack-kms-controller.v1.0.28
        version: 1.0.28
      - name: ack-kms-controller.v1.0.27
        version: 1.0.27
      - name: ack-kms-controller.v1.0.26
        version: 1.0.26
      - name: ack-kms-controller.v1.0.25
        version: 1.0.25
      - name: ack-kms-controller.v1.0.24
        version: 1.0.24
      - name: ack-kms-controller.v1.0.23
        version: 1.0.23
      - name: ack-kms-controller.v1.0.22
        version: 1.0.22
      - name: ack-kms-controller.v1.0.19
        version: 1.0.19
      - name: ack-kms-controller.v1.0.18
        version: 1.0.18
      - name: ack-kms-controller.v1.0.17
        version: 1.0.17
      - name: ack-kms-controller.v1.0.16
        version: 1.0.16
      - name: ack-kms-controller.v1.0.15
        version: 1.0.15
      - name: ack-kms-controller.v1.0.14
        version: 1.0.14
      - name: ack-kms-controller.v1.0.13
        version: 1.0.13
      - name: ack-kms-controller.v1.0.12
        version: 1.0.12
      - name: ack-kms-controller.v1.0.8
        version: 1.0.8
      - name: ack-kms-controller.v1.0.7
        version: 1.0.7
      - name: ack-kms-controller.v1.0.5
        version: 1.0.5
      - name: ack-kms-controller.v1.0.4
        version: 1.0.4
      - name: ack-kms-controller.v1.0.3
        version: 1.0.3
      - name: ack-kms-controller.v1.0.2
        version: 1.0.2
      - name: ack-kms-controller.v1.0.0
        version: 1.0.0
      - name: ack-kms-controller.v0.1.5
        version: 0.1.5
      - name: ack-kms-controller.v0.1.4
        version: 0.1.4
      - name: ack-kms-controller.v0.1.3
        version: 0.1.3
      - name: ack-kms-controller.v0.1.2
        version: 0.1.2
      - name: ack-kms-controller.v0.1.1
        version: 0.1.1
      - name: ack-kms-controller.v0.1.0
        version: 0.1.0
      - name: ack-kms-controller.v0.0.20
        version: 0.0.20
      - name: ack-kms-controller.v0.0.19
        version: 0.0.19
      - name: ack-kms-controller.v0.0.18
        version: 0.0.18
      - name: ack-kms-controller.v0.0.17
        version: 0.0.17
      - name: ack-kms-controller.v0.0.16
        version: 0.0.16
      - name: ack-kms-controller.v0.0.15
        version: 0.0.15
      - name: ack-kms-controller.v0.0.14
        version: 0.0.14
      - name: ack-kms-controller.v0.0.13
        version: 0.0.13
      - name: ack-kms-controller.v0.0.12
        version: 0.0.12
      - name: ack-kms-controller.v0.0.11
        version: 0.0.11
      - name: ack-kms-controller.v0.0.10
        version: 0.0.10
      - name: ack-kms-controller.v0.0.9
        version: 0.0.9
      - name: ack-kms-controller.v0.0.8
        version: 0.0.8
      - name: ack-kms-controller.v0.0.7
        version: 0.0.7
      - name: ack-kms-controller.v0.0.6
        version: 0.0.6
      - name: ack-kms-controller.v0.0.5
        version: 0.0.5
      - name: ack-kms-controller.v0.0.4
        version: 0.0.4
      - name: ack-kms-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kms-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: cloud-bulldozer
      provider-url: ""
    name: ripsaw
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ripsaw.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, Cloud Provider
          containerImage: quay.io/cloud-bulldozer/benchmark-operator:latest
          description: '[Benchmark-operator](https://github.com/cloud-bulldozer/ripsaw/)
            is a benchmarking tool to benchmark your kubernetes cluster. The intent
            of this Operator is to deploy common workloads to establish a performance
            baseline of Kubernetes cluster'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/cloud-bulldozer/benchmark-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Benchmark
            name: benchmarks.ripsaw.cloudbulldozer.io
            version: v1alpha1
        description: '[Benchmark-operator](https://github.com/cloud-bulldozer/benchmark-operator/)
          is a benchmarking tool to benchmark your kubernetes cluster. The intent
          of this Operator is to deploy common workloads to establish a performance
          baseline of Kubernetes cluster'
        displayName: benchmark-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - benchmark
        links:
        - name: Site
          url: https://github.com/cloud-bulldozer/benchmark-operator
        - name: Documentation
          url: https://github.com/cloud-bulldozer/benchmark-operator/tree/master/docs
        maintainers:
        - email: rsevilla@redhat.com
          name: rsevilla87
        - email: jtaleric@redhat.com
          name: jtaleric
        - email: rzaleski@redhat.com
          name: dry939
        - email: kwhitley@redhat.com
          name: whitleykeith
        maturity: alpha
        provider:
          name: cloud-bulldozer
        relatedImages:
        - k8s.gcr.io/redis:v1
        - quay.io/cloud-bulldozer/benchmark-operator:latest
        - quay.io/openshift-community-operators/ripsaw@sha256:ff8d796d5bd6c8cb0e155b3ced631122397758eb472470223e30352eeddba480
        version: 0.1.0
      entries:
      - name: ripsaw.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ripsaw
    provider:
      name: cloud-bulldozer
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: openshift-zero-trust-workload-identity-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zero-trust-workload-identity-manager.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "ZeroTrustWorkloadIdentityManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "zero-trust-workload-identity-manager",
                    "app.kubernetes.io/part-of": "zero-trust-workload-identity-manager"
                  },
                  "name": "zerotrustworkloadidentitymanager-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
          createdAt: 2025-06-16T11:15:37
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: zero-trust-workload-identity-manager
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/zero-trust-workload-identity-manager
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterFederatedTrustDomain
            name: clusterfederatedtrustdomains.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterSPIFFEID
            name: clusterspiffeids.spire.spiffe.io
            version: v1alpha1
          - kind: ClusterStaticEntry
            name: clusterstaticentries.spire.spiffe.io
            version: v1alpha1
          - kind: SpiffeCSIDriver
            name: spiffecsidrivers.operator.openshift.io
            version: v1alpha1
          - kind: SpireAgent
            name: spireagents.operator.openshift.io
            version: v1alpha1
          - kind: SpireOIDCDiscoveryProvider
            name: spireoidcdiscoveryproviders.operator.openshift.io
            version: v1alpha1
          - kind: SpireServer
            name: spireservers.operator.openshift.io
            version: v1alpha1
          - kind: ZeroTrustWorkloadIdentityManager
            name: zerotrustworkloadidentitymanagers.operator.openshift.io
            version: v1alpha1
        description: The Zero Trust Workload Identity Manager is an Openshift Day-2
          Operator designed to automate the setup and management of SPIFFE/SPIRE components
          (like spire-server, spire-agent, spiffe-csi-driver, and oidc-discovery-provider)
          within OpenShift clusters. It enables zero-trust security by dynamically
          issuing and rotating workload identities, enforcing strict identity-based
          authentication across workloads. This manager abstracts complex SPIRE configurations,
          streamlining workload onboarding with secure identities and improving the
          cluster's overall security posture.
        displayName: Zero Trust Workload Identity Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zero-trust-workload-identity-manager
        - ztwim
        - spire
        - spiffe-spire
        - security
        links:
        - name: Documentation
          url: https://github.com/openshift/zero-trust-workload-identity-manager/blob/master/README.md
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-agent-rhel9@sha256:0281d6ab233a685ba0f739629584aacd9254e0b13a697fe8708efa41289e746e
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-operator-bundle@sha256:1974903a3e2b692eedab2602395fdc4a6e2ffbc79ed4e633705c05ce7771e1b7
        - registry.redhat.io/zero-trust-workload-identity-manager/zero-trust-workload-identity-manager-rhel9@sha256:17f533beea953c0a41b4c0a15a6bfd1f959565361509ec521fa43281a4c08e2a
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-helper-rhel9@sha256:a39c80c6e1495ea166a858c8ed03304c6455420c077d937c99bf8732a6829790
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-controller-manager-rhel9@sha256:67f0e6670195c84515e7d78139a2951e463db11445740fe1d9a449a7f338556c
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-oidc-discovery-provider-rhel9@sha256:5ba7e94b7ea06852d32002b055df556ff2a2fdc9a2e8f793b837c4856d773782
        - registry.access.redhat.com/ubi9@sha256:861e833044a903f689ecfa404424494a7e387ab39cf7949c54843285d13a9774
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:7d170ba9b3de2e88dc0d0a6b0c13630ca3b23eed96ea5e3c725516046d7c7fc7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-csi-driver-rhel9@sha256:a16ac0a4363f7cffada39a2f17aa42f5de3faf78b3688ae5ebf2ecf3a4759cb7
        - registry.redhat.io/zero-trust-workload-identity-manager/spiffe-spire-server-rhel9@sha256:70f97c613feeaadfed99d37aabd62e1cb2247825292ad9be806bc2fd065c9166
        version: 0.1.0
      entries:
      - name: zero-trust-workload-identity-manager.v0.1.0
        version: 0.1.0
      name: tech-preview-v0.1
    defaultChannel: tech-preview-v0.1
    packageName: openshift-zero-trust-workload-identity-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: node-observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-observability-operator.v0.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "labels": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha1",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservability",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  },
                  "type": "crio-kubelet"
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityMachineConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "debug": {
                    "enableCrioProfiling": true
                  },
                  "nodeSelector": {
                    "kubernetes.io/hostname": "worker-a"
                  }
                }
              },
              {
                "apiVersion": "nodeobservability.olm.openshift.io/v1alpha2",
                "kind": "NodeObservabilityRun",
                "metadata": {
                  "name": "nodeobservabilityrun-sample",
                  "namespace": "node-observability-operator"
                },
                "spec": {
                  "nodeObservabilityRef": {
                    "name": "cluster"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
          createdAt: 2022-12-05T09:25:09
          olm.skipRange: <0.2.0
          operatorframework.io/suggested-namespace: node-observability-operator
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservability prepares a subset of worker nodes (identified
              by a label) for running node observability queries, such as profiling
            displayName: Node Observability
            kind: NodeObservability
            name: nodeobservabilities.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityMachineConfig is the Schema for the nodeobservabilitymachineconfigs
              API
            displayName: Node Observability Machine Config
            kind: NodeObservabilityMachineConfig
            name: nodeobservabilitymachineconfigs.nodeobservability.olm.openshift.io
            version: v1alpha2
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha1
          - description: NodeObservabilityRun is a request to run observability actions
              on the nodes previously selected in NodeObservability resource
            displayName: Node Observability Run
            kind: NodeObservabilityRun
            name: nodeobservabilityruns.nodeobservability.olm.openshift.io
            version: v1alpha2
        description: |-
          An Operator that will be able to gather debugging/profiling data over
          a custom period of time which would be helpful to troubleshoot and resolve issues
          for OpenShift customers.
        displayName: Node Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node-observability-operator
        links:
        - name: Source Code
          url: https://github.com/openshift/node-observability-operator
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/noo/node-observability-operator-bundle-rhel8@sha256:0fbbb0b32adadc5746251103d25bba6bd365b2ff60e289e4694061fc92c3ce6e
        - registry.redhat.io/noo/node-observability-rhel8-operator@sha256:53ba77f648154ac8aca1ae6c83815078dfde211dba1e42f597e69844e4954c01
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e3dad360d0351237a16593ca0862652809c41a2127c2f98b9e0a559568efbd10
        - registry.redhat.io/noo/node-observability-agent-rhel8@sha256:a8e60d9f63e2ed0cf1e27309aac54014e3f1efc947b51ca0a99376338d220875
        version: 0.2.0
      entries:
      - name: node-observability-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: node-observability-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Palo Alto Networks
      provider-url: ""
    name: pcc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pcc-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Console",
                "metadata": {
                  "name": "pcc-console",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "ConsoleDefender",
                "metadata": {
                  "name": "pcc-consoledefender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "consoleConfig": {
                    "serviceType": "ClusterIP"
                  },
                  "defenderConfig": {
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              },
              {
                "apiVersion": "pcc.paloaltonetworks.com/v1alpha1",
                "kind": "Defender",
                "metadata": {
                  "name": "pcc-defender",
                  "namespace": "twistlock"
                },
                "spec": {
                  "defenderConfig": {
                    "clusterAddress": "twistlock-console.example.com",
                    "consoleAddress": "https://twistlock-console.example.com:8083",
                    "docker": false
                  },
                  "namespace": "twistlock",
                  "version": "21_08_520"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
          description: Deploy Prisma Cloud Compute for cloud-native security in your
            clusters
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator
          support: pcc-integrations@paloaltonetworks.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Install the Prisma Cloud Compute Console and Defenders
            displayName: Console and Defenders
            kind: ConsoleDefender
            name: consoledefenders.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install the Prisma Cloud Compute Console
            displayName: Console
            kind: Console
            name: consoles.pcc.paloaltonetworks.com
            version: v1alpha1
          - description: Install Prisma Cloud Compute Defenders
            displayName: Defenders
            kind: Defender
            name: defenders.pcc.paloaltonetworks.com
            version: v1alpha1
        description: |
          ## Features
          - Deploy Console
              - Create initial local administrator user
              - Add license
          - Deploy Defenders
          - Support for offline deployments
          - Upgrade Console
          - Upgrade Defenders

          ## Installation
          See the [Prisma Cloud Compute Operator documentation](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator) for installation guides.

          ## Support
          This operator is available "as is," and should be seen as community-supported; however, Palo Alto Networks will continue contributing whenever possible.
          Palo Alto Networks does not provide technical support or help with using or troubleshooting the operator through normal support channels.
          We encourage you to open GitHub [issues](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/issues) and [pull requests](https://github.com/PaloAltoNetworks/prisma-cloud-compute-operator/pulls) to track bugs and feature requests.

          ## Other links
          [Prisma Cloud Compute product documentation](https://docs.paloaltonetworks.com/prisma/prisma-cloud/prisma-cloud-admin-compute.html)

          [Prisma Cloud Compute API documentation](https://prisma.pan.dev/api/cloud/cwpp)
        displayName: Prisma Cloud Compute Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - palo
        - palo alto
        - palo alto networks
        - paloalto
        - twistlock
        - compute
        - prisma
        - prisma cloud
        - prisma cloud compute
        - security
        - image
        - images
        - container
        - containers
        - docker
        - podman
        - vulnerability
        - vulnerabilities
        - compliance
        - runtime
        - scan
        - scanning
        - monitor
        - monitoring
        - alert
        - alerting
        links:
        - name: Prisma Cloud
          url: https://www.paloaltonetworks.com/prisma/cloud
        maintainers:
        - email: pcc-integrations@paloaltonetworks.com
          name: Wyatt Gill
        maturity: alpha
        provider:
          name: Palo Alto Networks
        relatedImages:
        - quay.io/prismacloud/pcc-operator@sha256:51a343d5ea0115e3c43ccaa546492e85ab74cf2cf2658d725c1ea7099fc89191
        - registry.connect.redhat.com/paloalto/pcc-operator@sha256:c7200a8c652ee544b0e0bb4d9427eab3a3b2dee6ca61ee88abb8052316382673
        version: 0.2.3
      entries:
      - name: pcc-operator.v0.2.3
        version: 0.2.3
      name: stable
    defaultChannel: stable
    packageName: pcc-operator
    provider:
      name: Palo Alto Networks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: External Secrets
      provider-url: https://external-secrets.io
    name: external-secrets-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.6.0-rc1
        version: 0.6.0-rc1
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.3
        version: 0.4.3
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      - name: external-secrets-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: external-secrets-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": "1234"
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "key": "provider-key",
                      "property": "provider-key-property",
                      "version": "provider-key-version"
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1alpha1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterExternalSecret",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "externalSecretName": "hello-world-es",
                  "externalSecretSpec": {
                    "data": [
                      {
                        "remoteRef": {
                          "key": "provider-key",
                          "property": "provider-key-property",
                          "version": "provider-key-version"
                        },
                        "secretKey": "secret-key-to-be-managed"
                      }
                    ],
                    "dataFrom": [
                      {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    ],
                    "refreshInterval": "1h",
                    "secretStoreRef": {
                      "kind": "SecretStore",
                      "name": "secret-store-name"
                    },
                    "target": {
                      "creationPolicy": "Merge",
                      "name": "my-secret",
                      "template": {
                        "data": {
                          "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                        },
                        "metadata": {
                          "annotations": {},
                          "labels": {}
                        },
                        "templateFrom": [
                          {
                            "configMap": {
                              "items": [
                                {
                                  "key": "alertmanager.yaml"
                                }
                              ],
                              "name": "alertmanager"
                            }
                          }
                        ],
                        "type": "kubernetes.io/dockerconfigjson"
                      }
                    }
                  },
                  "namespaceSelector": {
                    "matchLabels": {
                      "cool": "label"
                    }
                  },
                  "refreshTime": "1m"
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ClusterSecretStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret",
                            "namespace": "example"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "ExternalSecret",
                "metadata": {
                  "annotations": {
                    "acme.org/sha": 1234
                  },
                  "labels": {
                    "acme.org/owned-by": "q-team"
                  },
                  "name": "hello-world"
                },
                "spec": {
                  "data": [
                    {
                      "remoteRef": {
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      },
                      "secretKey": "secret-key-to-be-managed"
                    }
                  ],
                  "dataFrom": [
                    {
                      "extract": {
                        "conversionStrategy": "Default",
                        "key": "provider-key",
                        "property": "provider-key-property",
                        "version": "provider-key-version"
                      }
                    },
                    {
                      "find": {
                        "conversionStrategy": "Unicode",
                        "name": {
                          "regexp": ".*foobar.*"
                        },
                        "path": "path-to-filter",
                        "tags": {
                          "foo": "bar"
                        }
                      }
                    }
                  ],
                  "refreshInterval": "1h",
                  "secretStoreRef": {
                    "kind": "SecretStore",
                    "name": "secret-store-name"
                  },
                  "target": {
                    "creationPolicy": "Merge",
                    "deletionPolicy": "Retain",
                    "name": "my-secret",
                    "template": {
                      "data": {
                        "config.yml": "endpoints:\n- https://{{ .data.user }}:{{ .data.password }}@api.exmaple.com\n"
                      },
                      "metadata": {
                        "annotations": {},
                        "labels": {}
                      },
                      "templateFrom": [
                        {
                          "configMap": {
                            "items": [
                              {
                                "key": "alertmanager.yaml"
                              }
                            ],
                            "name": "alertmanager"
                          }
                        }
                      ],
                      "type": "kubernetes.io/dockerconfigjson"
                    }
                  }
                }
              },
              {
                "apiVersion": "external-secrets.io/v1beta1",
                "kind": "SecretStore",
                "metadata": {
                  "name": "example",
                  "namespace": "example-ns"
                },
                "spec": {
                  "controller": "dev",
                  "provider": {
                    "aws": {
                      "auth": {
                        "secretRef": {
                          "accessKeyIDSecretRef": {
                            "key": "access-key",
                            "name": "awssm-secret"
                          },
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-key",
                            "name": "awssm-secret"
                          }
                        }
                      },
                      "region": "eu-central-1",
                      "role": "iam-role",
                      "service": "SecretsManager"
                    },
                    "gcpsm": {
                      "auth": {
                        "secretRef": {
                          "secretAccessKeySecretRef": {
                            "key": "secret-access-credentials",
                            "name": "gcpsm-secret"
                          }
                        }
                      },
                      "projectID": "myproject"
                    },
                    "vault": {
                      "auth": {
                        "appRole": {
                          "path": "approle",
                          "roleId": "db02de05-fa39-4855-059b-67221c5c2f63",
                          "secretRef": {
                            "key": "vault-token",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          }
                        },
                        "kubernetes": {
                          "mountPath": "kubernetes",
                          "role": "demo",
                          "secretRef": {
                            "key": "vault",
                            "name": "my-secret",
                            "namespace": "secret-admin"
                          },
                          "serviceAccountRef": {
                            "name": "my-sa",
                            "namespace": "secret-admin"
                          }
                        },
                        "tokenSecretRef": {
                          "key": "vault-token",
                          "name": "my-secret",
                          "namespace": "secret-admin"
                        }
                      },
                      "caBundle": "...",
                      "caProvider": {
                        "key": "cert-key",
                        "name": "my-cert-secret",
                        "namespace": "my-cert-secret-namespace",
                        "type": "Secret"
                      },
                      "namespace": "a-team",
                      "path": "secret",
                      "server": "https://vault.acme.org",
                      "version": "v2"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.external-secrets.io/v1alpha1",
                "kind": "OperatorConfig",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "requeueInterval": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  },
                  "concurrent": 1,
                  "controllerClass": "",
                  "crds": {
                    "createClusterExternalSecret": true,
                    "createClusterSecretStore": true
                  },
                  "createOperator": true,
                  "deploymentAnnotations": {},
                  "extraArgs": {},
                  "extraEnv": [],
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/external-secrets/external-secrets",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "installCRDs": false,
                  "leaderElect": false,
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "priorityClassName": "",
                  "processClusterExternalSecret": true,
                  "processClusterStore": true,
                  "prometheus": {
                    "enabled": false,
                    "service": {
                      "port": 8080
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "scopedNamespace": "",
                  "scopedRBAC": false,
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "certCheckInterval": "5m",
                    "certDir": "/tmp/certs",
                    "create": true,
                    "deploymentAnnotations": {},
                    "extraArgs": {},
                    "extraEnv": [],
                    "fullnameOverride": "",
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/external-secrets/external-secrets",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nameOverride": "",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "podSecurityContext": {},
                    "priorityClassName": "",
                    "prometheus": {
                      "enabled": false,
                      "service": {
                        "port": 8080
                      }
                    },
                    "rbac": {
                      "create": true
                    },
                    "replicaCount": 1,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "create": true,
                      "name": ""
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
          createdAt: "2024-12-03T10:44:19Z"
          description: Operator to configure external-secrets helm-chart based operator
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/external-secrets/external-secrets-helm-operator
          support: external-secrets
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ACRAccessToken
            name: acraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster scoped resource that allows you to manage ExternalSecrets
              across namespaces
            displayName: ClusterExternalSecret
            kind: ClusterExternalSecret
            name: clusterexternalsecrets.external-secrets.io
            version: v1beta1
          - kind: ClusterGenerator
            name: clustergenerators.generators.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1alpha1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1alpha1
          - description: Cluster-wide resource that defines how to access the external
              API
            displayName: ClusterSecretStore v1beta1
            kind: ClusterSecretStore
            name: clustersecretstores.external-secrets.io
            version: v1beta1
          - kind: ECRAuthorizationToken
            name: ecrauthorizationtokens.generators.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1alpha1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1alpha1
          - description: Describes what data should be fetched
            displayName: ExternalSecret v1beta1
            kind: ExternalSecret
            name: externalsecrets.external-secrets.io
            version: v1beta1
          - kind: Fake
            name: fakes.generators.external-secrets.io
            version: v1alpha1
          - kind: GCRAccessToken
            name: gcraccesstokens.generators.external-secrets.io
            version: v1alpha1
          - kind: GithubAccessToken
            name: githubaccesstokens.generators.external-secrets.io
            version: v1alpha1
          - description: Configures external-secrets helm-chart based operator
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.external-secrets.io
            version: v1alpha1
          - kind: Password
            name: passwords.generators.external-secrets.io
            version: v1alpha1
          - kind: PushSecret
            name: pushsecrets.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1alpha1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1alpha1
          - description: Namespaced resource that defines how to access the external
              API
            displayName: SecretStore v1beta1
            kind: SecretStore
            name: secretstores.external-secrets.io
            version: v1beta1
          - kind: STSSessionToken
            name: stssessiontokens.generators.external-secrets.io
            version: v1alpha1
          - kind: UUID
            name: uuids.generators.external-secrets.io
            version: v1alpha1
          - kind: VaultDynamicSecret
            name: vaultdynamicsecrets.generators.external-secrets.io
            version: v1alpha1
          - kind: Webhook
            name: webhooks.generators.external-secrets.io
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK (Helm version) to configure **[official external-secrets operator helm chart](https://github.com/external-secrets/external-secrets)**, so it can be installed via OLM without having to do any change on current Helm Charts.

          Before any other resources provided by this Operator can be deployed, it is essential to create an
          OperatorConfig resource.

          The usual Helm Chart file `values.yaml`, like:
          ```yaml
          prometheus:
            enabled: true
            service:
              port: 8080
          resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          need to be encapsulated into a new custom resource called `OperatorConfig`:
          ```yaml
          apiVersion: operator.external-secrets.io/v1alpha1
          kind: OperatorConfig
          metadata:
            name: cluster
          spec:
            prometheus:
              enabled: true
              service:
                port: 8080
            resources:
             requests:
               cpu: 10m
               memory: 96Mi
             limits:
               cpu: 100m
               memory: 256Mi
          ```

          Once the OperatorConfig resource is deployed, the operator will create all helm chart resources, using the custom resource name as a preffix for all resources names, like for example a `Deployment` called `cluster-external-secrets`.

          ## Documentation

          * [Upstream Documentation](https://external-secrets.io)
          * [Helm Operator Install](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/install.md)
          * [Helm Operator Development](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/development.md)
          * [Helm Operator Release](https://github.com/external-secrets/external-secrets-helm-operator/blob/main/docs/release.md)

          ## Contributing

          You can contribute by:

          * Raising any issues you find using External Secrets Operator
          * Fixing issues by opening [Pull Requests](https://github.com/external-secrets/external-secrets-helm-operator/pulls)
          * Submitting a patch or opening a PR
          * Improving documentation
          * Talking about External Secrets Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/external-secrets/external-secrets-helm-operator/issues).

          ## License

          External Secrets Operator is under Apache 2.0 license. See the [LICENSE](LICENSE) file for details.
        displayName: External Secrets Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secrets
        - secrets-manager
        - vault
        - aws
        - azure
        - google
        - ibm
        - akeyless
        - yandex
        - gitlab
        - oracle
        links:
        - name: External Secrets Operator
          url: https://external-secrets.io
        - name: GitHub
          url: https://github.com/external-secrets/external-secrets
        - name: Slack
          url: https://kubernetes.slack.com/messages/external-secrets
        maintainers:
        - email: contact@external-secrets.io
          name: external-secrets community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: External Secrets
          url: https://external-secrets.io
        relatedImages:
        - ghcr.io/external-secrets/external-secrets-helm-operator@sha256:e7939e97e11e26d7368d9231c5dbf134139b29580badd1dcc7de0f621ff4e3b6
        - quay.io/community-operator-pipeline-prod/external-secrets-operator@sha256:6fa2ddca55d1d134b0938b47ea8a861de7408fa566ad806b200a379b8500a2a1
        version: 0.11.0
      entries:
      - name: external-secrets-operator.v0.11.0
        version: 0.11.0
      - name: external-secrets-operator.v0.10.7
        version: 0.10.7
      - name: external-secrets-operator.v0.10.6
        version: 0.10.6
      - name: external-secrets-operator.v0.10.5
        version: 0.10.5
      - name: external-secrets-operator.v0.10.4
        version: 0.10.4
      - name: external-secrets-operator.v0.10.3
        version: 0.10.3
      - name: external-secrets-operator.v0.10.2
        version: 0.10.2
      - name: external-secrets-operator.v0.10.1
        version: 0.10.1
      - name: external-secrets-operator.v0.10.0
        version: 0.10.0
      - name: external-secrets-operator.v0.9.20
        version: 0.9.20
      - name: external-secrets-operator.v0.9.19
        version: 0.9.19
      - name: external-secrets-operator.v0.9.18
        version: 0.9.18
      - name: external-secrets-operator.v0.9.17
        version: 0.9.17
      - name: external-secrets-operator.v0.9.16
        version: 0.9.16
      - name: external-secrets-operator.v0.9.15-2
        version: 0.9.15-2
      - name: external-secrets-operator.v0.9.14
        version: 0.9.14
      - name: external-secrets-operator.v0.9.13
        version: 0.9.13
      - name: external-secrets-operator.v0.9.12
        version: 0.9.12
      - name: external-secrets-operator.v0.9.11
        version: 0.9.11
      - name: external-secrets-operator.v0.9.10
        version: 0.9.10
      - name: external-secrets-operator.v0.9.9
        version: 0.9.9
      - name: external-secrets-operator.v0.9.8
        version: 0.9.8
      - name: external-secrets-operator.v0.9.7
        version: 0.9.7
      - name: external-secrets-operator.v0.9.6
        version: 0.9.6
      - name: external-secrets-operator.v0.9.5
        version: 0.9.5
      - name: external-secrets-operator.v0.9.4
        version: 0.9.4
      - name: external-secrets-operator.v0.9.3
        version: 0.9.3
      - name: external-secrets-operator.v0.9.2
        version: 0.9.2
      - name: external-secrets-operator.v0.9.1
        version: 0.9.1
      - name: external-secrets-operator.v0.9.0
        version: 0.9.0
      - name: external-secrets-operator.v0.8.7
        version: 0.8.7
      - name: external-secrets-operator.v0.8.6
        version: 0.8.6
      - name: external-secrets-operator.v0.8.5
        version: 0.8.5
      - name: external-secrets-operator.v0.8.4
        version: 0.8.4
      - name: external-secrets-operator.v0.8.3
        version: 0.8.3
      - name: external-secrets-operator.v0.8.2
        version: 0.8.2
      - name: external-secrets-operator.v0.8.1
        version: 0.8.1
      - name: external-secrets-operator.v0.8.0
        version: 0.8.0
      - name: external-secrets-operator.v0.7.2
        version: 0.7.2
      - name: external-secrets-operator.v0.7.1
        version: 0.7.1
      - name: external-secrets-operator.v0.7.0
        version: 0.7.0
      - name: external-secrets-operator.v0.7.0-rc1
        version: 0.7.0-rc1
      - name: external-secrets-operator.v0.6.1
        version: 0.6.1
      - name: external-secrets-operator.v0.6.0
        version: 0.6.0
      - name: external-secrets-operator.v0.5.9
        version: 0.5.9
      - name: external-secrets-operator.v0.5.8
        version: 0.5.8
      - name: external-secrets-operator.v0.5.7
        version: 0.5.7
      - name: external-secrets-operator.v0.5.6
        version: 0.5.6
      - name: external-secrets-operator.v0.5.5
        version: 0.5.5
      - name: external-secrets-operator.v0.5.4
        version: 0.5.4
      - name: external-secrets-operator.v0.5.3
        version: 0.5.3
      - name: external-secrets-operator.v0.5.2
        version: 0.5.2
      - name: external-secrets-operator.v0.5.1
        version: 0.5.1
      - name: external-secrets-operator.v0.5.0
        version: 0.5.0
      - name: external-secrets-operator.v0.4.4
        version: 0.4.4
      - name: external-secrets-operator.v0.4.2
        version: 0.4.2
      - name: external-secrets-operator.v0.4.1
        version: 0.4.1
      - name: external-secrets-operator.v0.3.11
        version: 0.3.11
      - name: external-secrets-operator.v0.3.10
        version: 0.3.10
      - name: external-secrets-operator.v0.3.9
        version: 0.3.9
      name: stable
    defaultChannel: alpha
    packageName: external-secrets-operator
    provider:
      name: External Secrets
      url: https://external-secrets.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Strimzi
      provider-url: ""
    name: strimzi-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: stable
    - currentCSV: strimzi-cluster-operator.v0.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "whitelist":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": 1,
                        "offset.storage.replication.factor": 1,
                        "status.storage.replication.factor": 1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "NetworkOutboundCapacityGoal",
                    "CpuCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
          createdAt: "2021-05-14 11:37:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  Starting with Strimzi 0.23, we support only the `v1beta2` API version of all Strimzi custom resources.  This is a required for migration to `apiextensions/v1` which is needed because Kubernetes 1.22 will remove support for `apiextensions/v1beta1`.  Migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs **before upgrading to Strimzi 0.23**. For more details about the CRD upgrades, see the [documentation](https://strimzi.io/docs/operators/0.23.0/deploying.html#assembly-upgrade-resources-str).
          ### New in 0.23
          * Migrate to CRD v1 (required by Kubernetes 1.22+)
          * Strimzi API versions `v1alpha1` and `v1beta1` were removed from all Strimzi custom resources apart from `KafkaTopic` and `KafkaUser` (use `v1beta2` versions instead)
          * Add support for Kafka 2.8.0 and 2.6.2, remove support for Kafka 2.5.x
          * Make it possible to configure maximum number of connections and maximum connection creation rate in listener configuration
          * Add support for configuring finalizers for `loadbalancer` type listeners
          * Use dedicated Service Account for Kafka Connect Build on Kubernetes
          * Remove direct ZooKeeper access for handling user quotas in the User Operator. Add usage of Admin Client API instead.
          * Support for configuring custom Authorizer implementation
          * Changed Reconciliation interval for Topic Operator from 90 to 120 seconds (to keep it the same as for other operators)
          * Changed Zookeeper session timeout default value to 18 seconds for Topic and User Operators (for improved resiliency)
          * Removed requirement for replicas and partitions KafkaTopic spec making these parameters optional
          * Support to configure a custom filter for parent CR's labels propagation into subresources
          * Allow disabling service links (environment variables describing Kubernetes services) in Pod template
          * Add support for separate control plane listener (disabled by default, available via the `ControlPlaneListener` feature gate)
          * Support for Dual Stack networking
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/master/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/master/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:699bff48ee445d8ba381a791cd0340015e3ecede8d169f41bc151095ab7fb5cc
        - quay.io/strimzi/kafka@sha256:d2745f48ed262290ce66c3362ffb9b9a8cfeb9c76beac40c995f72f3f06ad10e
        - quay.io/strimzi/kafka@sha256:dbc53bb374d2cb9513a5bb8f80fea751c91b249cabd36d8b3c04af5e5c961738
        - quay.io/strimzi/kafka@sha256:e90695b35d6dd5db20c9900dd0dab95d1ef908f3145b4f9f436b903303c88549
        - quay.io/strimzi/jmxtrans@sha256:b4ec1844f4cc5eb74cd0aac83d74b81b977e178bf670d1136d1af0b68663d0a3
        - quay.io/strimzi/kafka@sha256:641f46114c59d61577bcb3f4ef5af16aea3ed42c2a575f8d3641c82848df85e8
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka-bridge@sha256:bed07186355a4c04ee50e90048e3bd3a023264093a9ef28f0278cce44640e93f
        - quay.io/strimzi/operator@sha256:59d939f342f96d0e8a54c96fd187832c4b19cd4137bf55ace53d486728ae6c5f
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:68cdc37990bc7da8a6bbe3a4908378c8026e5ff980bf8c44c0277f689006fc51
        version: 0.23.0
      entries:
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.23.x
    - currentCSV: strimzi-cluster-operator.v0.24.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnectS2I",
                "metadata":{
                  "name":"my-connect-s2i-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 2,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster with Source 2 Image support
            displayName: Kafka Connect Source to Image
            kind: KafkaConnectS2I
            name: kafkaconnects2is.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.24 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.24 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.24\n* Add support
          for Kafka 2.7.1 and remove support for 2.6.0, 2.6.1, and 2.6.2\n* Add support
          for [Kubernetes Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-kubernetes-config-provider)\n*
          Use Red Hat UBI8 base image\n* Support for patching of service accounts
          and configuring their labels and annotations. The feature is disabled by
          default and enabled using the new `ServiceAccountPatching` feature gate.\n*
          Added support for configuring cluster-operator's worker thread pool size
          that is used for various sync and async tasks\n* Add Kafka Quotas plugin
          with produce, consume, and storage quotas\n* Support pausing reconciliation
          of KafkaTopic CR with annotation `strimzi.io/pause-reconciliation`\n* Update
          cruise control to 2.5.57\n* Update to Strimzi Kafka Bridge to 0.20.1\n*
          Support for broker load information added to the rebalance optimization
          proposal. Information on the load difference, before and after a rebalance
          is stored in a ConfigMap\n* Add support for selectively changing the verbosity
          of logging for individual CRs, using markers.\n* Added support for `controller_mutation_rate'
          quota. Creation/Deletion of topics and creation of partitions can be configured
          through this.\n* Use newer version of Kafka Exporter with different bugfixes
          \n### Supported Features\n* **Manages the Kafka Cluster** - Deploys and
          manages all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for morroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashabords\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\n\nAll bugs, tasks or enhancements are tracked as
          [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues).
          Issues which  might be a good start for new contributors are marked with
          [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\n\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\n\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\n\nIf you want to
          get in touch with us first before contributing, you can use:\n* [Strimzi
          mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:8959b7968ab8b3306906cdbff2ebb8d63329af37e58124a601843795c4ef5bd6
        - quay.io/strimzi/kafka@sha256:95cfe9000afda2f7def269ca46472d3803ee85146c521884884d8505a7187daf
        - quay.io/strimzi/kafka@sha256:fbb08410d9595029bc4a02ed859971264e6ce2dc85dd6a9855eaa7bb58b52a25
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:786ea69c77e0651f009a73a8b3ce0b3d41b3cd71b917dc0d9bdc61c725f494a6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:97762fcd8bfeec64fe4d3d43a6e60bc705ff2c8773354357e545adf31c8f85c5
        - quay.io/strimzi/jmxtrans@sha256:2c54adabf6f31c087dbb898057f9c5e200aaf4c5b39fb3a47bbaf7f6587df72d
        - quay.io/strimzi/kafka-bridge@sha256:537ea10e87619251204747ebc1005c632a6c6ed270f5ef06f86c0c3ae4a7553b
        version: 0.24.0
      entries:
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.24.x
    - currentCSV: strimzi-cluster-operator.v0.25.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"2.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"2.8",
                        "inter.broker.protocol.version":"2.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"2.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"2.8.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster with S2I support"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
          createdAt: "2021-06-24 13:13:13"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.25 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.25
          * Move from Scala 2.12 to Scala 2.13
          * Open Policy Agent authorizer updated to a new version supporting Scala 2.13.
            The new version of the authroizer uses different format of the input data sent to the OPA server and might require you to update your policies.
            See [release notes](https://github.com/strimzi/strimzi-kafka-operator/releases/tag/0.25.0) for more details.

          * Allow a custom password to be set for SCRAM-SHA-512 users by referencing a secret in the `KafkaUser` resource
          * Add support for [EnvVar Configuration Provider for Apache Kafka](https://github.com/strimzi/kafka-env-var-config-provider)
          * Add support for `tls-external` authentication to User Operator to allow management of ACLs and Quotas for TLS users with user certificates generated externally
          * Support for disabling the automatic generation of network policies by the Cluster Operator
          * Configure fixed size limit for `emptyDir` volumes used for temporary files
          * Update Strimzi Kafka Bridge to 0.20.2
          * The `KafkaConnectS2I` resource has been removed and is no longer supported by the operator.
            Please use the [migration guide](https://strimzi.io/docs/operators/0.24.0/full/using.html#proc-migrating-kafka-connect-s2i-str) to migrate your `KafkaConnectS2I` deployments to [`KafkaConnect` Build](https://strimzi.io/docs/operators/latest/full/deploying.html#creating-new-image-using-kafka-connect-build-str) instead.

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ee17eb3611019b0cfa5e9e196610edc1dbd7084471fb20990f055c89da8eef45
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/operator@sha256:afaf1dd1f2f696ae80674d74159661ea239ae609f78a73a1423c7a17a39e5044
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:4fa7a30eb3f679a495aeb96794056ffd003237686b2b60e693c8cabf6bc15bae
        - quay.io/strimzi/jmxtrans@sha256:af5eb8459ac9a0acbca51e67c6e0217f62ba404cba9f2251968c927dcef75252
        - quay.io/strimzi/kafka-bridge@sha256:a005e20d7fbe04d6f587f2093e21a00556ec62b850df757b9f7d9d57e3141c40
        - quay.io/strimzi/kafka@sha256:56a33815f7d50a97ea9bb4cce1e3a6d04d918b841b5aa285f16b792d8930059e
        - quay.io/strimzi/kafka@sha256:d8991608ec71f17e850c33714ee02ca87ba94bc0f077be62ad1ed127ca244823
        version: 0.25.0
      entries:
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.25.x
    - currentCSV: strimzi-cluster-operator.v0.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
          createdAt: "2021-12-14 00:16:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.26 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.26.1
          * **Updated Log4j2 to fix CVE-2021-44228**
          ### New in 0.26.0
          * Add support for Kafka 2.8.1 and 3.0.0; remove Kafka 2.7.0 and 2.7.1
          * Add support for exposing JMX in Zookeeper
          * Allow configuring labels and annotations for JMX authentication secrets
          * Enable Cruise Control anomaly.detection configurations
          * Add support for building connector images from the Maven coordinates
          * Allow Kafka Connect Build artifacts to be downloaded from insecure servers
          * Add option to specify pull secret in Kafka Connect Build on OpenShift
          * Configurable authentication, authorization, and SSL for Cruise Control API
          * Allow to configure `/tmp` volume size via Pod template. By default `1Mi` is used.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6dce519730a6eb73bc18bef8372cdde65c29af3cb52eb78afc808d8582526fd3
        - quay.io/strimzi/kafka@sha256:bd22a134b0c351ddfc78a3263850c2f16733c52f57ac92984cbe0ffff477c724
        - quay.io/strimzi/jmxtrans@sha256:bff5d3bab575d0b9ed95dd79d2d22518cc00c4d1c9dbb5614ed16149d4920a1b
        - quay.io/strimzi/kafka-bridge@sha256:d3899c79747e4cb3a6df5a9fcd80b81f038ca0634a97329c1d160499469876dd
        - quay.io/strimzi/kafka@sha256:6367b0fbba9593e3ac736040ea2d7ef71296b1ec144d3b54d9370352141b90e4
        - quay.io/strimzi/kaniko-executor@sha256:b0ca0b4fb16822e55ea1250c54c73c7c4248282908065422ededa49fa400436a
        - quay.io/strimzi/kafka@sha256:8c022a8535742b2eea691737f863af5ddef1e8df08b415b8139ecd02a79052a5
        - quay.io/strimzi/maven-builder@sha256:663eb81388ae8f824e7920c272f6d2e2274cf6c140d61416607261cdce9d50e2
        - quay.io/strimzi/operator@sha256:5d90caa66f57939ba62a9a50efa00fce319ba596e719a3f8a42170d7635ee3f4
        version: 0.26.1
      entries:
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.26.x
    - currentCSV: strimzi-cluster-operator.v0.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.0.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "log.message.format.version":"3.0",
                        "inter.broker.protocol.version":"3.0"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.0.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.0.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
          createdAt: "2022-01-14 20:00:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.27 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.27.1
          * Updated Log4j2 to version 2.17.1 and Fabric8 Kubernetes Client to 5.10.2
          ### New in 0.27.0
          * Updated Log4j2 to version 2.17.0 (adresses CVE-2021-44228, CVE-2021-45046, CVE-2021-45105)
          * (Experimental) support for AArch64 / ARM64 architecture
          * `ControlPlaneListener` and `ServiceAccountPatching` feature gates are now in the beta phase and are enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/operator@sha256:9ea49407a04006d0bddfaaee822adfb95f5857af32ffcee81ba60ce891b07ba8
        - quay.io/strimzi/maven-builder@sha256:9acd178cc504be8acaa70a39f3c17830aae1f0e93a656110790f9476b09260d5
        - quay.io/strimzi/kafka@sha256:a3ec567ca529df2fa41f8986d21e86dde66be94aa4baa56faa66db4dcb3b258e
        - quay.io/strimzi/kafka-bridge@sha256:55a52316aead02efa2004174e11ad4728bb8b6f1c1415f725f597f1d2f2da188
        - quay.io/strimzi/kafka@sha256:5682a6bfcc4ee13d53a30d587c8f96bc40abe3618880989336c5af274ea3cefa
        - quay.io/strimzi/kafka@sha256:f4d68bb94447c6612f70de7f6587e9e7ef712c83769ea1f11a8dbda0d241a059
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:9f2c01e88b29154ba7a0f70a1c69a49c86a567808cf057cb5d115159278cf2cf
        - quay.io/strimzi/jmxtrans@sha256:d2d2477517eb1ac98b51ad6d64d8c1c578141637afa8d46037076f0013183c0b
        version: 0.27.1
      entries:
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.27.x
    - currentCSV: strimzi-cluster-operator.v0.28.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.1.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.1"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.1.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.1.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
          createdAt: "2022-02-12 23:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.28 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.28.0
          * Support for Kafka 3.1.0
          * Support for `StrimziPodSet` resources (disabled by default through the `UseStrimziPodSets` feature gate)
          * Support custom authentication mechanisms in Kafka listeners
          * Intra-broker disk balancing using Cruise Control
          * Add support for disabling the FIPS mode in OpenJDK
          * Fix renewing your own CA certificates
          * Added the option `createBootstrapService` to disable the creation of the bootstrap service for the Load Balancer Type Listener
          * Dependency updates (Strimzi Kafka Bridge, Strimzi OAuth, Open Policy Agent Authroizer plugin and others)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for morroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributes partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashabords
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.

          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.

          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.

          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:c345c783d8fbbee4edeca2ebc47e7c454fcc2e26686820b8037de97369280ba0
        - quay.io/strimzi/kafka@sha256:06e9bda15495feb2b92d9950e834cdeed6ef20064f021a879610f1b59dd46950
        - quay.io/strimzi/kafka@sha256:7c04f65e216bb941458aeadffb85e9eeee43d5ca67da566017c2c93a0664c98e
        - quay.io/strimzi/kaniko-executor@sha256:a07ebe6600d6baba09980abdde2931b3b6abd2f75aeb2cac5c6a2d22a37aa807
        - quay.io/strimzi/maven-builder@sha256:3d5d664e089651944712361a6d8ba42b578dbd9bb3dfd9972f614520c9cfab75
        - quay.io/strimzi/operator@sha256:8b010c5fb8244b742a37f264feefc116d75bb2abd2d6895375cbeea250727b39
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:edec37e045212f74d160724ffc7b2c8ccf4e60eb95629aae80360494d78d5433
        - quay.io/strimzi/jmxtrans@sha256:3a55a81790dc88b90d1bf858e2cfd7e591ad0e79f763c6d1667b2d3eb0d0c79d
        version: 0.28.0
      entries:
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.28.x
    - currentCSV: strimzi-cluster-operator.v0.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
          createdAt: "2022-05-15 00:01:07"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.29 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.29.0
          * Add support for Apache Kafka 3.0.1, 3.1.1 and 3.2.0
          * Renew user certificates in User Operator only during maintenance windows
          * Remove TLS sidecar from Cruise Control pod. Cruise Control is now configured to not using ZooKeeper, so the TLS sidecar is not needed anymore.
          * Allow Cruise Control topic names to be configured
          * Add support for `spec.rack.topologyKey` property in Mirror Maker 2 to enable "fetch from the closest replica" feature.
          * Experiemntal support for the s390x platform
          * Added new rebalancing modes on the `KafkaRebalance` custom resource to add or remove brokers
          * Experimental KRaft mode (ZooKeeper-less Kafka). Use it for development and testing only!
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:9dfc09273ad8fc4bdc75cbe93b4a061a05a5defb5884a221ae1e178ba602787f
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/operator@sha256:afde94c6d2544207cb7f3c94b54455643558a9d5daaa13b9f736cc8da20e319c
        - quay.io/strimzi/maven-builder@sha256:0c096a59faf8fd224298c01415a895043db4647534c060cb7e53207e9448bf65
        - quay.io/strimzi/kafka@sha256:95678e843bc87cbe571f3f299045dd075725816b59e1b87821cd6025876ec559
        - quay.io/strimzi/jmxtrans@sha256:f89e3c9618430908a8b390f2490f5aa43687b12889b434472a962ce0170c6463
        - quay.io/strimzi/kafka@sha256:284cde40784b6ff7bfb81b6dd28928e1e8c598fe6a72504f025592e5b43fa94c
        - quay.io/strimzi/kafka@sha256:935caab03f1d0365b47ddefec0c1d8fa449fc3de715ccdd298aedef5ce968c73
        - quay.io/strimzi/kafka@sha256:ac190141120c1800cc7131d3a1f701e64efdcd4a5d87f0b2e0e2cadfdef8b438
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:465f0bef1f167b1d16ba9ae5f2d9a1133d11786b6809698a84c5c962387dbf6f
        - quay.io/strimzi/kafka-bridge@sha256:b7bc3dd90e87223766f608a0087b521f0bba79d1c21eaa833bc2708c71148b7f
        version: 0.29.0
      entries:
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.29.x
    - currentCSV: strimzi-cluster-operator.v0.30.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
          createdAt: "2022-07-16 00:00:10"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org)
          cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/)
          in various deployment configurations. See our [website](https://strimzi.io)
          for more details about the project.\n### CRD Upgrades\n**!!! IMPORTANT !!!**
          \ This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.
          \ If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed
          for all Strimzi CRDs and CRs before the upgrade to 0.30 is done!  If upgrading
          from Strimzi version earlier than 0.22, you need to first install the CRDs
          from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi
          CRDs and CRs before the upgrade to 0.30 is done!  For more details about
          the CRD upgrades, see the documentation.\n### New in 0.30.0\n* The `UseStrimziPodSets`
          feature gate moves to beta stage. By default, StrimziPodSets are used instead
          of StatefulSets. If needed, `UseStrimziPodSets` can be disabled in the feature
          gates configuration in the Cluster Operator.\n* The `ServiceAccountPatching`
          feature gate moves to GA. It cannot be disabled anymore and will be permanently
          enabled.\n* Remove support for Kafka 3.0.0 and 3.0.1\n* Add support for
          `simple` authorization and for the User Operator to the experimental `UseKRaft`
          feature gate. _(Note: Due to [KAFKA-13909](https://issues.apache.org/jira/browse/KAFKA-13909),
          broker restarts currently don't work when authorization is enabled.)_\n*
          Add CPU and network capacity overrides for Cruise Control capacity config\n*
          Use better encryption and digest algorithms when creating the PKCS12 stores.
          For existing clusters, the certificates will not be updated during upgrade
          but only next time the PKCS12 store is created. \n* Operator emits Kafka
          events to explain why it restarted a Kafka broker\n* Better configurability
          of the Kafka Admin client in the User Operator\n* Update Strimzi Kafka Bridge
          to 0.21.6\n### Supported Features\n* **Manages the Kafka Cluster** - Deploys
          and manages all of the components of this complex application, including
          dependencies like Apache ZooKeeper® that are traditionally hard to administer.\n*
          **Includes Kafka Connect** - Allows for configuration of common data sources
          and sinks to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          Strimzi Operator understands how to run and upgrade between a set of Kafka
          versions. When specifying a new version in your config, check to make sure
          you aren't using any features that may have been removed. See [the upgrade
          guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str)
          for more information.\n### Storage\nAn efficient data storage infrastructure
          is essential to the optimal performance of Apache Kafka®. Apache Kafka®
          deployed via Strimzi requires block storage. The use of file storage (for
          example, NFS) is not recommended.\nThe Strimzi Operator supports three types
          of data storage:\n* Ephemeral (Recommended for development only!)\n* Persistent\n*
          JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nStrimzi
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation to the current _main_
          branch as well as all releases can be found on our [website](https://strimzi.io/documentation).\n###
          Getting help\nIf you encounter any issues while using Strimzi, you can get
          help using:\n* [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n*
          [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          Contributing\nYou can contribute by:\n* Raising any issues you find using
          Strimzi\n* Fixing issues by opening Pull Requests\n* Improving documentation\n*
          Talking about Strimzi\nAll bugs, tasks or enhancements are tracked as [GitHub
          issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues
          which  might be a good start for new contributors are marked with [\"good-start\"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start)
          label.\nThe [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md)
          describes how to build Strimzi and how to  test your changes before submitting
          a patch or opening a PR.\nThe [Documentation Contributor Guide](https://strimzi.io/contributing/guide/)
          describes how to contribute to Strimzi documentation.\nIf you want to get
          in touch with us first before contributing, you can use:\n* [Strimzi mailing
          list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)\n* [Strimzi
          Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)\n###
          License\nStrimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).\n"
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:77fc68da171c74d3a4b75b1030eaaa5eae12e200002f7aa2ee1969385eba1d66
        - quay.io/strimzi/kafka@sha256:d9258ddd453c3a53ca8e737b989d27725818da95f7a4fc879c9f07fe743c0cef
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8be098242f971aefa61a4a24d2c6630ecaa7856ece842166c38887a66bce1bda
        - quay.io/strimzi/kaniko-executor@sha256:544c4987f7f3ca558b04e9ef4357be204f088212b7b5b9ed0f909401a8bc5238
        - quay.io/strimzi/maven-builder@sha256:f2cb45234f3a7a010824a370d03032b706271cd0d1d260da929f046f9259d7e0
        - quay.io/strimzi/jmxtrans@sha256:de1ca88b713b926b72df535aca2cfad24fb1b2bf40481065d6fa0c421b872320
        - quay.io/strimzi/operator@sha256:587db6f4389f56209f48853beedbfe249f6d590a4cb0000be2c42b57e43c2adb
        - quay.io/strimzi/kafka@sha256:5d881605425ab0b37d0b9cac6957d00ba904df0db185ca6181c810151bdab8e7
        - quay.io/strimzi/kafka@sha256:68d6f4cb5cc5ccad44de799bd9cf417c469a09e1024dbe42298942f25a09e004
        version: 0.30.0
      entries:
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.30.x
    - currentCSV: strimzi-cluster-operator.v0.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
          createdAt: "2022-09-20 21:07:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          **This is the last version of Strimzi with support for Kubernetes 1.16, 1.17 and 1.18!**
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from Strimzi 0.22, migration to `v1beta2` needs to be completed for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  If upgrading from Strimzi version earlier than 0.22, you need to first install the CRDs from Strimzi 0.22 and complete the migration to `v1beta2` for all Strimzi CRDs and CRs before the upgrade to 0.31 is done!  For more details about the CRD upgrades, see the documentation.
          ### New in 0.31.1
          * Kafka 3.1.2 and 3.2.3 (fixes CVE-2022-34917)
          * Make `sasl.server.max.receive.size` broker option user configurable
          * Update Strimzi Kafka Bridge to 0.22.1
          ### New in 0.31.0
          * Support for Apache Kafka 3.2.1
          * Pluggable Pod Security Profiles with built-in support for _restricted_ Kubernetes Security Profile
          * Support for leader election and running multiple operator replicas (1 active leader replicas and one or more stand-by replicas)
          * Support for IPv6 addresses being used in Strimzi issued certificates
          * `StrimziPodSet` reconciliation metrics
          * Updated Strimzi Kafka Bridge to 0.22.0 and Kafka Exporter to 1.6.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:6e5e207c70c25c595d3ce60268c0214c7518e95836c9b1212ef1eb2f008c7a76
        - quay.io/strimzi/jmxtrans@sha256:52dd107062d78c29bb58c905ac654c39df72bbe3c25613023aa33e834930606d
        - quay.io/strimzi/kaniko-executor@sha256:06581fb615f0b13ddfe27fae26ba5836b12c6e28afd55a81cba7685c9ca4bf79
        - quay.io/strimzi/maven-builder@sha256:670ca70900f2229ef140a592b77f2107934601f8b050b79250eab17ff62099b6
        - quay.io/strimzi/kafka@sha256:03cc48bc5cf367f898338665b253855bed7d6096fe577adfe344da2f2cddcfe4
        - quay.io/strimzi/kafka@sha256:b7c059bd72d8b063a96dde788d91d1ad87498bad3e73acba600929d595aeb7d6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:6afbc7ffc98918e3a557b0b3d9b2c3aa80caa47b9a73205fac9e6572c66f4df7
        - quay.io/strimzi/kafka@sha256:652924a1dba46aa716caba8381f1e791c09035e31ce62fdcc81f9b69af36cd02
        - quay.io/strimzi/kafka@sha256:9553aae22d35448681cbbf7a831a28149fb159f3df58ccf8c8ec4ca7eb9c1fdf
        - quay.io/strimzi/kafka-bridge@sha256:3ddac62a4f8c8a3ebb72d5841d3826f7fd3fbb0e00eb75fb40bced0a576b86a9
        - quay.io/strimzi/kafka@sha256:430b3eb70e47a91d7b5b6f230d5be024c2de853099cc0437e219b443e0ead9f5
        - quay.io/strimzi/kafka@sha256:83acc1e2cc1e64a88bdbb54b9818a6ad3658c3da60e72c4be365f824c5fcf57c
        version: 0.31.1
      entries:
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.31.x
    - currentCSV: strimzi-cluster-operator.v0.32.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
          createdAt: "2022-10-31 17:05:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **From the 0.32.0 release, Strimzi supports only Kubernetes 1.19 and newer! Kubernetes 1.16, 1.17 and 1.18 are not supported anymore.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.32.0/deploying.html#assembly-upgrade-str).
          ### New in 0.32.0
          * Add support for Kafka 3.3.1 and remove support for Kafka 3.1.0, 3.1.1, and 3.1.2
          * Update KafkaConnector CR status so the 'NotReady' condition is added if the connector or any tasks are reporting a 'FAILED' state.
          * Add auto-approval mechanism on KafkaRebalance resource when an optimization proposal is ready
          * The ControlPlaneListener feature gate moves to GA
          * Add client rack-awareness support to Strimzi Bridge pods
          * Add support for OpenTelemetry for distributed tracing
          * ZookeeperRoller considers unready pods
          * Support multiple operations per ACLRule
          * Add new listener type "cluster-ip" which is using per-broker services instead of the pod DNS names
          * Update Strimzi OAuth library to 0.11.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/operator@sha256:c5e0e5dca7503a8371b8d366db0f49e3fc68f62ea480262e39b6cc277991c97c
        - quay.io/strimzi/kafka@sha256:680ae1958dbcb9da8ee4128a67c1163a6ee2744221f7d29f73d8bcc237fd0173
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:593e9980b786ca8ad3223151c3e17f16c7652ee0e86d722af4c34b023038ecc2
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:eefdd05485167d490f0011d9e5d2bb6112db1daea25940c79e1f55b3af254548
        - quay.io/strimzi/jmxtrans@sha256:208928133b2854f60bb34912e07561a3511de895b815960e73d565be2d553918
        - quay.io/strimzi/kafka@sha256:8c9f38b45dd814e617c9c51cc90fd51e23d7dd23e8efe7e9eff013522bd6a5a0
        - quay.io/strimzi/kafka@sha256:bb97008b7f928ed98939b2008a7baf8909c88e726a59aaf607f2eeba55375cc0
        - quay.io/strimzi/kafka-bridge@sha256:aad0cee4381cc1c8c6638ab5732b135acd271b5889486ca9a33312cce67e4a02
        - quay.io/strimzi/kafka@sha256:4bacf5b54041bd0ab9199c8c5d51e8058758139db4b7abc380cf0400737716ef
        version: 0.32.0
      entries:
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.32.x
    - currentCSV: strimzi-cluster-operator.v0.33.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
          createdAt: "2023-02-18 21:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.33 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.33.2/deploying.html#assembly-upgrade-str).
          ### New in 0.33
          _Note: Strimzi 0.33.2 add support for Kafka 3.4.0 which fixes CVE-2023-25194._
          * Add support for Kafka 3.3.2
          * Automatic restarts of failed Connect or Mirror Maker 2 connectors
          * Redesign of Strimzi User Operator to improve its scalability
          * Move to Java 17
          * Improved FIPS support
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3a380d5f3ea509f90cdf95280d83b42f5de4f86ee3a512e02234e7639f50c9eb
        - quay.io/strimzi/kafka@sha256:116cf543fbadadd0b45007ce24d44a0c79a62280e8faa1fc80df40291fbdf05d
        - quay.io/strimzi/kafka@sha256:30945aecb9d63cf8863c672222f130469c89aefa07078c95aaed52c40fd16146
        - quay.io/strimzi/kafka@sha256:40a847f8455ec6ea9ff9fe1fff20b1ed01f848326dc13977d13ac911f0cb40b7
        - quay.io/strimzi/kafka@sha256:9eca149cfc436580e3cdd6cee0ca318366b8d0ec53c2cf007e4e2c614b988921
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:05e66583757da895fd23c3af9830ae1ac6a84d107107e5b6bbaece57e4d79c9a
        - quay.io/strimzi/kafka@sha256:301dc52ed5891ed4fa33e79ca3c344b1ac0eec7469dd46a5a7ac22064e053fc2
        - quay.io/strimzi/jmxtrans@sha256:5ebdafa7a5e3718ed3274d6254c9e655613c272340094b997b1c97fde2598b4a
        - quay.io/strimzi/kafka@sha256:fe9d3f8b029ff32d5d725e57384592914957314d63e92fda3abf4028fe4e66f6
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/maven-builder@sha256:de9e51365e7acc63feae745618804e6286fd730f41c3b31582cbb7319a93b483
        - quay.io/strimzi/operator@sha256:cf0d41fdb99b20e91063e23eb1d664817c18866c9a2f5c4430eac897e78d5d53
        version: 0.33.2
      entries:
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.33.x
    - currentCSV: strimzi-cluster-operator.v0.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
          createdAt: "2023-03-17 13:02:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.34 supports only Kubernetes 1.19 and newer! Kubernetes versions 1.16, 1.17 and 1.18 are not supported anymore since Strimzi 0.32.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.34.0/deploying.html#assembly-upgrade-str).
          ### New in 0.34
          * Add support for Apache Kafka 3.4.0 (fixes CVE-2023-25194) and remove support for Apache Kafka 3.2.x
          * Stable Pod identities for Kafka Connect and MirrorMaker 2 (disabled by default)
          * Use JDK HTTP client in the Kubernetes client instead of the OkHttp client
          * Add truststore configuration for HTTPS connections to Open Policy Agent server
          * Update Strimzi HTTP Bridge to 0.25.0 and strimzi Oauth to 0.12.0
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/maven-builder@sha256:f781b6b857d437c2de62585462e97b9fae357912c9f6711c1df39f431278878d
        - quay.io/strimzi/jmxtrans@sha256:62272e8a03da2144340bd7267f7a824d1ed70f4f3e30dc7aadfba9ee1b9baf76
        - quay.io/strimzi/kafka@sha256:1ace2cc0db9653ac1b388477fa56c3c1cd97c7b3846eba3c7824dc58ea179476
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:02f53d3395b2b98f92a96b466c83ca15c79a75d82451e2ed8ba921375e727c7f
        - quay.io/strimzi/kafka@sha256:1da35fd7b36b113d1343e254fdc6efd97ca8660cc7977d8985aa3a3547d5ebb1
        - quay.io/strimzi/kafka@sha256:37cf8d0fea1d2078511920190a72b49977ff4047b6ecbfed6b394c6f9e1e2ff0
        - quay.io/strimzi/kaniko-executor@sha256:dc27d8aeb851245c79fb7276febe98769cc098d678bea096b338caaaf111dd27
        - quay.io/strimzi/operator@sha256:f4ee3ca482a2dd19785f59634578a0850273444c1173a5699a99cf02f3a018fe
        - quay.io/strimzi/kafka-bridge@sha256:51babca0f7a328f4db005c6a16daa8ac4a4306c46fe3ca89382919134aa09179
        version: 0.34.0
      entries:
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.34.x
    - currentCSV: strimzi-cluster-operator.v0.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.4"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
          createdAt: "2023-06-07 16:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **Strimzi 0.35 is the last version with support for Kubernetes 1.19 and 1.20. From Strimzi 0.36, only Kubernetes 1.21 and newer will be supported.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.35
          * Move feature gate `UseStrimziPodSets` to GA and remove support for StatefulSets
          * Remove support for JMX Trans
          * Redesigned the Cluster and User Operator configuration to make it more efficient and flexible
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:39778b90c2b2afc30261e4ad5135805e1a10a2b60e2e53108fb9f80487f1208a
        - quay.io/strimzi/maven-builder@sha256:88a79eff3b3a386880a630658964b7754caed9e99dd6e645a4c0d23d0fdb47ee
        - quay.io/strimzi/operator@sha256:06a94a3021cf028ccc1a49271f35f79216029e344536e664f196c1725ff2c663
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8abe0994052e2a328e54c9d30c920dc331e9b816a4606234a3023b9ef65ec4ed
        - quay.io/strimzi/kafka-bridge@sha256:d6be183e492f8f88157ab9fe0af53950df8b6711a8a8c33da465de6064f6f86e
        - quay.io/strimzi/kafka@sha256:0d910e7138cb49e1cd8cd84cef88bce35698b93ddd683a3398f1d485a3162693
        - quay.io/strimzi/kafka@sha256:4de4874a7b722ad813f4dcc58acf509527bca0609999b81e70d81e3b38534d9d
        - quay.io/strimzi/kafka@sha256:54c6b25b31f51ef401c1b6e2a1b27432911e819bf3e502e05186f01be3f798e5
        version: 0.35.1
      entries:
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.35.x
    - currentCSV: strimzi-cluster-operator.v0.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
          createdAt: "2023-07-10 21:47:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!! Strimzi 0.36.1 supports only Kubernetes 1.21 and newer! Kubernetes versions 1.19 and 1.20 are not supported anymore since Strimzi 0.36.**
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.35.1/deploying.html#assembly-upgrade-str).
          ### New in 0.36
          * Add support for Apache Kafka 3.4.1, 3.5.0, and 3.5.1
          * Remove support for 3.3.1 and 3.3.2
          * Add support for _Kafka node pools_
          * Add support for _Unidirectional Topic Operator_
          * Enable SCRAM-SHA authentication in KRaft mode (supported in Apache Kafka 3.5.0 and newer)
          * Add support for insecure flag in Maven artifacts in Kafka Connect Build
          * Improve Kafka rolling update to avoid rolling broker in log recovery
          * Added support for Kafka Exporter topic exclude and consumer group exclude parameters
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:ef3f85e1227e886a131d3ddec3fb431aef6d3daed46115b288246b9ace0e69e2
        - quay.io/strimzi/kafka@sha256:9f35f40be3d47b0cb1f6628f7ffd9c6a687ab9f564829f341dd66fd0759c064f
        - quay.io/strimzi/kafka@sha256:a0da6a48b5bcbff355365f07df2d99ef6443d22c5690495f10a6417e761b3799
        - quay.io/strimzi/kafka@sha256:ea35817ff1ef295c682d70562f209ef658fcd87c61f467a4549d7a2f39e71700
        - quay.io/strimzi/kaniko-executor@sha256:5ff9fc7d0b7b694875ad0150fc787c582d42da3b4864d807ee8a6524e47b0e4e
        - quay.io/strimzi/maven-builder@sha256:2a469b40122027b981914db5b5c102be48280e039c3089963422c181174af226
        - quay.io/strimzi/kafka@sha256:67c2b563a65c85c9393608f9af7eaf12ea48dcc0ef5f726d51f27273a3845186
        - quay.io/strimzi/operator@sha256:e9e03b31007c0f9b67a6a33e6b644e8c1ba35bd9976804ee3d22a04e4bf511b1
        - quay.io/strimzi/kafka-bridge@sha256:fefee252895e9f94756ed5c9ea71749e3fa89281df01bc55878a244ca7d17697
        version: 0.36.1
      entries:
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.36.x
    - currentCSV: strimzi-cluster-operator.v0.37.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
          createdAt: "2023-09-03 17:45:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.37.0/deploying.html#assembly-upgrade-str).
          ### New in 0.37
          * The StableConnectIdentites feature gate moves to a beta stage.
          * Support for the ppc64le platform
          * Version fields to the Kafka custom resource status to track installation and upgrade state
          * Infinite auto-restarts of Kafka Connect and Kafka Mirror Maker 2 connectors
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:dfeb1eb7b94107c443b6bc82fcfee7fcf782516c83c6fc64451fb6a7e2de74a6
        - quay.io/strimzi/kaniko-executor@sha256:5729f1f0511da95e64e358cfabb3980ad9a660af2968118a87d6755f5d652a67
        - quay.io/strimzi/kafka-bridge@sha256:5b821f6f3396002ffa7c5804391b627ff9da3b494b2777dde518c55c2b73cefc
        - quay.io/strimzi/kafka@sha256:38bbadcdb196a373e1698be890caa1dc01134ce91763d8f95c71768a117e1932
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:8b9b06ead5e014b920c82034ded615413dca53cf36f804005408726c9bbda66f
        - quay.io/strimzi/kafka@sha256:1de6e073e8625a193fcfd11dd1a6717cc17596911c10bcebd792c00b08a7ebaf
        - quay.io/strimzi/operator@sha256:52f376e64b9b31d74162e26248c278e9bd301d2d8b4240ee5bcf780dfe2f8ef8
        - quay.io/strimzi/kafka@sha256:2997ba1c169eb792a6004d7a41b674560515f3ea6e62ab8b9a1fdb747eab40c1
        - quay.io/strimzi/maven-builder@sha256:995be06a50b86c992fcee9f00e44f6a21c3406975e4e35802f10e88b72ffb9df
        version: 0.37.0
      entries:
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.37.x
    - currentCSV: strimzi-cluster-operator.v0.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
          createdAt: "2023-10-14 18:18:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.38.0/deploying.html#assembly-upgrade-str).
          ### New in 0.38
          * Add support for Apache Kafka 3.6.0 and remove support for 3.4.0 and 3.4.1
          * Container signing and SBOMs
          * Stopping Connect and MirrorMaker 2 connectors
          * Manual rolling updates of Kafka Connect and Kafka Mirror Maker 2 pods using annotation
          * Prevent scale-down of brokers that still contain any replicas
          * Run ZooKeeper- and KRaft-based clusters in parallel
          * Unidirectional Topic Operator improvements
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:4fe7a3f3f78e93369284a5ab456873c911230926ca071d1d88167e63d8ce376c
        - quay.io/strimzi/maven-builder@sha256:922d91dab9e2b8e14b6db25d8b48d7f4e3521bedd435468cc22039599ac4f888
        - quay.io/strimzi/operator@sha256:aee45bbe439335e45a0423e6565bae0d0727d85c7b807812949bfa127d467fe6
        - quay.io/openshift-community-operators/strimzi-kafka-operator@sha256:51db956f1472ae0498c57b9488448c34cf7e24afab6b9997eb3689b001d73277
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:0cc01438e8596643407b6b8ff05fafcea7ef54fa9410d4ae2492a9dbf41118b1
        - quay.io/strimzi/kafka@sha256:20d451fb32ec82eab3c16891198ace46bb9b9193f1f5511bb60ce249080dd7d4
        - quay.io/strimzi/kafka@sha256:50e24c797455f9cdbb3cdde8e4f26bb8e1f0503469456da5d179017fd3e03ffe
        version: 0.38.0
      entries:
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.38.x
    - currentCSV: strimzi-cluster-operator.v0.39.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
          createdAt: "2023-12-14 20:28:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.39.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.39 is the last minor release with support for Kubernetes 1.21 and 1.22.** From Strimzi 0.40 on, only Kubernetes 1.23 and newer will be supported.
          ### New in 0.39
          * Add support for Apache Kafka 3.5.2 and 3.6.1
          * The `StableConnectIdentities` feature gate moves to GA stage and is now permanently enabled without the possibility to disable it.
            All Connect and Mirror Maker 2 operands will now use StrimziPodSets.

          * The `KafkaNodePools` feature gate moves to the beta stage and is enabled by default.
            If needed, `KafkaNodePools` can be disabled in the feature gates configuration in the Cluster Operator.

          * The `UnidirectionalTopicOperator` feature gate moves to the beta stage and is enabled by default.
            If needed, `UnidirectionalTopicOperator` can be disabled in the feature gates configuration in the Cluster Operator.

          * Improved Kafka Connect metrics and dashboard example files
          * Allow specifying and managing KRaft metadata version
          * Add support for KRaft to KRaft upgrades (Apache Kafka upgrades for the KRaft-based clusters)
          * Improved Kafka Mirror Maker 2 dashboard example file
          * Support for rolling updates of KRaft controller nodes
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:13838889cc20a3a382ea64e749cad94f997c0b928840588f2d1befcf19247bea
        - quay.io/strimzi/kafka@sha256:2badc65532b483bba6c1f6a8b0c9895e993ce24e9ebd3cd953d0d17fbe36d27a
        - quay.io/strimzi/kafka@sha256:84691b3afb67b312400e01572ff08b64c1e1b1d32a4276e0c21bba10c71c36fe
        - quay.io/strimzi/kaniko-executor@sha256:723b7add2473e73c45c81361e1708af4971c321b0391c1298a0557fb8ca4f7f6
        - quay.io/strimzi/maven-builder@sha256:6ce3c16284379d5db9e42a6f3e2b765c981091164a28fdd6bf927b6c5d4ab557
        - quay.io/strimzi/kafka@sha256:49b7530ada1175750ad5a706e545ba3594d85e1e3a90f98b0200a73fa9875057
        - quay.io/strimzi/kafka@sha256:53b960411e852f5cb182ecc5d395a3562865f83e15e6251f3a9af0eb413feeed
        - quay.io/strimzi/kafka@sha256:dd1706f36031c96dfcd0da7b3f804168e5f076220caa302973455fdcd76bd03e
        - quay.io/strimzi/operator@sha256:50b49f9b5e554cca517fc2794e11319d5aaba1676494b285baa3a74fc79eb501
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:392f9299c75a3ca4f593eaebdfe515145696f12db9f990bb004cfa58dd6bf0ef
        version: 0.39.0
      entries:
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.39.x
    - currentCSV: strimzi-cluster-operator.v0.40.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
          createdAt: "2024-03-09 21:27:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on  [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!**  You have to upgrade first to one of the previous versions of Strimzi.  You will also need to convert the CRD resources.  For more details, see the [documentation](https://strimzi.io/docs/operators/0.40.0/deploying.html#assembly-upgrade-str).

          **From Strimzi 0.40 on, we support only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.40
          * Add support for Apache Kafka 3.7.0. Remove support for Apache Kafka 3.5.0, 3.5.1, and 3.5.2.
          * The `UseKRaft` feature gate moves to beta stage and is enabled by default. If needed, `UseKRaft` can be disabled in the feature gates configuration in the Cluster Operator.
          * Add support for ZooKeeper to KRaft migration by enabling the users to move from using ZooKeeper to store metadata to KRaft.
          * Add support for moving from dedicated controller-only KRaft nodes to mixed KRaft nodes
          * Added support for Kafka Exporter `offset.show-all` parameter
          * Prevent removal of the `broker` process role from KRaft mixed-nodes that have assigned partition-replicas
          * Improve broker scale-down prevention to continue in reconciliation when scale-down cannot be executed
          * Added support for Tiered Storage by enabling the configuration of custom storage plugins through the Kafka custom resource.
          * Update HTTP bridge to the latest 0.28.0 release
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which  might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to  test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c491a95f9b9b58f406461faf222ef612ad8d2f4d7b53f7f3645b20c99d7d1583
        - quay.io/strimzi/kafka@sha256:e2b9f2835498831bbf30ac2d651de143d57088bcb23943101b066e941db92aca
        - quay.io/strimzi/kaniko-executor@sha256:3fb4ace22589fcf8d3f01b74c37b10b32d87e15d103ea59e0933b61c4e46ff4a
        - quay.io/strimzi/maven-builder@sha256:d02a9b48c77a44eeab6426c9af4adf75405217994bcb01ce0196e616041350d8
        - quay.io/strimzi/operator@sha256:2228d89b53e3524e69d1bcf5c2cd373884ce696d9005389a6c78dad23310e837
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:df4e4dce45f87e684e7e1044a7edb2456160141afb6ba57b9af28088ffbac407
        - quay.io/strimzi/kafka-bridge@sha256:73deb1af0d93e9fa8955a5a412d0e7138384bdc80e7574e815d21796e446b636
        - quay.io/strimzi/kafka@sha256:2c143eb10b66037ef877311eb767457e4b877d5ff2c47d393c76babec155caff
        version: 0.40.0
      entries:
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.40.x
    - currentCSV: strimzi-cluster-operator.v0.41.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
          createdAt: "2024-05-09 22:23:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.41.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.41 supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.41
          * Support for Apache Kafka 3.6.2
          * Metrics to monitor CA expiration
          * Support for JBOD storage in KRaft mode. (Note: JBOD support in KRaft mode is considered early access in Apache Kafka 3.7.x)
          * Support for changing topic replication factor by changing the `KafkaTopic` resources.
          * The `KafkaNodePools` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. To use the Kafka Node Pool resources, you still need to use the `strimzi.io/node-pools: enabled` annotation on the `Kafka` custom resources.
          * Support for configuring the `externalIPs` field in node port type services.
          * The `UnidirectionalTopicOperator` feature gate moves to GA stage and is permanently enabled without the possibility to disable it. If the topics whose names start with `strimzi-store-topic` and `strimzi-topic-operator` still exist, you can delete them.
          * Improve validation of `KafkaMirrorMaker2` resource to make sure the Connect cluster is set to the target cluster
          * Continue reconciliation after failed manual rolling update using the `strimzi.io/manual-rolling-update` annotation (when the `ContinueReconciliationOnManualRollingUpdateFailure` feature gate is enabled)
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:c3a346235ce4dea42986648ca07f2482d8fc0bf654c1d441b1301a23999304db
        - quay.io/strimzi/kafka@sha256:41811bb106d6182978a889eaedffeadffaa5ab1599c4cd8e4f5015b131c7ac30
        - quay.io/strimzi/kafka@sha256:bf49499980bbe5d4dd399de0dd396089b0fd9dd29bc46e14d39b20f1e826aec1
        - quay.io/strimzi/kaniko-executor@sha256:9404037bd5431b012c8c633ec3345907667453c5532e5bfaaedc78286de22c62
        - quay.io/strimzi/maven-builder@sha256:17f09b07fc41a059e6a048cb766046d43a072cfd3a8266c97388a37543dadee8
        - quay.io/strimzi/kafka@sha256:6243d98890c665ef65b5f5fa2ffd046412281ffef31bc7a07cac37bd77b2579d
        - quay.io/strimzi/operator@sha256:4c687a4fac96ffc22433ca6d00fb1bd31722767b44c9d87c465f95b57fb70089
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:fecad1a330421341e6946252f95735ed177a690eab6250be097413ea29ce9473
        - quay.io/strimzi/kafka-bridge@sha256:c833f930ead2d1e9054912a20689ee979d335f5e5ca8769e3c6b8a5514a0df96
        version: 0.41.0
      entries:
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.41.x
    - currentCSV: strimzi-cluster-operator.v0.42.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
          createdAt: "2024-07-03 12:34:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.42.0/deploying.html#assembly-upgrade-str).

          **From version 0.40.0, Strimzi supports only Kubernetes 1.23 and newer.** Kubernetes versions 1.21 and 1.22 are no longer supported.
          ### New in 0.42
          * Support for Apache Kafka 3.7.1
          * The UseKRaft feature gate moves to GA stage and is permanently enabled without the possibility of disabling it. To use KRaft (ZooKeeper-less Apache Kafka), you still need to use the strimzi.io/kraft: enabled annotation on the Kafka custom resources or migrate from an existing ZooKeeper-based cluster.
          * Update the base image used by Strimzi containers from UBI8 to UBI9
          * Add support for filename patterns when configuring trusted certificates
          * Enhance KafkaBridge resource with consumer inactivity timeout and HTTP consumer/producer enablement.
          * Add support for setting publishNotReadyAddresses on services for listener types other than internal.
          * Added support for configuring the quotas plugin with type strimzi or kafka in the Kafka custom resource. The Strimzi Quotas plugin version was updated to 0.3.1.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Aapche ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:bc49c10b3e839e161e515c3a1e411d1f8df9de5e2c8d71aa7b63ccdbbae9e8d2
        - quay.io/strimzi/operator@sha256:ca5c1b70e1196deec4bf519921bfa55a967aa1ea15bd4c182bc56bea08d2e114
        - quay.io/strimzi/kafka-bridge@sha256:fdddffb3de59ba48c05964bb0e760593088d24d1ee35b090112fe6cc6b553cd8
        - quay.io/strimzi/kafka@sha256:6527c426f753521c07f3a3df2ee432273a41b109594096b4c6cae2e8e609d669
        - quay.io/strimzi/maven-builder@sha256:342e8d0086c53cf7334127798e5bd3764b28f5e4909ba67d65f6d937dc3dba05
        - quay.io/strimzi/kaniko-executor@sha256:11821b0078a270612f1bcc37f5dc716ab273c974648c40f95fa4804352e647a7
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:9dcf24fecf3ee0d999601c276ac15e8634c275590f2c42b8a6bcb83984bf1445
        - quay.io/strimzi/kafka@sha256:c56c5abf94897f20f976603cba0678a33cdfbbcd6f9030a193f7684cb9894a6c
        - quay.io/strimzi/kafka@sha256:6c519531cdd5f40f326b295d8827344140efe2607a1826f7f1cab8c3427f43f7
        - quay.io/strimzi/kafka@sha256:841a46e9c10481df3ef702412caf362b55ae07c30a3a5927073356cd36dc049f
        version: 0.42.0
      entries:
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.42.x
    - currentCSV: strimzi-cluster-operator.v0.43.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### Upgrades
          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.43.0/deploying.html#assembly-upgrade-str).

          **Strimzi 0.43 is the last version with support for Kubernetes 1.23 and 1.24.** From Strimzi 0.44, only Kubernetes versions 1.25 and newer will be supported.
          ### New in 0.43
          * Support for Apache Kafka 3.8.0
          * Removed support for Apache Kafka 3.6.0, 3.6.1, and 3.6.2
          * Specifying additional volumes and volume mounts in Strimzi custom resources
          * Support for custom Cruise Control API users
          * Unregistration of KRaft nodes after scale-down
          * Kafka Exporter updated to 1.8.0 (and the Grafana dashboard updated to work with it)
          * Service account based authentication
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:6414f8f3985f221b0eda53867c00494140108d17579c4dcbbc93934bb00e5d49
        - quay.io/strimzi/operator@sha256:d3732245f34afaa6b571f1f912e1cd017891b22003d6367ae8ca059b2a5cb13c
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a9950264bfefc536b4e0a0adc2348f92aa41da55dd50a2324ddad0253291b67c
        - quay.io/strimzi/kafka-bridge@sha256:80f329ecd3ba96fac12e75be874dce0debc8395337737581de34238ce11c8fb1
        - quay.io/strimzi/kafka@sha256:ad06ead5125bce87498b1582a76e78143a6d21408ac4b12bb7e6007aa4cabd84
        - quay.io/strimzi/kafka@sha256:b2f4e8895e9ed66228512ce14aec685101fa75dc329128d47cb1c15e3aef3c9c
        - quay.io/strimzi/kafka@sha256:d27cf5bf3624e18ba705246b14f92d12b8a15eb8551b13abf027cf71b549baf0
        version: 0.43.0
      entries:
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.43.x
    - currentCSV: strimzi-cluster-operator.v0.44.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
          createdAt: "2024-08-22 15:19:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.44.0/deploying.html#assembly-upgrade-str).

          **From version 0.44, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with ZooKeeper support is expected to be Strimzi 0.45. Please plan your migration to KRaft (ZooKeeper-less Apache Kafka) accordingly. Follow the [documentation](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) for more details.

          **Kafka Mirror Maker 1 support will be soon removed from Apache Kafka and Strimzi.** Currently, the last Strimzi version with Mirror Maker 1 support is expected to be Strimzi 0.45. Please plan your migration to Mirror Maker 2 or another mirroring tool.

          ### New in 0.44
          * Managing offsets in Connect and Mirror Maker 2 connectors through custom resources
          * Templating `advertisedHost` and `host` fields in listener configuration
          * Configuring environment variables in operands based on Secrets and ConfigMaps
          * Disabling Pod disruption Budget management
          * Auto-rebalancing at scale-up/scale-down
          * `ContinueReconciliationOnManualRollingUpdateFailure` feature gate moves to the beta stage and is enabled by default
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:7b55f971a3b62923a9a61a61c82f542e07deeb6ffdbaede68244ba358a0853f2
        - quay.io/strimzi/operator@sha256:b07b81f7e282dea2e4e29a7c93cfdd911d4715a5c32fe4c4e3d7c71cba3091e8
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:4d59e60cc75de82bd3f3a063237cef8a5598d315f6d53e803d37589a94929b0e
        - quay.io/strimzi/kafka-bridge@sha256:cc95ce9f496057f3b18d5e8297b40a410cc66f9fd4a2227615971458a26bac25
        - quay.io/strimzi/kafka@sha256:238bb080825b7ecebaa819be09788bc3d056ae8162ebcd41444076dc22f85740
        - quay.io/strimzi/kafka@sha256:98d332b7fa5c9e48f60b516e00ed5812bd9268e014fd2337284e7cc3e664f1eb
        - quay.io/strimzi/kafka@sha256:9a4224635e09ece2c1d3ffe9d715813cf63037461c9f06c1b5f53554aed5ecdc
        version: 0.44.0
      entries:
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.44.x
    - currentCSV: strimzi-cluster-operator.v0.45.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.1",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
          createdAt: "2025-06-27 01:38:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.45.1/deploying.html#assembly-upgrade-str).

          **From version 0.45, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**

          **ZooKeeper and Kafka Mirror Maker 1 support will be removed from Apache Kafka and Strimzi.** Strimzi 0.45 is the last minor Strimzi version with support for ZooKeeper-based Apache Kafka clusters and MirrorMaker 1 deployments. Please make sure to [migrate to KRaft](https://strimzi.io/docs/operators/latest/full/deploying.html#assembly-kraft-mode-str) and MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          **Strimzi EnvVar Configuration Provider and Strimzi MirrorMaker 2 Extensions will be removed from Strimzi.** Strimzi 0.45 is the last Strimzi version to include the Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0). Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          ### New in 0.45.1
          * Add support for Kafka 3.9.1
          * Update HTTP bridge to 0.31.2
          * Update Strimzi Kafka OAuth library to 0.15.1
          * Fixed bug which may lead to metadata loss within the cluster when restarting a KRaft migration after a previous rollback due to missing `/migration` znode deletion.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka@sha256:ba52ed046b1dccdbd96f4e68057ce014d862a7c9c1fc670760c023b9aa09f23f
        - quay.io/strimzi/kafka@sha256:6a05924839a2c7da0431408e6d6223a055e75d88b3f057ac5caa7f9097227f56
        - quay.io/strimzi/kafka@sha256:e98fbe49cadbc5d3afbd8f17abea218f216c95d52f4fa1fe514a7a988624da24
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/kafka-bridge@sha256:c462441a966f0732855e684b4840563bfad45391345c9511fa49e1630ba9c3e3
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:40d17d4d0632c272779da152f710fb7c2de50acc3791eb0b9fe445a13c0cb102
        - quay.io/strimzi/kafka@sha256:173cb5793756f75739c51a83bbe06d7f0094fd003a2982160a5cc6e216cd6d28
        - quay.io/strimzi/maven-builder@sha256:770e410f1e7e0e92c31bae6c19831ba36d59dd11b44e90c159aef16a965aa998
        - quay.io/strimzi/operator@sha256:58e8722216b0b0e8b1dd79109725238fa4016a16cc6a9fca0e327bccb1506810
        version: 0.45.1
      entries:
      - name: strimzi-cluster-operator.v0.45.1
        version: 0.45.1
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.45.x
    - currentCSV: strimzi-cluster-operator.v0.46.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
          createdAt: "2025-06-26 11:58:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.

          **!!! IMPORTANT !!!**

          **Direct upgrade from Strimzi 0.22 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Strimzi. You will also need to convert the CRD resources. For more details, see the [documentation](https://strimzi.io/docs/operators/0.46.0/deploying.html#assembly-upgrade-str).

          **From version 0.46, Strimzi supports only Kuberneets 1.25 and newer. Kubernetes 1.23 and 1.24 are not supported anymore.**
          **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed. Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!
          **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images. Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.
          **Support for MirrorMaker 1 has been removed. Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.
          ### New in 0.46.1
          * Update CSV file for OLM installation to add missing permission for broker Pods to "get" nodes Kubernetes resource.
          * Allow the injection of annotations on the deployment object in Helm charts.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          You can contribute by:
          * Raising any issues you find using Strimzi
          * Fixing issues by opening Pull Requests
          * Improving documentation
          * Talking about Strimzi
          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/strimzi/strimzi-kafka-operator/issues). Issues which might be a good start for new contributors are marked with ["good-start"](https://github.com/strimzi/strimzi-kafka-operator/labels/good-start) label.
          The [Development guide](https://github.com/strimzi/strimzi-kafka-operator/blob/main/development-docs/DEV_GUIDE.md) describes how to build Strimzi and how to test your changes before submitting a patch or opening a PR.
          The [Documentation Contributor Guide](https://strimzi.io/contributing/guide/) describes how to contribute to Strimzi documentation.
          If you want to get in touch with us first before contributing, you can use:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:3e1815aed3f74b8573cebfced66b2c78ca0cf7c746c26db2002380cd39d09ae6
        - quay.io/strimzi/kafka@sha256:82273d873ce21a73f8f3118bf61627cb15734b43ea7ca92d43b290340b44e485
        - quay.io/strimzi/kafka@sha256:f3837f8cb2a60e906e84486128c41c398424f4085df2348e9e0486bfc12089fd
        - quay.io/strimzi/kaniko-executor@sha256:ef1a39c363e145041d80103c3c12da9429ce06cf21dff6fb1fb75d0c0ed9c35b
        - quay.io/strimzi/maven-builder@sha256:b532234c234fc30fecec992ad1a6154c5ef08ebf9a20ec8bc62e0c88d791ff36
        - quay.io/strimzi/operator@sha256:6a22e4968e9d409405423d12c04a0e3fecaea2ee2794bd27de7cf21f59453265
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:8e321dc1bf74ba331b62d21bff4d8993956ab935cf5be32f01e7f7aa6897bf22
        version: 0.46.1
      entries:
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.46.x
    - currentCSV: strimzi-cluster-operator.v0.47.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.0.0",
                      "metadataVersion": "4.0-IV3",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.0.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
          createdAt: "2025-07-12 17:21:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.47 is the last Strimzi version with support for Kubernetes 1.25 and 1.26.**
            From Strimzi 0.48.0 on, we will support only Kubernetes 1.27 and newer.

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.47.0
          * Support for Kafka 3.9.1
          * Progress tracking for Cruise Control rebalances
          * Support for Kubernetes Image Volumes to mount custom plugins
          * Support for [JsonTemplateLayout](https://logging.apache.org/log4j/2.x/manual/json-template-layout.html) in Operators, Cruise Control, and Kafka 4.0.0
          * Support for Strimzi Metrics Reporter for the Kafka brokers / controllers
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/strimzi/kafka-bridge@sha256:428c5587f547dabfc995a73745569d113380c7e6b8bf0b2504c6a9a00d6c4608
        - quay.io/strimzi/kafka@sha256:551882f47a6d2dc8300bcfca251fffffc3379acb2724a7efe16f17c0c4f529fb
        - quay.io/strimzi/kafka@sha256:8a793cdf5084cdb72c6663419530985158b16330a518720e9e9d1d96babbaf7e
        - quay.io/strimzi/kafka@sha256:978e2ed5f0d315e8bf0ab0503ee839acbd9da01ca427e6581b59ac25d36bc9c9
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:c67a3ad3d92420ac8b7f8435eeec53ebffaea4e894e65414a7c47c20e3faba27
        - quay.io/strimzi/operator@sha256:589c4d63641d9a944462dd682f6e5febe8b38ff55b9253949b061aca16feb154
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:16fbdaaf303ab40202d49438737f3307926594a5259081eeeb39595b7ea41f8e
        version: 0.47.0
      entries:
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.47.x
    - currentCSV: strimzi-cluster-operator.v0.48.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec":{
                  "kafka":{
                      "version":"4.1.0",
                      "metadataVersion": "4.1-IV1",
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"4.1.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"4.1.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Strimzi Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
          createdAt: "2025-09-25 10:32:00"
          description: Strimzi provides a way to run an Apache Kafka cluster on Kubernetes
            or OpenShift in various deployment configurations.
          features.operators.openshift.io/disconnected: “true”
          features.operators.openshift.io/fips-compliant: “true”
          features.operators.openshift.io/proxy-aware: “true”
          features.operators.openshift.io/tls-profiles: “false”
          features.operators.openshift.io/token-auth-aws: “false”
          features.operators.openshift.io/token-auth-azure: “false”
          features.operators.openshift.io/token-auth-gcp: “false”
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Strimzi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          Strimzi provides a way to run an [Apache Kafka®](https://kafka.apache.org) cluster on [Kubernetes](https://kubernetes.io/) or [OpenShift](https://www.openshift.com/) in various deployment configurations. See our [website](https://strimzi.io) for more details about the project.
          ### !!! IMPORTANT !!!
          * **Support for ZooKeeper-based clusters and for migration from ZooKeeper-based clusters to KRaft has been removed.**
            Please make sure all your clusters are using KRaft before upgrading to Strimzi 0.46.0 or newer!

          * **Strimzi 0.48.0 supports only Kubernetes 1.27 and newer.**
            Kubernetes 1.25 and 1.26 are not supported anymore

          * **Strimzi EnvVar Configuration Provider (deprecated in Strimzi 0.38.0) and Strimzi MirrorMaker 2 Extensions (deprecated in Strimzi 0.28.0) plugins were removed from Strimzi container images.**
            Please use the Apache Kafka EnvVarConfigProvider and Identity Replication Policy instead.

          * **Support for MirrorMaker 1 has been removed.**
            Please make sure to migrate to MirrorMaker 2 before upgrading to Strimzi 0.46 or newer.

          ### New in 0.48.0
          * Support for Kafka 4.1.0
          * Add monitoring of custom resources using kubernetes-state-metrics (KSM)
          * Support for type: custom client authentication
          * Added support for Strimzi Metrics Reporter to Kafka Connect, Mirror Maker 2 and Kafka Bridge.
          * Added support for KIP-745 in Kafka Connector and Mirror Maker 2, allowing the usage of includeTasks and onlyFailed arguments in Kafka connectors restart.
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Strimzi Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://strimzi.io/docs/operators/latest/deploying.html#assembly-upgrading-kafka-versions-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Strimzi requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Strimzi Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Strimzi also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation to the current _main_ branch as well as all releases can be found on our [website](https://strimzi.io/documentation).
          ### Getting help
          If you encounter any issues while using Strimzi, you can get help using:
          * [Strimzi mailing list on CNCF](https://lists.cncf.io/g/cncf-strimzi-users/topics)
          * [Strimzi Slack channel on CNCF workspace](https://cloud-native.slack.com/messages/strimzi)
          ### Contributing
          Learn more about contributing to Strimzi on our [Join Us page](https://strimzi.io/join-us/).
          ### License
          Strimzi is licensed under the [Apache License, Version 2.0](https://github.com/strimzi/strimzi-kafka-operator/blob/main/LICENSE).
        displayName: Strimzi
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Website
          url: https://strimzi.io/
        - name: Documentation
          url: https://strimzi.io/documentation/
        - name: Mailing list
          url: https://lists.cncf.io/g/cncf-strimzi-users/topics
        - name: Slack
          url: https://cloud-native.slack.com/messages/strimzi
        - name: GitHub
          url: https://github.com/strimzi/strimzi-kafka-operator
        maintainers:
        - email: cncf-strimzi-users@lists.cncf.io
          name: Strimzi
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Strimzi
        relatedImages:
        - quay.io/community-operator-pipeline-prod/strimzi-kafka-operator@sha256:a4a74cd4bb69717cc58d8996231fd49d8a18be3069a88489ad2409bde4b2a17e
        - quay.io/strimzi/kafka-bridge@sha256:53034f64f0b672f10b5bacea1c7a25132f118df7fd5c9032c4dbf702ed93796a
        - quay.io/strimzi/kafka@sha256:34450afc1a3399a9026fc3c6ec90e9f6adfec9faf6520545b4d84e90cda34964
        - quay.io/strimzi/kafka@sha256:48db05ad2e9fa39eb388089e7e25654bcbf4da1ddc3a5635d675056c28bbcb96
        - quay.io/strimzi/kaniko-executor@sha256:a5088c14d7b8bf1d336669cba047b971bf3ece8969647dae2c1e3a07a7be0c5e
        - quay.io/strimzi/maven-builder@sha256:64380efcb8da87fdc0e0de238a20456be9abb830174c2f2331882b7a84657e57
        - quay.io/strimzi/operator@sha256:73dc9555c4a73094b497ffb14d816de5ff40144ffc470efd75ccf128afa22778
        version: 0.48.0
      entries:
      - name: strimzi-cluster-operator.v0.48.0
        version: 0.48.0
      - name: strimzi-cluster-operator.v0.47.0
        version: 0.47.0
      - name: strimzi-cluster-operator.v0.46.1
        version: 0.46.1
      - name: strimzi-cluster-operator.v0.46.0
        version: 0.46.0
      - name: strimzi-cluster-operator.v0.45.0
        version: 0.45.0
      - name: strimzi-cluster-operator.v0.44.0
        version: 0.44.0
      - name: strimzi-cluster-operator.v0.43.0
        version: 0.43.0
      - name: strimzi-cluster-operator.v0.42.0
        version: 0.42.0
      - name: strimzi-cluster-operator.v0.41.0
        version: 0.41.0
      - name: strimzi-cluster-operator.v0.40.0
        version: 0.40.0
      - name: strimzi-cluster-operator.v0.39.0
        version: 0.39.0
      - name: strimzi-cluster-operator.v0.38.0
        version: 0.38.0
      - name: strimzi-cluster-operator.v0.37.0
        version: 0.37.0
      - name: strimzi-cluster-operator.v0.36.1
        version: 0.36.1
      - name: strimzi-cluster-operator.v0.36.0
        version: 0.36.0
      - name: strimzi-cluster-operator.v0.35.1
        version: 0.35.1
      - name: strimzi-cluster-operator.v0.35.0
        version: 0.35.0
      - name: strimzi-cluster-operator.v0.34.0
        version: 0.34.0
      - name: strimzi-cluster-operator.v0.33.2
        version: 0.33.2
      - name: strimzi-cluster-operator.v0.33.1
        version: 0.33.1
      - name: strimzi-cluster-operator.v0.33.0
        version: 0.33.0
      - name: strimzi-cluster-operator.v0.32.0
        version: 0.32.0
      - name: strimzi-cluster-operator.v0.31.1
        version: 0.31.1
      - name: strimzi-cluster-operator.v0.31.0
        version: 0.31.0
      - name: strimzi-cluster-operator.v0.30.0
        version: 0.30.0
      - name: strimzi-cluster-operator.v0.29.0
        version: 0.29.0
      - name: strimzi-cluster-operator.v0.28.0
        version: 0.28.0
      - name: strimzi-cluster-operator.v0.27.1
        version: 0.27.1
      - name: strimzi-cluster-operator.v0.27.0
        version: 0.27.0
      - name: strimzi-cluster-operator.v0.26.1
        version: 0.26.1
      - name: strimzi-cluster-operator.v0.26.0
        version: 0.26.0
      - name: strimzi-cluster-operator.v0.25.0
        version: 0.25.0
      - name: strimzi-cluster-operator.v0.24.0
        version: 0.24.0
      - name: strimzi-cluster-operator.v0.23.0
        version: 0.23.0
      name: strimzi-0.48.x
    defaultChannel: stable
    packageName: strimzi-kafka-operator
    provider:
      name: Strimzi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hnas-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hnas-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.csi.hitachi.com/v1",
                "kind": "HNAS",
                "metadata": {
                  "name": "hnas",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
          createdAt: "2025-07-30T15:12:23Z"
          description: An operator for managing Hitachi NAS CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HNAS is the Schema for the hnas API
            displayName: HNAS
            kind: HNAS
            name: hnas.charts.csi.hitachi.com
            version: v1
        description: |-
          The Hitachi NAS CSI driver operator deploys the Hitachi NAS CSI driver on Kubernetes and OpenShift environments.  The driver can be used to create and manage persistent storage hosted on Hitachi NAS products, including those in the Hitachi VSP One File 3x range and the HNAS range of products.

          For the complete documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the Hitachi NAS CSI Driver User Guide - MK-92HNAS091.  The guide covers prerequisites for the NAS system and covers the driver functionality and all configuration options available for the driver.
        displayName: Hitachi NAS CSI Driver Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        - NAS
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maturity: stable
        minKubeVersion: 1.28.0
        provider:
          name: Hitachi
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d5e46da8aff7d73d6f00c761dae94472bcda6e78f4f17b3802dc89d44de0111b
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-driver@sha256:6b6f6816c9f4e608e93e7fbc0347255415426a0ff0c9a53b7750e5ba54d91262
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:d7138bcc3aa5f267403d45ad4292c95397e421ea17a0035888850f424c7de25d
        - registry.connect.redhat.com/hitachi/hnas-csi-operator-bundle@sha256:79e6ba3db78aa55a9f3839a172843d284d91e9cdaae9cb2c8e7e705f3d03115d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:ab774734705a906561e15b67f6a96538f3ad520335d636f793aaafe87a3b5200
        - registry.k8s.io/sig-storage/csi-attacher@sha256:a399393ff5bd156277c56bae0c08389b1a1b95b7fd6ea44a316ce55e0dd559d7
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:dd788d79cf4c1b8edee6d9b80b8a1ebfc51a38a365c5be656986b129be9ac784
        - registry.hitachivantara.com/fpgroup-docker-public/hnas-csi-operator@sha256:9d3f9f10fc7172ee9cebbd9ee318906756d7e96909f2e66e0534e98bb53e0a7b
        - registry.k8s.io/sig-storage/livenessprobe@sha256:2c5f9dc4ea5ac5509d93c664ae7982d4ecdec40ca7b0638c24e5b16243b8360f
        version: 1.0.0
      entries:
      - name: hnas-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: hnas-csi-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: conjur-follower-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: conjur-follower-operator.v2.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0 <2.5.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:680a2044bc0e94fbc9376cd1eed9d067de91657f566c9752654480c939447841
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:75fb5de2c111084ffcc3d973e6c63f1adfae78e29f6fd02152419fbca553ce7b
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:60b293c23587072b3376f83afee95b0207228c5c7b80483b025adda572c3551d
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:cf46bdb4238cdb1dc51329822da55b7836958be2344d26f9e1819c0aa97fece6
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:b3096c17e7f3b9d723c906619dde293ea084e78c0da48c534335a8eee07b23c7
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:bee2cd31b3620a31e98f9d21326a2cdcecb36479099b06cc9660e4e83354d940
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:383776c23556a0a327a0bacc843d22c4b1516082ddd0cfb8a8275500f2eb841a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:a9b109c7f3bfba5553aa4366338fa211896564f8e33856d75b4918acf8429f1e
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:ed53aea1866b1e2262386ac38cb9e1e47cc3f9047ae47f49070e2f506af6ca5a
        version: 2.5.3
      entries:
      - name: conjur-follower-operator.v2.5.3
        version: 2.5.3
      name: "2.5"
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: preview
    - currentCSV: conjur-follower-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "conjur.cyberark.com/v1",
                "kind": "ConjurFollower",
                "metadata": {
                  "labels": {
                    "app": "conjur-follower"
                  },
                  "name": "my-conjur-follower",
                  "namespace": "cyberark-conjur"
                },
                "spec": {
                  "images": {
                    "configurator": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3",
                    "conjur": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a",
                    "failoverRebaser": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370",
                    "info": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89",
                    "nginx": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475",
                    "postgres": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f",
                    "syslogNg": "registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813"
                  },
                  "livenessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "master": {
                    "account": "demo",
                    "audience": "https://conjur.host.name/",
                    "authentication": "authn-jwt",
                    "authenticatorID": "my-authenticator-id",
                    "caCertificateFrom": {
                      "configMapKeyRef": {
                        "key": "ca.crt",
                        "name": "conjur-master-ca.crt"
                      }
                    },
                    "hostname": "conjur-master.mycompany.local"
                  },
                  "readinessProbeSettings": {
                    "postgres": {
                      "failureThreshold": 240,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1,
                      "timeoutSeconds": 3
                    }
                  },
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: ""
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: CyberArk
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConjurFollower is the Schema for the conjurfollowers API
            displayName: ConjurFollower
            kind: ConjurFollower
            name: conjurfollowers.conjur.cyberark.com
            version: v1
        description: |-
          Use this Operator for managing Conjur Followers in an OpenShift cluster.

          CyberArk products secure your most sensitive and high-value assets—and supporting your Identity Security goals is our top priority. We offer 24/7 service for high priority issues to all customers with resources across ten countries and in all continents.

          CyberArk Conjur makes secrets management simple. Conjur provides a seamless open source interface to securely authenticate, control, and audit non-human access across tools, applications, containers, and cloud environments via robust secrets management.
        displayName: Conjur Enterprise Follower Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - credential
        - cyber
        - cyberark
        - conjur
        - enterprise
        - management
        - password
        - secret
        - secrets management
        - security
        - vault
        links:
        - name: Conjur Secrets Manager Enterprise Documentation
          url: https://docs.cyberark.com/Product-Doc/OnlineHelp/AAM-DAP/Latest/en/Content/Resources/_TopNav/cc_Home.htm
        maintainers:
        - email: conj_maintainers@cyberark.com
          name: Conjur Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-nginx@sha256:5c85b99a5995ab09052325b58d734e3732de6f332c16354def7b7d8acd26d475
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator-bundle@sha256:ca749c884679bab810b04bae61976f904f31a54e2ac361410db88f2677de6f7c
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-operator@sha256:c5337633eb3cc272f989e3e70c2e9bace9c819bba9e817537c7a691a8b6ee596
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-conjur@sha256:f435e4af1080777a280fc7340b9af25a1f5d4f921afff14f1a511ee1ac387e0a
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-syslog-ng@sha256:a9211dbab74c58fe72e990c2910e4291e3be3c041850874782fab40eebcb2813
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-configurator@sha256:64a746d6a7b4b955daaf1f70d188f26705e9f80e8a2ded728d1d47a648fddea3
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-info@sha256:90c8b2549ffdd5b0fac3b1a0c9dcea5f27516df604218b5524660b88a9c2ab89
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-postgres@sha256:d8eb50c00e52ee089ef991aa5550733eab4c2fd7d1974709f843a5a42692be9f
        - registry.connect.redhat.com/cyberark/conjur-openshift-follower-failover-rebaser@sha256:90ed18b7665f3f449ff8d01bb3d31cb67568fcbbd21372324718e30ba13b8370
        version: 2.6.3
      entries:
      - name: conjur-follower-operator.v2.6.3
        version: 2.6.3
      - name: conjur-follower-operator.v2.6.2
        version: 2.6.2
      - name: conjur-follower-operator.v2.6.1
        version: 2.6.1
      - name: conjur-follower-operator.v2.6.0
        version: 2.6.0
      - name: conjur-follower-operator.v2.5.0
        version: 2.5.0
      - name: conjur-follower-operator.v2.4.4
        version: 2.4.4
      - name: conjur-follower-operator.v2.4.3
        version: 2.4.3
      - name: conjur-follower-operator.v2.4.2
        version: 2.4.2
      - name: conjur-follower-operator.v2.3.5
        version: 2.3.5
      - name: conjur-follower-operator.v2.3.4
        version: 2.3.4
      - name: conjur-follower-operator.v2.3.1-290
        version: 2.3.1
      - name: conjur-follower-operator.v2.3.0
        version: 2.3.0
      - name: conjur-follower-operator.v2.2.4
        version: 2.2.4
      - name: conjur-follower-operator.v2.2.2
        version: 2.2.2
      - name: conjur-follower-operator.v2.2.1
        version: 2.2.1
      name: stable
    defaultChannel: stable
    packageName: conjur-follower-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-rds-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-rds-controller.v1.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "rds.services.k8s.aws/v1alpha1",
                "kind": "DBSubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
          createdAt: "2025-09-22T17:27:42Z"
          description: AWS RDS controller is a service controller for managing RDS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DBClusterEndpoint represents the state of an AWS rds DBClusterEndpoint
              resource.
            displayName: DBClusterEndpoint
            kind: DBClusterEndpoint
            name: dbclusterendpoints.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterParameterGroup represents the state of an AWS rds
              DBClusterParameterGroup resource.
            displayName: DBClusterParameterGroup
            kind: DBClusterParameterGroup
            name: dbclusterparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBCluster represents the state of an AWS rds DBCluster resource.
            displayName: DBCluster
            kind: DBCluster
            name: dbclusters.rds.services.k8s.aws
            version: v1alpha1
          - description: DBClusterSnapshot represents the state of an AWS rds DBClusterSnapshot
              resource.
            displayName: DBClusterSnapshot
            kind: DBClusterSnapshot
            name: dbclustersnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBInstance represents the state of an AWS rds DBInstance
              resource.
            displayName: DBInstance
            kind: DBInstance
            name: dbinstances.rds.services.k8s.aws
            version: v1alpha1
          - description: DBParameterGroup represents the state of an AWS rds DBParameterGroup
              resource.
            displayName: DBParameterGroup
            kind: DBParameterGroup
            name: dbparametergroups.rds.services.k8s.aws
            version: v1alpha1
          - description: DBProxy represents the state of an AWS rds DBProxy resource.
            displayName: DBProxy
            kind: DBProxy
            name: dbproxies.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSnapshot represents the state of an AWS rds DBSnapshot
              resource.
            displayName: DBSnapshot
            kind: DBSnapshot
            name: dbsnapshots.rds.services.k8s.aws
            version: v1alpha1
          - description: DBSubnetGroup represents the state of an AWS rds DBSubnetGroup
              resource.
            displayName: DBSubnetGroup
            kind: DBSubnetGroup
            name: dbsubnetgroups.rds.services.k8s.aws
            version: v1alpha1
          - description: GlobalCluster represents the state of an AWS rds GlobalCluster
              resource.
            displayName: GlobalCluster
            kind: GlobalCluster
            name: globalclusters.rds.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Relational Database Service ("RDS") resources in AWS from within your Kubernetes cluster.

          **About Amazon RDS**

          Amazon Relational Database Service (Amazon RDS) makes it easy to set up, operate, and scale a relational database in the cloud. It provides cost-efficient and resizable capacity while automating time-consuming administration tasks such as hardware provisioning, database setup, patching and backups. It frees you to focus on your applications so you can give them the fast performance, high availability, security and compatibility they need.

          Amazon RDS is available on several database instance types - optimized for memory, performance or I/O -and provides you with six familiar database engines to choose from, including Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle Database, and SQL Server. You can use the AWS Database Migration Service to easily migrate or replicate your existing databases to Amazon RDS.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RDS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rds
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RDS Developer Resources
          url: https://aws.amazon.com/rds/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: rds maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/rds-controller:1.6.2
        - quay.io/community-operator-pipeline-prod/ack-rds-controller@sha256:80b0d6e33fc94471d05e05f6d370be3a84fcb5c2044486e007d57360eb69d9d7
        version: 1.6.2
      entries:
      - name: ack-rds-controller.v1.6.2
        version: 1.6.2
      - name: ack-rds-controller.v1.6.1
        version: 1.6.1
      - name: ack-rds-controller.v1.6.0
        version: 1.6.0
      - name: ack-rds-controller.v1.5.0
        version: 1.5.0
      - name: ack-rds-controller.v1.4.26
        version: 1.4.26
      - name: ack-rds-controller.v1.4.25
        version: 1.4.25
      - name: ack-rds-controller.v1.4.24
        version: 1.4.24
      - name: ack-rds-controller.v1.4.22
        version: 1.4.22
      - name: ack-rds-controller.v1.4.21
        version: 1.4.21
      - name: ack-rds-controller.v1.4.20
        version: 1.4.20
      - name: ack-rds-controller.v1.4.19
        version: 1.4.19
      - name: ack-rds-controller.v1.4.18
        version: 1.4.18
      - name: ack-rds-controller.v1.4.17
        version: 1.4.17
      - name: ack-rds-controller.v1.4.16
        version: 1.4.16
      - name: ack-rds-controller.v1.4.15
        version: 1.4.15
      - name: ack-rds-controller.v1.4.14
        version: 1.4.14
      - name: ack-rds-controller.v1.4.13
        version: 1.4.13
      - name: ack-rds-controller.v1.4.12
        version: 1.4.12
      - name: ack-rds-controller.v1.4.11
        version: 1.4.11
      - name: ack-rds-controller.v1.4.10
        version: 1.4.10
      - name: ack-rds-controller.v1.4.8
        version: 1.4.8
      - name: ack-rds-controller.v1.4.7
        version: 1.4.7
      - name: ack-rds-controller.v1.4.6
        version: 1.4.6
      - name: ack-rds-controller.v1.4.4
        version: 1.4.4
      - name: ack-rds-controller.v1.4.3
        version: 1.4.3
      - name: ack-rds-controller.v1.4.2
        version: 1.4.2
      - name: ack-rds-controller.v1.4.1
        version: 1.4.1
      - name: ack-rds-controller.v1.4.0
        version: 1.4.0
      - name: ack-rds-controller.v1.3.0
        version: 1.3.0
      - name: ack-rds-controller.v1.2.5
        version: 1.2.5
      - name: ack-rds-controller.v1.2.4
        version: 1.2.4
      - name: ack-rds-controller.v1.2.3
        version: 1.2.3
      - name: ack-rds-controller.v1.2.2
        version: 1.2.2
      - name: ack-rds-controller.v1.2.1
        version: 1.2.1
      - name: ack-rds-controller.v1.2.0
        version: 1.2.0
      - name: ack-rds-controller.v1.1.11
        version: 1.1.11
      - name: ack-rds-controller.v1.1.9
        version: 1.1.9
      - name: ack-rds-controller.v1.1.8
        version: 1.1.8
      - name: ack-rds-controller.v1.1.7
        version: 1.1.7
      - name: ack-rds-controller.v1.1.6
        version: 1.1.6
      - name: ack-rds-controller.v1.1.5
        version: 1.1.5
      - name: ack-rds-controller.v1.1.4
        version: 1.1.4
      - name: ack-rds-controller.v1.1.3
        version: 1.1.3
      - name: ack-rds-controller.v1.1.2
        version: 1.1.2
      - name: ack-rds-controller.v1.1.1
        version: 1.1.1
      - name: ack-rds-controller.v1.1.0
        version: 1.1.0
      - name: ack-rds-controller.v1.0.0
        version: 1.0.0
      - name: ack-rds-controller.v0.1.3
        version: 0.1.3
      - name: ack-rds-controller.v0.1.1
        version: 0.1.1
      - name: ack-rds-controller.v0.1.0
        version: 0.1.0
      - name: ack-rds-controller.v0.0.30
        version: 0.0.30
      - name: ack-rds-controller.v0.0.29
        version: 0.0.29
      - name: ack-rds-controller.v0.0.28
        version: 0.0.28
      - name: ack-rds-controller.v0.0.27
        version: 0.0.27
      - name: ack-rds-controller.v0.0.26
        version: 0.0.26
      - name: ack-rds-controller.v0.0.25
        version: 0.0.25
      - name: ack-rds-controller.v0.0.24
        version: 0.0.24
      - name: ack-rds-controller.v0.0.23
        version: 0.0.23
      - name: ack-rds-controller.v0.0.22
        version: 0.0.22
      - name: ack-rds-controller.v0.0.21
        version: 0.0.21
      - name: ack-rds-controller.v0.0.20
        version: 0.0.20
      - name: ack-rds-controller.v0.0.19
        version: 0.0.19
      - name: ack-rds-controller.v0.0.18
        version: 0.0.18
      - name: ack-rds-controller.v0.0.17
        version: 0.0.17
      - name: ack-rds-controller.v0.0.16
        version: 0.0.16
      - name: ack-rds-controller.v0.0.15
        version: 0.0.15
      - name: ack-rds-controller.v0.0.14
        version: 0.0.14
      - name: ack-rds-controller.v0.0.13
        version: 0.0.13
      - name: ack-rds-controller.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-rds-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-messaging-topology-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-messaging-topology-operator.v1.17.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{"apiVersion":"rabbitmq.com/v1beta1","kind":"Binding","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Exchange","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Federation","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Permission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Policy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"OperatorPolicy","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Queue","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"SchemaReplication","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Shovel","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1alpha1","kind":"SuperStream","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"User","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"Vhost","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}},{"apiVersion":"rabbitmq.com/v1beta1","kind":"TopicPermission","metadata":{"name":"test"},"spec":{"rabbitmqClusterReference":{"name":"rabbitmq-cluster"}}}]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
          createdAt: 2025-08-21T11:15:42
          description: Kubernetes operator to allow developers to create and manage
            RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative
            Kubernetes API.
          repository: https://github.com/rabbitmq/messaging-topology-operator
          support: VMware Tanzu
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding
            displayName: Binding
            kind: Binding
            name: bindings.rabbitmq.com
            version: v1beta1
          - description: Exchange
            displayName: Exchange
            kind: Exchange
            name: exchanges.rabbitmq.com
            version: v1beta1
          - description: Federation
            displayName: Federation
            kind: Federation
            name: federations.rabbitmq.com
            version: v1beta1
          - description: Permission
            displayName: Permission
            kind: Permission
            name: permissions.rabbitmq.com
            version: v1beta1
          - description: TopicPermission
            displayName: TopicPermission
            kind: TopicPermission
            name: topicpermissions.rabbitmq.com
            version: v1beta1
          - description: Policy
            displayName: Policy
            kind: Policy
            name: policies.rabbitmq.com
            version: v1beta1
          - description: OperatorPolicy
            displayName: OperatorPolicy
            kind: OperatorPolicy
            name: operatorpolicies.rabbitmq.com
            version: v1beta1
          - description: Queue
            displayName: Queue
            kind: Queue
            name: queues.rabbitmq.com
            version: v1beta1
          - description: Schema Replication
            displayName: SchemaReplication
            kind: SchemaReplication
            name: schemareplications.rabbitmq.com
            version: v1beta1
          - description: Shovel
            displayName: Shovel
            kind: Shovel
            name: shovels.rabbitmq.com
            version: v1beta1
          - description: Super Stream
            displayName: SuperStream
            kind: SuperStream
            name: superstreams.rabbitmq.com
            version: v1alpha1
          - description: User
            displayName: User
            kind: User
            name: users.rabbitmq.com
            version: v1beta1
          - description: Vhost
            displayName: Vhost
            kind: Vhost
            name: vhosts.rabbitmq.com
            version: v1beta1
          required:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: |
          ## About the managed application
          RabbitMQ is an open-source message-broker software that originally implemented the Advanced Message Queuing Protocol and has since been extended with a plug-in architecture to support Streaming Text Oriented Messaging Protocol, MQ Telemetry Transport, and other protocols.
          ## About this Operator
          Kubernetes operator to allow developers to create and manage RabbitMQ messaging topologies within a RabbitMQ cluster using a declarative Kubernetes API. A Messaging topology is the collection of objects such as exchanges, queues, bindings and policies that provides specific messaging or streaming scenario. This operator is used with RabbitMQ clusters deployed via the RabbitMQ Cluster Kubernetes Operator. This repository contains custom controllers and custom resource definitions (CRDs) enabling a declarative API for RabbitMQ messaging topologies.
          ## Prerequisites for enabling this Operator
          The operator works on top of the RabbitMQ cluster operator.
          Messaging Topology Operator is tested with the latest release of RabbitMQ Cluster Operator. It uses the generated default user secret from RabbitmqCluster (set in rabbitmqcluster.status.binding) to authenticate with RabbitMQ server.
        displayName: rabbitmq-messaging-topology-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: GitHub
          url: https://github.com/rabbitmq/messaging-topology-operator
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/install-topology-operator.html
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-messaging-topology-operator@sha256:4c62fb6d686a20a28437905d4e1a9b93ac05853b4dbf29fcce34421a43268724
        - quay.io/rabbitmqoperator/messaging-topology-operator:1.17.4
        version: 1.17.4
      entries:
      - name: rabbitmq-messaging-topology-operator.v1.17.4
        version: 1.17.4
      - name: rabbitmq-messaging-topology-operator.v1.16.0
        version: 1.16.0
      - name: rabbitmq-messaging-topology-operator.v1.15.0
        version: 1.15.0
      - name: rabbitmq-messaging-topology-operator.v1.14.2
        version: 1.14.2
      - name: rabbitmq-messaging-topology-operator.v1.14.1
        version: 1.14.1
      - name: rabbitmq-messaging-topology-operator.v1.13.0
        version: 1.13.0
      - name: rabbitmq-messaging-topology-operator.v1.12.2
        version: 1.12.2
      - name: rabbitmq-messaging-topology-operator.v1.12.1
        version: 1.12.1
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-messaging-topology-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-hana-express-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-hana-express-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "db.sap-redhat.io/v1alpha1",
                "kind": "HanaExpress",
                "metadata": {
                  "name": "hanaexpress-sample"
                },
                "spec": {
                  "credential": {
                    "key": "hxepasswd.json",
                    "name": "hxepasswd"
                  },
                  "pvcSize": "1Gi"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
          createdAt: "2023-07-16T20:53:02Z"
          description: Manage the lifecycle of SAP Hana Express
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HanaExpress is the Schema for the hanaexpresses API
            displayName: Hana Express
            kind: HanaExpress
            name: hanaexpresses.db.sap-redhat.io
            version: v1alpha1
        description: Operator for managing the lifecycle of SAP Hana Express on Kubernetes
          platform
        displayName: SAP Hana Express Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Database
        - SAP
        links:
        - name: Sap Hana Express Operator
          url: https://github.com/redhat-sap/sap-hana-express-operator
        maintainers:
        - email: mjiao@redhat.com
          name: RedHat@SAPLinuxLab
        maturity: alpha
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/sap-hana-express-operator@sha256:0fa4f29bd7669690be87a08405bed95d3ead2293a125b28149de2b10b02d2cd6
        - quay.io/redhat-sap-cop/sap-hanaexpress-operator:0.1.4
        version: 1.0.2
      entries:
      - name: sap-hana-express-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: sap-hana-express-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CrowdStrike
      provider-url: https://crowdStrike.com
    name: falcon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: falcon-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconContainer",
                "metadata": {
                  "name": "falcon-sidecar-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "sidecar"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  },
                  "registry": {
                    "type": "crowdstrike"
                  }
                }
              },
              {
                "apiVersion": "falcon.crowdstrike.com/v1alpha1",
                "kind": "FalconNodeSensor",
                "metadata": {
                  "name": "falcon-node-sensor"
                },
                "spec": {
                  "falcon": {
                    "tags": [
                      "daemonset"
                    ],
                    "trace": "none"
                  },
                  "falcon_api": {
                    "client_id": "PLEASE_FILL_IN",
                    "client_secret": "PLEASE_FILL_IN",
                    "cloud_region": "autodiscover"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security,Monitoring
          containerImage: quay.io/crowdstrike/falcon-operator:0.8.1
          createdAt: "2024-06-06T10:00:00Z"
          description: '[DEPRECATED] Use the CrowdStrike Falcon Operator from the
            certified channel instead'
          operatorframework.io/suggested-namespace: falcon-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/CrowdStrike/falcon-operator
          support: Community Only
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FalconContainer is the Schema for the falconcontainers API
            displayName: Falcon Container
            kind: FalconContainer
            name: falconcontainers.falcon.crowdstrike.com
            version: v1alpha1
          - description: FalconNodeSensor is the Schema for the falconnodesensors
              API
            displayName: Falcon Node Sensor
            kind: FalconNodeSensor
            name: falconnodesensors.falcon.crowdstrike.com
            version: v1alpha1
        description: |-
          ## Deprecated
          This CrowdStrike community operator is deprecated, use the CrowdStrike Falcon Operator from the certified channel instead.

          ## About the CrowdStrike Falcon Operator
          The CrowdStrike Falcon Operator deploys CrowdStrike Falcon Workload Protection to the cluster. The operator exposes 2 custom resources that allows you to deploy either the Falcon Container Sensor or Falcon Node Sensor.

          ## About Custom Resources

          | Custom Resource                                       | Description                                                      |
          | :--------                                             | :------------                                                    |
          | [FalconContainer](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/container/README.md) | Manages installation of Falcon Container Sensor on the cluster   |
          | [FalconNodeSensor](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/resources/node/README.md)     | Manages installation of Falcon Linux Sensor on the cluster nodes |

          ## Installation and Deployment

          For installation and deployment of the CrowdStrike Falcon Operator and its Custom Resources, please read the [Installation and Deployment Guide](./docs/install_guide.md) and choose the deployment method that is right for your target environment.

          ## Getting Help
          If you encounter any issues while using the Falcon Operator, you can create an issue on our [Github repo](https://github.com/CrowdStrike/falcon-operator) for bugs, enhancements, or other requests.

          ## Contributing
          You can contribute by:

          * Raising any issues you find using Falcon Operator
          * Fixing issues by opening [Pull Requests](https://github.com/CrowdStrike/falcon-operator/pulls)
          * Improving documentation
          * Talking about the CrowdStrike Falcon Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/CrowdStrike/falcon-operator/issues).

          ## Additional Resources
           - CrowdStrike Container Security: [Product Page](https://www.crowdstrike.com/products/cloud-security/falcon-cloud-workload-protection/container-security/)
           - So You Think Your Containers Are Secure? Four Steps to Ensure a Secure Container Deployment: [Blog Post](https://www.crowdstrike.com/blog/four-steps-to-ensure-a-secure-containter-deployment/)
           - Container Security With CrowdStrike: [Blog Post](https://www.crowdstrike.com/blog/tech-center/container-security/)
           - To learn more about Falcon Container Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/support/documentation/146/falcon-container-sensor-for-linux), [Release Notes](https://falcon.crowdstrike.com/support/news/release-notes-falcon-container-sensor-for-linux)
           - To learn more about Falcon Sensor for Linux: [Deployment Guide](https://falcon.crowdstrike.com/documentation/20/falcon-sensor-for-linux#kubernetes-support)
           - [Developer Documentation](https://github.com/CrowdStrike/falcon-operator/tree/main/docs/developer_guide.md)
        displayName: '[DEPRECATED] CrowdStrike Operator'
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - runtime-protection
        links:
        - name: Falcon Operator
          url: https://github.com/CrowdStrike/falcon-operator
        maintainers:
        - email: integrations@crowdstrike.com
          name: CrowdStrike Solutions Architecture
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: CrowdStrike
          url: https://crowdStrike.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/community-operator-pipeline-prod/falcon-operator@sha256:3916adf5083c83838564b1ed039decbadb2efa4ec288699ef0137a66983a3cc4
        - quay.io/crowdstrike/falcon-operator:0.8.1
        version: 0.8.2
      entries:
      - name: falcon-operator.v0.8.2
        version: 0.8.2
      - name: falcon-operator.v0.8.1
        version: 0.8.1
      name: alpha
    defaultChannel: alpha
    packageName: falcon-operator
    provider:
      name: CrowdStrike
      url: https://crowdStrike.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable
    - currentCSV: elasticsearch-operator.v5.8.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Elasticsearch",
                "metadata": {
                  "name": "elasticsearch"
                },
                "spec": {
                  "indexManagement": {
                    "mappings": [
                      {
                        "aliases": [
                          "infra",
                          "logs.infra"
                        ],
                        "name": "infra",
                        "policyRef": "infra-policy"
                      }
                    ],
                    "policies": [
                      {
                        "name": "infra-policy",
                        "phases": {
                          "delete": {
                            "minAge": "2d",
                            "namespaceSpec": [
                              {
                                "minAge": "5h",
                                "namespace": "openshift-monitoring"
                              }
                            ],
                            "pruneNamespacesInterval": "24h"
                          },
                          "hot": {
                            "actions": {
                              "rollover": {
                                "maxAge": "8h"
                              }
                            }
                          }
                        },
                        "pollInterval": "30m"
                      }
                    ]
                  },
                  "managementState": "Managed",
                  "nodeSpec": {
                    "resources": {
                      "limits": {
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "512Mi"
                      }
                    }
                  },
                  "nodes": [
                    {
                      "nodeCount": 1,
                      "roles": [
                        "client",
                        "data",
                        "master"
                      ],
                      "storage": {
                        "size": "20G"
                      }
                    }
                  ],
                  "redundancyPolicy": "ZeroRedundancy"
                }
              },
              {
                "apiVersion": "logging.openshift.io/v1",
                "kind": "Kibana",
                "metadata": {
                  "name": "kibana"
                },
                "spec": {
                  "managementState": "Managed",
                  "nodeSelector": {},
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "memory": "512Mi"
                    },
                    "requests": {
                      "memory": "512Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
          createdAt: "2024-09-19T15:14:33Z"
          description: |
            The Elasticsearch Operator for OCP provides a means for configuring and managing an Elasticsearch cluster for tracing and cluster logging.
            ## Prerequisites and Requirements
            ### Elasticsearch Operator Namespace
            The Elasticsearch Operator must be deployed to the global operator group namespace
            ### Memory Considerations
            Elasticsearch is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Elasticsearch cluster.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory. Each ES node can operate with a
            lower memory setting though this is not recommended for production deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.6.0-0 <5.8.21'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: AOS Cluster Logging, Jaeger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Elasticsearch cluster instance
            displayName: Elasticsearch
            kind: Elasticsearch
            name: elasticsearches.logging.openshift.io
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.logging.openshift.io
            version: v1
        description: "### IN OCP 4.16 AND FORWARD \n\nELASTICSEARCH IS ONLY SUPPORTED
          FOR **SERVICEMESH, TRACING, AND KIALI**, AND THIS OPERATOR **WILL BE REMOVED
          FROM THE OPENSHIFT OPERATOR CATALOG IN NOVEMBER 2025**.  ELASTICSEARCH IS
          **NO LONGER SUPPORTED FOR LOG STORAGE**. **KIBANA IS NO LONGER SUPPORTED**
          IN OCP 4.16 AND FORWARD. \n\n### About \n\nThe Elasticsearch Operator for
          OCP provides a means for configuring and managing an Elasticsearch cluster
          for use in tracing \nand cluster logging as well as a Kibana instance to
          connect to it.\nThis operator only supports OCP Cluster Logging and Jaeger.
          \ It is tightly coupled to each and is not currently capable of\nbeing used
          as a general purpose manager of Elasticsearch clusters running on OCP.\n\nPlease
          note: For a general purpose Elasticsearch operator, please use Elastic's
          Elasticsearch (ECK) Operator [here](https://catalog.redhat.com/software/containers/elastic/eck-operator/5fabf6d1ecb52450895164be?container-tabs=gti)\n\nIt
          is recommended that this operator be installed in the `openshift-operators-redhat`
          namespace to \nproperly support the Cluster Logging and Jaeger use cases.\n\nOnce
          installed, the operator provides the following features for **Elasticsearch**:\n*
          **Create/Destroy**: Deploy an Elasticsearch cluster to the same namespace
          in which the elasticsearch CR is created.\n* **Update**: Changes to the
          elasticsearch CR will be scheduled and applied to the cluster in a controlled
          manner (most often as a rolling upgrade).\n* **Cluster health**: The operator
          will periodically poll the cluster to evaluate its current health (such
          as the number of active shards and if any cluster nodes have reached their
          storage watermark usage).\n* **Redeploys**: In the case where the provided
          secrets are updated, the Elasticsearch Operator will schedule and perform
          a full cluster restart.\n* **Index management**: The Elasticsearch Operator
          will create cronjobs to perform index management such as roll over and deletion.\n\nOnce
          installed, the operator provides the following features for **Kibana**:\n*
          **Create/Destroy**: Deploy a Kibana instance to the same namespace in which
          the kibana CR is created (this should be the same namespace as the elasticsearch
          CR).\n* **Update**: Changes to the kibana CR will be scheduled and applied
          to the cluster in a controlled manner.\n* **Redeploys**: In the case where
          the provided secrets are updated, the Elasticsearch Operator will perform
          a restart.\n\n### Additionally provided features\n* Out of the box multitenancy
          that is integrated with OCP user access control.\n* Document Level Security\n*
          mTLS communication between Elasticsearch, Kibana, Index Management cronjobs,
          and CLO's Fluentd\n* OCP prometheus dashboard for Elasticsearch clusters\n*
          Prometheus Alerting rules  \n"
        displayName: OpenShift Elasticsearch Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - jaeger
        links:
        - name: Elastic
          url: https://www.elastic.co/
        - name: Elasticsearch Operator
          url: https://github.com/openshift/elasticsearch-operator
        maintainers:
        - email: aos-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/elasticsearch-operator-bundle@sha256:e1798fb819576860e1f3988e05b7254e6c6600dde51075e69cd0202b42363e6e
        - registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:b26d2218aa6b1efea83a1717d62f81009c97e8082434b918937c609f5cdab912
        - registry.redhat.io/openshift-logging/elasticsearch-rhel9-operator@sha256:78ffdf26b40a000d06eddfa4f66fdca9a11dc21c82a3492eac7cb9f4338dfcf3
        - registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8d603100978c655020a8f4d81c527162348f468d93989edbc00463b8297fb8a5
        - registry.redhat.io/openshift-logging/kibana6-rhel8@sha256:a2c200667738fdb079ba11ed3b0df3c5901373428d5e560bc1efea9d5e1ddd67
        - registry.redhat.io/openshift-logging/logging-curator5-rhel9@sha256:394e0f46fb8be44c720b29aed716443747a636d2758226c9386f930f95758470
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d646427f21f61e65037a5e45df73f73832370bb079252767ca1c02a8088d2abf
        version: 5.8.21
      entries:
      - name: elasticsearch-operator.v5.8.21
        version: 5.8.21
      - name: elasticsearch-operator.v5.8.20
        version: 5.8.20
      - name: elasticsearch-operator.v5.8.19
        version: 5.8.19
      - name: elasticsearch-operator.v5.8.18
        version: 5.8.18
      - name: elasticsearch-operator.v5.8.17
        version: 5.8.17
      - name: elasticsearch-operator.v5.8.16
        version: 5.8.16
      - name: elasticsearch-operator.v5.8.15
        version: 5.8.15
      - name: elasticsearch-operator.v5.8.14
        version: 5.8.14
      - name: elasticsearch-operator.v5.8.13
        version: 5.8.13
      - name: elasticsearch-operator.v5.8.12
        version: 5.8.12
      - name: elasticsearch-operator.v5.8.11
        version: 5.8.11
      - name: elasticsearch-operator.v5.8.10
        version: 5.8.10
      - name: elasticsearch-operator.v5.8.9
        version: 5.8.9
      - name: elasticsearch-operator.v5.8.8
        version: 5.8.8
      - name: elasticsearch-operator.v5.8.7
        version: 5.8.7
      - name: elasticsearch-operator.v5.8.6
        version: 5.8.6
      - name: elasticsearch-operator.v5.8.5
        version: 5.8.5
      - name: elasticsearch-operator.v5.8.4
        version: 5.8.4
      - name: elasticsearch-operator.v5.8.3
        version: 5.8.3
      - name: elasticsearch-operator.v5.8.2
        version: 5.8.2
      - name: elasticsearch-operator.v5.8.1
        version: 5.8.1
      - name: elasticsearch-operator.v5.8.0
        version: 5.8.0
      name: stable-5.8
    defaultChannel: stable-5.8
    packageName: elasticsearch-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/lightspeed-service
    name: lightspeed-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: alpha
    - currentCSV: lightspeed-operator.v1.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ols.openshift.io/v1alpha1",
                "kind": "OLSConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightspeed-operator",
                    "app.kubernetes.io/instance": "olsconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "olsconfig",
                    "app.kubernetes.io/part-of": "lightspeed-operator"
                  },
                  "name": "cluster"
                },
                "spec": {
                  "llm": {
                    "providers": [
                      {
                        "credentialsSecretRef": {
                          "name": "credentials"
                        },
                        "models": [
                          {
                            "name": "gpt-3.5-turbo-1106"
                          }
                        ],
                        "name": "OpenAI"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-08-28T16:37:51Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-lightspeed
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus", "OpenShift Kubernetes Engine", "OpenShift Virtualization
            Engine"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/lightspeed-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Red Hat OpenShift Lightspeed instance. OLSConfig is the Schema
              for the olsconfigs API
            displayName: OLSConfig
            kind: OLSConfig
            name: olsconfigs.ols.openshift.io
            version: v1alpha1
        description: |-
          OpenShift Lightspeed Operator provides generative AI-based virtual assistant which integrates into the OpenShift web console. OpenShift Lightspeed can answer natural language questions related to OpenShift Container Platform.

          To configure the OpenShift Lightspeed Operator, you need a Large Language Model (LLM) . You can host the LLM using RHELAI or RHOAI or use one of the models from the following Software-as-a-Service providers:

          - OpenAI
          - Microsoft Azure OpenAI
          - IBM WatsonX

          For more information, see [About OpenShift Lightspeed](https://docs.redhat.com/en/documentation/red_hat_openshift_lightspeed/1.0tp1/html/about/ols-about-openshift-lightspeed) in the official product documentation.

          Questions, comments, concerns, or just want to talk to the OpenShift Lightspeed team? [Email us](mailto:openshift-lightspeed-contact-requests@redhat.com)!
        displayName: OpenShift Lightspeed Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ai
        - assistant
        - openshift
        - llm
        links:
        - name: Lightspeed Operator
          url: https://github.com/openshift/lightspeed-operator
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/lightspeed-service
        relatedImages:
        - registry.redhat.io/openshift-lightspeed/lightspeed-service-api-rhel9@sha256:ec359967a87bc7ee2c1d6a61e553acbdb0dc85d6f240a1865bf57d3ba5dc7193
        - registry.redhat.io/openshift-lightspeed/lightspeed-console-plugin-rhel9@sha256:dfc42c2651ac8793f61ff0e54334ecb006dc4a77487bc7c6878cdf47bd4ac080
        - registry.redhat.io/openshift-lightspeed/lightspeed-rhel9-operator@sha256:b12803642df4d9b3ad0cf800315021483f543c6a8827b3901e57bdb65f10dc48
        - registry.redhat.io/openshift-lightspeed/openshift-mcp-server-rhel9@sha256:f1e4521ff2cc076f361c04d8c9aeac09ffcedb89a2f1dbe71f85caa27472adab
        - registry.redhat.io/rhel9/postgresql-16@sha256:6d2cab6cb6366b26fcf4591fe22aa5e8212a3836c34c896bb65977a8e50d658b
        - registry.redhat.io/openshift-lightspeed/lightspeed-operator-bundle@sha256:78627f5bed3fa4fc207439a6144cb28c7d8493119394ad2a75382473418a70ac
        version: 1.0.5
      entries:
      - name: lightspeed-operator.v1.0.5
        version: 1.0.5
      name: stable
    defaultChannel: stable
    packageName: lightspeed-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/lightspeed-service
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v21.1.10",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
          createdAt: "2021-10-19T15:47:03Z"
          description: CockroachDB Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:2664d78fbe88c368927ddbf80704f430e6ca2ce1f72cb4b7e1906dfe72be7cd0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fe0c615b4d1ef4ec6e089e85deae2b6bd85410be8ed525e2317c759669417f47
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:3edda46eccce8350440a0ea78dad2e41a3eaa0838fe57945ad5318739e968516
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:27ef61d9f320a52ee9617f2ab8a1d1a1eacc576a0add2d85d99f9413912b651f
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:87dcdbdc07904b217880e70484582c9d87dde7e4071e01cf9e2e6da43111190e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:38e940144e34b69dfb7d1cf31456e7a0016c5dd40d5626baa346446c7678ceb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1b56420753eac1126039b8ad4be82ddb59c89a4ec72fb7b6198cea88a10f3a91
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e39bce1d9b08a7d62eb05d53e29fabc668bf6a4e4ae3eb9994eebf5b642ff6cf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:cdab9fc62e07dd349703ae394adb48ab3d2281610df35e081f4b76e208b38c8e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:a5414e54c4c513c7877f925b09eadc4c28eb592d3e120bcf9c65ab10e07cd607
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:162a356cc8b423926634f83a1372b6321104d8de7a2c5b9b7b2681c769a88cc7
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:fec6b0f56f2b94f141c8553a63b2e08b9494d4e6c5dc109d73268621ca445c56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ab47ddd22328e0a1564814f049616c3e3569d24c0a290abee9308392c6a0de23
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:b10c121395f5bc46d61fdeddd35a0c8d9ae4e67a9a77d8ada90e0ea53666323e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ec1672ce8091b677689a31cf09dbde5d1697f4caddae0151040b0fb156722e9d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c1a07a8686bbadcb8c68ffba4e40e1285e581e24d43e5811f38a22d57fb9cc56
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0bd22d69db5ac570f30d3379afc67f6c2e46fc5b027c8aab42c3f562085d4672
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:de94036e9cc8618e0493313512d880ae145e94d3bff54ed51b0de66a88da1319
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:5eb59093ee81f8115d82ec6a8a1c35eefad055cbb3a175978cef7afe9196e6a0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:b02549e5c3673b8154441990afa05a478d1ecc2bd0a70af64fd693cb6b469cb0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:98b76cf63aabc39d026317905d6ba72eec3f0d35c0f770ebb8b789c3fa0709ee
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d9a756df2ed2536b938d7c24fbf3de149cd1a0f44a3a1a0cc5d042fe8980362e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:49421968126f9c98499bc0cf0c65729370ab566da3fbd1588c100b20052b972e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:6667919a41d304d5d4ade3ded4f11b42d722a995a4283e11e15320529f7f9abf
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1ed18af1d3ffc65c24946c817d7d04d5f01025d3fb6c9eb78940e343b989bed0
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:e67299fc0e79cff5b6caf9c0df154180dbaedc57b84f072acf68b7b2c958668d
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:c56f980b87a066d1e65ac8cff3d75040af9bd13093494cf4b4d30b5234d56edc
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:288ae92ebdfc848540ff80ef682b74e50809e9742cafce22b028112326d66b65
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:ba9da9aa9d662c0fa9fd16a40551eedf100869b008cf9c0c09cbb87a09fda697
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:1cea12847a0d236437f7fc28fba476fe512c343cd020d21bb50316e50da6ffd6
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:25791b76b0d4b7220dadab3a405b3fdc330ba264ef6c9fbfdfb924c154c4bb5e
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:9f5dfcde260804cd2d6976a6b3544ddabdd899365be2104210dc78ba13755ec3
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:bcf3f976437c4ee52e922aa5d7e382a86a031d780b711c08c6c158dbf2d26cfa
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:d290099496d070f135b5cc1e6bd59607512524d399df81698f11cfa09b8dca4a
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:0ef0234b6fd42977b8a9eda2c59ca8194cc1c8cf6425b99d7bf8fde8feb826c5
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroach@sha256:867c46bb4d8ca3f2fb8dc94063facc99758344c7b9d04f10e50932023ef5d262
        - registry.marketplace.redhat.com/rhm/cockroachdb/cockroachdb-operator@sha256:9fba7cd5fdf589317911925cb69338508bf9de928b1817c6f117c612635bbab3
        version: 2.3.0
      entries:
      - name: cockroach-operator.v2.3.0
        version: 2.3.0
      name: beta
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:56Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cockroachdb-certified-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle-marketplace@sha256:9d84d46dd1eb74afa18466c73766402b46c2885ef483600aff83065f8065d8d1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified-rhmp
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eap
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eap-operator.v3.2.2
      currentCSVDesc:
        annotations:
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "false"
          containerImage: registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
          createdAt: "2025-07-10T14:38:00Z"
          description: Operator that creates and manages Java applications running
            on JBoss EAP.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Enterprise Application
            Platform", "Red Hat Runtimes", "Red Hat Integration", "Red Hat Process
            Automation"]'
          repository: https://github.com/wildfly/wildfly-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An application running on EAP application runtime.
            displayName: WildFlyServer
            kind: WildFlyServer
            name: wildflyservers.wildfly.org
            version: v1alpha1
        description: "# Red Hat JBoss Enterprise Application Platform (JBoss EAP)\n\nRed
          Hat® JBoss® Enterprise Application Platform (JBoss EAP) delivers enterprise-grade
          security, performance, and scalability in any environment.  Some features
          of JBoss EAP are:\n\n**Optimized for cloud and containers**: with simplified
          deployment and full Jakarta EE performance for applications in any environment.\n\n**Lightweight,
          flexible architecture**: which is well-suited for microservices and traditional
          applications.\n\n**Support and standardize microservices development**:
          With the JBoss Enterprise Application Platform expansion pack that supports
          Eclipse MicroProfile application programming interfaces (APIs).\n\n# JBoss
          EAP Operator\nOperators are software components that extend the Kubernetes
          API to customize how to manage applications and their components.\n\nThe
          JBoss EAP Operator facilitates the deployment and management of JBoss EAP
          on container orchestration platforms such as Red Hat OpenShift. The operator
          acts as an automated management tool, simplifying the process of provisioning,
          scaling, and updating JBoss EAP applications.\n\n# JBoss EAP Operator capabilities\nThe
          JBoss EAP Operator deploys and manages multiple instances of the JBoss EAP
          Java application in a Red Hat OpenShift cluster. \n\n**Simplifies management**:
          Creates, configures, manages, and seamlessly upgrades instances of complex
          stateful applications. No more manual configuration, making management a
          breeze.\n\n**Efficient scaling**: Effortlessly scales the instances of your
          JBoss EAP Java application based on demand.\n\n**Faster updates**: Seamlessly
          updates, enabling you to roll out new operator features and bug fixes without
          downtime\n\n**Safe transaction recovery**: Provides safe transaction recovery
          in your application cluster by verifying all transactions are completed
          before scaling down the replicas and marking a pod as clean for termination.\n\n#
          Operator installation and configuration\nLet's walk through the steps to
          install the JBoss EAP Operator on your OpenShift Cluster and deploy your
          first example application:\n\n**Step 1: Install JBoss EAP from OperatorHub**\n1.
          Click Install in the JBoss EAP operator window.\n2. In the Install Operator
          window, for Installation mode, select All namespaces on the cluster The
          operator is available in openshift-operators namespace.\n\n**Step 2: Build
          an application image**\n\nThe JBoss EAP Operator deploys an application
          image. If you do not have your JBoss EAP application image built yet, you
          can use the following example to build the image using Helm Charts.\n\n1.
          In the main navigation, click the dropdown menu (perspective switcher) and
          select Developer.\n2. Click Add > Helm Chart.\n3. Search for \"JBoss EAP
          8\" (Provided by Red Hat) and create a new Helm Chart release.\n4. On the
          Release name text box, type “eap-application”. \n5. Switch to the YAML view
          and replace the existing code with the following:\n\n    ```\n    build:\n
          \     uri: https://github.com/jboss-eap-up-and-running/eap8-getting-started\n
          \   deploy:\n      enabled: false\n    ```\n\n6. Click Create to build the
          application image.\n\n    The application image is created with the name
          eap-application:latest.\n\n**Step 3: Deploy your application with JBoss
          EAP Operator**\n\n1. In the main navigation, click the dropdown menu (perspective
          switcher) and select Developer.\n2. Click Add > Operator Backed > WildFlyServer
          > Create.\n3. Switch to the YAML view and replace the existing code with
          the following:\n\n    ```\n    apiVersion: wildfly.org/v1alpha1\n    kind:
          WildFlyServer\n    metadata:\n      name: eap-application\n    spec:\n      applicationImage:
          eap-application:latest\n      replicas: 1\n    ```\n\n 4. Click Create to
          deploy the application with JBoss EAP Operator.\n"
        displayName: JBoss EAP
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eap
        - java
        - open source
        - application runtime
        links:
        - name: JBoss EAP Homepage
          url: https://www.redhat.com/en/technologies/jboss-middleware/application-platform
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_jboss_enterprise_application_platform/
        - name: Blog
          url: https://developers.redhat.com/blog/
        - name: Quickstart
          url: https://github.com/jboss-developer/jboss-eap-quickstarts
        - name: Red Hat Customer Portal
          url: https://access.redhat.com/support
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat, Inc.
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/jboss-eap/eap-rhel9-operator@sha256:2361e73b8c517177e80f60d204996ebf65b166e4e8b60b329323a2b409c54f6c
        - registry.redhat.io/jboss-eap/eap-operator-bundle@sha256:0db70367917668964e2f4f5efdcef864ccc3f065848023e52caf4abdec728d9d
        version: 3.2.2
      entries:
      - name: eap-operator.v3.2.2
        version: 3.2.2
      - name: eap-operator.v3.2.1
        version: 3.2.1
      - name: eap-operator.v3.2.0
        version: 3.2.0
      - name: eap-operator.v3.1.9
        version: 3.1.9
      - name: eap-operator.v3.1.8
        version: 3.1.8
      - name: eap-operator.v3.1.7
        version: 3.1.7
      - name: eap-operator.v3.1.6
        version: 3.1.6
      - name: eap-operator.v3.1.5
        version: 3.1.5
      - name: eap-operator.v3.1.4
        version: 3.1.4
      - name: eap-operator.v3.1.3
        version: 3.1.3
      - name: eap-operator.v3.1.2
        version: 3.1.2
      - name: eap-operator.v3.1.1
        version: 3.1.1
      - name: eap-operator.v3.1.0
        version: 3.1.0
      - name: eap-operator.v3.0.2
        version: 3.0.2
      - name: eap-operator.v3.0.1
        version: 3.0.1
      - name: eap-operator.v3.0.0
        version: 3.0.0
      - name: eap-operator.v2.4.3
        version: 2.4.3
      - name: eap-operator.v2.4.2
        version: 2.4.2
      - name: eap-operator.v2.4.1
        version: 2.4.1
      - name: eap-operator.v2.4.0
        version: 2.4.0
      - name: eap-operator.v2.3.10
        version: 2.3.10
      - name: eap-operator.v2.3.9
        version: 2.3.9
      - name: eap-operator.v2.3.8
        version: 2.3.8
      - name: eap-operator.v2.3.7
        version: 2.3.7
      - name: eap-operator.v2.3.6
        version: 2.3.6
      - name: eap-operator.v2.3.5
        version: 2.3.5
      - name: eap-operator.v2.3.4
        version: 2.3.4
      - name: eap-operator.v2.3.3
        version: 2.3.3
      - name: eap-operator.v2.3.2
        version: 2.3.2
      - name: eap-operator.v2.3.1
        version: 2.3.1
      - name: eap-operator.v2.3.0
        version: 2.3.0
      - name: eap-operator.v2.2.2
        version: 2.2.2
      - name: eap-operator.v2.2.1
        version: 2.2.1
      - name: eap-operator.v2.2.0
        version: 2.2.0
      name: stable
    defaultChannel: stable
    packageName: eap
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cockroach Labs
      provider-url: ""
    name: cockroachdb-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroach-operator.v2.18.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "crdb.cockroachlabs.com/v1alpha1",
                "kind": "CrdbCluster",
                "metadata": {
                  "name": "crdb-tls-example",
                  "namespace": "placeholder"
                },
                "spec": {
                  "cockroachDBVersion": "v25.2.2",
                  "dataStore": {
                    "pvc": {
                      "spec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "10Gi"
                          }
                        },
                        "volumeMode": "Filesystem"
                      }
                    }
                  },
                  "nodes": 3,
                  "tlsEnabled": true
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
          createdAt: "2025-07-22T11:30:55Z"
          description: CockroachDB Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/cockroachdb/cockroach-operator
          support: Cockroach Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrdbCluster is the CRD for the cockroachDB clusters API
            displayName: CockroachDB Operator
            kind: CrdbCluster
            name: crdbclusters.crdb.cockroachlabs.com
            version: v1alpha1
        description: This operator launches and manages a 3+ node cluster of CockroachDB.
        displayName: CockroachDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - cockroachDB
        links:
        - name: Documentation
          url: https://docs.cockroachlabs.com
        - name: Operator Source
          url: https://github.com/cockroachdb/cockroach-operator
        - name: Cockroach Labs
          url: https://www.cockroachlabs.com
        maintainers:
        - email: support@cockroachlabs.com
          name: Cockroach Labs Support
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Cockroach Labs
        relatedImages:
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4b3aa590623a38f435b075de4c09f8a343d8ece8b673e6bf7c6df5d5eb4bdf4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9b5a2e0c006eb4be8e1444734a98743551f278a24e81c20bebc96f0f3b631eb0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d78c2e0709214f96794564f13386861f4b541e55b5bd316ad495cf104f7d40b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2af60025ecb3bb933b61328be4c2b2bfd0e7d26f53b72430629208505e45c6d2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2d89398f03736f8ffaa82b10fed0f19fba166d3d2145b26f25cb71f3c39acf07
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db4e5c2f27ba23e95a363c8174c01f0f9c6b6a27e0e581c86a38f3a1f68dc43c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d467383af41aa80c26172964c34152abeba45121599804e502984655b72179f0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8737b67b91983817e3e4ff3f47f0b070bd36162d309b60993529426c6979cb24
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:022c6e1a0de875474bc953718a481a0af526c49dc910094dcdce0b5677aced18
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b83eaff2e217c06098509709cead525b55b00ccd48019d8357e6231892008055
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator-bundle@sha256:096946dd896b6d9d937eb6529c3a23a84ccb129649c5a0389d0ab83989c28b53
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:31b161f9bf9028a1a62cd2eef6abda01257d8889bdc3159822642549a75e3429
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9492894eaa5f43d00ac80b60029b32a0b0a85ae88ba609b46f98702c9349a5bb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d14f77f70aea0422a906510c1ac2652ff76e131ed705cb002f80e13e887b0cd9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2a128b570352a5a7f82b27353f9b774caccce83e0e7bb2673e4802d96ad0bd51
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:82d8104a37349c513447e68ea3790efb256370632353176e5fc13c2e8df1bd44
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf7d3bea47638b763b312c7c11ca034a7ac62d993d75a13edceec74799b97e52
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b06bd7f3802e8fa2af14bb0b07804cad4cc0fbbc97da18d09a9106bf074ed6d8
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cbc74889e9ca5edbd5aaa243d7649bca0c4da2719239f8d0ec9827868112a771
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1304d7f871a95c8b087bad1b548217562792edc66c8fe07c621e05242b8cd53e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e23af8f9d32c3f80a52d79ceb4b359dda1d75ba0dd4cb8ed4b66ac2f4e40c69b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da3bd2e9ff2b08d5bb86b1f5faa95a865f85c5e8912df00b6b79190c1e518a2f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bd336715332f776a72aaf5c3ca52e62dd53a9e8521c1feb6e135cc868d694a39
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0e2c7023f1f97acc20c46a4944635a9c5f57bb9bb3dd534cf41fea1a9cfdfdbe
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:310310515625f099a928545865f7096997871ee4a16650a01c76c3799a18b684
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08496b84b3cc691909227be17e67bcd6f4d5185a93712f44d8a0985bd2c5fcd7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6a2ee8d62c586eb74ebbef6bef256e06e1ba5cb51c1ae55b1654adafc2ca8859
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:65923c37fecda2d3174212fa9d9a6f34241d065389336ea69a7be99ba16cedc2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0ec1dc184a2199736b4bd7de3241a357be7427caea5da1c931b08bc06f5c3dc0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b046aa0a42dae992a03685ed82a20835532d6d4bf048e2366ff7382be4b3138
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8fc8de4f3671a73f9aa747e5b23ec973d192fff6f2869c38985cd656063c1e3d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cc20270212f01fc231ff788e20b0476b8228c440bfc40430c0836d5eb702a042
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:db07461824e73410456ef540b0bce3c7fe550aa67a73d6174300ca74b8f75276
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4d8e804b35b73d7978c112748aea2d2bb68f65081b90bca62c39f2b6e2989378
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a59cafff8d4ff2c62204ece15880117011c30978823c232902c7182d747ec827
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e151bded670f4cbd165a8ef864324805667d3e84ce3ace0cdf8f9aad087fccf1
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1558a1f5eab79f1a8c6c85d81740cc539dc23025ea457c0019d93a72dc05d57a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a6e2fe24ad52913e2160f85b9fe665a4e4c4d79b164b4b899f69dd9d92f999e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5d6901b6b6fd6594d42bfc1842d050c7cffcb0867a5591a67469f0f860b1f21b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6ca927f137be28481e3a226e6c6bcb7c5ac27664984f17ffc6c1419cd7d08eb7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e0f21abea34825e55622c432a306f2e8aafd64d913eb78f2f601c51dc878a1da
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1a3ac544d87a9d8e7c92897c4c23f803753c7ab31467db91b5c51244afc2c6be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:10b9df672260c023dfb37166e156c0d7e0e13144dc3e485af922e8151fd3ab05
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cf001173bbf00cdf443b8d64a3ad724389f1cf83e3628d0b609019c8cddad35a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:468b0a40a33be1dd5a2a89ce745d7e9012bc7371f2f01cc0f8f12c0ac887db43
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:266165ab87b19a05d6796e8b15ff0eba632d6ac96e6393400464c6136dd500ec
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e18d73af3257c8cdf00245f2f718b56667524c1cc69d3db28e9b868a539ae16f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:226d7f1c3f0d70ac6532ad9490bebcdef44a40959b6aa7c11b0f2d1383a46217
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f53b07212320c6939d3bb4ddd4e6a0bce311ab5b63858eb1a8728531f0dbd510
        - registry.connect.redhat.com/cockroachdb/cockroachdb-operator@sha256:1eba1f4e0e09cc2205e28db8b90dbb4cf75cacaa10f7a397e657c8595d56fa82
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b0057de903654ea0315a2c96affcc8f50a92ede8ad121ed30edb3aaf0fb7b554
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fd6c9b8ca82ff4f2d0fe29166d1fc9bca6439cb40907be60f4d8f1b9735de2f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4b56c0311af14220a466aa11f25fd10fa38280825539bd02215c945bbafb861c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a6f1c43486d044fed9f59406118f429c8037b389bd8c8d7a1c7aeb89ed2661f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7ad12df8e258b03c4099d375d02d272bd774bf188d108901d5cf6cb865e57ef3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8e8c081c231b5b15eca483ceec16ea9190ba73aae6e421865162446eb584519a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5181196fff3bf8a5ec8fcfe4f1958cc69f285f0beb4f0b476b4b5e0912f3db42
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce3d94a3fbdcde5858beb3c265987759971d991c397c6486e19904bd4706f44d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8f1f2fd1aca914a1ef007117c85ac944055bbfb873a468a7374d239ef20b065e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4635550222775fc75737be5e5c0359303e3a23c4d1e1c296d7408cd15ad94aad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6bb15b36d64eebb6e4c1db5a5466e108b271d53383c58e0b6c78cec214756a9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:56109e57ee0379cf48644bcf8226a9238e01139cacc5499002c99f973f121911
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:fe4a40006d1e8439423de1417cc9bc55716d790b882485074bb2a5ec68d7d1ad
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d4f753151af4c3d3f5a1df9014461316249a6d0621b8fba5c4430307b3355251
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc1ab77a9537ccf236ce6dc370940dfedcfaa91b6e3db37bf2d5eade7ff84708
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:363d8feb560324a848c5771edd224dc19020013af7641979ee81ae8ee536d7e3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3bf017b2b2506724500f5e29017d1b956825dd0f61a7c0049b0ff084c3554b5b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:cad60044ad6573bd80b65e892a9ec0510dbe4fcbbfc6b51010ecf419f56f1024
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9097e135159f46396ef4649f724d61bb5b9113c488f989d404586fbb7486fd85
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5a54d564e08baefc2786f4b742b2507d31984664cd00333406aa7a3aa5b9c913
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:28ae9661b38a72decf413c6a6c3758e61098debf5162197c27a12a070c7a096d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1b6d934a7b6c42a61e5ad51380e69af341a575b42205299fd58bc0c6ff306f6a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5240f9f591263569fad7ee934e81fa8a1617de3c25da94ddd59ee5b2b3407f90
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:818f592c50d4ed5599faa053344f7ecd963003352f1faa3b72dc0f5e00a0e78b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bd75aa918355e77cd671c3e11ee807861378e96245245f4fdf243cba5749f9a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7bf36dab9b3257a4bba24d986f1451fc51a3a37023fbb988281f1941fd3faedd
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bc4b75ed71845d5b18e1d9163d480fd9d0d3cb5ebbcfed00a2e4e1b174c0a5de
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2b5c028f293c6f394e58a5495cb3dbfbedc215d0d03778d0217e045586d55a46
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f29c31bb267451b389608af3e8a4a28d6ca0691e67b77e8f57bb5e461b064893
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d8cfb221f67516675fdfde061f6ec5dfee1224ca7eb3d8be375c27d483f306af
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f42fa22aac3cc41ce72e144a72530ffbee765a392b24593dafd06cc098f8410b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a0faa52a999289b308dc4caf050c6d9fe2f42e112aa30e4af282ed1f21c5c346
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5458a8fd89aa664b60e59c35bd933975cd37d18e19b11589e4ca0c2844ea0ed4
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:91f7e59f9e35a0395756eafed906963c535e886125294d671d930ce959a8f511
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0fea6ed8a78ea624240ee6ffb3573d5cf1115186aa9180fcd7d1273351deaaa3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8a70ba12f610ac0e62066e417a0833bc1d01e91ac99d0dd794a2c0cd2a350164
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:46d9287ef11e3b21a36abfa31a26da3b93040bb6163b9223972a5dd1930fecff
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c74cf927a2b2c5098ad74ad0880996f8b042186a9911d888a231ef5b9e2a0715
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:44a0f40a6c46ba08297b482f3d96defe87b181c944803e5381c9acc42cabf97f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9d0bfb642090ff55a3cc3da0f72663cae18063622a4ec785ba99bc71ef4db116
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:87db6e079b96fb392d98ef5fc7063836bff10717efd75061016ce688be0ba38e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1e5d85f5db10fa79d09263db90daa44549bcc3b025e9f3f98cf0d4e51766394b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:53c7b960a2ed70a0998a4f8e28a4f78f7f676e01cffebe665e91e3a7629d88ed
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acc083f7dde1fbc557d858d03fc26b68b2056997fc1a457e4695650214d5b303
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:72ea3d0ef72972dba93362dc31366c5946a9898f5c16780b877316a97068ef8f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:8443c15eb1653dd9e5ec7aad54d4b8b092d41a0977e42b76e372e5e21280746e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e8f26fd132affbaeb8ef6b1807a646c83b301e5dc334a592dfb2cf8dc3dbaf10
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:743cda1eda6e273c798e3405a9a97e07dd6aa5ad4d5a52789be3b0f39a32a039
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f6371377718325235a2720dba0c97eb4321ee29b31a28cb893f76abd62baa339
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:32fdca575c334822e4356aab36a7ed97b685c065925fe85f1b8ba8425c57159b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1f6b1ec92251397dbcf7ed4a82f1f05619d64676aae2717f4eff1c33e4c2c27a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:5b3de46526d169fcd1318354545b3a18861843fa2022a23ed04d795493825287
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:29abb169ffefe36f6d69edd438a7135ee23059cedab8f87ad3b2d52361b96b35
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e82684b828256b3c2814930210031b3b70bc818514d4359a25f87ffdbf22be6e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:08cec123813304ab03bd66d72a09949667a874a42320bedafda391ebcdb2e56c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:086d7b435993fdf21fd59b0093f52105a9028d6b769398b0033f98a65dfb7e79
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f5a0ccc02dc9e938e484d5b5282ff650d1890d5f754c30a0c0d717989ed5d600
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:2171e1c75d0c665fe33e1ea334e605604c51de6dbc22d88a3a0ac604259c448f
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6408f89b0a88ec154f7ee98c314be574971c4c9aaf906b473d10124d822889be
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9ab968e4ca62f23bf9848e5475ef4076df7318a8560f63d5b2090a32af6fe4e2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:dd521d7bad62da054eec54179be75c643a5492d981e11c0e66db41902cd2182c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ec6eb7c28c213cc83b2d7919cd87988f9a07f12276eb7351d0915f1567a5b095
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f438fbbf9c87ca0d54c170e2c5105b1e265182e1c2ede48adea370e2be9dc895
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c9a31bdf19175880d0981e21dd19a1458b6e8cde8d4acbcb554f3bea3578a64e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1d590946b2f7c7ec0c7ec057a338e81f864d6a9e3d3ac4e0769aa5756a8e13fc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:ce1209ca6cc8724baee981c97c66e2852e512423bc0c0a16cdd80a6838b54031
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f74c8a5dea4560ad59c6f6f6ab1e699f063334ee8db704460f870508dbeaeff0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f3ffebd812881a830ff46c860984814dd61edcd51b478ebd3667759fb7710682
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:79d688bebe0643ef0dd6ba7108cdb34c0442470c0c20ae0df88b5d02a3476cd0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:c2a62179e8ca293d7e30b9da599e48795345cf07a72abd3a83063107b53b03cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:b77b80eed7ce149d2a3f228e5f188e76d73f32bc50ecde20bba9cfa7bdc4714c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:37bf9e1ec7e3eff47755fe7521223f31ec6aefc6e63901db3881b30898cf78f7
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7481c0cf3b579c8d98d94d4aa0334ca94bf2ccfcd25ad70b57c42b22bbf4bcee
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e5814748c5ab229a1dea5662a007063c0df06fbbfcfe041fd9a6df9ef67e07cc
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3f587e7b9f5a611013af2608e2151ed782859a52501b00adbc95739c56a8ff1e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d978678b3c254abd52f69f226b0aacc4b02c2aaca328c54ef10f6f9bb453582d
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7b907d03b001b758d8a4b5e898725c73dda2ec2aa4b7425b685320b947154d11
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:da367cf0ac52045002e1c186f8e6964267ad87d5c25f9e72fd2c9b9a98a32702
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:85a0e085901c2a73192556d5d70795d4b66cbcfe01d24c73b6fe6db8d5f2e5ca
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:4e5f7df1dc1e1db398c36d590431e7e5782897b209972d8e9e4671971c10d1b6
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:35ede65457d845e8c063ad7d9515d3a69f98e0751424ecbac484963678a0b1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:e50aab39722af22ccbf3d1db19e37972d52b0fc9a40998a64618dc966b2bac57
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:50393ad005fed19618d04f0716c6d55e2f1afce651e785c396638e852cac83b3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:d0014b1c19890b9e8ef1a4f3efe7eb5b269be5610ba6ababd4258b52d2f85e8a
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6366ba29f301a82215c05e60d78e1de66bcee85440ce2fc3c692712d4da32685
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:744193037d6b64cb73fecab15ebe022abc2b79ab0335b76ec1415ac6379b7dfa
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1374cbdc4ecf38e4b4226328f532e830b8932b34d9449bf2fbb36e60d0310521
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:3b78a4d3864e16d270979d9d7756012abf09e6cdb7f1eb4832f6497c26281099
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9f42e929f2cc7e9b6d24dd61e19c5d93dcfeabd34b9c8b935afbe78577efb18b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:1ce71e80c0cd0e0063eba275ab91f77982f2e92a8aef8e89cab99b93feab0a88
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:90cf21c36d19d87b97d06edd1c0912a3bb288c769e6159651877db3206aa7355
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f045fe5037d44aab1a0b2209241b3931ec7cb9897cb3c6a96a9f8d5280e55dd2
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:aeac9cd8e8acdbd2b0c339c48b85079c159b04f0ade24df8e3f9cae3f48b4c70
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:bed27f3f2c2d7627d1af4500d7ea0302da5c3fd3d875a6480fb6d98da481312e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:42c9930d6861f6a2147ed6e01827eddd730279966763ec68868567a19cc164ab
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:6cd9f9b9fcb6e8f9a931a6ad01cb66329a74d40b22f9029fbc50e57248bad7cb
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7d6320ac7fb0925f10cf6f524a13de629cbf13ccc371912e4fe8c777bc15d1b5
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:9c85c25d5392817d17b68714dca6195ad53255d43e9c65b021d75b12d97e0579
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:858dc82977b3c9e989a3a20626671b0c67c834c97132094a9212bb227ab55929
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:f05824bb8eaa5a93e54059449eaa236dd83d729049ad00065ef3242f5f74c91e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:68fe64fe3e112288ccfc2a7dffd69b9b2786a5da3d3bdedd6addf41659d460c9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:a4f932c3e9ff3aeb70ba1f7a92aa4642bd14cbb7d37d04ff750ed517b916cdb9
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:740ba10f1ef0a6c5fa6ad31ab18855fc75f8e024ce4c23b0676c5a9c5b62cd75
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:58de9d0b2ce005c7ad3e99152100e59b7677398557028a7a59e56e669fabb81c
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:742a6f69a3a209a7080a1e911fd96467d2a91aacb3f91d6be80a6a6e2e54b2ae
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:acbfee4492327a720fe7da76929a86192d8af2442bbb76f0a089f6a81a890be3
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:61e602b0e3582bf00460eb9328808ccf66e3ff1b1f6df07ee6c5c76a91846876
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0754f7901feea9017bc5041cb91b7ac1cdbca10c2064f37c3648a2c395f24f83
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:0b0e0588c1aadf0c3826cd584bd2b3a7d6781882dcb5c5e037b47ff5cb84509b
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:02e1ef01a023da0eee653f63b314786fd66a78d184a97a314422f6f9c3ef8e7e
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:19f36f53f7da67755eb86da77098c5eeb84b381f9c18aa29ba573793dc498564
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:40f00476b63cb9e370fe85f01407173693213a9242fc117d28b06378ca0d98e0
        - registry.connect.redhat.com/cockroachdb/cockroach@sha256:7daaf1368ab2c25694cdea0a598e1de59247b04a4b4b4b8348ed6edbaa194f9d
        version: 2.18.2
      entries:
      - name: cockroach-operator.v2.18.2
        version: 2.18.2
      name: stable
    defaultChannel: stable
    packageName: cockroachdb-certified
    provider:
      name: Cockroach Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: maximilianopizarro
      provider-url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
    name: jhipster-online-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jhipster-online-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maximilianopizarro.github.io/v1alpha1",
                "kind": "JhipsterOnline",
                "metadata": {
                  "name": "jhipsteronline-sample"
                },
                "spec": {
                  "affinity": {},
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 100,
                    "minReplicas": 1,
                    "targetCPUUtilizationPercentage": 80
                  },
                  "env": {
                    "APPLICATION_GITHUB_CLIENT-ID": "CLIENT-ID",
                    "APPLICATION_GITHUB_CLIENT-SECRET": "CLIENT-SECRET",
                    "APPLICATION_GITHUB_HOST": "https://github.com",
                    "APPLICATION_JHIPSTER-CMD_CMD": "jhipster-quarkus",
                    "JAVA_APP_JAR": "jhonline-2.33.0.war",
                    "SPRING_DATASOURCE_PASSWORD": "jhipster",
                    "SPRING_DATASOURCE_URL": "jdbc:mariadb://mariadb:3306/jhipsteronline",
                    "SPRING_DATASOURCE_USERNAME": "jhipster"
                  },
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "quay.io/maximilianopizarro/jhipster-online",
                    "tag": "quarkus"
                  },
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "livenessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {},
                  "readinessProbe": {
                    "httpGet": {
                      "path": "/jdl-studio/",
                      "port": 8080
                    }
                  },
                  "replicaCount": 1,
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "securityContext": {},
                  "service": {
                    "port": 8080,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automount": false,
                    "create": false,
                    "name": ""
                  },
                  "tolerations": [],
                  "volumeMounts": [],
                  "volumes": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
          createdAt: "2025-05-30T22:42:19Z"
          description: JHipster Online provides an intuitive web interface for generating
            JHipster projects, allowing you to define your application architecture
            using a JDL (JHipster Domain Language) model and generate the code directly
            into a GitHub repository.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/maximilianoPizarro/jhipster-online-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: JhipsterOnline
            name: jhipsteronlines.maximilianopizarro.github.io
            version: v1alpha1
        description: |
          JHipster Online Operator: This repository hosts the JHipster Online Operator, designed to seamlessly deploy JHipster Online on Red Hat OpenShift and Kubernetes environments.

          JHipster Online provides an intuitive web interface for generating JHipster projects, allowing you to define your application architecture using a JDL (JHipster Domain Language) model and generate the code directly into a GitHub repository.

          Overview: The JHipster Online Operator manages the deployment and lifecycle of the JHipster Online application. It leverages the power of Kubernetes Operators to provide a native Kubernetes experience for JHipster Online users.

          This artifact includes:
          - JHipster 8.8.0. for generate Spring Boot 3.4.1 projects.
          - generator-jhipster-quarkus 3.4.0 for generate Quarkus 3.11.1 projects.
          - JDL Studio for add JDL models files by PR on your repo.

          Usage: Once the operator is running, you can create instances of JHipster Online by defining `JhipsterOnline` custom resources. An example `JhipsterOnline` custom resource is provided in the `alm-examples`. You would apply this (or a similar) YAML file to your cluster to create an instance of JHipster Online. Feel free to customize the `JhipsterOnline` Custom Resource Definition (CRD) instance to suit your needs, such as configuring GitHub integration, database connections, and resource allocations.

          Configuration: The `JhipsterOnline` custom resource supports various configuration options including `replicaCount`, `image`, `imagePullSecrets`, `serviceAccount`, `podAnnotations`, `podLabels`, `podSecurityContext`, `securityContext`, `service` type and port, `ingress` settings (enabled, `className`, `annotations`, `hosts`, `tls`), OpenShift `Route` (enabled), `autoscaling` (enabled, `minReplicas`, `maxReplicas`, `targetCPUUtilizationPercentage`), `resource` requests and limits, `nodeSelector`, `tolerations`, `affinity`, `liveness` and `readiness` probes, `env` variables, `volumes`, and `volumeMounts`. Refer to the detailed `README.md` for a full table of parameters.
        displayName: JHipster Online Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - java
        - spring-boot
        - jhipster
        - quarkus
        links:
        - name: Jhipster Online Operator
          url: https://maximilianopizarro.github.io/jhipster-online-operator/
        maintainers:
        - email: maximiliano.pizarro.5@gmail.com
          name: Maximiliano Pizarro
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: maximilianopizarro
          url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/jhipster-online-operator@sha256:5f2436297c026832e5c15241d60090da581daa5502e352ea33d77f22ef3e7a52
        - quay.io/maximilianopizarro/jhipster-online-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: jhipster-online-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: jhipster-online-operator
    provider:
      name: maximilianopizarro
      url: https://maximilianopizarro.github.io/jhipster-online-helm-chart/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Argo CD Community
      provider-url: ""
    name: argocd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: argocd-operator.v0.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "AppProject",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "Application",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ApplicationSet",
                "metadata": {
                  "name": "example"
                },
                "spec": null
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "ArgoCDExport",
                "metadata": {
                  "name": "argocdexport-sample"
                },
                "spec": {
                  "argocd": "argocd-sample"
                }
              },
              {
                "apiVersion": "argoproj.io/v1alpha1",
                "kind": "NotificationsConfiguration",
                "metadata": {
                  "name": "default-notifications-configuration"
                },
                "spec": {
                  "templates": {
                    "template.app-created": "email:\n  subject: Application {{.app.metadata.name}} has been created.\nmessage: Application {{.app.metadata.name}} has been created.\nteams:\n  title: Application {{.app.metadata.name}} has been created.",
                    "template.app-deleted": "email:\n  subject: Application {{.app.metadata.name}} has been deleted.\nmessage: Application {{.app.metadata.name}} has been deleted.\nteams:\n  title: Application {{.app.metadata.name}} has been deleted.",
                    "template.app-deployed": "email:\n  subject: New version of an application {{.app.metadata.name}} is up and running.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} is now running new version of deployments manifests.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Revision\",\n        \"value\": \"{{.app.status.sync.revision}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    },\n    {\n      \"name\": \"Revision\",\n      \"value\": \"{{.app.status.sync.revision}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: New version of an application {{.app.metadata.name}} is up and running.",
                    "template.app-health-degraded": "email:\n  subject: Application {{.app.metadata.name}} has degraded.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} has degraded.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\": \"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#f4c030\",\n      \"fields\": [\n      {\n        \"title\": \"Health Status\",\n        \"value\": \"{{.app.status.health.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Health Status\",\n      \"value\": \"{{.app.status.health.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Application {{.app.metadata.name}} has degraded.",
                    "template.app-sync-failed": "email:\n  subject: Failed to sync application {{.app.metadata.name}}.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}}  The sync operation of application {{.app.metadata.name}} has failed at {{.app.status.operationState.finishedAt}} with the following error: {{.app.status.operationState.message}}\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Failed at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#FF0000'\n  title: Failed to sync application {{.app.metadata.name}}.",
                    "template.app-sync-running": "email:\n  subject: Start syncing application {{.app.metadata.name}}.\nmessage: |\n  The sync operation of application {{.app.metadata.name}} has started at {{.app.status.operationState.startedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#0DADEA\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Started at\",\n      \"value\": \"{{.app.status.operationState.startedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Operation\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Start syncing application {{.app.metadata.name}}.",
                    "template.app-sync-status-unknown": "email:\n  subject: Application {{.app.metadata.name}} sync status is 'Unknown'\nmessage: |\n  {{if eq .serviceType \"slack\"}}:exclamation:{{end}} Application {{.app.metadata.name}} sync is 'Unknown'.\n  Application details: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}.\n  {{if ne .serviceType \"slack\"}}\n  {{range $c := .app.status.conditions}}\n      * {{$c.message}}\n  {{end}}\n  {{end}}\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#E96D76\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Application\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  title: Application {{.app.metadata.name}} sync status is 'Unknown'",
                    "template.app-sync-succeeded": "email:\n  subject: Application {{.app.metadata.name}} has been successfully synced.\nmessage: |\n  {{if eq .serviceType \"slack\"}}:white_check_mark:{{end}} Application {{.app.metadata.name}} has been successfully synced at {{.app.status.operationState.finishedAt}}.\n  Sync operation details are available at: {{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true .\nslack:\n  attachments: |\n    [{\n      \"title\": \"{{ .app.metadata.name}}\",\n      \"title_link\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}\",\n      \"color\": \"#18be52\",\n      \"fields\": [\n      {\n        \"title\": \"Sync Status\",\n        \"value\": \"{{.app.status.sync.status}}\",\n        \"short\": true\n      },\n      {\n        \"title\": \"Repository\",\n        \"value\": \"{{.app.spec.source.repoURL}}\",\n        \"short\": true\n      }\n      {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"title\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\",\n        \"short\": true\n      }\n      {{end}}\n      ]\n    }]\n  deliveryPolicy: Post\n  groupingKey: \"\"\n  notifyBroadcast: false\nteams:\n  facts: |\n    [{\n      \"name\": \"Sync Status\",\n      \"value\": \"{{.app.status.sync.status}}\"\n    },\n    {\n      \"name\": \"Synced at\",\n      \"value\": \"{{.app.status.operationState.finishedAt}}\"\n    },\n    {\n      \"name\": \"Repository\",\n      \"value\": \"{{.app.spec.source.repoURL}}\"\n    }\n    {{range $index, $c := .app.status.conditions}}\n      {{if not $index}},{{end}}\n      {{if $index}},{{end}}\n      {\n        \"name\": \"{{$c.type}}\",\n        \"value\": \"{{$c.message}}\"\n      }\n    {{end}}\n    ]\n  potentialAction: |-\n    [{\n      \"@type\":\"OpenUri\",\n      \"name\":\"Operation Details\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.context.argocdUrl}}/applications/{{.app.metadata.name}}?operation=true\"\n      }]\n    },\n    {\n      \"@type\":\"OpenUri\",\n      \"name\":\"Open Repository\",\n      \"targets\":[{\n        \"os\":\"default\",\n        \"uri\":\"{{.app.spec.source.repoURL | call .repo.RepoURLToHTTPS}}\"\n      }]\n    }]\n  themeColor: '#000080'\n  title: Application {{.app.metadata.name}} has been successfully synced"
                  },
                  "triggers": {
                    "trigger.on-created": "- description: Application is created.\n  oncePer: app.metadata.name\n  send:\n  - app-created\n  when: \"true\"",
                    "trigger.on-deleted": "- description: Application is deleted.\n  oncePer: app.metadata.name\n  send:\n  - app-deleted\n  when: app.metadata.deletionTimestamp != nil",
                    "trigger.on-deployed": "- description: Application is synced and healthy. Triggered once per commit.\n  oncePer: app.status.operationState.syncResult.revision\n  send:\n  - app-deployed\n  when: app.status.operationState.phase in ['Succeeded'] and app.status.health.status\n      == 'Healthy'",
                    "trigger.on-health-degraded": "- description: Application has degraded\n  send:\n  - app-health-degraded\n  when: app.status.health.status == 'Degraded'",
                    "trigger.on-sync-failed": "- description: Application syncing has failed\n  send:\n  - app-sync-failed\n  when: app.status.operationState.phase in ['Error', 'Failed']",
                    "trigger.on-sync-running": "- description: Application is being synced\n  send:\n  - app-sync-running\n  when: app.status.operationState.phase in ['Running']",
                    "trigger.on-sync-status-unknown": "- description: Application status is 'Unknown'\n  send:\n  - app-sync-status-unknown\n  when: app.status.sync.status == 'Unknown'",
                    "trigger.on-sync-succeeded": "- description: Application syncing has succeeded\n  send:\n  - app-sync-succeeded\n  when: app.status.operationState.phase in ['Succeeded']"
                  }
                }
              },
              {
                "apiVersion": "argoproj.io/v1beta1",
                "kind": "ArgoCD",
                "metadata": {
                  "name": "argocd-sample"
                },
                "spec": {
                  "controller": {
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2048Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "1024Mi"
                      }
                    }
                  },
                  "ha": {
                    "enabled": false,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "redis": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "repo": {
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "512Mi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "server": {
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "125m",
                        "memory": "128Mi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "sso": {
                    "dex": {
                      "resources": {
                        "limits": {
                          "cpu": "500m",
                          "memory": "256Mi"
                        },
                        "requests": {
                          "cpu": "250m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "provider": "dex"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
          createdAt: "2025-09-11T06:29:33Z"
          description: Argo CD is a declarative, GitOps continuous delivery tool for
            Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/argoproj-labs/argocd-operator
          support: Argo CD
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Application is a group of Kubernetes resources as defined
              by a manifest.
            displayName: Application
            kind: Application
            name: applications.argoproj.io
            version: v1alpha1
          - description: An ApplicationSet is a group or set of Application resources.
            displayName: ApplicationSet
            kind: ApplicationSet
            name: applicationsets.argoproj.io
            version: v1alpha1
          - description: An AppProject is a logical grouping of Argo CD Applications.
            displayName: AppProject
            kind: AppProject
            name: appprojects.argoproj.io
            version: v1alpha1
          - description: ArgoCDExport is the Schema for the argocdexports API
            displayName: Argo CDExport
            kind: ArgoCDExport
            name: argocdexports.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1alpha1
          - description: ArgoCD is the Schema for the argocds API
            displayName: Argo CD
            kind: ArgoCD
            name: argocds.argoproj.io
            version: v1beta1
          - description: NotificationsConfiguration is the Schema for the notificationsconfiguration
              API
            displayName: NotificationsConfiguration
            kind: NotificationsConfiguration
            name: notificationsconfigurations.argoproj.io
            version: v1alpha1
        description: |
          ## Overview

          The Argo CD Operator manages the full lifecycle for [Argo CD](https://argoproj.github.io/argo-cd/) and it's
          components. The operator's goal is to automate the tasks required when operating an Argo CD cluster.

          Beyond installation, the operator helps to automate the process of upgrading, backing up and restoring as needed and
          remove the human as much as possible. In addition, the operator aims to provide deep insights into the Argo CD
          environment by configuring Prometheus and Grafana to aggregate, visualize and expose the metrics already exported by
          Argo CD.

          The operator aims to provide the following, and is a work in progress.

          * Easy configuration and installation of the Argo CD components with sane defaults to get up and running quickly.
          * Provide seamless upgrades to the Argo CD components.
          * Ability to back up and restore an Argo CD cluster from a point in time or on a recurring schedule.
          * Aggregate and expose the metrics for Argo CD and the operator itself using Prometheus and Grafana.
          * Autoscale the Argo CD components as necessary to handle variability in demand.

          ## Usage

          Deploy a basic Argo CD cluster by creating a new ArgoCD resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCD
          metadata:
            name: example-argocd
          spec: {}
          ```

          ## Backup

          Backup the cluster above by creating a new ArgoCDExport resource in the namespace where the operator is installed.

          ```
          apiVersion: argoproj.io/v1alpha1
          kind: ArgoCDExport
          metadata:
            name: example-argocdexport
          spec:
            argocd: example-argocd
          ```

          See the [documentation](https://argocd-operator.readthedocs.io) and examples on
          [GitHub](https://github.com/argoproj-labs/argocd-operator) for more information.
        displayName: Argo CD
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - kubernetes
        links:
        - name: Argo CD Project
          url: https://argoproj.github.io/argo-cd/
        - name: Operator Documentation
          url: https://argocd-operator.readthedocs.io
        - name: Operator Source Code
          url: https://github.com/argoproj-labs/argocd-operator
        maintainers:
        - email: aveerama@redhat.com
          name: Abhishek Veeramalla
        maturity: alpha
        provider:
          name: Argo CD Community
        relatedImages:
        - quay.io/argoprojlabs/argocd-operator@sha256:223ad82fb697471ede79201c4a042115c9ca5c00a172d5bdcd5a780e58c46c79
        - quay.io/community-operator-pipeline-prod/argocd-operator@sha256:483854621b95dcefa03d18a970600b271726422ae3362479f73dcfca9085c92d
        version: 0.15.0
      entries:
      - name: argocd-operator.v0.15.0
        version: 0.15.0
      - name: argocd-operator.v0.14.1
        version: 0.14.1
      - name: argocd-operator.v0.14.0
        version: 0.14.0
      - name: argocd-operator.v0.13.0
        version: 0.13.0
      - name: argocd-operator.v0.12.0
        version: 0.12.0
      - name: argocd-operator.v0.11.0
        version: 0.11.0
      - name: argocd-operator.v0.10.1
        version: 0.10.1
      - name: argocd-operator.v0.10.0
        version: 0.10.0
      - name: argocd-operator.v0.9.1
        version: 0.9.1
      - name: argocd-operator.v0.9.0
        version: 0.9.0
      - name: argocd-operator.v0.8.0
        version: 0.8.0
      - name: argocd-operator.v0.7.0
        version: 0.7.0
      - name: argocd-operator.v0.5.0
        version: 0.5.0
      - name: argocd-operator.v0.4.0
        version: 0.4.0
      - name: argocd-operator.v0.3.0
        version: 0.3.0
      - name: argocd-operator.v0.2.1
        version: 0.2.1
      - name: argocd-operator.v0.2.0
        version: 0.2.0
      - name: argocd-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: argocd-operator
    provider:
      name: Argo CD Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtr-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtr-operator.1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "namespace": "mtr",
                 "name": "mtr",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
          createdAt: "2024-09-03T17:22:23Z"
          description: Migration Toolkit for Runtimes is an analysis tool that supports
            the modernization and migration of Java applications.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: mtr
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/windup/windup-operator/
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create an MTR Web Console application.  \n\nTo access the
              MTR Web Console application from the \"Developer\" perspective, please
              go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nIn
              order to connect with default login credentials, please use \"migration\"
              as username and \"password\" as password.  \n\nKnown issue  \nIf you
              want to customize the MTR Web Console instance's parameters and you
              can not see any in the `Form view`, please switch to the `YAML view`
              and change them as needed.\n"
            displayName: Migration Toolkit for Runtimes
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: "The Migration Toolkit for Runtimes (MTR) is a web console application
          that supports large-scale Java application modernization and migration projects
          across a broad range of transformations and use cases.  \n\nIt analyzes
          application code, supports effort estimation, accelerates code migration,
          and enables users to move applications to containers.  \n\nFor more information
          please refer to the [https://developers.redhat.com/products/mtr/overview](https://developers.redhat.com/products/mtr/overview)
          page.\n"
        displayName: Migration Toolkit for Runtimes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - metric
        - troubleshooting
        - run-time
        - migration
        - modernization
        links:
        - name: Website
          url: https://developers.redhat.com/products/mtr/overview
        - name: GitHub
          url: https://github.com/windup/windup-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_runtimes
        maintainers:
        - email: migrate@redhat.com
          name: Migration Toolkit for Runtimes
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/mtr/mtr-operator-bundle@sha256:190a6fa2a6a9b6b182cdf9cab18814d05c203b54ca18dc533e553fa5dcca779e
        - registry.redhat.io/mtr/mtr-rhel8-operator@sha256:4df7c4ad353339e8b958ffc651e811667767c8a0dcdd61f89352c926da52a513
        - registry.redhat.io/mtr/mtr-web-container-rhel8@sha256:cb88a5453b506ca70563537ec6db2f44ff8cb4050c9652ec86eb5128aa313003
        - registry.redhat.io/mtr/mtr-web-executor-container-rhel8@sha256:2f0622774367f33efc321926196681322ab1535144e6a0785608c8050fc59f2e
        version: 1.2.7
      entries:
      - name: mtr-operator.1.2.7
        version: 1.2.7
      - name: mtr-operator.1.2.6
        version: 1.2.6
      - name: mtr-operator.1.2.5
        version: 1.2.5
      - name: mtr-operator.1.2.4
        version: 1.2.4
      - name: mtr-operator.1.2.3
        version: 1.2.3
      - name: mtr-operator.1.2.2
        version: 1.2.2
      - name: mtr-operator.1.2.1
        version: 1.2.1
      - name: mtr-operator.1.2.0
        version: 1.2.0
      - name: mtr-operator.1.1.1
        version: 1.1.1
      - name: mtr-operator.1.1.0
        version: 1.1.0
      - name: mtr-operator.1.0.3
        version: 1.0.3
      - name: mtr-operator.1.0.2
        version: 1.0.2
      - name: mtr-operator.1.0.1
        version: 1.0.1
      - name: mtr-operator.1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: mtr-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: sandboxed-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sandboxed-containers-operator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
          createdAt: 2021-06-07-13:38:06
          olm.skipRange: <1.0.2
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Features & benefits
          - **Isolated Developer Environments & Priviliges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as CAP_ADMIN or CAP_BPF. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select preview-1.0 from the list of available Update Channel options.
            This ensures that you install the version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              openshift-sandboxed-containers-operator namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              openshift-sandboxed-containers-operator causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a KataConfig CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:051bd7f1dad8cc3251430fee32184be8d64077aba78580184cef0255d267bdcf
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:c6c589d5e47ba9564c66c84fc2bc7e5e046dae1d56a3dc99d7343f01e42e4d31
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:2808a0397495982b4ea0001ede078803a043d5c9b0285662b08044fe4c11f243
        version: 1.0.2
      deprecation:
        message: |
          The preview-1.0 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.0
    - currentCSV: sandboxed-containers-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
          createdAt: "2021-10-08T13:20:20.250789Z"
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift’s Defense-in-Depth strategy. For more information
          [see](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5).

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `preview-1.1` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CRD instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.8/sandboxed_containers/understanding-sandboxed-containers.html).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kernel-isolated containers
        links:
        - name: OpenShift sandboxed containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        provider:
          name: Red Hat
          url: https://www.github.com/openshift/sandboxed-containers-operator
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-rhel8-operator@sha256:b1e824e126c579db0f56d04c3d1796d82ed033110c6bc923de66d95b67099611
        - registry.redhat.io/openshift-sandboxed-containers-tech-preview/osc-operator-bundle@sha256:a91cee14f47824ce49759628d06bf4e48276e67dae00b50123d3233d78531720
        version: 1.1.0
      deprecation:
        message: |
          The preview-1.1 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.1.0
        version: 1.1.0
      - name: sandboxed-containers-operator.v1.0.2
        version: 1.0.2
      name: preview-1.1
    - currentCSV: sandboxed-containers-operator.v1.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-12T14:30:28Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.10.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["peerpods.confidentialcontainers.org","peerpodconfigs.confidentialcontainers.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
          - kind: PeerPod
            name: peerpods.confidentialcontainers.org
            version: v1alpha1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `candidate` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the [OpenShift sandboxed containers documentation](https://docs.redhat.com/en/documentation/openshift_sandboxed_containers/).
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-webhook-rhel9@sha256:6802f00a939f24ecd6dad150433db5ea3fd72d171f0c410cb894d3af910fc7fd
        - registry.redhat.io/openshift-sandboxed-containers/osc-dm-verity-image@sha256:e633bc27853f811d919e64461d0c917e659f5430458abe2304a4c85b0d871c80
        - registry.redhat.io/openshift-sandboxed-containers/osc-monitor-rhel9@sha256:037cbc221ae8fda83cdf1093127795ea806ecda53893deb49cdf981e7863774c
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-builder-rhel9@sha256:c15f36a639fa07fee11f74c6563696ec8ef4da3fb1a99b7bd3f308dd04881339
        - registry.redhat.io/openshift-sandboxed-containers/osc-podvm-payload-rhel9@sha256:8d6b1262c6cedd04e33441264539842d82ed057f64eda5c4bce59566bcb72558
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel9-operator@sha256:1ffb432d9a8833854fa5e5262dcbf42530dfbaadfddf6978f7bc7804b84d2e3f
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:4477810692429ac498b6ec1b295e6ead0f8fea25e91f6ddf17255d7e349b82e1
        - registry.redhat.io/openshift-sandboxed-containers/osc-cloud-api-adaptor-rhel9@sha256:7d1fa1e85af0d34c23e75868b1388eff17abe633c04a8dc55fbefcc470195c05
        version: 1.10.3
      entries:
      - name: sandboxed-containers-operator.v1.10.3
        version: 1.10.3
      - name: sandboxed-containers-operator.v1.10.2
        version: 1.10.2
      - name: sandboxed-containers-operator.v1.10.1
        version: 1.10.1
      - name: sandboxed-containers-operator.v1.10.0
        version: 1.10.0
      - name: sandboxed-containers-operator.v1.9.0
        version: 1.9.0
      - name: sandboxed-containers-operator.v1.8.1
        version: 1.8.1
      - name: sandboxed-containers-operator.v1.7.0
        version: 1.7.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.6.0
        version: 1.6.0
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.3
        version: 1.5.3
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.2
        version: 1.5.2
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.1
        version: 1.5.1
      - deprecation:
          message: |
            This bundle is deprecated. Please use the latest stable bundle instead.
        name: sandboxed-containers-operator.v1.5.0
        version: 1.5.0
      name: stable
    - currentCSV: sandboxed-containers-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "false"
          containerImage: registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
          createdAt: "2021-11-22T14:44:50.817336Z"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.2` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/4.9/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:ff2bb666c2696fed365df55de78141a02e372044647b8031e6d06e7583478af4
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:8f30a9129d817c3f4e404d2c43fb47e196d8c8da3badba4c48f65d440a4d7584
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:a660f0b54b9139bed9a3aeef3408001c0d50ba60648364a98a09059b466fbcc1
        version: 1.2.2
      deprecation:
        message: |
          The stable-1.2 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.2.2
        version: 1.2.2
      name: stable-1.2
    - currentCSV: sandboxed-containers-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kataconfiguration.openshift.io/v1",
                "kind": "KataConfig",
                "metadata": {
                  "name": "example-kataconfig"
                }
              }
            ]
          capabilities: Seamless Upgrades
          olm.skipRange: '>=1.1.0 <1.3.3'
          operatorframework.io/suggested-namespace: openshift-sandboxed-containers-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/sandboxed-containers-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The kataconfig CR represent a installation of Kata in a cluster
              and its current state.
            kind: KataConfig
            name: kataconfigs.kataconfiguration.openshift.io
            version: v1
        description: |-
          OpenShift sandboxed containers, based on the Kata Containers open source
          project, provides an Open Container Initiative (OCI) compliant container
          runtime using lightweight virtual machines, running your workloads in their own
          isolated kernel and therefore contributing an additional layer of isolation
          back to OpenShift's Defense-in-Depth strategy. Click [this link](https://catalog.redhat.com/software/operators/detail/5ee0d499fdbe7cddc2c91cf5) for
          more information.

          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Features & benefits
          - **Isolated Developer Environments & Privileges Scoping**
            As a developer working on debugging an application using state-of-the-art
            tooling you might need elevated privileges such as `CAP_ADMIN` or `CAP_BPF`. With
            OpenShift sandboxed containers, any impact will be limited to a separate
            dedicated kernel.

          - **Legacy Containerized Workload Isolation**
            You are mid-way in converting a containerized monolith into cloud-native
            microservices. However, the monolith still runs on your cluster unpatched and
            unmaintained. OpenShift sandboxed containers helps isolate it in its own kernel
            to reduce risk.

          - **Safe Multi-tenancy & Resource Sharing (CI/CD Jobs, CNFs, ..)**
            If you are providing a service to multiple tenants, it could mean that the
            service workloads are sharing the same resources (e.g., worker node). By
            deploying in a dedicated kernel, the impact of these workloads have on one
            another is greatly reduced.

          - **Additional Isolation with Native Kubernetes User Experience**
            OpenShift sandboxed containers is used as a compliant OCI runtime.
            Therefore, many operational patterns used with normal containers are still
            preserved including but not limited to image scanning, GitOps, Imagestreams,
            and so on.

          # How to install
            Read the information about the Operator and click Install.

            On the Install Operator page:

            - Select `stable-1.3` from the list of available Update Channel options.
            This ensures that you install the latest version of OpenShift sandboxed containers
            that is compatible with your OpenShift Container Platform version.

            - For Installed Namespace, ensure that the Operator recommended namespace
              option is selected. This installs the Operator in the mandatory
              `openshift-sandboxed-containers-operator` namespace, which is automatically
              created if it does not exist. Attempting to install the OpenShift
              sandboxed containers Operator in a namespace other than
              `openshift-sandboxed-containers-operator` causes the installation to fail.

            - For Approval Strategy, ensure that Automatic, which is the default value,
              is selected. OpenShift sandboxed containers automatically updates when a new
              z-stream release is available.

            - Click Install to make the Operator available to the OpenShift sandboxed
              containers namespace.

            - The OpenShift sandboxed containers Operator is now installed on your
              cluster. You can trigger the Operator by enabling the runtime on your cluster.
              You can do this by creating a `KataConfig` CustomResourceDefinition(CRD) instance. For this click
              on "create instance" on the operator overview page.

          # Documentation
          See the official documentation [here](https://docs.openshift.com/container-platform/latest/sandboxed_containers/understanding-sandboxed-containers.html)
        displayName: OpenShift sandboxed containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sandboxed-containers
        - kata
        links:
        - name: Sandboxed Containers Operator
          url: https://www.github.com/openshift/sandboxed-containers-operator
        maintainers:
        - email: support@redhat.com'
          name: '''Red Hat'
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cb836456974e510eb4bccbffadbc6d99d5f57c36caec54c767a158ffd8a025d5
        - registry.redhat.io/openshift-sandboxed-containers/osc-operator-bundle@sha256:8da62ba1c19c905bc1b87a6233ead475b047a766dc2acb7569149ac5cfe7f0f1
        - registry.redhat.io/openshift-sandboxed-containers/osc-rhel8-operator@sha256:5d2b03721043e5221dfb0cf164cf59eba396ba3aae40a56c53aa3496c625eea0
        version: 1.3.3
      deprecation:
        message: |
          The stable-1.3 channel is deprecated. Please use the 'stable' channel for the latest supported version.
      entries:
      - name: sandboxed-containers-operator.v1.3.3
        version: 1.3.3
      name: stable-1.3
    defaultChannel: stable
    packageName: sandboxed-containers-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-load-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-load-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Loadcore",
                "metadata": {
                  "name": "loadcore-sample"
                },
                "spec": {
                  "accessRoute": {
                    "enabled": true,
                    "host": "loadcore.apps.cluster.perf.keysight.lab"
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "reportTemplates": {
                    "enabled": false
                  },
                  "wirelessDataModel": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:42:44Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Loadcore
            name: loadcores.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Operator
        displayName: Keysight LoadCore Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Load Core
          url: https://keysight-load-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-load-core-bundle@sha256:f62a7f8eb74d0d5309fede6c9c0201d1f0c3736f93f1aeec7446089d64fce28f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-load-core@sha256:245d9be60ff6e9a6e5e42a026a4ee7b8168f391a0491de23ae827d1906af88e5
        version: 0.1.21
      entries:
      - name: keysight-load-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-load-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Sigstore Authors
      provider-url: https://sigstore.dev
    name: sigstore-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sigstore-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.sigstore.dev/v1alpha1",
                "kind": "Rekor",
                "metadata": {
                  "name": "rekor"
                },
                "spec": {
                  "server": {
                    "ingress": {
                      "hostname": "rekor.example.com"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          containerImage: ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
          createdAt: "2021-09-01T21:20:28Z"
          description: Helm Charts for Sigstore Project.
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/sigstore/sigstore-helm-operator
          support: The Sigstore Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Rekor
            name: rekors.helm.sigstore.dev
            version: v1alpha1
        description: |-
          Helm Charts for the sigstore project

          ## Charts

          The following charts are available:

          * [Rekor](https://github.com/sigstore/helm-charts/tree/main/charts/rekor)
        displayName: sigstore
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        links:
        - name: Sigstore Helm Operator
          url: https://github.com/sabre1041/sigstore-helm-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: The Sigstore Authors
          url: https://sigstore.dev
        relatedImages:
        - ghcr.io/sigstore/sigstore-helm-operator@sha256:d9c6eb162b97a1aa1ad3bf515db65d01a96e55b6f8bcd646864d3e64a3ee2b9e
        - quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/openshift-community-operators/sigstore-helm-operator@sha256:ff4e548a8a003bd23e5076362bafb2e22f77b8445c5e3585d99cd5ce26fcaf56
        version: 0.0.1
      entries:
      - name: sigstore-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: sigstore-helm-operator
    provider:
      name: The Sigstore Authors
      url: https://sigstore.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: mtv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mtv-operator.v2.7.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "<base64-encoded-playbook>"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
          createdAt: "2024-10-11T14:05:47Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.7.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.5 is supported on OpenShift 4.12 and 4.13

          Migration Toolkit for Virtualization 2.6 is supported on OpenShift 4.14 and 4.15

          Migration Toolkit for Virtualization 2.7 is supported on OpenShift 4.16 and 4.17

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/articles/6001751).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift
          url: http://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:8337924489a1747d4bbca6dd6ac42210731cf721b549b5830846b271621e2ac1
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:6d0aeffbccaf9df83b42c530bf82707aa7c1cc16f72a5ef30b66fbd562c01ab3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:8c05a303be5c536f4051ef291475b578d4d38ca0946fb4b4dd8bca376ef432aa
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:e91c2e5473f3e24c185dd25e9843f333e6d7419d9c5f4d4396669717b5969683
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:c543dbbb66fcac02345e1d3c12333780721bf049dca0f0a27d4c55b15a5d3e14
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:a65780c5680a6fbbd0e3d0eed423b14a9c9b3d7d17f801c1eb6eb6b777384ca5
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:fd6fb83bbeb9010f7e91f0afc91beac0bf38cb79c2de8f533a2bd07b1b238262
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:936fd68d7659683de7789fcf34d2e5873bd4bd1534c01c2d80501647f7c3c442
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel8-operator@sha256:8dbeace89a921e0d11f85378099e39cc357702d2f20dfb9d14cebbe150d2500d
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:1dd8f0f08ff1323d910fea8adfa13cef64c2fd2f01968e4d9d55f71cf42835d3
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:365115407964d182de7e7564438c6fd089e2c0bcc0f94ec48621ca0cc2bf0fe4
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4bea07e101740afc36a1617363a79b1c474efcacd584f6e475a53b3f2395a079
        version: 2.7.2
      entries:
      - name: mtv-operator.v2.7.2
        version: 2.7.2
      name: release-v2.7
    - currentCSV: mtv-operator.v2.8.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "konveyor-forklift"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
          createdAt: "2025-08-26T14:06:48Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:3d0409c1a0b17c51d3dbd5aa0371edd597e3c5ba194aad987fe867dd804d1923
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:e4145efb22dfc336d8b81f24f5b488db602d6d547479d615bf82b4fdce96c784
        version: 2.8.7
      entries:
      - name: mtv-operator.v2.8.7
        version: 2.8.7
      - name: mtv-operator.v2.8.6
        version: 2.8.6
      - name: mtv-operator.v2.8.5
        version: 2.8.5
      - name: mtv-operator.v2.8.4
        version: 2.8.4
      - name: mtv-operator.v2.8.3
        version: 2.8.3
      - name: mtv-operator.v2.8.2
        version: 2.8.2
      - name: mtv-operator.v2.8.1
        version: 2.8.1
      - name: mtv-operator.v2.8.0
        version: 2.8.0
      name: release-v2.8
    - currentCSV: mtv-operator.v2.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "ForkliftController",
                "metadata": {
                  "name": "forklift-controller",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "feature_ui_plugin": "true",
                  "feature_validation": "true",
                  "feature_volume_populator": "true"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Hook",
                "metadata": {
                  "name": "example-hook",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "image": "quay.io/konveyor/hook-runner",
                  "playbook": "\u003cbase64-encoded-playbook\u003e"
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Host",
                "metadata": {
                  "name": "example-host",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "id": "",
                  "ipAddress": "",
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Migration",
                "metadata": {
                  "name": "example-migration",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "plan": {
                    "name": "example-plan",
                    "namespace": "openshift-mtv"
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "NetworkMap",
                "metadata": {
                  "name": "example-networkmap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "name": "",
                        "namespace": "",
                        "type": "pod"
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OpenstackVolumePopulator",
                "metadata": {
                  "name": "example-openstack",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "identityUrl": "",
                  "imageId": "",
                  "secretName": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "OvirtVolumePopulator",
                "metadata": {
                  "name": "example-ovirt-imageio",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "diskId": "",
                  "engineSecretName": "",
                  "engineUrl": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Plan",
                "metadata": {
                  "name": "example-plan",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "provider": {
                    "destination": {
                      "name": "",
                      "namespace": ""
                    },
                    "source": {
                      "name": "",
                      "namespace": ""
                    }
                  }
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                  "name": "example-provider",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "secret": {
                    "name": "",
                    "namespace": ""
                  },
                  "type": "vmware",
                  "url": ""
                }
              },
              {
                "apiVersion": "forklift.konveyor.io/v1beta1",
                "kind": "StorageMap",
                "metadata": {
                  "name": "example-storagemap",
                  "namespace": "openshift-mtv"
                },
                "spec": {
                  "map": [
                    {
                      "destination": {
                        "storageClass": ""
                      },
                      "source": {
                        "id": ""
                      }
                    }
                  ],
                  "provider": {
                    "name": "",
                    "namespace": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "true"
          containerImage: registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
          createdAt: "2025-09-09T18:06:51Z"
          description: Facilitates migration of VM workloads to OpenShift Virtualization
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/fipsmode: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "forklift.konveyor.io/v1beta1",
              "kind": "ForkliftController",
              "metadata": {
                "name": "forklift-controller",
                "namespace": "openshift-mtv"
              },
              "spec": {
                "feature_ui_plugin": "true",
                "feature_validation": "true",
                "feature_volume_populator": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mtv
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/kubev2v/forklift
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VM migration controller
            displayName: ForkliftController
            kind: ForkliftController
            name: forkliftcontrollers.forklift.konveyor.io
            version: v1beta1
          - description: Hook schema for the hooks API
            displayName: Hook
            kind: Hook
            name: hooks.forklift.konveyor.io
            version: v1beta1
          - description: VM host
            displayName: Host
            kind: Host
            name: hosts.forklift.konveyor.io
            version: v1beta1
          - description: VM migration
            displayName: Migration
            kind: Migration
            name: migrations.forklift.konveyor.io
            version: v1beta1
          - description: VM network map
            displayName: NetworkMap
            kind: NetworkMap
            name: networkmaps.forklift.konveyor.io
            version: v1beta1
          - description: OpenStack Volume Populator
            displayName: OpenstackVolumePopulator
            kind: OpenstackVolumePopulator
            name: openstackvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: oVirt Volume Populator
            displayName: OvirtVolumePopulator
            kind: OvirtVolumePopulator
            name: ovirtvolumepopulators.forklift.konveyor.io
            version: v1beta1
          - description: VM migration plan
            displayName: Plan
            kind: Plan
            name: plans.forklift.konveyor.io
            version: v1beta1
          - description: VM provider
            displayName: Provider
            kind: Provider
            name: providers.forklift.konveyor.io
            version: v1beta1
          - description: VM storage map
            displayName: StorageMap
            kind: StorageMap
            name: storagemaps.forklift.konveyor.io
            version: v1beta1
          - description: VSphere Xcopy Volume Populator
            displayName: VSphereXcopyVolumePopulator
            kind: VSphereXcopyVolumePopulator
            name: vspherexcopyvolumepopulators.forklift.konveyor.io
            version: v1beta1
        description: |
          The Migration Toolkit for Virtualization Operator manages the deployment and life cycle of Migration Toolkit for Virtualization on [OpenShift](https://www.openshift.com/) Container Platform.

          ### Installation

          OpenShift Virtualization must be installed on an OpenShift migration target cluster before you can use MTV to transfer any VMs to that cluster

          Once you have successfully installed the Operator, proceed to deploy components by creating the required ForkliftController CR.

          By default, the Operator installs the following components on a target cluster:

          * Controller, to coordinate migration processes.
          * UI, the web console to manage migrations.
          * Validation, a service to validate migration workflows.

          ### Compatibility

          Migration Toolkit for Virtualization 2.8 is supported on OpenShift 4.16, 4.17 and 4.18

          Migration Toolkit for Virtualization 2.9 is supported on OpenShift 4.17, 4.18 and 4.19

          More information on compatibility in the [MTV Lifecycle document](https://access.redhat.com/support/policy/updates/migration-toolkit-for-virtualization).

          ### Documentation
          Documentation can be found on the [Red Hat Customer Portal](https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization/).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Virtualization Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Creating a case in the [Red Hat Customer Portal](https://access.redhat.com/support/cases/) with any issues you find using Migration Toolkit for Application and its Operator.
          * Fixing issues by opening Pull Requests in the [KubeV2V](https://github.com/kubev2v/) under Forklift Projects.
          * Improving Forklift upstream [documentation](https://github.com/kubev2v/forklift-documentation/).
        displayName: Migration Toolkit for Virtualization Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - forklift
        - konveyor
        - mtv
        links:
        - name: Migration Toolkit for Virtualization Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_virtualization
        - name: Forklift Operator
          url: https://github.com/kubev2v/forklift
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/migration-toolkit-virtualization/mtv-validation-rhel9@sha256:41cc528bf93000095b4eb56aaff9a6dfb124d1711a2f31f0745163ff934ac96e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-controller-rhel9@sha256:0c8ebfdd52053dbcc221a3df7d603163a821dd9e0ac6f54977e081adc6cb4504
        - registry.redhat.io/migration-toolkit-virtualization/mtv-must-gather-rhel8@sha256:a299acb18821691cddba52904140cef8bf12632fbf2e752b33118a574a5ff394
        - registry.redhat.io/migration-toolkit-virtualization/mtv-operator-bundle@sha256:ad3b548405e8c4c075ef7d4c0eb0d0430684066ed7eab4563a1f8378eaf3d939
        - registry.redhat.io/migration-toolkit-virtualization/mtv-virt-v2v-rhel9@sha256:4b37b1aecbbf087257be5ba2bb559f2de005aacb4cd371192cdde1a9891739cf
        - registry.redhat.io/migration-toolkit-virtualization/mtv-vsphere-xcopy-volume-populator-rhel9@sha256:c0a4a232ff8adfb7d57e0761251ad0d79a0f02d27c4bd813f62f6b943deb6585
        - registry.redhat.io/migration-toolkit-virtualization/mtv-console-plugin-rhel9@sha256:f90f4eaaaa02b289704b19dd171282f3d11f370ed6cea992e55964729ad5116e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-ova-provider-server-rhel9@sha256:258efbf073f78f8b5b48fdda22ef055e2315046cfecc9644da51959289c836d7
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhel9-operator@sha256:ce1ef8fac4648de8413a592455e0f9eb2da56de5f26914cbae0e7069c5663709
        - registry.redhat.io/migration-toolkit-virtualization/mtv-api-rhel9@sha256:d57b39674260f3bb1ae4c164022c2c7be3cc9645f1c46a83e43f40965633451a
        - registry.redhat.io/migration-toolkit-virtualization/mtv-openstack-populator-rhel9@sha256:aeaa9ffcb0ffbb0f69b2712a3031f7e2dfc0824273b5c9c0f6349bc701a0e89e
        - registry.redhat.io/migration-toolkit-virtualization/mtv-populator-controller-rhel9@sha256:e438c5fccc957ad26ecfc3633561342d3b6f2a6d5a4c7839109564008de62222
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:83a523392a9c715972bba7044c7c3d3944594cc89778449e1f59fbd6a2659baf
        version: 2.9.4
      entries:
      - name: mtv-operator.v2.9.4
        version: 2.9.4
      - name: mtv-operator.v2.9.3
        version: 2.9.3
      - name: mtv-operator.v2.9.2
        version: 2.9.2
      - name: mtv-operator.v2.9.1
        version: 2.9.1
      - name: mtv-operator.v2.9.0
        version: 2.9.0
      name: release-v2.9
    defaultChannel: release-v2.9
    packageName: mtv-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: ""
    name: red-hat-hawtio-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: latest
    - currentCSV: red-hat-hawtio-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              },
              {
                "apiVersion": "hawt.io/v2",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24,
                    "internalSSL": true
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Cluster",
                  "version": "1.12"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
          createdAt: "2025-08-07T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v2
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:9e7121c29c3bd26d48c2e4d1a169b10cd09d82501d6ad7fdf76c76de797d8482
        - registry.redhat.io/rhbac-4/hawtio-rhel9-operator@sha256:87b100dbe1340bfc1d88a0640be9b7a7d42b43223f5d5d112dc01080a77c86ea
        version: 1.2.1
      entries:
      - name: red-hat-hawtio-operator.v1.2.1
        version: 1.2.1
      - name: red-hat-hawtio-operator.v1.2.0-0.1754874242.p
        version: 1.2.0+0.1754874242.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754338551.p
        version: 1.2.0+0.1754338551.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1754317392.p
        version: 1.2.0+0.1754317392.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1753863723.p
        version: 1.2.0+0.1753863723.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752679472.p
        version: 1.2.0+0.1752679472.p
      - name: red-hat-hawtio-operator.v1.2.0-0.1752452299.p
        version: 1.2.0+0.1752452299.p
      - name: red-hat-hawtio-operator.v1.2.0
        version: 1.2.0
      - name: red-hat-hawtio-operator.v1.1.1
        version: 1.1.1
      - name: red-hat-hawtio-operator.v1.1.0
        version: 1.1.0
      - name: red-hat-hawtio-operator.v1.1.0-0.1742200932.p
        version: 1.1.0+0.1742200932.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1740358952.p
        version: 1.1.0+0.1740358952.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739994212.p
        version: 1.1.0+0.1739994212.p
      - name: red-hat-hawtio-operator.v1.1.0-0.1739992853.p
        version: 1.1.0+0.1739992853.p
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1
    - currentCSV: red-hat-hawtio-operator.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              },
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "hawtio-online"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "200Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace",
                  "version": 1.12
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
          createdAt: "2024-01-30T12:18:50Z"
          description: HawtIO eases the discovery and management of Java applications
            deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HawtIO eases the discovery and management of hawtio-enabled
              applications
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1
          - kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          Red Hat build of HawtIO
          ==============

          HawtIO eases the discovery and management of Java applications deployed on OpenShift.

          To secure the communication between HawtIO and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `hawtio-online.hawtio.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/manual/agents.html#agents-jvm).
        displayName: HawtIO Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: HawtIO Web site
          url: https://hawt.io
        maintainers:
        - email: hawtio@googlegroups.com
          name: The HawtIO team
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbac-4/hawtio-operator-bundle@sha256:0ee0bd2449f183277af845e312257b4a1ced4de865831b15e1758a397d8d29eb
        - registry.redhat.io/rhbac-4/hawtio-rhel8-operator@sha256:f04d81d24c757db7b678464382ad8c5182b1580e714123a9f4f5fc9fbd9b8c57
        version: 1.0.3
      entries:
      - name: red-hat-hawtio-operator.v1.0.3
        version: 1.0.3
      - name: red-hat-hawtio-operator.v1.0.2
        version: 1.0.2
      - name: red-hat-hawtio-operator.v1.0.1
        version: 1.0.1
      - name: red-hat-hawtio-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1
    packageName: red-hat-hawtio-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:2cf058c819b40512094f84f4d374457455370ec4994d34cee6e39b9d87841f05
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-karavan-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-karavan-operator.v3.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Karavan",
                "metadata": {
                  "labels": {
                    "app": "karavan"
                  },
                  "name": "karavan"
                },
                "spec": {
                  "instances": 1,
                  "auth": "public",
                  "environment": "dev",
                  "runtimes": "quarkus,spring-boot",
                  "gitPullInterval": "10s"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/apache/camel-karavan-operator:3.20.1
          repository: https://github.com/apache/camel-karavan
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Karavan
            name: karavans.camel.apache.org
            version: v1
        description: |
          Apache Camel Karavan
          ====================
          Karavan is an Integration Toolkit for Apache Camel, which makes integration easy and fun through the visualization of pipelines,
          integration with runtimes and package, image build and deploy to kubernetes out-of-the-box.

          ## Installation

          1. Install Apache Camel Karavan Operator
          2. Install Tekton operator
          3. Create namespace, ex: `karavan`
          4. Create `Secret` in the namespace, ex:
          ```
          kind: Secret
          apiVersion: v1
          metadata:
            name: karavan
            namespace: karavan
          type: Opaque
          stringData:
            master-password: karavan
            oidc-secret: secret
            oidc-server-url: https://hostname/auth/realms/karavan
            oidc-frontend-url: https://hostname/auth
            git-repository: https://github.com/repository/projects.git
            git-password: password
            git-username: username
            git-branch: main
            image-registry: image-registry.openshift-image-registry.svc:5000
          ```
          5. Create `Karavan` instance in the namespace, ex:
          ```
          apiVersion: camel.apache.org/v1
          kind: Karavan
          metadata:
            name: karavan
            namespace: karavan
          spec:
            instances: 1
            auth: public
            environment: dev
            runtimes: quarkus,spring-boot
            gitPullInterval: 10s
          ```
        displayName: Camel Karavan Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - camel
        - karavan
        - integration
        - microservices
        - low-code
        links:
        - name: Apache Camel Karavan source code repository
          url: https://github.com/apache/camel-karavan
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - ghcr.io/apache/camel-karavan-operator:3.20.1
        - quay.io/openshift-community-operators/camel-karavan-operator@sha256:6dc78ecdf6e1936872626c02d41986d6553c1500002d85692c97e86fa77c129d
        version: 3.20.1
      entries:
      - name: camel-karavan-operator.v3.20.1
        version: 3.20.1
      - name: camel-karavan-operator.v3.20.0
        version: 3.20.0
      - name: camel-karavan-operator.v3.18.5
        version: 3.18.5
      name: alpha
    defaultChannel: alpha
    packageName: camel-karavan-operator
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:5a5e2f775760aa68c85d0f69b700d23ef2a6be7bb415bd96401d03eb6077f90c
        - quay.io/redhat-developer/servicebinding-operator@sha256:b355d2370926063a3de8c7a09c0ca8c2348a8d158d6a7095d0e5e8381cc95be5
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:79a9191f0cd5d11af004ead87ba9eb5359c67fd658c346db4c3cea3b16b23987
        - quay.io/redhat-developer/servicebinding-operator@sha256:c9c9e45321db2d4d3cdc2d364337af844f69c9893d0cc74d2fa56c846ffcf3da
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/service-binding-operator@sha256:54ea11add0a1d988e0c965968f9a65e011978190f76fa2f826f1378474bac1d4
        - quay.io/redhat-developer/servicebinding-operator@sha256:ac47f496fb7ecdcbc371f8c809fad2687ec0c35bbc8c522a7ab63b3e5ffd90ea
        version: 1.4.0
      entries:
      - name: service-binding-operator.v1.4.0
        version: 1.4.0
      - name: service-binding-operator.v1.3.4
        version: 1.3.4
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.2
        version: 1.3.2
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm-hub
      app.kubernetes.io/name: kmm-hub
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management-hub
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management-hub.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:888a3acad339cf38a324c718e0cbe1237ca5bf2f4636b01d001752658e4cca1b
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:f182078b0077b6df6d47ddcab0204e963c1475486bd61b8eccb73b46eb8b2995
        version: 2.0.2
      entries:
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management-hub.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c8b3505ed491c1bf6a913f161715c8ac483d8d9adf1fd29916ea4edad987cd67
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:d1c1490d4c20dc2ec31effa1af067676c4d264e3b7d4670fc11c2baf8a400001
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        version: 2.1.1
      entries:
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management-hub.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:90c1946b3f577cde0435402b04baa52dc54e5aab3bb84bd834ac026a57453296
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:af5a13351f960c0938c68604b772f24743c0e997701838a7784c53406471e72a
        version: 2.2.1
      entries:
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management-hub.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "jobNamespace": "default",
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: Kernel Module Management Hub
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:0d9f5293d69eebd91fffd83ae371141e640cab8f9e2d559a83714ce0f94fee08
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:f1d8f3c95f621c3faf41867c5dd4bc13fa2b9b291422d6184e6e42f1d1460259
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        version: 2.3.0
      entries:
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management-hub.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hub.kmm.sigs.x-k8s.io/v1beta1",
                "kind": "ManagedClusterModule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm-hub",
                    "app.kubernetes.io/name": "kmm-hub",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "managedclustermodule-sample"
                },
                "spec": {
                  "moduleSpec": {
                    "moduleLoader": {
                      "container": {
                        "kernelMappings": [],
                        "modprobe": {
                          "moduleName": "sample"
                        }
                      }
                    },
                    "selector": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "selector": {
                    "cluster.open-cluster-management.io/clusterset": "default"
                  },
                  "spokeNamespace": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm-hub
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ManagedClusterModule describes how to load a kernel module
              on managed clusters
            displayName: Managed Cluster Module
            kind: ManagedClusterModule
            name: managedclustermodules.hub.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Hub component for Kernel Module Management
        displayName: Kernel Module Management (Hub)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - hub
        - kernel
        - kmm
        - kmm-hub
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-hub-rhel9-operator@sha256:2efb81137ad2981b994c62d460c917749308133a7dcad5b9a2b251e18f3052ec
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-hub-operator-bundle@sha256:c4dc5f967c59e6371f344aa840fac24d8a17c127c71a2babf67d26cf6526acc5
        version: 2.4.1
      entries:
      - name: kernel-module-management-hub.v2.4.1
        version: 2.4.1
      - name: kernel-module-management-hub.v2.4.0
        version: 2.4.0
      - name: kernel-module-management-hub.v2.3.0
        version: 2.3.0
      - name: kernel-module-management-hub.v2.2.1
        version: 2.2.1
      - name: kernel-module-management-hub.v2.2.0
        version: 2.2.0
      - name: kernel-module-management-hub.v2.1.1
        version: 2.1.1
      - name: kernel-module-management-hub.v2.1.0
        version: 2.1.0
      - name: kernel-module-management-hub.v2.0.2
        version: 2.0.2
      - name: kernel-module-management-hub.v2.0.1
        version: 2.0.1
      - name: kernel-module-management-hub.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management-hub
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Couchbase
      provider-url: ""
    name: couchbase-enterprise-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: couchbase-operator.v2.6.4-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator@sha256:1f0a6865c360a14ba98706fd6fe8dff3bdbdede32e3622f93f19c82e377ca33f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:d6675617c19b4ac7c1913078eaf9ffbe3d03761b570dffa002d9de11f337f19f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.6.4-7
      entries:
      - name: couchbase-operator.v2.6.4-7
        version: 2.6.4-7
      name: 2.6.4
    - currentCSV: couchbase-operator.v2.7.0-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator@sha256:707571a7881bc12d5f1b0b1a94eb3c675d5c0ea0114c994a47ceb5aa79df4013
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:2a35a0ef852cb4d76e963df50449712a06ff13391c30675aaf931e21cb561b8a
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.7.0-5
      entries:
      - name: couchbase-operator.v2.7.0-5
        version: 2.7.0-5
      name: 2.7.0
    - currentCSV: couchbase-operator.v2.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:00f7ce0148988e93679d4b53b45ae8d2ad54f308b908940e7b28504318f718c6
        - registry.connect.redhat.com/couchbase/operator@sha256:a79e1b6781d5f8badf664e7efea6e2889e43fae6e0d37169d766685dcd6ff394
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        version: 2.7.1-1
      entries:
      - name: couchbase-operator.v2.7.1-1
        version: 2.7.1-1
      name: 2.7.1
    - currentCSV: couchbase-operator.v2.8.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:c352f5633beef2eaaa324dda66b9b484e5ea79efbbc3b781079f2e9d10c880a9
        - registry.connect.redhat.com/couchbase/operator@sha256:df61928b064014d851c6e7dff6dcd63a433a27fa5623254b6e389c3eea170e36
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        version: 2.8.0-2
      entries:
      - name: couchbase-operator.v2.8.0-2
        version: 2.8.0-2
      name: 2.8.0
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: 2.8.1
    - currentCSV: couchbase-operator.v2.8.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCluster",
                "metadata": {
                  "name": "cb-example"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a",
                  "cluster": {
                    "clusterName": "cb-example",
                    "dataServiceMemoryQuota": "256Mi",
                    "indexServiceMemoryQuota": "256Mi",
                    "searchServiceMemoryQuota": "256Mi",
                    "eventingServiceMemoryQuota": "256Mi",
                    "analyticsServiceMemoryQuota": "1Gi",
                    "indexStorageSetting": "memory_optimized",
                    "autoFailoverTimeout": "120s",
                    "autoFailoverMaxCount": 3,
                    "autoFailoverOnDataDiskIssues": true,
                    "autoFailoverOnDataDiskIssuesTimePeriod": "120s",
                    "autoFailoverServerGroup": false
                  },
                  "upgradeStrategy": "RollingUpgrade",
                  "hibernate": false,
                  "hibernationStrategy": "Immediate",
                  "recoveryPolicy": "PrioritizeDataIntegrity",
                  "onlineVolumeExpansionTimeoutInMins": 20,
                  "security": {
                    "adminSecret": "cb-example-auth",
                    "rbac": {
                      "managed": true,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                    }
                  },
                  "xdcr": {
                     "managed": false,
                      "selector": {
                        "matchLabels": {
                          "cluster": "cb-example"
                        }
                      }
                   },
                  "backup": {
                    "image": "registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f",
                    "managed": false,
                    "serviceAccountName": "couchbase-backup",
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "monitoring": {
                    "prometheus": {
                      "enabled": false,
                      "image": "registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1"
                    }
                  },
                  "networking": {
                    "exposeAdminConsole": true,
                    "adminConsoleServices": [
                      "data"
                    ],
                    "exposedFeatures": [
                      "xdcr"
                    ],
                    "exposedFeatureServiceType": "NodePort",
                    "adminConsoleServiceType": "NodePort",
                    "cloudNativeGateway": {
                      "image": "registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc"
                    }
                  },
                  "buckets": {
                    "managed": true,
                    "selector": {
                      "matchLabels": {
                        "cluster": "cb-example"
                      }
                    }
                  },
                  "logRetentionTime": "604800s",
                  "logRetentionCount": 20,
                  "enablePreviewScaling": false,
                  "servers": [
                    {
                      "size": 3,
                      "name": "all_services",
                      "services": [
                        "data",
                        "index",
                        "query",
                        "search",
                        "eventing",
                        "analytics"
                      ]
                    }
                  ]
                },
                "status": {
                  "size": 0
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBucket",
                "metadata": {
                  "name": "default",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "memoryQuota": "100Mi",
                  "replicas": 2,
                  "ioPriority": "low",
                  "evictionPolicy": "valueOnly",
                  "conflictResolution": "lww",
                  "enableFlush": false,
                  "enableIndexReplica": true,
                  "compressionMode": "passive"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseEphemeralBucket",
                 "metadata": {
                    "name": "ephemeral-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "replicas": 2,
                    "ioPriority": "low",
                    "evictionPolicy": "noEviction",
                    "conflictResolution": "lww",
                    "enableFlush": false,
                    "compressionMode": "passive"
                 },
                 "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseMemcachedBucket",
                 "metadata": {
                    "name": "memcached-bucket",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "memoryQuota": "100Mi",
                    "enableFlush": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "fullName": "My User",
                  "authDomain": "local",
                  "authSecret": "cb-example-auth"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseGroup",
                "metadata": {
                  "name": "my-group",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "roles": [
                    {
                      "name": "bucket_admin",
                      "bucket": "default"
                    }
                  ]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseRoleBinding",
                "metadata": {
                  "name": "my-role-binding"
                },
                "spec": {
                  "subjects": [
                    {
                      "kind": "CouchbaseUser",
                      "name": "my-user"
                    }
                  ],
                  "roleRef": {
                    "kind": "CouchbaseGroup",
                    "name": "my-group"
                  }
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseReplication",
                 "metadata": {
                    "name": "my-replication",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "bucket": "default",
                    "remoteBucket": "default",
                    "compressionType": "Snappy",
                    "filterExpression": "",
                    "paused": false
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseBackup",
                "metadata": {
                  "name": "cb-backup",
                  "labels": {
                    "cluster": "cb-example"
                  }
                },
                "spec": {
                  "strategy": "full_incremental",
                  "full": {
                    "schedule": "0 3 * * 6"
                  },
                  "incremental": {
                    "schedule": "0 3 * * 1-6"
                  },
                  "successfulJobsHistoryLimit": 1,
                  "failedJobsHistoryLimit": 3,
                  "backOffLimit": 2,
                  "backupRetention": "24h",
                  "logRetention": "24h",
                  "size": "5Gi"
                },
                "status": {}
              },
              {
                 "apiVersion": "couchbase.com/v2",
                 "kind": "CouchbaseBackupRestore",
                 "metadata": {
                    "name": "cb-restore",
                    "labels": {
                       "cluster": "cb-example"
                    }
                 },
                 "spec": {
                    "backup": "cb-backup",
                    "repo": "cb-example-2020-10-29T19_00_03",
                    "start": {
                       "int": 1
                    },
                    "end": {
                       "int": 1
                    },
                    "backOffLimit": 2,
                    "logRetention": "24h"
                 },
                 "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollectionGroup",
                "metadata": {
                  "name": "my-collection-group"
                },
                "spec": {
                  "maxTTL": "",
                  "names": ["my-collection"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseCollection",
                "metadata": {
                  "name": "my-collection"
                },
                "spec": {
                  "maxTTL": "",
                  "name": "my-collection"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScopeGroup",
                "metadata": {
                  "name": "my-scope-group"
                },
                "spec": {
                  "collections": {},
                  "names": ["my-scope"]
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseScope",
                "metadata": {
                  "name": "my-scope"
                },
                "spec": {
                  "collections": {},
                  "name": "my-scope"
                },
                "status": {}
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseMigrationReplication",
                "metadata": {
                  "name": "default-migration"
                },
                "migrationMapping": {
                  "mappings": []
                },
                "spec": {
                  "bucket": "default",
                  "remoteBucket": "default"
                }
              },
              {
                "apiVersion": "couchbase.com/v2",
                "kind": "CouchbaseAutoscaler",
                "metadata": {
                  "name": "do.not.create.internal.only"
                },
                "spec": {
                  "servers": "internal",
                  "size": 2
                },
                "status": {
                  "labelSelector": "",
                  "size": 2
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          createdAt: 2023/04/28
          description: The Couchbase Autonomous Operator allows users to easily deploy,
            manage, and maintain Couchbase deployments
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/couchbase-enterprise-certified-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          support: Couchbase, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages Clusters
            displayName: Couchbase Cluster
            kind: CouchbaseCluster
            name: couchbaseclusters.couchbase.com
            version: v2
          - description: Manages Buckets
            displayName: Couchbase Bucket
            kind: CouchbaseBucket
            name: couchbasebuckets.couchbase.com
            version: v2
          - description: Manages Ephemeral Buckets
            displayName: Couchbase Ephemeral Bucket
            kind: CouchbaseEphemeralBucket
            name: couchbaseephemeralbuckets.couchbase.com
            version: v2
          - description: Manages Memcached Buckets
            displayName: Couchbase Memcached Bucket
            kind: CouchbaseMemcachedBucket
            name: couchbasememcachedbuckets.couchbase.com
            version: v2
          - description: Manages RBAC Users
            displayName: Couchbase User
            kind: CouchbaseUser
            name: couchbaseusers.couchbase.com
            version: v2
          - description: Manages RBAC Groups
            displayName: Couchbase Groups
            kind: CouchbaseGroup
            name: couchbasegroups.couchbase.com
            version: v2
          - description: Manages RBAC Binding
            displayName: Couchbase Role Binding
            kind: CouchbaseRoleBinding
            name: couchbaserolebindings.couchbase.com
            version: v2
          - description: Manages Couchbase Replications
            displayName: Couchbase Replications
            kind: CouchbaseReplication
            name: couchbasereplications.couchbase.com
            version: v2
          - description: Manages Backups
            displayName: Couchbase Backup
            kind: CouchbaseBackup
            name: couchbasebackups.couchbase.com
            version: v2
          - description: Manages Backup Restores
            displayName: Couchbase Backup Restore
            kind: CouchbaseBackupRestore
            name: couchbasebackuprestores.couchbase.com
            version: v2
          - description: Manages RBAC Collection Groups
            displayName: Couchbase Collection Groups
            kind: CouchbaseCollectionGroup
            name: couchbasecollectiongroups.couchbase.com
            version: v2
          - description: Manages RBAC Collections
            displayName: Couchbase Collection
            kind: CouchbaseCollection
            name: couchbasecollections.couchbase.com
            version: v2
          - description: Manages RBAC Scope Groups
            displayName: Couchbase Scope Group
            kind: CouchbaseScopeGroup
            name: couchbasescopegroups.couchbase.com
            version: v2
          - description: Manages RBAC Scopes
            displayName: Couchbase Scope
            kind: CouchbaseScope
            name: couchbasescopes.couchbase.com
            version: v2
          - description: Special migration mapping within XDCR to take a filtered
              list from the default scope and collection of the source to target bucket.
            displayName: Couchbase Migration Replication
            kind: CouchbaseMigrationReplication
            name: couchbasemigrationreplications.couchbase.com
            version: v2
          - description: Internal Autoscaling Management Resource
            displayName: Couchbase Autoscaler
            kind: CouchbaseAutoscaler
            name: couchbaseautoscalers.couchbase.com
            version: v2
        description: |
          The Couchbase Autonomous Operator allows users to easily deploy, manage, and maintain Couchbase deployments on OpenShift. By installing this integration you will be able to deply Couchbase Server clusters with a single command.

          ## Supported Features

          * **Automated cluster provisioning** - Deploying a Couchbase Cluster has never been easier. Fill out a Couchbase specific configuration and let the Couchbase Operator take care of provisioning nodes and setting up cluster to your exact specification.

          * **On-demand scalability** - Automatically scale your cluster up or down by changing a simple configuration parameter and let the Couchbase Operator handle provisioning of new nodes and joining them into the cluster.

          * **Auto-recovery** - Detect Couchbase node failures, rebalance out bad nodes, and bring the cluster back up to the desired capacity. Auto-recovery is completely automated so you can sleep easy through the night knowing that the Couchbase Operator will handle any failures.

          * **Geo-distribution** - Replicate your data between datacenters to move data closer to the users who consume it and protect against disaster scenarios where an entire datacenter becomes unavailable.

          * **Persistent storage** - Define persistent network-attached storage for each node in your cluster to allow pods to be recovered even if the node they were running on is no longer available.

          * **Rack/zone awareness** - Tell the Couchbase Operator about availability zones in your datacenter and let the operator take care of ensuring that nodes in your cluster are deployed equally across each zone.

          * **Supportability** - When things go wrong, use the cbopinfo tool provided with the Couchbase Operator to collect relevant data about your Couchbase deployment so that you can quickly address issues.

          * **Centralized configuration management** - Manage your configuration centrally with OpenShift. Updates to the configuration are watched by the Couchbase Operator and actions are taken to make the target cluster match the desired configuration.
          ## Required Parameters
          * `authSecret` - provide the name of a secret that contains two keys for the `username` and `password` of the super user ([documentation](https://docs.couchbase.com/operator/1.2/couchbase-cluster-config.html))

          ## About Couchbase Server

          Built on the most powerful NoSQL technology, Couchbase Server delivers unparalleled performance at scale, in any cloud. With features like memory-first architecture, geo-distributed deployments, and workload isolation, Couchbase Server excels at supporting mission-critical applications at scale while maintaining submillisecond latencies and 99.999% availability. Plus, with the most comprehensive SQL-compatible query language (N1QL), migrating from RDBMS to Couchbase Server is easy with ANSI joins.
        displayName: Couchbase Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - couchbase
        - database
        - key value
        - nosql
        - open source
        links:
        - name: Couchbase
          url: https://www.couchbase.com
        - name: Documentation
          url: https://docs.couchbase.com/operator/current/overview.html
        - name: Downloads
          url: https://www.couchbase.com/downloads
        maintainers:
        - email: support@couchbase.com
          name: Couchbase
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Couchbase
        relatedImages:
        - registry.connect.redhat.com/couchbase/server@sha256:4e0d41b559c5d536aa06709457d6c788fd039b0c915c14f2d429213bdfeff824
        - registry.connect.redhat.com/couchbase/server@sha256:643272bd58b0b584863400d281f1dcb147099c7f2120a947347be4b1f97391a7
        - registry.connect.redhat.com/couchbase/server@sha256:033c0b2b8eb12ac63a158754dce20d9787a470c77252c60bd4d9e11f5b2dbc7a
        - registry.connect.redhat.com/couchbase/operator@sha256:52a11164a9cc79588790765af7a74ebad14ae75e81681a23bb65bd393bf42096
        - registry.connect.redhat.com/couchbase/cloud-native-gateway@sha256:50bbc00cce14ec968c9b8e9ef06a6c5bd2860d38d973cf7af7271d4ff66fe9cc
        - registry.connect.redhat.com/couchbase/operator-rhmp@sha256:5213e1c50c238e30bcc50c98b337ea3a145234f79e4bdc7a4ed46504a14f914a
        - registry.connect.redhat.com/couchbase/server@sha256:437cd5aa05ff057b17f47d2500dec7cefca21682eb0f4badff4538dc164bc1a7
        - registry.connect.redhat.com/couchbase/exporter@sha256:de3bacf8233553452db1d9cec7c98b28d12c41aa025d830d6b7c5de0da9723b1
        - registry.connect.redhat.com/couchbase/operator-backup@sha256:382511f532b0206d2bb0cecc8f91b055809b5ccf89136b6d37d55a40e266812f
        version: 2.8.1-1
      entries:
      - name: couchbase-operator.v2.8.1-1
        version: 2.8.1-1
      name: stable
    defaultChannel: stable
    packageName: couchbase-enterprise-certified-rhmp
    provider:
      name: Couchbase
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxrm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxrm-operator-certified.v3.58.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "example-nexusrepo"
                },
                "spec": {
                  "config": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/sonatype-nexus-conf"
                  },
                  "deployment": {
                    "additionalContainers": null,
                    "additionalVolumeMounts": null,
                    "additionalVolumes": null,
                    "annotations": {},
                    "initContainers": null,
                    "postStart": {
                      "command": null
                    },
                    "preStart": {
                      "command": null
                    },
                    "terminationGracePeriodSeconds": 120
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "path": "/",
                    "rules": null,
                    "tls": {
                      "enabled": true,
                      "secretName": "nexus-tls"
                    }
                  },
                  "nexus": {
                    "dockerPort": 5003,
                    "env": [
                      {
                        "name": "INSTALL4J_ADD_VM_PARAMS",
                        "value": "-Xms2703M -Xmx2703M -XX:MaxDirectMemorySize=2703M -XX:+UnlockExperimentalVMOptions -XX:+UseCGroupMemoryLimitForHeap"
                      }
                    ],
                    "hostAliases": [],
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "nexusPort": 8081,
                    "podAnnotations": {},
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 30,
                      "path": "/",
                      "periodSeconds": 30
                    },
                    "resources": {},
                    "securityContext": {
                    },
                    "service": {
                      "type": "NodePort"
                    }
                  },
                  "nexusProxyRoute": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null
                  },
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "enabled": true,
                    "storageSize": "8Gi"
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": null,
                    "enabled": false,
                    "labels": null,
                    "name": "docker",
                    "portName": "docker"
                  },
                  "secret": {
                    "data": null,
                    "enabled": false,
                    "mountPath": "/etc/secret-volume",
                    "readOnly": true
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "statefulset": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
          createdAt: "2023-07-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          olm.skipRange: <3.58.1-2
          repository: https://github.com/sonatype/operator-nxrm3
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          High Availability Clustering (HA-C) is not supported for Nexus Repository Pro for OpenShift.

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to that service for nexus.port (8081).

          By default, the Nexus Repository starts up in OSS mode until a license is installed.

          The Nexus Repository can be further configured via the NexusRepo custom resource definition:

          | Parameter                                   | Description                         | Default                                 |
          | ------------------------------------------  | ----------------------------------  | ----------------------------------------|
          | `statefulset.enabled`                       | Use statefulset instead of deployment | `false` |
          | `deploymentStrategy.type`                   | Deployment Strategy     |  `Recreate` |
          | `nexus.env`                                 | Nexus environment variables         | `See example.` |
          | `nexus.resources`                           | Nexus resource requests and limits  | `{}`                                    |
          | `nexus.dockerPort`                          | Port to access docker               | `5003`                                  |
          | `nexus.nexusPort`                           | Internal port for Nexus service     | `8081`                                  |
          | `nexus.service.type`                        | Service for Nexus                   |`NodePort`                                |
          | `nexus.service.clusterIp`                   | Specific cluster IP when service type is cluster IP. Use None for headless service |`nil`   |
          | `nexus.securityContext`                     | Security Context |
          | `nexus.labels`                              | Service labels                      | `{}`                                    |
          | `nexus.podAnnotations`                      | Pod Annotations                     | `{}`
          | `nexus.livenessProbe.initialDelaySeconds`   | LivenessProbe initial delay         | 30                                      |
          | `nexus.livenessProbe.periodSeconds`         | Seconds between polls               | 30                                      |
          | `nexus.livenessProbe.failureThreshold`      | Number of attempts before failure   | 6                                       |
          | `nexus.livenessProbe.timeoutSeconds`        | Time in seconds after liveness probe times out    | `nil`                     |
          | `nexus.livenessProbe.path`                  | Path for LivenessProbe              | /                                       |
          | `nexus.readinessProbe.initialDelaySeconds`  | ReadinessProbe initial delay        | 30                                      |
          | `nexus.readinessProbe.periodSeconds`        | Seconds between polls               | 30                                      |
          | `nexus.readinessProbe.failureThreshold`     | Number of attempts before failure   | 6                                       |
          | `nexus.readinessProbe.timeoutSeconds`       | Time in seconds after readiness probe times out    | `nil`                    |
          | `nexus.readinessProbe.path`                 | Path for ReadinessProbe             | /                                       |
          | `nexus.hostAliases`                         | Aliases for IPs in /etc/hosts       | []                                      |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `tolerations`                               | tolerations list                    | `[]`                                    |
          | `config.enabled`                            | Enable configmap                    | `false`                                 |
          | `config.mountPath`                          | Path to mount the config            | `/sonatype-nexus-conf`                  |
          | `config.data`                               | Configmap data                      | `nil`                                   |
          | `deployment.terminationGracePeriodSeconds`  | Time to allow for clean shutdown    | 120                                     |
          | `deployment.annotations`                    | Annotations to enhance deployment configuration  | `{}`                       |
          | `deployment.initContainers`                 | Init containers to run before main containers  | `nil`                        |
          | `deployment.postStart.command`              | Command to run after starting the nexus container  | `nil`                    |
          | `deployment.preStart.command`               | Command to run before starting the nexus container  | `nil`                   |
          | `deployment.additionalContainers`           | Add additional Container         | `nil`                                      |
          | `deployment.additionalVolumes`              | Add additional Volumes           | `nil`                                      |
          | `deployment.additionalVolumeMounts`         | Add additional Volume mounts     | `nil`                                      |
          | `secret.enabled`                            | Enable secret                    | `false`                                    |
          | `secret.mountPath`                          | Path to mount the secret         | `/etc/secret-volume`                       |
          | `secret.readOnly`                           | Secret readonly state            | `true`                                     |
          | `secret.data`                               | Secret data                      | `nil`                                      |
          | `service.enabled`                           | Enable additional service        | `nil`                                      |
          | `service.name`                              | Service name                     | `nil`                                      |
          | `service.portName`                          | Service port name                | `nil`                                      |
          | `service.labels`                            | Service labels                   | `nil`                                      |
          | `service.annotations`                       | Service annotations              | `nil`                                      |
          | `service.loadBalancerSourceRanges`          | Service LoadBalancer source IP whitelist | `nil`                              |
          | `service.targetPort`                        | Service port                     | `nil`                                      |
          | `service.port`                              | Port for exposing service        | `nil`                                      |
          | `route.enabled`         | Set to true to create route for additional service | `false` |
          | `route.name`            | Name of route                                      | `docker` |
          | `route.portName`        | Target port name of service                        | `docker` |
          | `route.labels`          | Labels to be added to route                        | `{}` |
          | `route.annotations`     | Annotations to be added to route                   | `{}` |
          | `route.path`            | Host name of Route e.g jenkins.example.com         | nil |
        displayName: Nexus Repository Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:06fe31f40ff3fd7556fc7b8f6b2efe80ada1d9079ec9aaad60d1fdc0ee0677f0
        - registry.connect.redhat.com/sonatype/nxrm-operator-bundle@sha256:47a7bafff2715fbef6130ca0b84560b6c2358275821b72ae54ad46020e02bee5
        - registry.connect.redhat.com/sonatype/nxrm-operator-certified@sha256:4945ad993a10e27b358b2d9869d05f23ca761162ec088e5a833939022bc23192
        version: 3.58.1-2
      entries:
      - name: nxrm-operator-certified.v3.58.1-2
        version: 3.58.1-2
      name: stable
    defaultChannel: stable
    packageName: nxrm-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-dcap-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-dcap-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "aesmd": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/aesmd-dcap",
                      "tag": "latest"
                    },
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "multiPlatformRegistration": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/multi-platform-registration-ubi8",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities, in-kernel driver
          installed and fsgsbase support.\n\n## Installation\n * Choose a namespace
          to install the Operator. \n Operator will be available in all namespaces.\n*
          Select Approval Strategy: Available options are Automatic or Manual. Automatic
          approval will automatically upgrade the running instance of your Operator
          without human intervention when a new version is available. In case of Manual
          approval, when a new version is available, OLM will create an update request.
          You must manually approve the request to update the Operator to a new version.\n*
          Generate your CCM Account Join Token: Please log in to https://ccm.fortanix.com,
          and in the Management Console tab, click the + ENROLL NODE button. In the
          ENROLL NODE screen, a Join Token will be generated in the text box for \"Get
          a join token to register an SGX compute node\". This Join Token is used
          by the compute node to authenticate itself.\n* Base64 encode the Join Token:\n```shell\necho
          -n <join-token> | openssl base64 -A\n```\n* Create an instance of CcmNodeAgent
          by setting the `spec.ccmJoinToken.Token` to the base64 encoded CCM account
          join token.\n* After CcmNodeAgent resource is installed successfully you
          should be able to see all the worker nodes of the cluster in your CCM account
          under Compute Node Tab.\n\n## Getting Started\nNow to run your application
          inside SGX enclaves on your cluster follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent DCAP Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent DCAP Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - docker.io/fortanix/ccm-node-agent-dcap-operator@sha256:4fd28bea26aee461eb9c1800cdf5f041b64ae3ea6652203f8ad9feb680aaefee
        - registry.connect.redhat.com/fortanix/ccm-node-agent-dcap-operator-bundle@sha256:545724c273dfd6a1fa923d468b2d054f39ec3186d01586a8b3fe494c0f22f8dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        version: 0.0.2
      entries:
      - name: ccm-node-agent-dcap-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-dcap-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NdmSpc project
      provider-url: ""
    name: ndmspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ndmspc-operator.v0.20250209.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.ndmspc.io/v1alpha1",
                "kind": "NdmSpcConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ndmspc-operator",
                    "app.kubernetes.io/instance": "ndmspcconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ndmspcconfig",
                    "app.kubernetes.io/part-of": "ndmspc-operator"
                  },
                  "name": "ndmspc-sample"
                },
                "spec": {
                  "executor": {
                    "image": "registry.gitlab.com/ndmspc/api:v0.20240215.2",
                    "name": "executor"
                  },
                  "executors": 1,
                  "ingressclasstype": "nginx",
                  "ingressenabled": true,
                  "ingresshostdomain": "127.0.0.1.sslip.io",
                  "ingressnostname": "ndmspc",
                  "ingressssltraefik": false,
                  "naf": {
                    "image": "registry.gitlab.com/ndmspc/naf:v0.20250129.0",
                    "name": "naf"
                  },
                  "ndmvr": {
                    "image": "registry.gitlab.com/ndmspc/ndmvr/shared:v0.20230320.0",
                    "name": "ndmvr"
                  },
                  "salsa": {
                    "image": "registry.gitlab.com/ndmspc/user/ndmspc:v0.20240412.0"
                  },
                  "storage": {
                    "image": "registry.gitlab.com/ndmspc/storage-server:v0.20250205.0",
                    "mongodb": {
                      "image": "docker.io/library/mongo:latest",
                      "password": "admin",
                      "port": 27017,
                      "username": "admin"
                    },
                    "name": "storage",
                    "port": 8080
                  },
                  "webapp": {
                    "image": "registry.gitlab.com/ndmspc/react-ndmspc:v0.20240412.0",
                    "name": ""
                  },
                  "webappcontainerport": 8080,
                  "webapps": 1,
                  "worker": {
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "1Gi"
                      }
                    }
                  },
                  "workers": 2,
                  "workerslots": 1,
                  "zmq2ws": {
                    "image": "registry.gitlab.com/ndmspc/zmq2ws:v1.20231012.0",
                    "name": "zmq2ws"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
          createdAt: "2025-02-09T12:49:06Z"
          description: The NdmSpc operator automates the deployment and management
            of Ndmspc suite on Kubernetes cluster.
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/ndmspc/ndmspc-operator
          support: NDMSPC Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NdmSpc configuration
            displayName: NdmSpc Config
            kind: NdmSpcConfig
            name: ndmspcconfigs.apps.ndmspc.io
            version: v1alpha1
        description: "The NdmSpc operator automates the deployment and management
          of Ndmspc suite on Kubernetes cluster. This operator allows user to deploy
          SALSA cluster with ZMQ2WS and job executor service and specify custom webapp
          page.\n\n### Capabilities\n\n- SALSA cluster\n- ZMQ2WS api service\n- Job
          executor service \n- Custom WebApp page\n"
        displayName: NdmSpc operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ndmspc
        links:
        - name: Ndmspc Operator
          url: https://ndmspc.gitlab.io/docs
        maintainers:
        - email: ndmspc@gmail.com
          name: NdmSpc Support
        maturity: alpha
        provider:
          name: NdmSpc project
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/ndmspc-operator@sha256:b1f47ce0cd6930d902dc3538a1a87140dbe9257dd50773a518fc2350185f90cc
        - registry.gitlab.com/ndmspc/ndmspc-operator:0.20250209.0
        version: 0.20250209.0
      entries:
      - name: ndmspc-operator.v0.20250209.0
        version: 0.20250209.0
      - name: ndmspc-operator.v0.20250131.0
        version: 0.20250131.0
      - name: ndmspc-operator.v0.20240412.0
        version: 0.20240412.0
      - name: ndmspc-operator.v0.11.6
        version: 0.11.6
      - name: ndmspc-operator.v0.11.4
        version: 0.11.4
      - name: ndmspc-operator.v0.11.3
        version: 0.11.3
      - name: ndmspc-operator.v0.11.2
        version: 0.11.2
      - name: ndmspc-operator.v0.11.1
        version: 0.11.1
      - name: ndmspc-operator.v0.10.2
        version: 0.10.2
      - name: ndmspc-operator.v0.10.1
        version: 0.10.1
      - name: ndmspc-operator.v0.10.0
        version: 0.10.0
      - name: ndmspc-operator.v0.9.0
        version: 0.9.0
      - name: ndmspc-operator.v0.7.4
        version: 0.7.4
      - name: ndmspc-operator.v0.7.1
        version: 0.7.1
      - name: ndmspc-operator.v0.6.0
        version: 0.6.0
      - name: ndmspc-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: ndmspc-operator
    provider:
      name: NdmSpc project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: klusterlet
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: klusterlet.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "Klusterlet",
                "metadata": {
                  "name": "klusterlet"
                },
                "spec": {
                  "clusterName": "cluster1",
                  "deployOption": {
                    "mode": "Singleton"
                  },
                  "externalServerURLs": [
                    {
                      "url": "https://localhost"
                    }
                  ],
                  "imagePullSpec": "quay.io/open-cluster-management/registration-operator:v0.13.1",
                  "namespace": "open-cluster-management-agent",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the Klusterlet.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Klusterlet defines the configurations of agents running on
              the managed cluster
            displayName: Klusterlet
            kind: Klusterlet
            name: klusterlets.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/klusterlet. \n\nThe
          Klusterlet proivdes the registation to the Hub clusters as a managed cluster.
          This operator supports the installation and upgrade of Klusterlet.\n\n##
          Prerequisites\nYou need a Hub cluster which has installed ClusterManager.\n\n##
          How to Install\n* Install Klusterlet operator by following instructions
          in top right button `Install`.\n* Optional: Create an agent namespace used
          to deploy Klusterlet. The default namespace is `open-cluster-management-agent`.\n*
          Create a secret `bootstrap-hub-kubeconfig` with the kubeconfig file of hub
          cluster in the agent namespace.\n\n  ```\n  $ kubectl create secret generic
          bootstrap-hub-kubeconfig --from-file=kubeconfig=<kubeconfig file of hub
          clsuter> -n <agent namespace>\n  ```\n\n* Create an instance of the Klusterlet
          resource to install the Klusterlet.\n\nThe pods of Klusterlet are created
          in agent namespace.\n\n```\n$ kubectl get pods -n open-cluster-management-agent\n
          \ NAME                                READY   STATUS    RESTARTS   AGE\n
          \ klusterlet-agent-55c6574876-nfxm9   1/1     Running   0          8m38s\n```\n\nThere
          is a ManagedCluster created in the Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster1               false                                                        27s\n```\n"
        displayName: Klusterlet
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - Klusterlet
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        - quay.io/community-operator-pipeline-prod/klusterlet@sha256:9efef9a034c9f12505a0d81bc9697c4a3ab7e3d05d1e7aaa4180490d70ddf6a2
        version: 0.13.1
      entries:
      - name: klusterlet.v0.13.1
        version: 0.13.1
      - name: klusterlet.v0.13.0
        version: 0.13.0
      - name: klusterlet.v0.12.0
        version: 0.12.0
      - name: klusterlet.v0.11.0
        version: 0.11.0
      - name: klusterlet.v0.10.0
        version: 0.10.0
      - name: klusterlet.v0.9.1
        version: 0.9.1
      - name: klusterlet.v0.8.0
        version: 0.8.0
      - name: klusterlet.v0.7.0
        version: 0.7.0
      - name: klusterlet.v0.6.0
        version: 0.6.0
      - name: klusterlet.v0.5.0
        version: 0.5.0
      - name: klusterlet.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: klusterlet
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "maxUnhealthy": "49%",
                  "remediationTemplate": {
                    "apiVersion": "poison-pill.medik8s.io/v1alpha1",
                    "kind": "PoisonPillRemediationTemplate",
                    "name": "group-x",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
          createdAt: "2021-04-20 12:00:00"
          description: An operator to monitor node healthcheck and remdiate via 3rd
            party providers like poison-pill.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: This operator deploys the Node Health Check controller and its
          artifacts
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://node-healthcheck-operator.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:30264b9f49912cc0e6bed59001869e3168caace20b292c9dad4cc2d4947d289d
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:0.1.0-latest
        version: 0.1.1
      entries:
      - name: node-healthcheck-operator.v0.1.1
        version: 0.1.1
      name: alpha
    - currentCSV: node-healthcheck-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.3.0
          createdAt: "2022-08-24 11:40:38"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.0'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: medik8s team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:d7cc6d2575eb252fb3e2468106e801f192da7844b0b153dc3c6a12cf7cc0a8fe
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/node-healthcheck-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: quay.io/medik8s/node-healthcheck-operator:v0.6.1
          createdAt: "2023-11-14T15:45:34Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.1.0 <0.6.1'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/medik8s/node-healthcheck-operator:v0.6.1
        - quay.io/medik8s/node-remediation-console:v0.6.1
        - quay.io/openshift-community-operators/node-healthcheck-operator@sha256:60e99bb91309e448c652898dac070a6d920fda164f27879d94632c88a9e61ffa
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      - name: node-healthcheck-operator.v0.6.0
        version: 0.6.0
      - name: node-healthcheck-operator.v0.5.0
        version: 0.5.0
      - name: node-healthcheck-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v1.2.4953-c248a51
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:c248a51ae1
          createdAt: "2025-09-22T13:41:19Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeploymentCustomization is the Schema for clusterdeploymentcustomizations
              API.
            displayName: ClusterDeploymentCustomization
            kind: ClusterDeploymentCustomization
            name: clusterdeploymentcustomizations.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: |-
              ClusterPool represents a pool of clusters that should be kept ready to be given out to users. Clusters are removed
              from the pool once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: |-
              MachinePoolNameLease is the Schema for the MachinePoolNameLeases API. This resource is mostly empty
              as we're primarily relying on the name to determine if a lease is available.
              Note that not all cloud providers require the use of a lease for naming, at present this
              is only required for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * IBM Cloud
          * Red Hat OpenStack
          * vSphere

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Troubleshooting](https://github.com/openshift/hive/blob/master/docs/troubleshooting.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
              logLevel: debug
              targetNamespace: hive
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/hive-operator:1.2.4953-c248a51
        - quay.io/openshift-hive/hive:c248a51ae1
        version: 1.2.4953-c248a51
      entries:
      - name: hive-operator.v1.2.4953-c248a51
        version: 1.2.4953-c248a51
      - name: hive-operator.v1.2.4928-76a65b6
        version: 1.2.4928-76a65b6
      - name: hive-operator.v1.2.4902-4929f16
        version: 1.2.4902-4929f16
      - name: hive-operator.v1.2.4896-a29a238
        version: 1.2.4896-a29a238
      - name: hive-operator.v1.2.4868-45ae912
        version: 1.2.4868-45ae912
      - name: hive-operator.v1.2.4863-e9ad0fa
        version: 1.2.4863-e9ad0fa
      - name: hive-operator.v1.2.4774-87bff59
        version: 1.2.4774-87bff59
      - name: hive-operator.v1.2.4766-6a82857
        version: 1.2.4766-6a82857
      - name: hive-operator.v1.2.4746-d25e0db
        version: 1.2.4746-d25e0db
      - name: hive-operator.v1.2.4738-b19773e
        version: 1.2.4738-b19773e
      - name: hive-operator.v1.2.4731-41f153f
        version: 1.2.4731-41f153f
      - name: hive-operator.v1.2.4703-02f3f00
        version: 1.2.4703-02f3f00
      - name: hive-operator.v1.2.4693-ec3bc72
        version: 1.2.4693-ec3bc72
      - name: hive-operator.v1.2.4686-553c6c2
        version: 1.2.4686-553c6c2
      - name: hive-operator.v1.2.4682-406d6ca
        version: 1.2.4682-406d6ca
      - name: hive-operator.v1.2.4672-af54e2f
        version: 1.2.4672-af54e2f
      - name: hive-operator.v1.2.4668-3ccd44d
        version: 1.2.4668-3ccd44d
      - name: hive-operator.v1.2.4664-00f84cc
        version: 1.2.4664-00f84cc
      - name: hive-operator.v1.2.4658-c13e207
        version: 1.2.4658-c13e207
      - name: hive-operator.v1.2.4654-3268a18
        version: 1.2.4654-3268a18
      - name: hive-operator.v1.2.4650-531d549
        version: 1.2.4650-531d549
      - name: hive-operator.v1.2.4644-365f074
        version: 1.2.4644-365f074
      - name: hive-operator.v1.2.4638-56d6d64
        version: 1.2.4638-56d6d64
      - name: hive-operator.v1.2.4625-389d717
        version: 1.2.4625-389d717
      - name: hive-operator.v1.2.4615-52100dd
        version: 1.2.4615-52100dd
      - name: hive-operator.v1.2.4605-b41ca3f
        version: 1.2.4605-b41ca3f
      - name: hive-operator.v1.2.4584-95efcb7
        version: 1.2.4584-95efcb7
      - name: hive-operator.v1.2.4574-1ec27ad
        version: 1.2.4574-1ec27ad
      - name: hive-operator.v1.2.4568-b3eca52
        version: 1.2.4568-b3eca52
      - name: hive-operator.v1.2.4564-bdf9d08
        version: 1.2.4564-bdf9d08
      - name: hive-operator.v1.2.4552-bf2d689
        version: 1.2.4552-bf2d689
      - name: hive-operator.v1.2.4532-cddd4f9
        version: 1.2.4532-cddd4f9
      - name: hive-operator.v1.2.4511-1fce770
        version: 1.2.4511-1fce770
      - name: hive-operator.v1.2.4500-b1ac27b
        version: 1.2.4500-b1ac27b
      - name: hive-operator.v1.2.4485-47df0d9
        version: 1.2.4485-47df0d9
      - name: hive-operator.v1.2.4469-7be6433
        version: 1.2.4469-7be6433
      - name: hive-operator.v1.2.4463-d7ead60
        version: 1.2.4463-d7ead60
      - name: hive-operator.v1.2.4461-97686bb
        version: 1.2.4461-97686bb
      - name: hive-operator.v1.2.4459-f910c9b
        version: 1.2.4459-f910c9b
      - name: hive-operator.v1.2.4451-83aedb9
        version: 1.2.4451-83aedb9
      - name: hive-operator.v1.2.4446-e11a23b
        version: 1.2.4446-e11a23b
      - name: hive-operator.v1.2.4441-fdbfaf0
        version: 1.2.4441-fdbfaf0
      - name: hive-operator.v1.2.4436-44323bc
        version: 1.2.4436-44323bc
      - name: hive-operator.v1.2.4428-ceaafa8
        version: 1.2.4428-ceaafa8
      - name: hive-operator.v1.2.4426-fb27114
        version: 1.2.4426-fb27114
      - name: hive-operator.v1.2.4420-fe0e302
        version: 1.2.4420-fe0e302
      - name: hive-operator.v1.2.4418-eb5403f
        version: 1.2.4418-eb5403f
      - name: hive-operator.v1.2.4416-fba8d52
        version: 1.2.4416-fba8d52
      - name: hive-operator.v1.2.4408-9c5806a
        version: 1.2.4408-9c5806a
      - name: hive-operator.v1.2.4404-410fed9
        version: 1.2.4404-410fed9
      - name: hive-operator.v1.2.4400-8c54fc9
        version: 1.2.4400-8c54fc9
      - name: hive-operator.v1.2.4386-6f1fd3e
        version: 1.2.4386-6f1fd3e
      - name: hive-operator.v1.2.4382-0e95d5b
        version: 1.2.4382-0e95d5b
      - name: hive-operator.v1.2.4366-df65e7e
        version: 1.2.4366-df65e7e
      - name: hive-operator.v1.2.4360-de169dd
        version: 1.2.4360-de169dd
      - name: hive-operator.v1.2.4355-e5f809f
        version: 1.2.4355-e5f809f
      name: alpha
    - currentCSV: hive-operator.v2.5.3516-a2ed9b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
          createdAt: "2022-06-03T13:24:12Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:be43412a2e4bad2cd9d15e4fbabe0292585c5bbbd65da12d519f30f17d2fe594
        - quay.io/openshift-hive/hive:v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      entries:
      - name: hive-operator.v2.5.3516-a2ed9b3
        version: 2.5.3516-a2ed9b3
      - name: hive-operator.v2.5.3514-4052426
        version: 2.5.3514-4052426
      - name: hive-operator.v2.5.3508-6cb94c6
        version: 2.5.3508-6cb94c6
      - name: hive-operator.v2.5.3506-f42617c
        version: 2.5.3506-f42617c
      - name: hive-operator.v2.5.3495-5cfdd44
        version: 2.5.3495-5cfdd44
      - name: hive-operator.v2.5.3491-6feded1
        version: 2.5.3491-6feded1
      - name: hive-operator.v2.5.3489-6dec9c3
        version: 2.5.3489-6dec9c3
      name: mce-2.0
    - currentCSV: hive-operator.v2.3.3039-60b8a9a
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
          createdAt: "2022-08-15T17:23:39Z"
          description: OpenShift cluster provisioning and management at scale.
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/hive-operator@sha256:822fd8e65c241cebedd835bfebf67645199561cfdd6148d5afffb8b35e499ae3
        - quay.io/openshift-hive/hive:v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      entries:
      - name: hive-operator.v2.3.3039-60b8a9a
        version: 2.3.3039-60b8a9a
      name: ocm-2.3
    - currentCSV: hive-operator.v2.4.3231-ba51985
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hive.openshift.io/v1","kind":"HiveConfig","metadata":{"name":"hive"},"spec":{"managedDomains":[{"aws":{"credentialsSecretRef":{"name":"my-route53-creds"}},"domains":["my-base-domain.example.com"]}]}}]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-hive/hive:v2.4.3231-ba51985
          createdAt: "2022-08-15T11:53:28Z"
          description: OpenShift cluster provisioning and management at scale.
          operators.operatorframework.io/internal-objects: '["checkpoints.hive.openshift.io","clusterdeprovisions.hive.openshift.io","clusterprovisions.hive.openshift.io","clusterstates.hive.openshift.io","machinepoolnameleases.hive.openshift.io","clustersyncleases.hiveinternal.openshift.io","clustersyncs.hiveinternal.openshift.io","fakeclusterinstalls.hiveinternal.openshift.io"]'
          repository: https://github.com/openshift/hive
          support: Hive Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Checkpoint is the Schema for the backup of Hive objects.
            displayName: Checkpoint
            kind: Checkpoint
            name: checkpoints.hive.openshift.io
            version: v1
          - description: ClusterClaim represents a claim to a cluster from a cluster
              pool.
            displayName: ClusterClaim
            kind: ClusterClaim
            name: clusterclaims.hive.openshift.io
            version: v1
          - description: ClusterDeployment is the Schema for the clusterdeployments
              API
            displayName: ClusterDeployment
            kind: ClusterDeployment
            name: clusterdeployments.hive.openshift.io
            version: v1
          - description: ClusterDeprovision is the Schema for the clusterdeprovisions
              API
            displayName: ClusterDeprovision
            kind: ClusterDeprovision
            name: clusterdeprovisions.hive.openshift.io
            version: v1
          - description: ClusterImageSet is the Schema for the clusterimagesets API
            displayName: ClusterImageSet
            kind: ClusterImageSet
            name: clusterimagesets.hive.openshift.io
            version: v1
          - description: ClusterPool represents a pool of clusters that should be
              kept ready to be given out to users. Clusters are removed from the pool
              once claimed and then automatically replaced with a new one.
            displayName: ClusterPool
            kind: ClusterPool
            name: clusterpools.hive.openshift.io
            version: v1
          - description: ClusterProvision is the Schema for the clusterprovisions
              API
            displayName: ClusterProvision
            kind: ClusterProvision
            name: clusterprovisions.hive.openshift.io
            version: v1
          - description: ClusterRelocate is the Schema for the ClusterRelocates API
            displayName: ClusterRelocate
            kind: ClusterRelocate
            name: clusterrelocates.hive.openshift.io
            version: v1
          - description: ClusterState is the Schema for the clusterstates API
            displayName: ClusterState
            kind: ClusterState
            name: clusterstates.hive.openshift.io
            version: v1
          - description: DNSZone is the Schema for the dnszones API
            displayName: DNSZone
            kind: DNSZone
            name: dnszones.hive.openshift.io
            version: v1
          - description: HiveConfig is the Schema for the hives API
            displayName: HiveConfig
            kind: HiveConfig
            name: hiveconfigs.hive.openshift.io
            version: v1
          - description: MachinePoolNameLease is the Schema for the MachinePoolNameLeases
              API. This resource is mostly empty as we're primarily relying on the
              name to determine if a lease is available. Note that not all cloud providers
              require the use of a lease for naming, at present this is only required
              for GCP where we're extremely restricted on name lengths.
            displayName: MachinePoolNameLease
            kind: MachinePoolNameLease
            name: machinepoolnameleases.hive.openshift.io
            version: v1
          - description: MachinePool is the Schema for the machinepools API
            displayName: MachinePool
            kind: MachinePool
            name: machinepools.hive.openshift.io
            version: v1
          - description: SelectorSyncIdentityProvider is the Schema for the SelectorSyncSet
              API
            displayName: SelectorSyncIdentityProvider
            kind: SelectorSyncIdentityProvider
            name: selectorsyncidentityproviders.hive.openshift.io
            version: v1
          - description: SelectorSyncSet is the Schema for the SelectorSyncSet API
            displayName: SelectorSyncSet
            kind: SelectorSyncSet
            name: selectorsyncsets.hive.openshift.io
            version: v1
          - description: SyncIdentityProvider is the Schema for the SyncIdentityProvider
              API
            displayName: SyncIdentityProvider
            kind: SyncIdentityProvider
            name: syncidentityproviders.hive.openshift.io
            version: v1
          - description: SyncSet is the Schema for the SyncSet API
            displayName: SyncSet
            kind: SyncSet
            name: syncsets.hive.openshift.io
            version: v1
          - description: ClusterSyncLease is a record of the last time that SyncSets
              and SelectorSyncSets were applied to a cluster.
            displayName: ClusterSyncLease
            kind: ClusterSyncLease
            name: clustersyncleases.hiveinternal.openshift.io
            version: v1alpha1
          - description: ClusterSync is the status of all of the SelectorSyncSets
              and SyncSets that apply to a ClusterDeployment.
            displayName: ClusterSync
            kind: ClusterSync
            name: clustersyncs.hiveinternal.openshift.io
            version: v1alpha1
          - description: FakeClusterInstall represents a fake request to provision
              an agent based cluster.
            displayName: FakeClusterInstall
            kind: FakeClusterInstall
            name: fakeclusterinstalls.hiveinternal.openshift.io
            version: v1alpha1
        description: |-
          Hive for Red Hat OpenShift is an operator that runs on top of Kubernetes/OpenShift. Hive can be used to provision
          and perform initial configuration of OpenShift clusters.

          For provisioning OpenShift, Hive uses the [OpenShift installer](https://github.com/openshift/installer).

          ### Supported cloud providers
          * AWS
          * Azure
          * Google Cloud Platform
          * Red Hat OpenStack

          In the future Hive will support more cloud providers.

          ## Documentation

          * [Quick Start Guide](https://github.com/openshift/hive/blob/master/docs/quick_start.md)
          * [Using Hive](https://github.com/openshift/hive/blob/master/docs/using-hive.md)
          * [Hiveutil CLI](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)
          * [Frequently Asked Questions](https://github.com/openshift/hive/blob/master/docs/FAQs.md)
          * [Architecture](https://github.com/openshift/hive/blob/master/docs/architecture.md)

          See the [project README](https://github.com/openshift/hive#documentation) for more documentation.

          ## Post Install Configuration

          After installing the Hive for Red Hat OpenShift operator, create a cluster-scoped `HiveConfig` CR to configure Hive.
          Upon creation of `HiveConfig`, the operator will create the necessary Kubernetes resources to launch Hive.

          Example `HiveConfig`:
          ```yaml
          ---
            apiVersion: hive.openshift.io/v1
            kind: HiveConfig
            metadata:
              name: hive
            spec:
              managedDomains:
              - aws:
                  credentialsSecretRef:
                    name: my-route53-creds
                domains:
                - my-base-domain.example.com
          ```

          ## Create a cluster

          To create a cluster with Hive, create a `ClusterDeployment` CR. You can also use the
          [`hiveutil` tool](https://github.com/openshift/hive/blob/master/docs/hiveutil.md)'s `create-cluster` command
          to create clusters.
        displayName: Hive for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - openshift
        - multi-cluster
        - cluster
        links:
        - name: Hive GitHub
          url: https://github.com/openshift/hive
        - name: 'Hive: Cluster-as-a-Service'
          url: https://www.openshift.com/blog/openshift-hive-cluster-as-a-service
        - name: OpenShift
          url: https://www.openshift.com/
        maintainers:
        - email: openshift-hive-team@redhat.com
          name: Hive Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-hive/hive:v2.4.3231-ba51985
        - quay.io/openshift-community-operators/hive-operator@sha256:9a3cb682afc66ad662045b3d8ff214203cc00be60f66f74564820f6dcf99cd69
        version: 2.4.3231-ba51985
      entries:
      - name: hive-operator.v2.4.3231-ba51985
        version: 2.4.3231-ba51985
      - name: hive-operator.v2.4.3227-74831db
        version: 2.4.3227-74831db
      - name: hive-operator.v2.4.3222-d68346c
        version: 2.4.3222-d68346c
      - name: hive-operator.v2.4.3206-b0714a7
        version: 2.4.3206-b0714a7
      - name: hive-operator.v2.4.3203-672cfe7
        version: 2.4.3203-672cfe7
      - name: hive-operator.v2.4.3201-8c50440
        version: 2.4.3201-8c50440
      - name: hive-operator.v2.4.3199-cce4cb9
        version: 2.4.3199-cce4cb9
      - name: hive-operator.v2.4.3197-6c9c8e4
        version: 2.4.3197-6c9c8e4
      - name: hive-operator.v2.4.3194-f82b7d8
        version: 2.4.3194-f82b7d8
      - name: hive-operator.v2.4.3192-fae0111
        version: 2.4.3192-fae0111
      - name: hive-operator.v2.4.3189-026033b
        version: 2.4.3189-026033b
      - name: hive-operator.v2.4.3180-7299056
        version: 2.4.3180-7299056
      - name: hive-operator.v1.2.3179-0756b70e6
        version: 1.2.3179-0756b70e6
      - name: hive-operator.v1.1.16
        version: 1.1.16
      - name: hive-operator.v1.1.15
        version: 1.1.15
      - name: hive-operator.v1.1.14
        version: 1.1.14
      - name: hive-operator.v1.1.13
        version: 1.1.13
      - name: hive-operator.v1.1.12
        version: 1.1.12
      - name: hive-operator.v1.1.11
        version: 1.1.11
      - name: hive-operator.v1.1.10
        version: 1.1.10
      - name: hive-operator.v1.1.9
        version: 1.1.9
      - name: hive-operator.v1.1.8
        version: 1.1.8
      - name: hive-operator.v1.1.7
        version: 1.1.7
      - name: hive-operator.v1.1.6
        version: 1.1.6
      name: ocm-2.4
    defaultChannel: alpha
    packageName: hive-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:3d5ad015a953708cc53ac2ea92b1b1d99c870a801b26e790d41458adc0a958e5
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/datadog-operator-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        - registry.connect.redhat.com/datadog/operator-bundle-marketplace@sha256:fea5b977249013181616f60b95c5e6d224cd0f248717171d946ff0e4d6eed6c0
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified-rhmp
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-memorydb-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-memorydb-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ACL",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "ParameterGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "Snapshot",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "SubnetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "memorydb.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
          createdAt: "2025-09-19T19:36:49Z"
          description: Amazon MemoryDB for Redis controller is a service controller
            for managing MemoryDB resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ACL represents the state of an AWS memorydb ACL resource.
            displayName: ACL
            kind: ACL
            name: acls.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Cluster represents the state of an AWS memorydb Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.memorydb.services.k8s.aws
            version: v1alpha1
          - description: ParameterGroup represents the state of an AWS memorydb ParameterGroup
              resource.
            displayName: ParameterGroup
            kind: ParameterGroup
            name: parametergroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: Snapshot represents the state of an AWS memorydb Snapshot
              resource.
            displayName: Snapshot
            kind: Snapshot
            name: snapshots.memorydb.services.k8s.aws
            version: v1alpha1
          - description: SubnetGroup represents the state of an AWS memorydb SubnetGroup
              resource.
            displayName: SubnetGroup
            kind: SubnetGroup
            name: subnetgroups.memorydb.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS memorydb User resource.
            displayName: User
            kind: User
            name: users.memorydb.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon MemoryDB for Redis resources in AWS from within your Kubernetes cluster.

          **About Amazon MemoryDB for Redis**

          MemoryDB for Redis is a durable, in-memory database service that delivers ultra-fast performance. It is purpose-built for modern applications with microservices architectures.
          MemoryDB is compatible with Redis, a popular open source data store, enabling you to quickly build applications using the same flexible and friendly Redis data structures, APIs, and commands that they already use today. With MemoryDB, all of your data is stored in memory, which enables you to achieve microsecond read and single-digit millisecond write latency and high throughput. MemoryDB also stores data durably across multiple Availability Zones (AZs) using a Multi-AZ transactional log to enable fast failover, database recovery, and node restarts.
          Delivering both in-memory performance and Multi-AZ durability, MemoryDB can be used as a high-performance primary database for your microservices applications, eliminating the need to separately manage both a cache and durable database.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - MemoryDB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memorydb
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon MemoryDB for Redis Developer Resources
          url: https://docs.aws.amazon.com/memorydb/?id=docs_gateway
        maintainers:
        - email: ack-maintainers@amazon.com
          name: memorydb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/memorydb-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-memorydb-controller@sha256:7f2e65d817479fcc5658cac745cef325b2bfd6256eeed8b819f1713afd59fe68
        version: 1.1.1
      entries:
      - name: ack-memorydb-controller.v1.1.1
        version: 1.1.1
      - name: ack-memorydb-controller.v1.1.0
        version: 1.1.0
      - name: ack-memorydb-controller.v1.0.15
        version: 1.0.15
      - name: ack-memorydb-controller.v1.0.13
        version: 1.0.13
      - name: ack-memorydb-controller.v1.0.12
        version: 1.0.12
      - name: ack-memorydb-controller.v1.0.11
        version: 1.0.11
      - name: ack-memorydb-controller.v1.0.10
        version: 1.0.10
      - name: ack-memorydb-controller.v1.0.9
        version: 1.0.9
      - name: ack-memorydb-controller.v1.0.8
        version: 1.0.8
      - name: ack-memorydb-controller.v1.0.7
        version: 1.0.7
      - name: ack-memorydb-controller.v1.0.6
        version: 1.0.6
      - name: ack-memorydb-controller.v1.0.5
        version: 1.0.5
      - name: ack-memorydb-controller.v1.0.4
        version: 1.0.4
      - name: ack-memorydb-controller.v1.0.3
        version: 1.0.3
      - name: ack-memorydb-controller.v1.0.2
        version: 1.0.2
      - name: ack-memorydb-controller.v1.0.1
        version: 1.0.1
      - name: ack-memorydb-controller.v1.0.0
        version: 1.0.0
      - name: ack-memorydb-controller.v0.0.12
        version: 0.0.12
      - name: ack-memorydb-controller.v0.0.11
        version: 0.0.11
      - name: ack-memorydb-controller.v0.0.10
        version: 0.0.10
      - name: ack-memorydb-controller.v0.0.9
        version: 0.0.9
      - name: ack-memorydb-controller.v0.0.8
        version: 0.0.8
      - name: ack-memorydb-controller.v0.0.7
        version: 0.0.7
      - name: ack-memorydb-controller.v0.0.6
        version: 0.0.6
      - name: ack-memorydb-controller.v0.0.5
        version: 0.0.5
      - name: ack-memorydb-controller.v0.0.4
        version: 0.0.4
      - name: ack-memorydb-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-memorydb-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: patch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: patch-operator.v0.1.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Patch",
                "metadata": {
                  "name": "test-complex-patch"
                },
                "spec": {
                  "patches": {
                    "test-complex-patch": {
                      "patchTemplate": "metadata:\n  annotations:\n    {{ (index . 1).metadata.name }}: {{ (index . 2).metadata.name }}    \n",
                      "patchType": "application/strategic-merge-patch+json",
                      "sourceObjectRefs": [
                        {
                          "apiVersion": "v1",
                          "kind": "Namespace",
                          "name": "test-patch-operator"
                        },
                        {
                          "apiVersion": "v1",
                          "kind": "ServiceAccount",
                          "name": "default",
                          "namespace": "test-patch-operator"
                        }
                      ],
                      "targetObjectRef": {
                        "apiVersion": "v1",
                        "kind": "ServiceAccount",
                        "name": "test",
                        "namespace": "test-patch-operator"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
          createdAt: "2024-05-10T14:07:19Z"
          description: An operator to apply patches to Kubernetes objects in a declarative
            way.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: patch-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/patch-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Patch
            name: patches.redhatcop.redhat.io
            version: v1alpha1
        description: "The patch operator helps with defining patches in a declarative
          way. This operator has two main features:\n\n1. [ability to patch an object
          at creation time via a mutating webhook](#creation-time-patch-injection)\n2.
          [ability to enforce patches on one or more objects via a controller](#runtime-patch-enforcement)\n
          \n## Creation-time patch injection\n\nWhy apply a patch at creation time
          when you could directly create the correct object? The reason is that sometime
          the correct value depends on configuration set on the specific cluster in
          which the object is being deployed. For example, an ingress/route hostname
          might depend on the specific cluster. Consider the following example based
          on cert-manager:\n \n```yaml\napiVersion: cert-manager.io/v1\nkind: ClusterIssuer\nmetadata:\n
          \ name: letsencrypt-issuer\nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nIn this
          example the fields: `<< access_key >>`, `<< region >>` and `<< hosted_zone_id
          >>` are dependent on the specific region in which the cluster is being deployed
          and in many cases they are discoverable from other configurations already
          present in the cluster. If you want to deploy the above Cluster Issuer object
          with a gitops approach, then there is no easy way to discover those values.
          The solution so far is to manually discover those values and create a different
          gitops configuration for each cluster. But consider if you could look up
          values at deploy time based on the cluster you are deploying to. Here is
          how this object might look:\n\n```yaml\napiVersion: cert-manager.io/v1\nkind:
          ClusterIssuer\nmetadata:\n  name: letsencrypt-issuer\n  namespace: {{ .Release.Namespace
          }}\n  annotations:\n    \"redhat-cop.redhat.io/patch\": |\n      spec:\n
          \       acme:\n        - dns01:\n            route53:\n              accessKeyID:
          {{ (lookup \"v1\" \"Secret\" .metadata.namespace \"cert-manager-dns-credentials\").data.aws_access_key_id
          | b64dec }}\n              secretAccessKeySecretRef:\n                name:
          cert-manager-dns-credentials\n                key: aws_secret_access_key\n
          \             region: {{ (lookup \"config.openshift.io/v1\" \"Infrastructure\"
          \"\" \"cluster\").status.platformStatus.aws.region }}\n              hostedZoneID:
          {{ (lookup \"config.openshift.io/v1\" \"DNS\" \"\" \"cluster\").spec.publicZone.id
          }} \nspec:\n  acme:\n    server: 'https://acme-v02.api.letsencrypt.org/directory'\n
          \   email: {{ .Values.letsencrypt.email }}\n    privateKeySecretRef:\n      name:
          letsencrypt-staging\n    solvers:  \n    - dns01:\n        route53:\n          accessKeyID:
          << access_key >>\n          secretAccessKeySecretRef:\n            name:
          cert-manager-dns-credentials\n            key: aws_secret_access_key\n          region:
          << region >>\n          hostedZoneID: << hosted_zone_id >>\n ```\n\nThe
          annotation specifies a patch that will be applied by a [MutatingWebhook](https://kubernetes.io/docs/reference/access-authn-authz/extensible-admission-controllers/),
          as you can see the three values are now looked up from different configurations
          in the cluster.\n\nTwo annotations influences the behavior of this MutatingWebhook\n
          \n1. \"redhat-cop.redhat.io/patch\" : this is the patch itself. The patch
          is evaluated as a template with the object itself as it's only parameter.
          The template is expressed in golang template notation and supports the same
          functions as helm template including the [lookup](https://helm.sh/docs/chart_template_guide/functions_and_pipelines/#using-the-lookup-function)
          function which plays a major role here. The patch must be expressed in yaml
          for readability. It will be converted to json by the webhook logic.\n2.
          \"redhat-cop.redhat.io/patch-type\" : this is the type of json patch. The
          possible values are: `application/json-patch+json`, `application/merge-patch+json`
          and `application/strategic-merge-patch+json`. If this annotation is omitted
          it defaults to strategic merge.\n\n### Security Considerations\n \nThe lookup
          function, if used by the template, is executed with a client which impersonates
          the user issuing the object creation/update request. This should prevent
          security permission leakage.\n\n### Installing the creation time webhook\n
          \nThe creation time webhook is not installed by the operator. This is because
          there is no way to know which specific object type should be intercepted
          and intercepting all of the types would be too inefficient. It's up to the
          administrator then to install the webhook. Here is some guidance.\n\nIf
          you installed the operator via OLM, use the following webhook template:\n\n```yaml\napiVersion:
          admissionregistration.k8s.io/v1\n kind: MutatingWebhookConfiguration\nmetadata:\n
          \ name: patch-operator-inject\n  annotations:\n    service.beta.openshift.io/inject-cabundle:
          \"true\"\nwebhooks:\n - admissionReviewVersions:\n  - v1\n  clientConfig:\n
          \   service:\n      name: patch-operator-webhook-service\n      namespace:
          patch-operator\n      path: /inject\n  failurePolicy: Fail\n  name: patch-operator-inject.redhatcop.redhat.io\n
          \ rules:\n  - << add your intercepted objects here >>\n  sideEffects: None\n```\n\nIf
          you installed the operator via the Helm chart and are using cert-manager,
          use the following webhook template:\n\n```yaml\napiVersion: admissionregistration.k8s.io/v1\n
          kind: MutatingWebhookConfiguration\nmetadata:\n  name: patch-operator-inject\n
          \ annotations:\n    cert-manager.io/inject-ca-from: '{{ .Release.Namespace
          }}/webhook-server-cert'\n webhooks:\n- admissionReviewVersions:\n  - v1\n
          \ clientConfig:\n    service:\n      name: patch-operator-webhook-service\n
          \     namespace: patch-operator\n      path: /inject\n  failurePolicy: Fail\n
          \ name: patch-operator-inject.redhatcop.redhat.io\n  rules:\n  - << add
          your intercepted objects here >>\n  sideEffects: None\n```  \n\nYou should
          need to enable the webhook only for `CREATE` operations. So for example
          to enable the webhook on configmaps:\n\n```yaml\n  rules:\n  - apiGroups:\n
          \   - \"\"\n    apiVersions:\n    - v1\n    operations:\n    - CREATE\n
          \   resources:\n    - configmaps\n```\n\n## Runtime patch enforcement\n\nThere
          are situations when we need to patch pre-existing objects. Again this is
          a use case that is hard to model with gitops operators which will work only
          on object that they own. Especially with sophisticated Kubernetes distributions,
          it is not uncommon that a Kubernetes instance, at installation time, is
          configured with some default settings. Changing those configurations means
          patching those objects. For example, let's take the case of OpenShift Oauth
          configuration. This object is present by default and it is expected to be
          patched with any newly enabled authentication mechanism. This is how it
          looks like after installation:\n\n```yaml\napiVersion: config.openshift.io/v1\n
          kind: OAuth\nmetadata:\n  name: cluster\n  ownerReferences:\n    - apiVersion:
          config.openshift.io/v1\n      kind: ClusterVersion\n      name: version\n
          \     uid: 9a9d450b-3076-4e30-ac05-a889d6341fc3\n  resourceVersion: '20405124'\n
          spec: []\n```\n\nIf we need to patch it we can use the patch controller
          and the `Patch` object as follows:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: gitlab-ocp-oauth-provider\n  namespace:
          openshift-config\nspec:\n  serviceAccountRef:\n    name: default\n  patches:\n
          \   gitlab-ocp-oauth-provider:\n      targetObjectRef:\n        apiVersion:
          config.openshift.io/v1\n        kind: OAuth\n        name: cluster\n      patchTemplate:
          |\n        spec:\n          identityProviders:\n          - name: my-github
          \n            mappingMethod: claim \n            type: GitHub\n            github:\n
          \             clientID: \"{{ (index . 1).data.client_id | b64dec }}\" \n
          \             clientSecret: \n                name: ocp-github-app-credentials\n
          \             organizations: \n              - my-org\n              teams:
          []            \n      patchType: application/merge-patch+json\n      sourceObjectRefs:\n
          \     - apiVersion: v1\n        kind: Secret\n        name: ocp-github-app-credentials\n
          \       namespace: openshift-config\n```\n\nThis will cause the OAuth object
          to be patched and the patch to be enforced. That means that if anything
          changes on the secret that we use a parameter (which may be rotated) or
          in Oauth object itself, the patch will be reapplied. In this case we are
          adding a gitlab authentication provider.\n\n A `patch` has the following
          fields:\n\n`targetObjectRef` this refers to the object(s) receiving the
          patch. Mutliple object can be selected based on the following rules:\n \n|
          Namespaced Type | Namespace | Name | Selection type |\n| --- | --- | ---
          | --- |\n| yes | null | null | multiple selection across namespaces |\n|
          yes | null | not null | multiple selection across namespaces where the name
          corresponds to the passed name |\n| yes | not null | null | multiple selection
          within a namespace |\n| yes | not null | not nul | single selection |\n|
          no | N/A | null | multiple selection  |\n| no | N/A | not null | single
          selection |\n\nSelection can be further narrowed down by filtering by labels
          and/or annotations using the `labelSelector` and `annotationSelector` fields.
          The patch will be applied to all of the selected instances.\n\n`sourceObjectRefs`
          these are the objects that will be watched and become part of the parameters
          of the patch template. Name and Namespace of sourceRefObjects are interpreted
          as golang templates with the current target instance and the only parameter.
          This allows to select different source object for each target object.\n
          \nSo, for example, with this patch:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: Patch\nmetadata:\n  name: multiple-namespaced-targets-patch\nspec:\n
          \ serviceAccountRef:\n    name: default\n  patches:\n    multiple-namespaced-targets-patch:\n
          \     targetObjectRef:\n        apiVersion: v1\n        kind: ServiceAccount\n
          \       name: deployer\n      patchTemplate: |\n        metadata:\n          annotations:\n
          \           {{ (index . 1).metadata.uid }}: {{ (index . 1) }}\n      patchType:
          application/strategic-merge-patch+json\n      sourceObjectRefs:\n      -
          apiVersion: v1\n        kind: ServiceAccount\n        name: default\n        namespace:
          \"{{ .metadata.namespace }}\"\n        fieldPath: $.metadata.uid\n```\n\nThe
          `deployer` service accounts from all namespaces are selected as target of
          this patch, each patch template will receive a different parameter and that
          is the `default` service account of the same namespace as the namespace
          of the `deployer` service account being processed.\n\n`sourceObjectRefs`
          also have the `fieldPath` field which can contain a jsonpath expression.
          If a value is passed the jsonpath expression will be calculate for the current
          source object and the result will be passed as parameter of the template.\n\n`patchTemplate`
          This is the the template that will be evaluated. The result must be a valid
          patch compatible with the requested type and expressed in yaml for readability.
          The parameters passed to the template are the target object and then the
          all of the source object. So if you want to refer to the target object in
          the template you can use this expression `(index . 0)`. Higher indexes refer
          to the sourceObjectRef array. The template is expressed in golang template
          notation and supports the same functions as helm template.\n\n`patchType`
          is the type of the json patch. The possible values are: `application/json-patch+json`,
          `application/merge-patch+json` and `application/strategic-merge-patch+json`.
          If this annotation is omitted it defaults to strategic merge.\n\n### Patch
          Controller Security Considerations\n\nThe patch enforcement enacted by the
          patch controller is executed with a client which uses the service account
          referenced by the `serviceAccountRef` field. So before a patch object can
          actually work an administrator must have granted the needed permissions
          to a service account in the same namespace. The `serviceAccountRef` will
          default to the `default` service account if not specified.\n\n### Patch
          Controller Performance Considerations\n\nThe patch controller will create
          a controller-manager and per `Patch` object and a reconciler for each of
          the `PatchSpec` defined in the array on patches in the `Patch` object.\nThese
          reconcilers share the same cached client. In order to be able to watch changes
          on target and source objects of a `PatchSpec`, all of the target and source
          object type instances will be cached by the client. This is a normal behavior
          of a controller-manager client, but it implies that if you create patches
          on object types that have many instances in etcd (Secrets, ServiceAccounts,
          Namespaces for example), the patch operator instance will require a significant
          amount of memory. A way to contain this issue is to try to aggregate together
          `PatchSpec` that deal with the same object types. This will cause those
          object type instances to cached only once.  \n"
        displayName: Patch Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - patch
        - enforcing
        - declarative
        links:
        - name: Patch Operator
          url: https://github.com/redhat-cop/patch-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/patch-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/patch-operator@sha256:437f5b7c661d3e9bc2c8ac13b0c8f3565f685b9e5ef52791312bc64d55e2e476
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/patch-operator@sha256:030ade9b9428deca3ac6390618dc4d6e6d728577d1467a22dc4d4b68c9848886
        version: 0.1.11
      entries:
      - name: patch-operator.v0.1.11
        version: 0.1.11
      - name: patch-operator.v0.1.10
        version: 0.1.10
      - name: patch-operator.v0.1.9
        version: 0.1.9
      - name: patch-operator.v0.1.8
        version: 0.1.8
      - name: patch-operator.v0.1.7
        version: 0.1.7
      - name: patch-operator.v0.1.6
        version: 0.1.6
      - name: patch-operator.v0.1.5
        version: 0.1.5
      - name: patch-operator.v0.1.4
        version: 0.1.4
      - name: patch-operator.v0.1.3
        version: 0.1.3
      - name: patch-operator.v0.1.2
        version: 0.1.2
      - name: patch-operator.v0.1.1
        version: 0.1.1
      - name: patch-operator.v0.1.0
        version: 0.1.0
      - name: patch-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: patch-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: redis-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis"},"spec":{"mode":"cluster","size":3,"global":{"image":"quay.io/opstree/redis:v6.2","imagePullPolicy":"IfNotPresent","password":"Opstree@1234","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"master":{"service":{"type":"ClusterIP"},"redisConfig":{}},"slave":{"service":{"type":"ClusterIP"},"redisConfig":{}},"service":{"type":"ClusterIP"},"redisConfig":{},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:1.0","imagePullPolicy":"Always","resources":{"requests":{"cpu":"100m","memory":"128Mi"},"limits":{"cpu":"100m","memory":"128Mi"}}},"storage":{"volumeClaimTemplate":{"spec":{"storageClassName":"buildpiper-storage","accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.4.0
          createdAt: 06-02-2021
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: |+
          A Golang based redis operator that will make/oversee Redis standalone/cluster mode setup on top of the Kubernetes. It can create a redis cluster setup with best practices on Cloud as well as the Bare metal environment. Also, it provides an in-built monitoring capability using redis-exporter.

          ### Supported Features

          Here the features which are supported by this operator:-
          - Redis cluster/standalone mode setup
          - Inbuilt monitoring with prometheus exporter
          - Dynamic storage provisioning with pvc template
          - Resources restrictions with k8s requests and limits
          - Password/Password-less setup
          - Node selector and affinity
          - Priority class to manage setup priority
          - SecurityContext to manipulate kernel parameters
          ### Documentation

          [Documentation](https://docs.opstreelabs.in/redis-operator/)

          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.

          ### Contributing

          You can contribute to this project by:-
          - Raising Bugs and Feature Requests
          - Fixing issue by opening a Pull Request
          - Improving Documentation

          ### License

          Logging Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)


        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://docs.opstreelabs.in/redis-operator/
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.11.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:21f4648e0e71b33ef24b4921ab34b9e2b5e90eac25ade735d51331ac4f7c4cf3
        - quay.io/opstree/redis-operator:v0.4.0
        version: 0.4.0
      entries:
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: preview
    - currentCSV: redis-operator.v0.15.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisCluster","metadata":{"name":"redis-cluster"},"spec":{"clusterSize":3,"clusterVersion":"v7","securityContext":{"runAsUser":1000,"fsGroup":1000},"persistenceEnabled":true,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"Redis","metadata":{"name":"redis-standalone"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"IfNotPresent"},"securityContext":{"runAsUser":1000,"fsGroup":1000}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisSentinel","metadata":{"name":"redis-sentinel"},"spec":{"clusterSize":3,"podSecurityContext":{"runAsUser":1000,"fsGroup":1000},"redisSentinelConfig":{"redisReplicationName":"redis-replication"},"kubernetesConfig":{"image":"quay.io/opstree/redis-sentinel:v7.0.7","imagePullPolicy":"IfNotPresent","resources":{"requests":{"cpu":"101m","memory":"128Mi"},"limits":{"cpu":"101m","memory":"128Mi"}}}}},{"apiVersion":"redis.redis.opstreelabs.in/v1beta1","kind":"RedisReplication","metadata":{"name":"redis-replication"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/redis:v7.0.5","imagePullPolicy":"IfNotPresent"},"storage":{"volumeClaimTemplate":{"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"1Gi"}}}}},"redisExporter":{"enabled":true,"image":"quay.io/opstree/redis-exporter:v1.44.0","imagePullPolicy":"Always"},"podSecurityContext":{"runAsUser":1000,"fsGroup":1000}}}]'
          capabilities: Seamless Upgrades
          categories: Database, Storage
          certified: "false"
          containerImage: quay.io/opstree/redis-operator:v0.15.1
          createdAt: 24-09-2023
          description: A Golang based redis operator that will make/oversee Redis
            standalone/cluster mode setup on top of the Kubernetes.
          repository: https://github.com/ot-container-kit/redis-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Redis Cluster
            displayName: RedisCluster
            kind: RedisCluster
            name: redisclusters.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Replication
            displayName: RedisReplication
            kind: RedisReplication
            name: redisreplications.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis Sentinel
            displayName: RedisSentinel
            kind: RedisSentinel
            name: redissentinels.redis.redis.opstreelabs.in
            version: v1beta1
          - description: Redis
            displayName: Redis
            kind: Redis
            name: redis.redis.redis.opstreelabs.in
            version: v1beta1
        description: "A Golang based redis operator that will make/oversee Redis\nstandalone/cluster
          mode setup on top of the Kubernetes. It can create a\nredis cluster setup
          with best practices on Cloud as well as the Bare metal\nenvironment. Also,
          it provides an in-built monitoring capability using\nredis-exporter.\n\nOrganizations
          that are using Redis Operator to manage their redis workload \ncan be found
          here. If your organization is also using Redis Operator, please \nfree to
          add by creating a PR https://github.com/OT-CONTAINER-KIT/redis-operator/pulls.\n\nThis
          operator only supports versions of redis =>6.\n\n### Supported Features\n\nHere
          the features which are supported by this operator:-\n- Redis cluster and
          standalone mode setup\n- Redis cluster failover and recovery\n- Redis leader
          and follower setup with Sentinel for High Availability\n- Inbuilt monitoring
          with redis exporter\n- Password and password-less setup of redis\n- TLS
          support for additional security layer\n- Ipv4 and Ipv6 support for redis
          setup\n- External configuration support for leader and follower\n- Detailed
          monitoring grafana dashboard\n\n### Documentation\n\n[Documentation](https://ot-redis-operator.netlify.app/)\n\n###
          Getting Help\nThis project is managed by OpsTree Solutions. If you have
          any queries or\nsuggestions, mail us at opensource@opstree.com.\n\n### Contributing\n\nYou
          can contribute to this project by:-\n- Raising Bugs and Feature Requests\n-
          Fixing issue by opening a Pull Request\n- Improving Documentation\n\n###
          License\n\nLogging Operator is licensed under [Apache License, Version\n2.0](https://github.com/OT-CONTAINER-KIT/redis-operator/blob/master/LICENSE)\n"
        displayName: Redis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - Redis
        - Opstree
        - CRD
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-redis-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: beta
        minKubeVersion: 1.20.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/redis-operator@sha256:2ed6bfdf37e36c63cde3633443553dee410201200ad351e6d91a5651f8e20ecc
        - quay.io/opstree/redis-operator:v0.15.1
        version: 0.15.1
      entries:
      - name: redis-operator.v0.15.1
        version: 0.15.1
      - name: redis-operator.v0.15.0
        version: 0.15.0
      - name: redis-operator.v0.14.0
        version: 0.14.0
      - name: redis-operator.v0.13.0
        version: 0.13.0
      - name: redis-operator.v0.8.0
        version: 0.8.0
      - name: redis-operator.v0.6.0
        version: 0.6.0
      - name: redis-operator.v0.5.0
        version: 0.5.0
      - name: redis-operator.v0.4.0
        version: 0.4.0
      - name: redis-operator.v0.3.0
        version: 0.3.0
      name: stable
    defaultChannel: stable
    packageName: redis-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: integrity-shield-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: integrity-shield-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "provenanece": true,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  audit:\n    inform: true\n  admissionControl:\n    enforce: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.0
          createdAt: "2021-10-14T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:11edce56874ff1ee666752ec5fa521f3f6ae78c52cf1f289ceb97a86e6c083aa
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.0
      entries:
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.0
    - currentCSV: integrity-shield-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/open-cluster-management/integrity-shield-observer",
                    "imagePullPolicy": "Always",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  resp.status_code == 200\n  result := json.unmarshal(resp.raw_body)\n  result.allow == false\n  not is_detect_mode\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"tls_insecure_skip_verify\": true\n  })\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: detect\n  admissionOnly: false\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/open-cluster-management/integrity-shield-api",
                    "imagePullPolicy": "Always",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "512Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/open-cluster-management/integrity-shield-operator:0.3.1
          createdAt: "2021-10-22T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.

          [K8s Integrity Shield](https://github.com/open-cluster-management/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.

          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Detect (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster

          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).

          ## Preparations before installation

          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).


          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.

          If you want to change the settings such as default run mode (detection/enforcement) or audit interval,  please check [here](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).

          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-7b7f768bf7-fhrpg                           1/1     Running   0          20s
          integrity-shield-observer-5bc66f75f7-tn8fw                      1/1     Running   0          25s
          integrity-shield-operator-controller-manager-65b7fb58f7-j25zd   2/2     Running   0          3h5m
          ```

          After installation, you can protect Kubernetes resources by following this [document](https://github.com/open-cluster-management/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).

          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:

          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)

          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/open-cluster-management/integrity-shield
        - name: Operator Documentation
          url: https://github.com/open-cluster-management/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/open-cluster-management/integrity-shield-operator:0.3.1
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:5e4c319ce871a0dcbab7fff23325e6233c9dbcb2592fdeb1a9ba7c57ac2a63b2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        version: 0.3.1
      entries:
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.1
    - currentCSV: integrity-shield-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": true,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "resultDetailConfigKey": "config.yaml",
                    "resultDetailConfigName": "verify-resource-result",
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.2
          createdAt: "2022-01-25T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:d86bb21fdeec35b1fc9789023a558750da76e036e0790dcfce97dcd137dcb6fb
        - quay.io/stolostron/integrity-shield-operator:0.3.2
        version: 0.3.2
      entries:
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.2
    - currentCSV: integrity-shield-operator.v0.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apis.integrityshield.io/v1",
                "kind": "IntegrityShield",
                "metadata": {
                  "finalizers": [
                    "cleanup.finalizers.integrityshield.io"
                  ],
                  "name": "integrity-shield"
                },
                "spec": {
                  "affinity": {},
                  "labels": {
                    "app": "ishield-api"
                  },
                  "observer": {
                    "enabled": true,
                    "exportDetailResult": false,
                    "image": "quay.io/stolostron/integrity-shield-observer",
                    "imagePullPolicy": "IfNotPresent",
                    "interval": "5",
                    "logLevel": "info",
                    "name": "integrity-shield-observer",
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-observer"
                    }
                  },
                  "rego": "package integrityshieldcheck\nviolation[{\"msg\": msg}] {\n  not is_allowed_kind\n  not is_excluded\n  is_create_or_update\n  ishield_input := {\"parameters\":input.parameters, \"request\":input.review}\n  reqdata := json.marshal(ishield_input)\n  url := \"https://integrity-shield-api.REPLACE_WITH_SERVER_NAMESPSCE.svc:8123/api/request\"\n  resp := http_post(url, reqdata)\n  should_be_denied(resp)\n  not is_detect_mode\n  result := get_result(resp)\n  msg := sprintf(\"denied; %v\", [result])\n}\n\nhttp_post(url, postdata) = resp {\n  resp := http.send({\n    \"url\": url,\n    \"method\": \"POST\",\n    \"headers\": {\n      \"Accept\": \"application/json\",\n      \"Content-type\": \"application/json\",\n    },\n    \"raw_body\": postdata,\n    \"timeout\": \"2500ms\",\n    \"raise_error\": false,\n    \"tls_ca_cert\": REPLACE_WITH_CA_CERT\n  })\n}\n\n# get decision result from API response\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  not is_null(raw_body)\n  result := json.unmarshal(raw_body)\n}\n# if error, return it as result\nget_result(resp) = result {\n  raw_body := object.get(resp, \"raw_body\", null)\n  is_null(raw_body)\n  error_msg := object.get(resp, \"error\", \"\")\n  result := {\"http_error\":error_msg}\n}\n\n# if API response is null, deny\nshould_be_denied(resp) {\n  is_null(resp)\n}\n\n# if API response has status_code other than 200, deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  resp.status_code != 200\n}\n\n# if decision result is \"allow == false\" or not set, then deny\nshould_be_denied(resp) {\n  not is_null(resp)\n  result := get_result(resp)\n  allow := object.get(result, \"allow\", false)\n  allow == false\n}\n\n# request check\nis_create_or_update { is_create }\nis_create_or_update { is_update }\nis_create { input.review.operation == \"CREATE\" }\nis_update { input.review.operation == \"UPDATE\" }\n\n# shield config: allow\nis_allowed_kind { skip_kinds[_].kind == input.review.kind.kind }\n# shield config: inScopeNamespaceSelector\nis_excluded { exclude_namespaces[_] = input.review.namespace}\n\n# detect mode\nis_detect_mode { enforce_mode == \"detect\" }\n\n################### \n# Default setting #\n###################\n\n# Mode whether to deny a invalid request [enforce/detect]\nenforce_mode = \"enforce\"\n\n# kinds to be skipped\nskip_kinds = [\n          {\n            \"kind\": \"Event\"\n          },\n          {\n            \"kind\": \"Lease\"\n          },\n          {\n            \"kind\": \"Endpoints\"\n          },\n          {\n            \"kind\": \"TokenReview\"\n          },\n          {\n            \"kind\": \"SubjectAccessReview\"\n          },\n          {\n            \"kind\": \"SelfSubjectAccessReview\"\n          }\n        ]\n\n# exclude namespaces\nexclude_namespaces = [\n                      \"kube-node-lease\",\n                      \"kube-public\",\n                      \"kube-storage-version-migrator-operator\",\n                      \"kube-system\",\n                      \"open-cluster-management\",\n                      \"open-cluster-management-hub\",\n                      \"open-cluster-management-agent\",\n                      \"open-cluster-management-agent-addon\",\n                      \"openshift\",\n                      \"openshift-apiserver\",\n                      \"openshift-apiserver-operator\",\n                      \"openshift-authentication\",\n                      \"openshift-authentication-operator\",\n                      \"openshift-cloud-credential-operator\",\n                      \"openshift-cluster-csi-drivers\",\n                      \"openshift-cluster-machine-approver\",\n                      \"openshift-cluster-node-tuning-operator\",\n                      \"openshift-cluster-samples-operator\",\n                      \"openshift-cluster-storage-operator\",\n                      \"openshift-cluster-version\",\n                      \"openshift-compliance\",\n                      \"openshift-config\",\n                      \"openshift-config-managed\",\n                      \"openshift-config-operator\",\n                      \"openshift-console\",\n                      \"openshift-console-operator\",\n                      \"openshift-console-user-settings\",\n                      \"openshift-controller-manager\",\n                      \"openshift-controller-manager-operator\",\n                      \"openshift-dns\",\n                      \"openshift-dns-operator\",\n                      \"openshift-etcd\",\n                      \"openshift-etcd-operator\",\n                      \"openshift-gatekeeper-system\",\n                      \"openshift-image-registry\",\n                      \"openshift-infra\",\n                      \"openshift-ingress\",\n                      \"openshift-ingress-canary\",\n                      \"openshift-ingress-operator\",\n                      \"openshift-insights\",\n                      \"openshift-kni-infra\",\n                      \"openshift-kube-apiserver\",\n                      \"openshift-kube-apiserver-operator\",\n                      \"openshift-kube-controller-manager\",\n                      \"openshift-kube-controller-manager-operator\",\n                      \"openshift-kube-scheduler\",\n                      \"openshift-kube-scheduler-operator\",\n                      \"openshift-kube-storage-version-migrator\",\n                      \"openshift-kube-storage-version-migrator-operator\",\n                      \"openshift-kubevirt-infra\",\n                      \"openshift-machine-api\",\n                      \"openshift-machine-config-operator\",\n                      \"openshift-marketplace\",\n                      \"openshift-monitoring\",\n                      \"openshift-multus\",\n                      \"openshift-network-diagnostics\",\n                      \"openshift-network-operator\",\n                      \"openshift-node\",\n                      \"openshift-oauth-apiserver\",\n                      \"openshift-openstack-infra\",\n                      \"openshift-operators\",\n                      \"openshift-operator-lifecycle-manager\",\n                      \"openshift-ovirt-infra\",\n                      \"openshift-ovn-kubernetes\",\n                      \"openshift-sdn\",\n                      \"openshift-service-ca\",\n                      \"openshift-service-ca-operator\",\n                      \"openshift-user-workload-monitoring\",\n                      \"openshift-vsphere-infra\"\n                  ]\n",
                  "reporter": {
                    "image": "quay.io/stolostron/integrity-shield-reporter",
                    "imagePullPolicy": "IfNotPresent",
                    "intervalSeconds": "10",
                    "name": "integrity-shield-reporter",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "requestHandlerConfig": "defaultConstraintAction:\n  mode: inform\nsideEffect: \n  createDenyEvent: true\nlog:\n  level: info\n  manifestSigstoreLogLevel: info\n  format: json\ndecisionReporterConfig:\n  enabled: true\nrequestFilterProfile: \n  skipObjects:\n  - kind: ConfigMap\n    name: kube-root-ca.crt\n  - kind: ConfigMap\n    name: openshift-service-ca.crt\n  ignoreFields:\n  - fields:\n    - spec.host\n    objects:\n    - kind: Route\n  - fields:\n    - metadata.namespace\n    objects:\n    - kind: ClusterServiceVersion\n  - fields:\n    - metadata.labels.app.kubernetes.io/instance\n    - metadata.managedFields.*\n    - metadata.resourceVersion\n    - metadata.selfLink\n    - metadata.annotations.control-plane.alpha.kubernetes.io/leader\n    - metadata.annotations.kubectl.kubernetes.io/last-applied-configuration\n    - metadata.finalizers*\n    - metadata.annotations.namespace\n    - metadata.annotations.deprecated.daemonset.template.generation\n    - metadata.creationTimestamp\n    - metadata.uid\n    - metadata.generation\n    - status\n    - metadata.annotations.deployment.kubernetes.io/revision\n    - metadata.annotations.cosign.sigstore.dev/imageRef\n    - metadata.annotations.cosign.sigstore.dev/bundle\n    - metadata.annotations.cosign.sigstore.dev/message\n    - metadata.annotations.cosign.sigstore.dev/certificate\n    - metadata.annotations.cosign.sigstore.dev/signature\n    objects:\n    - name: '*'\n  - fields:\n    - secrets.*.name\n    - imagePullSecrets.*.name\n    objects:\n    - kind: ServiceAccount\n  - fields:\n    - spec.ports.*.nodePort\n    - spec.clusterIP\n    - spec.clusterIPs.0\n    objects:\n    - kind: Service\n  - fields:\n    - metadata.labels.olm.api.*\n    - metadata.labels.operators.coreos.com/*\n    - metadata.annotations.*\n    - spec.install.spec.deployments.*.spec.template.spec.containers.*.resources.limits.cpu\n    - spec.cleanup.enabled\n    objects:\n    - kind: ClusterServiceVersion\n  skipUsers:\n  - users: \n    - system:admin\n    - system:apiserver\n    - system:kube-scheduler\n    - system:kube-controller-manager\n    - system:serviceaccount:kube-system:generic-garbage-collector\n    - system:serviceaccount:kube-system:attachdetach-controller\n    - system:serviceaccount:kube-system:certificate-controller\n    - system:serviceaccount:kube-system:clusterrole-aggregation-controller\n    - system:serviceaccount:kube-system:cronjob-controller\n    - system:serviceaccount:kube-system:disruption-controller\n    - system:serviceaccount:kube-system:endpoint-controller\n    - system:serviceaccount:kube-system:horizontal-pod-autoscaler\n    - system:serviceaccount:kube-system:ibm-file-plugin\n    - system:serviceaccount:kube-system:ibm-keepalived-watcher\n    - system:serviceaccount:kube-system:ibmcloud-block-storage-plugin\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:namespace-controller\n    - system:serviceaccount:kube-system:node-controller\n    - system:serviceaccount:kube-system:job-controller\n    - system:serviceaccount:kube-system:pod-garbage-collector\n    - system:serviceaccount:kube-system:pv-protection-controller\n    - system:serviceaccount:kube-system:pvc-protection-controller\n    - system:serviceaccount:kube-system:replication-controller\n    - system:serviceaccount:kube-system:resourcequota-controller\n    - system:serviceaccount:kube-system:service-account-controller\n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ControllerRevision\n    - kind: Pod\n    users: \n    - system:serviceaccount:kube-system:daemon-set-controller\n  - objects: \n    - kind: Pod\n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:persistent-volume-binder\n  - objects: \n    - kind: ReplicaSet\n    users: \n    - system:serviceaccount:kube-system:deployment-controller\n  - objects: \n    - kind: Pod\n    users:  \n    - system:serviceaccount:kube-system:replicaset-controller\n  - objects: \n    - kind: PersistentVolumeClaim\n    users: \n    - system:serviceaccount:kube-system:statefulset-controller\n  - objects: \n    - kind: ServiceAccount\n    users: \n    - system:kube-controller-manager\n  - objects: \n    - kind: EndpointSlice\n    users: \n    - system:serviceaccount:kube-system:endpointslice-controller\n  - objects: \n    - kind: Secret\n    users: \n    - system:kube-controller-manager\n  - users: \n    - system:serviceaccount:openshift-marketplace:marketplace-operator\n    - system:serviceaccount:openshift-monitoring:cluster-monitoring-operator\n    - system:serviceaccount:openshift-network-operator:default\n    - system:serviceaccount:openshift-monitoring:prometheus-operator\n    - system:serviceaccount:openshift-cloud-credential-operator:default\n    - system:serviceaccount:openshift-machine-config-operator:default\n    - system:serviceaccount:openshift-infra:namespace-security-allocation-controller\n    - system:serviceaccount:openshift-cluster-version:default\n    - system:serviceaccount:openshift-authentication-operator:authentication-operator\n    - system:serviceaccount:openshift-apiserver-operator:openshift-apiserver-operator\n    - system:serviceaccount:openshift-kube-scheduler-operator:openshift-kube-scheduler-operator\n    - system:serviceaccount:openshift-kube-controller-manager-operator:kube-controller-manager-operator\n    - system:serviceaccount:openshift-controller-manager:openshift-controller-manager-sa\n    - system:serviceaccount:openshift-controller-manager-operator:openshift-controller-manager-operator\n    - system:serviceaccount:openshift-kube-apiserver-operator:kube-apiserver-operator\n    - system:serviceaccount:openshift-sdn:sdn-controller\n    - system:serviceaccount:openshift-machine-api:cluster-autoscaler-operator\n    - system:serviceaccount:openshift-machine-api:machine-api-operator\n    - system:serviceaccount:openshift-machine-config-operator:machine-config-controller\n    - system:serviceaccount:openshift-machine-api:machine-api-controllers\n    - system:serviceaccount:openshift-cluster-storage-operator:csi-snapshot-controller-operator\n    - system:serviceaccount:openshift-kube-controller-manager:localhost-recovery-client\n    - system:serviceaccount:openshift-kube-storage-version-migrator-operator:kube-storage-version-migrator-operator\n    - system:serviceaccount:openshift-etcd-operator:etcd-operator\n    - system:serviceaccount:openshift-service-ca:service-ca\n    - system:serviceaccount:openshift-config-operator:openshift-config-operator\n    - system:serviceaccount:openshift-kube-apiserver:localhost-recovery-client\n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects:\n    - namespace: openshift-service-ca, openshift-network-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:configmap-cabundle-injector-sa\n  - objects: \n    - namespace: openshift-service-ca-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca-operator:service-ca-operator\n  - objects: \n    - namespace: openshift-service-catalog-controller-manager-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-controller-manager-operator:openshift-service-catalog-controller-manager-operator\n  - objects: \n    - namespace: openshift-console-operator, openshift-console\n    users: \n    - system:serviceaccount:openshift-console-operator:console-operator\n  - objects: \n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:apiservice-cabundle-injector-sa\n    - namespace: openshift-service-ca\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-ca:service-serving-cert-signer-sa\n  - objects: \n    - namespace: openshift-service-catalog-apiserver-operator\n      kind: ConfigMap\n    users: \n    - system:serviceaccount:openshift-service-catalog-apiserver-operator:openshift-service-catalog-apiserver-operator\n  - objects: \n    - namespace: openshift-operator-lifecycle-manager\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - namespace: openshift-cluster-node-tuning-operator\n      kind: ConfigMap,DaemonSet\n    users: \n    - system:serviceaccount:openshift-cluster-node-tuning-operator:cluster-node-tuning-operator\n  - objects: \n    - namespace: openshift\n      kind: Secret\n    users: \n    - system:serviceaccount:openshift-cluster-samples-operator:cluster-samples-operator\n  - objects: \n    - namespace: openshift-ingress\n      kind: Deployment\n    users: \n    - system:serviceaccount:openshift-ingress-operator:ingress-operator\n  - objects: \n    - kind: ServiceAccount, Secret\n    users: \n    - system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller\n  - objects: \n    - namespace: openshift-marketplace\n      kind: Pod\n    users: \n    - system:node:*\n  - objects: \n    - kind: ServiceAccount, InstallPlan, OperatorGroup, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n  - objects: \n    - kind: InstallPlan, Role, RoleBinding, Deployment\n    users: \n    - system:serviceaccount:openshift-operator-lifecycle-manager:olm-operator-serviceaccount\n",
                  "requestHandlerConfigKey": "config.yaml",
                  "requestHandlerConfigName": "request-handler-config",
                  "security": {
                    "observerRole": "integrity-shield-observer-role",
                    "observerRoleBinding": "integrity-shield-observer-rolebinding",
                    "observerServiceAccountName": "integrity-shield-observer-sa",
                    "podSecurityPolicyName": "integrity-shield-psp",
                    "role": "integrity-shield-role",
                    "roleBinding": "integrity-shield-rolebinding",
                    "securityContext": {},
                    "serviceAccountName": "integrity-shield-sa"
                  },
                  "shieldApi": {
                    "image": "quay.io/stolostron/integrity-shield-api",
                    "imagePullPolicy": "IfNotPresent",
                    "name": "integrity-shield-api",
                    "port": 8080,
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "256Mi"
                      }
                    },
                    "selector": {
                      "app": "integrity-shield-api"
                    }
                  },
                  "shieldApiServiceName": "integrity-shield-api",
                  "shieldApiServicePort": 8123,
                  "shieldApiTlsSecretName": "integrity-shield-api-tls",
                  "useGatekeeper": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: quay.io/stolostron/integrity-shield-operator:0.3.3
          createdAt: "2022-03-18T18:00:00Z"
          description: K8s Integrity Shield is a tool for built-in preventive integrity
            control for regulated cloud workloads.
          operators.operatorframework.io/builder: operator-sdk-v1.15.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/integrity-shield
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IntegrityShield is the Schema for the integrityshields API
            displayName: Integrity Shield
            kind: IntegrityShield
            name: integrityshields.apis.integrityshield.io
            version: v1
        description: |-
          Kubernetes resources are represented as YAML files, which are applied to clusters when you create and update the resource. The YAML content is designed carefully to achieve the application desired state and should not be tampered with. If the YAML content is modified maliciously or accidentally, and applied to a cluster without notice, the cluster moves to an unexpected state.
          [K8s Integrity Shield](https://github.com/stolostron/integrity-shield) provides preventive control for enforcing signature verification for any requests to create or update resources. This operator supports the installation and management of K8s Integrity Shield on cluster.
          Two modes are selectively enabled on your cluster.
          - Enforce (Admission Control): Block to deploy unauthorized Kubernetes resources. K8s Integrity Shield works with [OPA/Gatekeeper](https://github.com/open-policy-agent/gatekeeper) to enable admission control based on signature verification for Kubernetes resources.
          - Inform (Continuous Monitoring): monitor Kubernetes resource integrity and report if unauthorized Kubernetes resources are deployed on cluster
          X509, PGP and Sigstore signing are supported for singing Kubernetes manifest YAML. K8s Integrity Shield supports Sigstore signing by using [k8s-manifest-sigstore](https://github.com/sigstore/k8s-manifest-sigstore).
          ## Preparations before installation
          OPA/Gatekeeper should be deployed before installing K8s Integrity Shield.
          The installation instructions to deploy OPA/Gatekeeper components is [here](https://open-policy-agent.github.io/gatekeeper/website/docs/install/).
          ## Installation
          Install K8s Integrity Shield Operator by following the instruction after clicking Install button at the top right. Then you can create the operator Custom Resource `IntegrityShield` to complete installation.
          If you want to change the settings such as default run mode (inform/enforce) or audit interval,  please check [here](https://github.com/stolostron/integrity-shield/blob/master/docs/README_ISHIELD_OPERATOR_CR.md).
          To verify that installation was completed successfully,
          run the following command.
          The following three pods will be installed with default CR.
          ```
          $ kubectl get pod -n integrity-shield-operator-system
          NAME                                                            READY   STATUS    RESTARTS   AGE
          integrity-shield-api-5799f69896-zbcjs                           2/2     Running   0          88s
          integrity-shield-observer-757f58f7d7-7pf5h                      1/1     Running   0          55s
          integrity-shield-operator-controller-manager-6f94d7d684-dg5bt   2/2     Running   0          4m27s
          ```
          After installation, you can protect Kubernetes resources by following this [document](https://github.com/stolostron/integrity-shield/blob/master/docs/README_GETTING-STARTED-TUTORIAL.md).
          ## Supported Versions
          ### Platform
          K8s Integrity Shield can be deployed with the operator. We have verified the feasibility on the following platforms:
          - [RedHat OpenShift 4.7.1 and 4.9.0](https://www.openshift.com)
          - [Kuberenetes v1.19.7 and v1.21.1](https://kubernetes.io)
          ### OPA/Gatekeeper
          - [gatekeeper-operator v0.2.0](https://github.com/open-policy-agent/gatekeeper)
          - [gatekeeper v3.5.2 and v3.6.0](https://github.com/open-policy-agent/gatekeeper)
        displayName: K8s Integrity Shield
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - integrity
        links:
        - name: Integrity Shield Operator
          url: https://github.com/stolostron/integrity-shield
        - name: Operator Documentation
          url: https://github.com/stolostron/integrity-shield/tree/master/docs
        maintainers:
        - email: acm-contact@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/integrity-shield-operator@sha256:7d94521a5cbdd34f8307653a3e1b596ec86415b2165fabdbabc8ed8bed35473d
        - quay.io/stolostron/integrity-shield-operator:0.3.3
        version: 0.3.3
      entries:
      - name: integrity-shield-operator.v0.3.3
        version: 0.3.3
      - name: integrity-shield-operator.v0.3.2
        version: 0.3.2
      - name: integrity-shield-operator.v0.3.1
        version: 0.3.1
      - name: integrity-shield-operator.v0.3.0
        version: 0.3.0
      name: alpha-0.3.3
    defaultChannel: alpha-0.3.3
    packageName: integrity-shield-operator
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
          createdAt: "2023-10-30 13:59:28"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.1.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:394d122efcf0da16bb68bfef1292395cda1bdb21909ae55f8fcc54a16814f1cf
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:420a4d4a8610a3694fd3ef9809c55a4f140ec8cea7f09219647f82a84d1f411b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: machine-deletion-remediation.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
          createdAt: "2024-08-11 06:47:20"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:62f93c5989d1d8217a3bac076addac05c0906a1a1f1eeaacb87bcf256812fdf9
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:1445e6df8f49672aa729806f0e257043ecc22e19e666bb5a799286c6b878b1dd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f4381aaf002b761a2ba614332a2af973f4de5f6a4c84dd96b8808725e561fa89
        version: 0.3.1
      entries:
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
          createdAt: "2023-04-24 08:22:17"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:ea8c060444aa8e89dfbd90d61d6ec3900ab4d6634f6aa8067face1e6974b9065
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel8-operator@sha256:7e1cad02f1e74a1355c168fc53925bd0fbcc55f2535a46f734c8de082eb49400
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-29 09:35:10"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.3.1 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.6/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/machine-deletion-remediation-operator-bundle@sha256:f1dd190022941300f1e886776eef6d9329a408cd2a7a02c24a0fc45aec9324db
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/machine-deletion-remediation-rhel9-operator@sha256:dbe9095d494ebc7b56982cb1f74329d7a7fd0cbb7544e9f90207eb35de7997dc
        version: 0.4.1
      entries:
      - name: machine-deletion-remediation.v0.4.1
        version: 0.4.1
      - name: machine-deletion-remediation.v0.4.0
        version: 0.4.0
      - name: machine-deletion-remediation.v0.3.1
        version: 0.3.1
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-notifications
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-notifications.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Notifications",
                "metadata": {
                  "name": "notifications-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp",
                    "test": "wap-rhel8-toolbox:8.7.0-3747-36-ocp",
                    "testRepository": "prod.harbor.keysight.digital/loadcore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:47Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Notifications
            name: notifications.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Notifications Operator
        displayName: Keysight WAP Notifications Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Notifications
          url: https://keysight-wap-notifications.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-notifications@sha256:50c24180b3c16cf3eb82f6f4eeb61b91a62a198c122f7e59aa680a25bc488c2d
        - registry.connect.redhat.com/loadcore/keysight-wap-notifications-bundle@sha256:9c2ebb98476bf3374b71a838c394017e7d31aaa045bd52b3a66924b6f6b7b067
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-notifications.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-notifications
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-licensing-operator
      app.kubernetes.io/managed-by: ibm-licensing-operator
      app.kubernetes.io/name: ibm-licensing
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      intent: projected
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-licensing-operator-app
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-licensing-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicensing",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-licensing-operator",
                    "app.kubernetes.io/managed-by": "ibm-licensing-operator",
                    "app.kubernetes.io/name": "ibm-licensing"
                  },
                  "name": "instance"
                },
                "spec": {
                  "version": "4.2.14",
                  "apiSecretToken": "ibm-licensing-token",
                  "datasource": "datacollector",
                  "httpsEnable": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
          createdAt: "2025-02-25T12:16:11Z"
          description: The IBM Licensing Operator provides a Kubernetes CRD-Based
            API to monitor the license usage of products.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-licensing-operator-app
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-licensing
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["ibmlicensingmetadatas.operator.ibm.com",
            "ibmlicensingdefinitions.operator.ibm.com", "ibmlicensingquerysources.operator.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'IBMLicensing custom resource is used to create an instance
              of the License Service, used to collect information about license usage
              of IBM containerized products and IBM Cloud Paks per cluster. You can
              retrieve license usage data through a dedicated API call and generate
              an audit snapshot on demand. Documentation: For additional details regarding
              install parameters check: https://ibm.biz/icpfs39install. License: Please
              refer to the IBM Terms website (ibm.biz/lsvc-lic) to find the license
              terms for the particular IBM product for which you are deploying this
              component.'
            displayName: IBM License Service
            kind: IBMLicensing
            name: ibmlicensings.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingDefinition
            name: ibmlicensingdefinitions.operator.ibm.com
            version: v1
          - description: 'IBMLicensingMetadata is the schema for IBM License Service.
              Thanks to IBMLicensingMetadata, you can track the license usage of Virtual
              Processor Core (VPC) metric by Pods that are managed by automation,
              for which licensing annotations are not defined at the time of deployment,
              such as the open source OpenLiberty. / For more information, see documentation:
              https://ibm.biz/icpfs39install. By installing the operator for IBM License
              Service, you accept the license terms: https://ibm.biz/icpfs39license.'
            displayName: IBMLicensing Metadata
            kind: IBMLicensingMetadata
            name: ibmlicensingmetadatas.operator.ibm.com
            version: v1alpha1
          - kind: IBMLicensingQuerySource
            name: ibmlicensingquerysources.operator.ibm.com
            version: v1
        description: |-
          **Important:**
          - If you are using the IBM Licensing Operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For the link to your IBM Cloud Pak documentation, see [IBM Cloud Paks that use Common Services](https://ibm.biz/BdyGwb).
          - If you are using a stand-alone IBM Container Software, you can use the IBM Licensing Operator directly. For more information, see [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).

          **IBM Licensing Operator overview**

          IBM Licensing Operator installs License Service. You can use License Service to collect information about license usage of IBM Containerized products and IBM Cloud Paks per cluster. You can retrieve license usage data through a dedicated API call and generate an audit snapshot on demand.

          **Supported platforms**

           Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64, Linux on Power (ppc64le), Linux on IBM Z and LinuxONE.

          **Prerequisites**

           Prerequisites depend on the integration of the License Service with an IBM Cloud Pak or IBM Containerized Software. For more information, see the applicable IBM Cloud Pak documentation or [ibm-licensing-operator for stand-alone IBM Containerized Software](https://ibm.biz/BdyGwh).
        displayName: IBM Licensing
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBMLicensing
        - IBM
        - Cloud
        - License
        - Service
        - License Service
        links:
        - name: GitHub
          url: https://github.com/IBM/ibm-licensing-operator
        maintainers:
        - email: talk2sam@us.ibm.com
          name: talk2sam
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.connect.redhat.com/cpopen/ibm-licensing-operator@sha256:a9f68d204b6acb50ac5023e5f2314ad0d06e06483f87290381d1e92476a79579
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        version: 4.2.14
      entries:
      - name: ibm-licensing-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-licensing-operator-app
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Dell Technologies
      provider-url: https://github.com/dell/csm-operator
    name: dell-csm-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dell-csm-operator-certified.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "authorization",
                  "namespace": "authorization"
                },
                "spec": {
                  "modules": [
                    {
                      "components": [
                        {
                          "enabled": true,
                          "name": "nginx"
                        },
                        {
                          "enabled": true,
                          "name": "cert-manager"
                        },
                        {
                          "authorizationController": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231",
                          "authorizationControllerReplicas": 1,
                          "certificate": "",
                          "controllerReconcileInterval": "5m",
                          "enabled": true,
                          "hostname": "csm-authorization.com",
                          "leaderElection": true,
                          "name": "proxy-server",
                          "opa": "docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095",
                          "opaKubeMgmt": "docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517",
                          "openTelemetryCollectorAddress": "",
                          "privateKey": "",
                          "proxyServerIngress": [
                            {
                              "annotations": {},
                              "hosts": [],
                              "ingressClassName": "nginx"
                            }
                          ],
                          "proxyService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682",
                          "proxyServiceReplicas": 1,
                          "roleService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c",
                          "roleServiceReplicas": 1,
                          "storageService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5",
                          "storageServiceReplicas": 1,
                          "tenantService": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358",
                          "tenantServiceReplicas": 1
                        },
                        {
                          "commander": "docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a",
                          "name": "redis",
                          "redis": "docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f",
                          "redisCommander": "rediscommander",
                          "redisName": "redis-csm",
                          "redisReplicas": 5,
                          "redisSecretProviderClass": [
                            {
                              "redisPasswordKey": "",
                              "redisSecretName": "",
                              "redisUsernameKey": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "sentinel": "sentinel"
                        },
                        {
                          "configSecretProviderClass": [
                            {
                              "configSecretName": "",
                              "secretProviderClassName": ""
                            }
                          ],
                          "name": "config"
                        },
                        {
                          "name": "storage-system-credentials",
                          "secretProviderClasses": {
                            "vault": [
                              "secret-provider-class-1",
                              "secret-provider-class-2"
                            ]
                          }
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": true,
                      "forceRemoveModule": true,
                      "name": "authorization-proxy-server"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "isilon",
                  "namespace": "isilon"
                },
                "spec": {
                  "driver": {
                    "authSecret": "isilon-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VERBOSE",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_ISI_PORT",
                          "value": "8080"
                        },
                        {
                          "name": "X_CSI_ISI_PATH",
                          "value": "/ifs/data/csi"
                        },
                        {
                          "name": "X_CSI_ISI_NO_PROBE_ON_START",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_AUTOPROBE",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_AUTH_TYPE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_CUSTOM_TOPOLOGY_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOISILON_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "AZ_RECONCILE_INTERVAL",
                          "value": "1h"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_ISI_QUOTA_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_ISI_ACCESS_ZONE",
                          "value": "System"
                        },
                        {
                          "name": "X_CSI_ISI_VOLUME_PATH_PERMISSIONS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ISI_IGNORE_UNRESOLVABLE_HOSTS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "isilon",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_PATH_LIMIT",
                          "value": "192"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerscale"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSCALE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSCALE_CLUSTER_PERFORMANCE_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSCALE_QUOTA_CAPACITY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "ISICLIENT_INSECURE",
                              "value": "true"
                            },
                            {
                              "name": "ISICLIENT_AUTH_TYPE",
                              "value": "1"
                            },
                            {
                              "name": "ISICLIENT_VERBOSE",
                              "value": "0"
                            },
                            {
                              "name": "POWERSCALE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSCALE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0",
                          "name": "metrics-powerscale"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-isilon",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-isilon/csi_sock",
                            "--mode=node",
                            "--driverPath=csi-isilon.dellemc.com",
                            "--driver-config-params=/csi-isilon-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powermax",
                  "namespace": "powermax"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powermax-creds",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_MANAGED_ARRAYS",
                          "value": "000000000001,000000000002"
                        },
                        {
                          "name": "X_CSI_POWERMAX_PORTGROUPS",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_TRANSPORT_PROTOCOL",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_K8S_CLUSTER_PREFIX",
                          "value": "CSM"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "X_CSI_VSPHERE_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VSPHERE_PORTGROUP",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VSPHERE_HOSTNAME",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_VCENTER_HOST",
                          "value": ""
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "X_CSI_POWERMAX_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "X_CSI_POWERMAX_DEBUG",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_REVPROXY_USE_SECRET",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powermax",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERMAX_ISCSI_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_TOPOLOGY_CONTROL_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": [
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/memory-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/disk-pressure",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "key": "node.kubernetes.io/network-unavailable",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REVPROXY_TLS_SECRET",
                              "value": "csirevproxy-tls-secret"
                            },
                            {
                              "name": "X_CSI_REVPROXY_PORT",
                              "value": "2222"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            },
                            {
                              "name": "DeployAsSidecar",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265",
                          "name": "csipowermax-reverseproxy"
                        }
                      ],
                      "configVersion": "v2.14.0",
                      "name": "csireverseproxy"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powermax"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERMAX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERMAX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERMAX_CAPACITY_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_PERFORMANCE_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERMAX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERMAX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            },
                            {
                              "name": "X_CSI_CONFIG_MAP_NAME",
                              "value": "powermax-reverseproxy-config"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414",
                          "name": "metrics-powermax"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powermax",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/powermax.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powermax-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powermax.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "powerstore",
                  "namespace": "powerstore"
                },
                "spec": {
                  "driver": {
                    "authSecret": "powerstore-config",
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_NODE_NAME_PREFIX",
                          "value": "csi-node"
                        },
                        {
                          "name": "X_CSI_FC_PORTS_FILTER_FILE_PATH",
                          "value": "/etc/fc-ports-filter"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOPOWERSTORE_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSM_AUTH_ENABLED",
                          "value": "false"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_MULTI_NAS_FAILURE_THRESHOLD",
                          "value": "5"
                        },
                        {
                          "name": "X_CSI_MULTI_NAS_COOLDOWN_PERIOD",
                          "value": "5m"
                        },
                        {
                          "name": "X_CSI_NFS_ACLS",
                          "value": "0777"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerstore",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_POWERSTORE_ENABLE_CHAP",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERSTORE_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-powerstore",
                            "--arrayConnectivityPollRate=60",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--csisock=unix:/var/lib/kubelet/plugins/csi-powerstore.dellemc.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/powerstore-config-params/driver-config-params.yaml",
                            "--driverPath=csi-powerstore.dellemc.com"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerstore"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERSTORE_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERSTORE_SPACE_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_ARRAY_POLL_FREQUENCY",
                              "value": "300"
                            },
                            {
                              "name": "POWERSTORE_FILE_SYSTEM_POLL_FREQUENCY",
                              "value": "20"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERSTORE_TOPOLOGY_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERSTORE_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERSTORE_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "ZIPKIN_URI",
                              "value": ""
                            },
                            {
                              "name": "ZIPKIN_SERVICE_NAME",
                              "value": "metrics-powerstore"
                            },
                            {
                              "name": "ZIPKIN_PROBABILITY",
                              "value": "0.0"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60",
                          "name": "metrics-powerstore"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    }
                  ]
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "unity",
                  "namespace": "unity"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_UNITY_ALLOW_MULTI_POD_ACCESS",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_EPHEMERAL_STAGING_PATH",
                          "value": "/var/lib/kubelet/plugins/kubernetes.io/csi/pv/"
                        },
                        {
                          "name": "X_CSI_ISCSI_CHROOT",
                          "value": "/noderoot"
                        },
                        {
                          "name": "X_CSI_UNITY_SYNC_NODEINFO_INTERVAL",
                          "value": "15"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "info"
                        },
                        {
                          "name": "GOUNITY_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOUNITY_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_FORMAT",
                          "value": "TEXT"
                        },
                        {
                          "name": "TENANT_NAME",
                          "value": ""
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "1"
                        },
                        {
                          "name": "X_CSI_UNITY_SKIP_CERTIFICATE_VALIDATION",
                          "value": "true"
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "ReadWriteOnceWithFSType",
                      "storageCapacity": true
                    },
                    "csiDriverType": "unity",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_ALLOWED_NETWORKS",
                          "value": ""
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 2,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "external-health-monitor"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "storage.dell.com/v1",
                "kind": "ContainerStorageModule",
                "metadata": {
                  "name": "vxflexos",
                  "namespace": "vxflexos"
                },
                "spec": {
                  "driver": {
                    "common": {
                      "envs": [
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLELISTVOLUMESNAPSHOT",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_VXFLEXOS_ENABLESNAPSHOTCGDELETE",
                          "value": "false"
                        },
                        {
                          "name": "CSI_LOG_LEVEL",
                          "value": "INFO"
                        },
                        {
                          "name": "GOSCALEIO_DEBUG",
                          "value": "false"
                        },
                        {
                          "name": "GOSCALEIO_SHOWHTTP",
                          "value": "false"
                        },
                        {
                          "name": "KUBELET_CONFIG_DIR",
                          "value": "/var/lib/kubelet"
                        },
                        {
                          "name": "CERT_SECRET_COUNT",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_QUOTA_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "INTERFACE_NAMES",
                          "value": null
                        }
                      ],
                      "image": "registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    "configVersion": "v2.15.0",
                    "controller": {
                      "envs": [
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_POWERFLEX_EXTERNAL_ACCESS",
                          "value": null
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "csiDriverSpec": {
                      "fSGroupPolicy": "File",
                      "storageCapacity": true
                    },
                    "csiDriverType": "powerflex",
                    "dnsPolicy": "ClusterFirstWithHostNet",
                    "forceRemoveDriver": true,
                    "initContainers": [
                      {
                        "envs": [
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          },
                          {
                            "name": "REPO_ADDRESS",
                            "value": "sftp://0.0.0.0"
                          },
                          {
                            "name": "REPO_USER",
                            "value": "sdcRepoUser"
                          },
                          {
                            "name": "MODULE_SIGCHECK",
                            "value": "0"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "imagePullPolicy": "IfNotPresent",
                        "name": "sdc"
                      }
                    ],
                    "node": {
                      "envs": [
                        {
                          "name": "X_CSI_SDC_ENABLED",
                          "value": "true"
                        },
                        {
                          "name": "X_CSI_APPROVE_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_HEALTH_MONITOR_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_ENABLED",
                          "value": "false"
                        },
                        {
                          "name": "X_CSI_RENAME_SDC_PREFIX",
                          "value": ""
                        },
                        {
                          "name": "X_CSI_MAX_VOLUMES_PER_NODE",
                          "value": "0"
                        },
                        {
                          "name": "X_CSI_SDC_SFTP_REPO_ENABLED",
                          "value": "false"
                        }
                      ],
                      "nodeSelector": null,
                      "tolerations": null
                    },
                    "replicas": 1,
                    "sideCars": [
                      {
                        "args": [
                          "--volume-name-prefix=csivol"
                        ],
                        "image": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                        "name": "provisioner"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                        "name": "attacher"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                        "name": "registrar"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474",
                        "name": "resizer"
                      },
                      {
                        "image": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d",
                        "name": "snapshotter"
                      },
                      {
                        "image": "registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1",
                        "name": "csi-metadata-retriever"
                      },
                      {
                        "enabled": false,
                        "envs": [
                          {
                            "name": "HOST_PID",
                            "value": "1"
                          },
                          {
                            "name": "MDM",
                            "value": "10.xx.xx.xx,10.xx.xx.xx"
                          }
                        ],
                        "image": "quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540",
                        "name": "sdc-monitor"
                      },
                      {
                        "args": [
                          "--monitor-interval=60s"
                        ],
                        "enabled": false,
                        "image": "registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea",
                        "name": "csi-external-health-monitor-controller"
                      }
                    ]
                  },
                  "modules": [
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "PROXY_HOST",
                              "value": "csm-authorization.com"
                            },
                            {
                              "name": "SKIP_CERTIFICATE_VALIDATION",
                              "value": "true"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da",
                          "name": "karavi-authorization-proxy"
                        }
                      ],
                      "configVersion": "v2.3.0",
                      "enabled": false,
                      "name": "authorization"
                    },
                    {
                      "components": [
                        {
                          "certificate": "",
                          "enabled": false,
                          "envs": [
                            {
                              "name": "NGINX_PROXY_IMAGE",
                              "value": "quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403"
                            }
                          ],
                          "image": "ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd",
                          "name": "otel-collector",
                          "privateKey": ""
                        },
                        {
                          "enabled": false,
                          "name": "cert-manager"
                        },
                        {
                          "enabled": false,
                          "envs": [
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_TOPOLOGY_METRICS_POLL_FREQUENCY",
                              "value": "30"
                            },
                            {
                              "name": "POWERFLEX_MAX_CONCURRENT_QUERIES",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_SDC_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_METRICS_ENABLED",
                              "value": "true"
                            },
                            {
                              "name": "POWERFLEX_SDC_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_VOLUME_IO_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_STORAGE_POOL_POLL_FREQUENCY",
                              "value": "10"
                            },
                            {
                              "name": "POWERFLEX_LOG_LEVEL",
                              "value": "INFO"
                            },
                            {
                              "name": "POWERFLEX_LOG_FORMAT",
                              "value": "TEXT"
                            },
                            {
                              "name": "COLLECTOR_ADDRESS",
                              "value": "otel-collector:55680"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d",
                          "name": "metrics-powerflex"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "observability"
                    },
                    {
                      "components": [
                        {
                          "envs": [
                            {
                              "name": "X_CSI_REPLICATION_PREFIX",
                              "value": "replication.storage.dell.com"
                            },
                            {
                              "name": "X_CSI_REPLICATION_CONTEXT_PREFIX",
                              "value": "powerflex"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16",
                          "name": "dell-csi-replicator"
                        },
                        {
                          "envs": [
                            {
                              "name": "REPLICATION_CTRL_LOG_LEVEL",
                              "value": "debug"
                            },
                            {
                              "name": "REPLICATION_CTRL_REPLICAS",
                              "value": "1"
                            },
                            {
                              "name": "RETRY_INTERVAL_MIN",
                              "value": "1s"
                            },
                            {
                              "name": "RETRY_INTERVAL_MAX",
                              "value": "5m"
                            },
                            {
                              "name": "DISABLE_PVC_REMAP",
                              "value": "false"
                            },
                            {
                              "name": "REPLICATION_ALLOW_PVC_CREATION_ON_TARGET",
                              "value": "false"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa",
                          "name": "dell-replication-controller-manager"
                        }
                      ],
                      "configVersion": "v1.13.0",
                      "enabled": false,
                      "name": "replication"
                    },
                    {
                      "components": [
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--skipArrayConnectionValidation=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--arrayConnectivityConnectionLossThreshold=3",
                            "--csisock=unix:/var/run/csi/csi.sock",
                            "--mode=controller",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-controller"
                        },
                        {
                          "args": [
                            "--labelvalue=csi-vxflexos",
                            "--leaderelection=false",
                            "--driverPodLabelValue=dell-storage",
                            "--ignoreVolumelessPods=false",
                            "--arrayConnectivityPollRate=5",
                            "--csisock=unix:/var/lib/kubelet/plugins/vxflexos.emc.dell.com/csi_sock",
                            "--mode=node",
                            "--driver-config-params=/vxflexos-config-params/driver-config-params.yaml"
                          ],
                          "envs": [
                            {
                              "name": "X_CSI_PODMON_API_PORT",
                              "value": "8083"
                            }
                          ],
                          "image": "registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe",
                          "imagePullPolicy": "IfNotPresent",
                          "name": "podmon-node"
                        }
                      ],
                      "configVersion": "v1.14.0",
                      "enabled": false,
                      "name": "resiliency"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
          createdAt: "2025-09-05T13:18:42Z"
          description: Easily install and manage Dell’s CSI Drivers and CSM
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/dell/csm-operator
          support: Dell Technologies
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ContainerStorageModule is the Schema for the containerstoragemodules
              API
            displayName: Container Storage Module
            kind: ContainerStorageModule
            name: containerstoragemodules.storage.dell.com
            version: v1
        description: "Dell Container Storage Modules (CSM) Operator is a Kubernetes
          Operator which can be \nused to install and manage Dell’s CSI drivers and
          CSM modules. \nBy using Dell CSM Operator, enterprises can quickly and easily
          deploy the \nCSM modules for CSI drivers making it easy for DevOps \nteams
          to build and optimize their Kubernetes workloads using enterprise grade
          storage \nwhile consistently aligning with the desired outcomes of the organization.\nThis
          operator is available as both community and certified operator for Openshift
          \nand as a community operator for upstream Kubernetes.\nOperator can be
          deployed via OperatorHub.io. using Operator Lifecycle Manager (OLM)\n\n###
          About Dell CSM\n\nDell CSM brings powerful enterprise storage features and
          functionality to Kubernetes for \neasier adoption of cloud native workloads,
          improved productivity and scalable operations. \n\nCSM delivers software
          modules for storage management that provide developers access to \nbuild
          automation for enhanced IT needs and other critical enterprise storage features.
          \n\nCurrently, the following modules are available as part of Dell CSM.
          \n*\tObservability - Delivers a high-level view of the storage capacity
          \n  and performance usage via Grafana dashboards to the Kubernetes users\n*\tReplication
          - Enables array replication capabilities for Kubernetes users\n*\tAuthorization
          - Provides both storage and Kubernetes administrators the ability \n  to
          apply RBAC for Dell CSI Drivers\n*\tResiliency - Designed to make Kubernetes
          applications more resilient to \n  node failures\n\nThese modules are available
          for various storage platforms in the Dell portfolio, \nsuch as Dell PowerStore,
          Dell PowerMax, Dell PowerFlex, Dell PowerScale, and Dell Unity XT.\n\nFor
          more information, visit [Dell’s Container Storage Module web page](https://www.dell.com/en-us/dt/solutions/kubernetes-containers/kubernetes-storage.htm)\n\nAdditional
          Materials\n*\t[Documentation](https://dell.github.io/csm-docs)\n*\t[CSM
          Solution Brief](https://www.delltechnologies.com/asset/en-us/products/storage/briefs-summaries/csm-solution-brief.pdf)\n"
        displayName: Dell Container Storage Modules
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dell Container Storage Modules
        - Dell CSI Driver
        - Dell CSM Modules
        - Powerflex
        - Powerscale
        - Powerstore
        - Unity
        - Authorization
        - Observability
        - Replication
        links:
        - name: Documentation
          url: https://dell.github.io/csm-docs
        maintainers:
        - email: container.storage.modules@dell.com
          name: Dell Container Storage Modules
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Dell Technologies
          url: https://github.com/dell/csm-operator
        relatedImages:
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-proxy@sha256:a265ec058b102d1cf175deb2b80718d28f6728c094ee3b5367a07473a0dc2682
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-sidecar@sha256:fadcba26fe3464925b7b8857d470204ba43a72e75edd32ffa83675c1db6530da
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-tenant@sha256:687faaa76ac70ab5f9f452c8a70d5ac45b1968ea2adc60198f2fd3e09b41e358
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:59335acaca515fceed5f0350c46cfcfd6037475f0cf35ae8ab41d37333039e2e
        - registry.connect.redhat.com/dell-emc/dell-csm-unity@sha256:2eeac19ad6633fba486f807791103dbd5a74ebcd109c3b4a8d12c238346ac5a3
        - docker.io/openpolicyagent/kube-mgmt@sha256:f5bf223771237756a56168729defb608ec2941c9b701d1f1af96f9ae82966517
        - docker.io/rediscommander/redis-commander@sha256:19cd0c49f418779fa2822a0496c5e6516d0c792effc39ed20089e6268477e40a
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerscale@sha256:a1b24f964308efc7d21e0496dbbe602819ac9ef1115899646a058665f3fb83a0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:bc7be893ecc3ad524194aa6573b2f5c06cd469bdf21a500ab6c99c2ba1c4d64d
        - registry.connect.redhat.com/dell-emc/dell-csm-powerscale@sha256:7959791ea0baf0337b87aaae46b7c7ff9f17e042bacc885791bce68b197ae1cf
        - quay.io/nginx/nginx-unprivileged@sha256:f9dfa9c20b2b0b7c5cc830374f22f23dee3f750b6c5291ca7e0330b5c88e6403
        - registry.connect.redhat.com/dell-emc/dell-csm-podmon@sha256:68780493ea9718faa399babd40cf09e1ace43e6a63a878d37612fec377067ebe
        - registry.connect.redhat.com/dell-emc/dell-container-storage-modules-operator@sha256:43ae28056fe1c8b6a0ce1635cf733cafd6448f8afe3cc0fd23e55ab0a1e65f0d
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ce054c6fade575e9d4dbd4c3d65b9c5d1b05160aacfb9cf8d8cac51d73f3ccea
        - registry.connect.redhat.com/dell-emc/dell-csm-powerflex@sha256:9c7e4af10c2f17e151cc0605f7cff9592fe6d63116e45b73a71a71b50fc52b5d
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerflex@sha256:ed6f8ad822792400a30ecd57489453a57580495b10236390a4ec24471771d01d
        - docker.io/library/redis@sha256:7521abdff715d396aa482183942f3fe643344287c29ccb66eee16ac08a92190f
        - registry.connect.redhat.com/dell-emc/dell-csm-replication-controller-manager@sha256:8cad45a81bd05be95170944850dd1b8b5fb7c8c5ee0397420d04cd2155ba52fa
        - registry.connect.redhat.com/dell-emc/dell-csm-operator-bundle@sha256:4512f786fed0d9b2285b5a2827ec3b597452f90c217e27398f2f0f552377d844
        - ghcr.io/open-telemetry/opentelemetry-collector-releases/opentelemetry-collector@sha256:6d260efde7406a1f7d731a5d9663cb6ce25fa634f3ef3ef4a69266aaae1680fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.connect.redhat.com/dell-emc/dell-csm-replicator@sha256:b427913a72121a261161cf2f81bb991dfd383fa1703d20f1f1ff4fb5743eba16
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powermax@sha256:baab9e0c17eb30a1a33832b8384d82bfff03ad378d432897231eab83dc679414
        - quay.io/dell/storage/powerflex/sdc@sha256:4aca94f895636efcc7308aeb8b083cb2f15133e255185b8db0805b9649ca8540
        - registry.connect.redhat.com/dell-emc/dell-csm-metadata-retriever@sha256:6de94d91a17a401b5f2e5cdf7bb50cd053521deaf1e189340d21c4249e8c4bf1
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax-reverseproxy@sha256:10da8c0f77d1694f31aa023a3bf793108a3363968fecab500eaf58fba3538265
        - registry.connect.redhat.com/dell-emc/dell-csm-powermax@sha256:5d7ffb9b4f6b3724861a9afb90085bf1ca867c58452d3fd8666928fa6b25aa4e
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/dell-emc/dell-csm-powerstore@sha256:f952c06f4d6eebae4247cb003c003555e433de90c00bf4122ec35886cc817334
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-storage@sha256:ecaafb211be66f090b468bce366fe870877cb921a14f53fc63ffb1a3f26405d5
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-controller@sha256:6de38a3c833ce514dd8bf496a23157ff9f11aeca6359095c5ac41a38ba3ec231
        - docker.io/openpolicyagent/opa@sha256:2b1c4b231f6dfa6a2baea4af4109203b10cd4b65e836d72f03bfe5d8dfce7095
        - registry.connect.redhat.com/dell-emc/dell-csm-authorization-role@sha256:d09ffef1866f24bffab2a0d38af11fbbc528b1eae04db1246617b5f366ae545c
        - registry.connect.redhat.com/dell-emc/dell-csm-metrics-powerstore@sha256:4d288473ab69dff348842d0a04c9c57aa7b76cc18c394e98906d09bfd08e0b60
        version: 1.10.0
      entries:
      - name: dell-csm-operator-certified.v1.10.0
        version: 1.10.0
      - name: dell-csm-operator-certified.v1.9.1
        version: 1.9.1
      - name: dell-csm-operator-certified.v1.9.0
        version: 1.9.0
      - name: dell-csm-operator-certified.v1.8.1
        version: 1.8.1
      - name: dell-csm-operator-certified.v1.8.0
        version: 1.8.0
      - name: dell-csm-operator-certified.v1.7.0
        version: 1.7.0
      - name: dell-csm-operator-certified.v1.6.1
        version: 1.6.1
      name: stable
    defaultChannel: stable
    packageName: dell-csm-operator-certified
    provider:
      name: Dell Technologies
      url: https://github.com/dell/csm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Helm Community
      provider-url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
    name: cockroachdb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cockroachdb.v5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "conf": {
                    "attrs": [],
                    "cache": "25%",
                    "cluster-name": "",
                    "disable-cluster-name-verification": false,
                    "http-port": 8080,
                    "join": [],
                    "locality": "",
                    "logtostderr": "INFO",
                    "max-sql-memory": "25%",
                    "port": 26257,
                    "single-node": false,
                    "sql-audit-dir": ""
                  },
                  "image": {
                    "credentials": {},
                    "pullPolicy": "IfNotPresent",
                    "repository": "cockroachdb/cockroach",
                    "tag": "v20.2.4"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "labels": {},
                    "paths": [
                      "/"
                    ],
                    "tls": []
                  },
                  "init": {
                    "affinity": {},
                    "annotations": {},
                    "labels": {
                      "app.kubernetes.io/component": "init"
                    },
                    "nodeSelector": {},
                    "resources": {},
                    "tolerations": []
                  },
                  "labels": {},
                  "networkPolicy": {
                    "enabled": false,
                    "ingress": {
                      "grpc": [],
                      "http": []
                    }
                  },
                  "service": {
                    "discovery": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      }
                    },
                    "ports": {
                      "grpc": {
                        "external": {
                          "name": "grpc",
                          "port": 26257
                        },
                        "internal": {
                          "name": "grpc-internal",
                          "port": 26257
                        }
                      },
                      "http": {
                        "name": "http",
                        "port": 8080
                      }
                    },
                    "public": {
                      "annotations": {},
                      "labels": {
                        "app.kubernetes.io/component": "cockroachdb"
                      },
                      "type": "ClusterIP"
                    }
                  },
                  "serviceMonitor": {
                    "annotations": {},
                    "enabled": false,
                    "interval": "10s",
                    "labels": {}
                  },
                  "statefulset": {
                    "annotations": {},
                    "args": [],
                    "budget": {
                      "maxUnavailable": 1
                    },
                    "env": [],
                    "labels": {
                      "app.kubernetes.io/component": "cockroachdb"
                    },
                    "nodeAffinity": {},
                    "nodeSelector": {},
                    "podAffinity": {},
                    "podAntiAffinity": {
                      "topologyKey": "kubernetes.io/hostname",
                      "type": "soft",
                      "weight": 100
                    },
                    "podManagementPolicy": "Parallel",
                    "priorityClassName": "",
                    "replicas": 3,
                    "resources": {},
                    "secretMounts": [],
                    "tolerations": [],
                    "topologySpreadConstraints": {
                      "maxSkew": 1,
                      "topologyKey": "topology.kubernetes.io/zone",
                      "whenUnsatisfiable": "ScheduleAnyway"
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    }
                  },
                  "storage": {
                    "hostPath": "",
                    "persistentVolume": {
                      "annotations": {},
                      "enabled": true,
                      "labels": {},
                      "size": "100Gi",
                      "storageClass": ""
                    }
                  },
                  "tls": {
                    "certs": {
                      "clientRootSecret": "cockroachdb-root",
                      "nodeSecret": "cockroachdb-node",
                      "provided": false,
                      "tlsSecret": false
                    },
                    "enabled": false,
                    "init": {
                      "image": {
                        "credentials": {},
                        "pullPolicy": "IfNotPresent",
                        "repository": "cockroachdb/cockroach-k8s-request-cert",
                        "tag": "0.4"
                      }
                    },
                    "serviceAccount": {
                      "create": true,
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/helmoperators/cockroachdb:v5.0.4
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/dmesser/cockroachdb-operator
          support: dmesser
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports the same parameters like the official [CockroachDB Helm Chart](https://github.com/cockroachdb/helm-charts/tree/master/cockroachdb).

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-5.0.4
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: dmesser@redhat.com
          name: Daniel Messer
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - quay.io/helmoperators/cockroachdb:v5.0.4
        - quay.io/openshift-community-operators/cockroachdb@sha256:f42337e7b85a46d83c94694638e2312e10ca16a03542399a65ba783c94a32b63
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 5.0.4
      entries:
      - name: cockroachdb.v5.0.4
        version: 5.0.4
      - name: cockroachdb.v5.0.3
        version: 5.0.3
      name: stable-5.x
    - currentCSV: cockroachdb.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1alpha1",
                "kind": "Cockroachdb",
                "metadata": {
                  "name": "cockroachdb-sample"
                },
                "spec": {"image":{"repository":"cockroachdb/cockroach","tag":"v22.2.4","pullPolicy":"IfNotPresent","credentials":{}},"labels":{},"clusterDomain":"cluster.local","conf":{"attrs":[],"cache":"25%","cluster-name":"","disable-cluster-name-verification":false,"join":[],"log":{"enabled":false,"config":{}},"logtostderr":"INFO","max-sql-memory":"25%","locality":"","single-node":false,"sql-audit-dir":"","port":26257,"http-port":8080,"path":"cockroach-data","store":{"enabled":false,"type":null,"size":null,"attrs":null}},"statefulset":{"replicas":3,"updateStrategy":{"type":"RollingUpdate"},"podManagementPolicy":"Parallel","budget":{"maxUnavailable":1},"args":[],"env":[],"secretMounts":[],"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{},"nodeAffinity":{},"podAffinity":{},"podAntiAffinity":{"topologyKey":"kubernetes.io/hostname","type":"soft","weight":100},"nodeSelector":{},"priorityClassName":"","tolerations":[],"topologySpreadConstraints":{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway"},"resources":{},"customLivenessProbe":{},"customReadinessProbe":{}},"service":{"ports":{"grpc":{"external":{"port":26257,"name":"grpc"},"internal":{"port":26257,"name":"grpc-internal"}},"http":{"port":8080,"name":"http"}},"public":{"type":"ClusterIP","labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}},"discovery":{"labels":{"app.kubernetes.io/component":"cockroachdb"},"annotations":{}}},"ingress":{"enabled":false,"labels":{},"annotations":{},"paths":["/"],"hosts":[],"tls":[]},"prometheus":{"enabled":true},"serviceMonitor":{"enabled":false,"labels":{},"annotations":{},"interval":"10s","namespaced":false},"storage":{"hostPath":"","persistentVolume":{"enabled":true,"size":"100Gi","storageClass":"","labels":{},"annotations":{}}},"init":{"labels":{"app.kubernetes.io/component":"init"},"jobAnnotations":{},"annotations":{},"affinity":{},"nodeSelector":{},"tolerations":[],"resources":{},"provisioning":{"enabled":false,"clusterSettings":null,"users":[],"databases":[]}},"tls":{"enabled":true,"serviceAccount":{"create":true,"name":""},"copyCerts":{"image":"busybox"},"certs":{"provided":false,"clientRootSecret":"cockroachdb-root","nodeSecret":"cockroachdb-node","tlsSecret":false,"selfSigner":{"enabled":true,"caProvided":false,"caSecret":"","minimumCertDuration":"624h","caCertDuration":"43800h","caCertExpiryWindow":"648h","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h","rotateCerts":true,"readinessWait":"30s","podUpdateTimeout":"2m"},"certManager":false,"certManagerIssuer":{"group":"cert-manager.io","kind":"Issuer","name":"cockroachdb","clientCertDuration":"672h","clientCertExpiryWindow":"48h","nodeCertDuration":"8760h","nodeCertExpiryWindow":"168h"},"useCertManagerV1CRDs":false},"selfSigner":{"image":{"repository":"cockroachlabs-helm-charts/cockroach-self-signer-cert","tag":"1.3","pullPolicy":"IfNotPresent","credentials":{},"registry":"gcr.io"}}},"networkPolicy":{"enabled":false,"ingress":{"grpc":[],"http":[]}},"iap":{"enabled":false}}
              }
            ]
          capabilities: Basic Install
          categories: Database
          certified: "false"
          containerImage: quay.io/cockroachdb/cockroach-helm-operator:6.0.0
          createdAt: 2021-01-23T10-00-00Z
          description: CockroachDB Operator based on the CockroachDB helm chart
          olm.skipRange: <6.0.0
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/cockroachdb/helm-charts
          support: CockroachDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a CockroachDB cluster
            displayName: CockroachDB
            kind: Cockroachdb
            name: cockroachdbs.charts.operatorhub.io
            version: v1alpha1
        description: |
          CockroachDB is a scalable, survivable, strongly-consistent SQL database.
          ## About this Operator
          This Operator is based on a Helm chart for CockroachDB. It supports reconfiguration for some parameters, but notably does not handle scale down of the replica count in a seamless manner. Scale up works great.

          ## Core capabilities
          * **StatefulSet** - Sets up a dynamically scalable CockroachDB cluster using a Kubernetes StatefulSet
          * **Expand Replicas** - Supports expanding the set of replicas by simply editing your object
          * **Dashboard** - Installs the CockroachDB user interface to administer your cluster. Easily expose it via an Ingress rule.
          Review all of the [confiuguration options](https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration) to best run your database instance. The example configuration is derived from the chart's [`values.yaml`](https://github.com/helm/charts/blob/master/stable/cockroachdb/values.yaml).
          ## Using the cluster
          The resulting cluster endpoint can be consumed from a `Service` that follows the pattern: `<StatefulSet-name>-public`. For example to connect using the command line client, use something like the following to obtain the name of the service:
          ```
          kubectl get service -l chart=cockroachdb-6.0.0
          NAME                                           TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)              AGE
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb          ClusterIP   None             <none>        26257/TCP,8080/TCP   24m
          example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public   ClusterIP   10.106.249.134   <none>        26257/TCP,8080/TCP   24m
          ```
          Then you can use the CockroachDB command line client to connect to the database cluster:
          ```
          kubectl run -it --rm cockroach-client --image=cockroachdb/cockroach --restart=Never --command -- ./cockroach sql --insecure --host example-9f8ngwzrxbxrulxqmdqfhn51h-cdb-public
          ```
          ## Before you start
          This Operator requires a cluster with PV support in order to run correctly.

          Note that for a production cluster, you will likely want to override the following parameters in [`values.yaml`](values.yaml) with your own values.

          - `statefulset.resources.requests.memory` and `statefulset.resources.limits.memory` allocate memory resources to CockroachDB pods in your cluster.
          - `conf.cache` and `conf.max-sql-memory` are memory limits that we recommend setting to 1/4 of the above resource allocation. When running CockroachDB, you must set these limits explicitly to avoid running out of memory.
          - `storage.persistentVolume.size` defaults to `100Gi` of disk space per pod, which you may increase or decrease for your use case.
          - `storage.persistentVolume.storageClass` uses the default storage class for your environment. We strongly recommend that you specify a storage class which uses an SSD.
          - `tls.enabled` must be set to `yes`/`true` to deploy in secure mode.

          For more information on overriding the `values.yaml` parameters, please see:
          > <https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes.html#step-2-start-cockroachdb>

          If you are running in secure mode (with configuration parameter `tls.enabled` set to `yes`/`true`) and `tls.certs.provided` set to `no`/`false`), then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          When `tls.certs.provided` is set to `yes`/`true`, this chart will use certificates created outside of Kubernetes. You may want to use this if you want to use a different certificate authority from the one being used by Kubernetes or if your Kubernetes cluster doesn't fully support certificate-signing requests. To use this, first set up your certificates and load them into your Kubernetes cluster as Secrets using the commands below:

          ```
          mkdir certs
          mkdir my-safe-directory
          cockroach cert create-ca --certs-dir=certs --ca-key=my-safe-directory/ca.key
          cockroach cert create-client root --certs-dir=certs --ca-key=my-safe-directory/ca.key
          kubectl create secret generic cockroachdb-root --from-file=certs
          cockroach cert create-node --certs-dir=certs --ca-key=my-safe-directory/ca.key localhost 127.0.0.1 eerie-horse-cockroachdb-public eerie-horse-cockroachdb-public.default eerie-horse-cockroachdb-public.default.svc.cluster.local *.eerie-horse-cockroachdb *.eerie-horse-cockroachdb.default *.eerie-horse-cockroachdb.default.svc.cluster.local
          kubectl create secret generic cockroachdb-node --from-file=certs
          ```

          Set `tls.certs.tlsSecret` to `yes/true` if you make use of [cert-manager][3] in your cluster.

          [cert-manager][3] stores generated certificates in dedicated TLS secrets. Thus, they are always named:

          * `ca.crt`
          * `tls.crt`
          * `tls.key`

          On the other hand, CockroachDB also demands dedicated certificate filenames:

          * `ca.crt`
          * `node.crt`
          * `node.key`
          * `client.root.crt`
          * `client.root.key`

          By activating `tls.certs.tlsSecret` we benefit from projected secrets and convert the TLS secret filenames to their according CockroachDB filenames.

          If you are running in secure mode, then you will have to manually approve the cluster's security certificates as the pods are created. You can see the pending CSRs (certificate signing requests) by running `kubectl get csr`, and approve them by running `kubectl certificate approve <csr-name>`. You'll have to approve one certificate for each CockroachDB node (e.g., `default.node.my-release-cockroachdb-0` and one client certificate for the job that initializes the cluster (e.g., `default.node.root`).

          Confirm that all pods are `Running` successfully and init has been completed:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS      RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running     0          1m
          my-release-cockroachdb-1            1/1       Running     0          1m
          my-release-cockroachdb-2            1/1       Running     0          1m
          my-release-cockroachdb-init-k6jcr   0/1       Completed   0          1m
          ```

          Confirm that persistent volumes are created and claimed for each pod:

          ```shell
          kubectl get pv
          ```

          ```
          NAME                                       CAPACITY   ACCESS MODES   RECLAIM POLICY   STATUS    CLAIM                                      STORAGECLASS   REASON    AGE
          pvc-64878ebf-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-0   standard                 51s
          pvc-64945b4f-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-1   standard                 51s
          pvc-649d920d-f3f0-11e8-ab5b-42010a8e0035   100Gi      RWO            Delete           Bound     default/datadir-my-release-cockroachdb-2   standard                 51s
          ```

          ## Upgrading the cluster

          ### Chart version 3.0.0 and after

          Launch a temporary interactive pod and start the built-in SQL client:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          Set `cluster.preserve_downgrade_option`, where `$current_version` is the CockroachDB version currently running (e.g., `19.2`):

          ```sql
          > SET CLUSTER SETTING cluster.preserve_downgrade_option = '$current_version';
          ```

          Exit the shell and delete the temporary pod:

          ```sql
          > \q
          ```

          Kick off the upgrade process by changing the new Docker image, where `$new_version` is the CockroachDB version to which you are upgrading:

          ```shell
          kubectl delete job my-release-cockroachdb-init
          ```

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"image":{"tag":"<new version tag here>"}}}' --type merge
          ```

          Kubernetes will carry out a safe [rolling upgrade](https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets) of your CockroachDB nodes one-by-one. Monitor the cluster's pods until all have been successfully restarted:

          ```shell
          kubectl get pods
          ```

          ```
          NAME                                READY     STATUS              RESTARTS   AGE
          my-release-cockroachdb-0            1/1       Running             0          2m
          my-release-cockroachdb-1            1/1       Running             0          3m
          my-release-cockroachdb-2            1/1       Running             0          3m
          my-release-cockroachdb-3            0/1       ContainerCreating   0          25s
          my-release-cockroachdb-init-nwjkh   0/1       ContainerCreating   0          6s
          ```

          ```shell
          kubectl get pods \
          -o jsonpath='{range .items[*]}{.metadata.name}{"\t"}{.spec.containers[0].image}{"\n"}'
          ```

          ```
          my-release-cockroachdb-0    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-1    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-2    cockroachdb/cockroach:v20.2.4
          my-release-cockroachdb-3    cockroachdb/cockroach:v20.2.4
          ```

          Resume normal operations. Once you are comfortable that the stability and performance of the cluster is what you'd expect post-upgrade, finalize the upgrade:

          ```shell
          kubectl run cockroachdb --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host=my-release-cockroachdb-public
          ```

          ```sql
          > RESET CLUSTER SETTING cluster.preserve_downgrade_option;
          > \q
          ```

          ## Configuration

          The following table lists the configurable parameters of the CockroachDB chart and their default values.
          For details see the [`values.yaml`](values.yaml) file.

          | Parameter                                                 | Description                                                     | Default                                          |
          | ---------                                                 | -----------                                                     | -------                                          |
          | `clusterDomain`                                           | Cluster's default DNS domain                                    | `cluster.local`                                  |
          | `conf.attrs`                                              | CockroachDB node attributes                                     | `[]`                                             |
          | `conf.cache`                                              | Size of CockroachDB's in-memory cache                           | `25%`                                            |
          | `conf.cluster-name`                                       | Name of CockroachDB cluster                                     | `""`                                             |
          | `conf.disable-cluster-name-verification`                  | Disable CockroachDB cluster name verification                   | `no`                                             |
          | `conf.join`                                               | List of already-existing CockroachDB instances                  | `[]`                                             |
          | `conf.max-disk-temp-storage`                              | Max storage capacity for temp data                              | `0`                                              |
          | `conf.max-offset`                                         | Max allowed clock offset for CockroachDB cluster                | `500ms`                                          |
          | `conf.max-sql-memory`                                     | Max memory to use processing SQL querie                         | `25%`                                            |
          | `conf.locality`                                           | Locality attribute for this deployment                          | `""`                                             |
          | `conf.single-node`                                        | Disable CockroachDB clustering (standalone mode)                | `no`                                             |
          | `conf.sql-audit-dir`                                      | Directory for SQL audit log                                     | `""`                                             |
          | `conf.port`                                               | CockroachDB primary serving port in Pods                        | `26257`                                          |
          | `conf.http-port`                                          | CockroachDB HTTP port in Pods                                   | `8080`                                           |
          | `image.repository`                                        | Container image name                                            | `cockroachdb/cockroach`                          |
          | `image.tag`                                               | Container image tag                                             | `v20.2.4`                                        |
          | `image.pullPolicy`                                        | Container pull policy                                           | `IfNotPresent`                                   |
          | `image.credentials`                                       | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `statefulset.replicas`                                    | StatefulSet replicas number                                     | `3`                                              |
          | `statefulset.updateStrategy`                              | Update strategy for StatefulSet Pods                            | `{"type": "RollingUpdate"}`                      |
          | `statefulset.podManagementPolicy`                         | `OrderedReady`/`Parallel` Pods creation/deletion order          | `Parallel`                                       |
          | `statefulset.budget.maxUnavailable`                       | k8s PodDisruptionBudget parameter                               | `1`                                              |
          | `statefulset.args`                                        | Extra command-line arguments                                    | `[]`                                             |
          | `statefulset.env`                                         | Extra env vars                                                  | `[]`                                             |
          | `statefulset.secretMounts`                                | Additional Secrets to mount at cluster members                  | `[]`                                             |
          | `statefulset.labels`                                      | Additional labels of StatefulSet and its Pods                   | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `statefulset.annotations`                                 | Additional annotations of StatefulSet Pods                      | `{}`                                             |
          | `statefulset.nodeAffinity`                                | [Node affinity rules][2] of StatefulSet Pods                    | `{}`                                             |
          | `statefulset.podAffinity`                                 | [Inter-Pod affinity rules][1] of StatefulSet Pods               | `{}`                                             |
          | `statefulset.podAntiAffinity`                             | [Anti-affinity rules][1] of StatefulSet Pods                    | auto                                             |
          | `statefulset.podAntiAffinity.topologyKey`                 | The topologyKey for auto [anti-affinity rules][1]               | `kubernetes.io/hostname`                         |
          | `statefulset.podAntiAffinity.type`                        | Type of auto [anti-affinity rules][1]                           | `soft`                                           |
          | `statefulset.podAntiAffinity.weight`                      | Weight for `soft` auto [anti-affinity rules][1]                 | `100`                                            |
          | `statefulset.nodeSelector`                                | Node labels for StatefulSet Pods assignment                     | `{}`                                             |
          | `statefulset.priorityClassName`                           | [PriorityClassName][4] for StatefulSet Pods                     | `""`                                             |
          | `statefulset.tolerations`                                 | Node taints to tolerate by StatefulSet Pods                     | `[]`                                             |
          | `statefulset.topologySpreadConstraints`                   | [Topology Spread Constraints rules][5] of StatefulSet Pods      | auto                                             |
          | `statefulset.topologySpreadConstraints.maxSkew`           | Degree to which Pods may be unevenly distributed                | `1`                                              |
          | `statefulset.topologySpreadConstraints.topologyKey`       | The key of node labels                                          | `topology.kubernetes.io/zone`                    |
          | `statefulset.topologySpreadConstraints.whenUnsatisfiable` | `ScheduleAnyway`/`DoNotSchedule` for unsatisfiable constraints  | `ScheduleAnyway`                                 |
          | `statefulset.resources`                                   | Resource requests and limits for StatefulSet Pods               | `{}`                                             |
          | `service.ports.grpc.external.port`                        | CockroachDB primary serving port in Services                    | `26257`                                          |
          | `service.ports.grpc.external.name`                        | CockroachDB primary serving port name in Services               | `grpc`                                           |
          | `service.ports.grpc.internal.port`                        | CockroachDB inter-communication port in Services                | `26257`                                          |
          | `service.ports.grpc.internal.name`                        | CockroachDB inter-communication port name in Services           | `grpc-internal`                                  |
          | `service.ports.http.port`                                 | CockroachDB HTTP port in Services                               | `8080`                                           |
          | `service.ports.http.name`                                 | CockroachDB HTTP port name in Services                          | `http`                                           |
          | `service.public.type`                                     | Public Service type                                             | `ClusterIP`                                      |
          | `service.public.labels`                                   | Additional labels of public Service                             | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.public.annotations`                              | Additional annotations of public Service                        | `{}`                                             |
          | `service.discovery.labels`                                | Additional labels of discovery Service                          | `{"app.kubernetes.io/component": "cockroachdb"}` |
          | `service.discovery.annotations`                           | Additional annotations of discovery Service                     | `{}`                                             |
          | `ingress.enabled`                                         | Enable ingress resource for CockroachDB                         | `false`                                          |
          | `ingress.labels`                                          | Additional labels of Ingress                                    | `{}`                                             |
          | `ingress.annotations`                                     | Additional annotations of Ingress                               | `{}`                                             |
          | `ingress.paths`                                           | Paths for the default host                                      | `[/]`                                            |
          | `ingress.hosts`                                           | CockroachDB Ingress hostnames                                   | `[]`                                             |
          | `ingress.tls[0].hosts`                                    | CockroachDB Ingress tls hostnames                               | `nil`                                            |
          | `ingress.tls[0].secretName`                               | CockroachDB Ingress tls secret name                             | `nil`                                            |
          | `serviceMonitor.enabled`                                  | Create [ServiceMonitor](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/design.md#servicemonitor) Resource for scraping metrics using [PrometheusOperator](https://github.com/prometheus-operator/prometheus-operator/blob/master/Documentation/user-guides/getting-started.md#prometheus-operator)                     | `false`                                             |
          | `serviceMonitor.labels`                                   | Additional labels of ServiceMonitor                             | `{}`                                             |
          | `serviceMonitor.annotations`                              | Additional annotations of ServiceMonitor                        | `{}`                                             |
          | `serviceMonitor.interval`                                 | ServiceMonitor scrape metrics interval                          | `10s`                                            |
          | `serviceMonitor.scrapeTimeout`                            | ServiceMonitor scrape timeout                                   | `nil`                                            |
          | `storage.hostPath`                                        | Absolute path on host to store data                             | `""`                                             |
          | `storage.persistentVolume.enabled`                        | Whether to use PersistentVolume to store data                   | `yes`                                            |
          | `storage.persistentVolume.size`                           | PersistentVolume size                                           | `100Gi`                                          |
          | `storage.persistentVolume.storageClass`                   | PersistentVolume class                                          | `""`                                             |
          | `storage.persistentVolume.labels`                         | Additional labels of PersistentVolumeClaim                      | `{}`                                             |
          | `storage.persistentVolume.annotations`                    | Additional annotations of PersistentVolumeClaim                 | `{}`                                             |
          | `init.labels`                                             | Additional labels of init Job and its Pod                       | `{"app.kubernetes.io/component": "init"}`        |
          | `init.annotations`                                        | Additional labels of the Pod of init Job                        | `{}`                                             |
          | `init.affinity`                                           | [Affinity rules][2] of init Job Pod                             | `{}`                                             |
          | `init.nodeSelector`                                       | Node labels for init Job Pod assignment                         | `{}`                                             |
          | `init.tolerations`                                        | Node taints to tolerate by init Job Pod                         | `[]`                                             |
          | `init.resources`                                          | Resource requests and limits for the Pod of init Job            | `{}`                                             |
          | `tls.enabled`                                             | Whether to run securely using TLS certificates                  | `no`                                             |
          | `tls.serviceAccount.create`                               | Whether to create a new RBAC service account                    | `yes`                                            |
          | `tls.serviceAccount.name`                                 | Name of RBAC service account to use                             | `""`                                             |
          | `tls.certs.provided`                                      | Bring your own certs scenario, i.e certificates are provided    | `no`                                             |
          | `tls.certs.clientRootSecret`                              | If certs are provided, secret name for client root cert         | `cockroachdb-root`                               |
          | `tls.certs.nodeSecret`                                    | If certs are provided, secret name for node cert                | `cockroachdb-node`                               |
          | `tls.certs.tlsSecret`                                     | Own certs are stored in TLS secret                              | `no`                                             |
          | `tls.init.image.repository`                               | Image to use for requesting TLS certificates                    | `cockroachdb/cockroach-k8s-request-cert`         |
          | `tls.init.image.tag`                                      | Image tag to use for requesting TLS certificates                | `0.4`                                            |
          | `tls.init.image.pullPolicy`                               | Requesting TLS certificates container pull policy               | `IfNotPresent`                                   |
          | `tls.init.image.credentials`                              | `registry`, `user` and `pass` credentials to pull private image | `{}`                                             |
          | `networkPolicy.enabled`                                   | Enable NetworkPolicy for CockroachDB's Pods                     | `no`                                             |
          | `networkPolicy.ingress.grpc`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |
          | `networkPolicy.ingress.http`                              | Whitelist resources to access gRPC port of CockroachDB's Pods   | `[]`                                             |


          Override the default parameters using the `CockroachDB` custom resource.

          ## Deep dive

          ### Connecting to the CockroachDB cluster

          Once you've created the cluster, you can start talking to it by connecting to its `-public` Service. CockroachDB is PostgreSQL wire protocol compatible, so there's a [wide variety of supported clients](https://www.cockroachlabs.com/docs/install-client-drivers.html). As an example, we'll open up a SQL shell using CockroachDB's built-in shell and play around with it a bit, like this (likely needing to replace `my-release-cockroachdb-public` with the name of the `-public` Service that was created with your installed chart):

          ```shell
          kubectl run cockroach-client --rm -it \
          --image=cockroachdb/cockroach \
          --restart=Never \
          -- sql --insecure --host my-release-cockroachdb-public
          ```

          ```
          Waiting for pod default/cockroach-client to be running, status is Pending,
          pod ready: false
          If you don't see a command prompt, try pressing enter.
          root@my-release-cockroachdb-public:26257> SHOW DATABASES;
          +--------------------+
          |      Database      |
          +--------------------+
          | information_schema |
          | pg_catalog         |
          | system             |
          +--------------------+
          (3 rows)
          root@my-release-cockroachdb-public:26257> CREATE DATABASE bank;
          CREATE DATABASE
          root@my-release-cockroachdb-public:26257> CREATE TABLE bank.accounts (id INT
          PRIMARY KEY, balance DECIMAL);
          CREATE TABLE
          root@my-release-cockroachdb-public:26257> INSERT INTO bank.accounts VALUES
          (1234, 10000.50);
          INSERT 1
          root@my-release-cockroachdb-public:26257> SELECT * FROM bank.accounts;
          +------+---------+
          |  id  | balance |
          +------+---------+
          | 1234 | 10000.5 |
          +------+---------+
          (1 row)
          root@my-release-cockroachdb-public:26257> \q
          Waiting for pod default/cockroach-client to terminate, status is Running
          pod "cockroach-client" deleted
          ```

          > If you are running in secure mode, you will have to provide a client certificate to the cluster in order to authenticate, so the above command will not work. See [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/client-secure.yaml) for an example of how to set up an interactive SQL shell against a secure cluster or [here](https://github.com/cockroachdb/cockroach/blob/master/cloud/kubernetes/example-app-secure.yaml) for an example application connecting to a secure cluster.

          ### Cluster health

          Because our pod spec includes regular health checks of the CockroachDB processes, simply running `kubectl get pods` and looking at the `STATUS` column is sufficient to determine the health of each instance in the cluster.

          If you want more detailed information about the cluster, the best place to look is the Admin UI.

          ### Accessing the Admin UI

          If you want to see information about how the cluster is doing, you can try pulling up the CockroachDB Admin UI by port-forwarding from your local machine to one of the pods (replacing `my-release-cockroachdb-0` with the name of one of your pods:

          ```shell
          kubectl port-forward my-release-cockroachdb-0 8080
          ```

          You should then be able to access the Admin UI by visiting <http://localhost:8080/> in your web browser.

          ### Failover

          If any CockroachDB member fails, it is restarted or recreated automatically by the Kubernetes infrastructure, and will re-join the cluster automatically when it comes back up. You can test this scenario by killing any of the CockroachDB pods:

          ```shell
          kubectl delete pod my-release-cockroachdb-1
          ```

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          After a while:

          ```shell
          kubectl get pods -l "app.kubernetes.io/instance=my-release,app.kubernetes.io/component=cockroachdb"
          ```

          ```
          NAME                      READY     STATUS        RESTARTS   AGE
          my-release-cockroachdb-0  1/1       Running       0          5m
          my-release-cockroachdb-1  1/1       Running       0          20s
          my-release-cockroachdb-2  1/1       Running       0          5m
          ```

          You can check the state of re-joining from the new pod's logs:

          ```shell
          kubectl logs my-release-cockroachdb-1
          ```

          ```
          [...]
          I161028 19:32:09.754026 1 server/node.go:586  [n1] node connected via gossip and
          verified as part of cluster {"35ecbc27-3f67-4e7d-9b8f-27c31aae17d6"}
          [...]
          cockroachdb-0.my-release-cockroachdb.default.svc.cluster.local:26257
          build:      beta-20161027-55-gd2d3c7f @ 2016/10/28 19:27:25 (go1.7.3)
          admin:      http://0.0.0.0:8080
          sql:
          postgresql://root@my-release-cockroachdb-1.my-release-cockroachdb.default.svc.cluster.local:26257?sslmode=disable
          logs:       cockroach-data/logs
          store[0]:   path=cockroach-data
          status:     restarted pre-existing node
          clusterID:  {35ecbc27-3f67-4e7d-9b8f-27c31aae17d6}
          nodeID:     2
          [...]
          ```

          ### NetworkPolicy

          To enable NetworkPolicy for CockroachDB, install [a networking plugin that implements the Kubernetes NetworkPolicy spec](https://kubernetes.io/docs/tasks/administer-cluster/declare-network-policy#before-you-begin), and set `networkPolicy.enabled` to `yes`/`true`.

          For Kubernetes v1.5 & v1.6, you must also turn on NetworkPolicy by setting the `DefaultDeny` Namespace annotation. Note: this will enforce policy for _all_ pods in the Namespace:

          ```shell
          kubectl annotate namespace default "net.beta.kubernetes.io/network-policy={\"ingress\":{\"isolation\":\"DefaultDeny\"}}"
          ```

          For more precise policy, set `networkPolicy.ingress.grpc` and `networkPolicy.ingress.http` rules. This will only allow pods that match the provided rules to connect to CockroachDB.

          ### Scaling

          Scaling should be managed via the `helm upgrade` command. After resizing your cluster on your cloud environment (e.g., GKE or EKS), run the following command to add a pod. This assumes you scaled from 3 to 4 nodes:

          ```shell
          kubectl patch cockroachdbs example -p '{"spec": {"statefulset":{"replicas":"4"}}}' --type merge
          ```

          Note, that if you are running in secure mode (`tls.enabled` is `yes`/`true`) and increase the size of your cluster, you will also have to approve the CSR (certificate-signing request) of each new node (using `kubectl get csr` and `kubectl certificate approve`).

          [1]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [2]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#node-affinity
          [3]: https://cert-manager.io/
          [4]: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
          [5]: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/
        displayName: CockroachDB Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cockroach
        - cockroachdb
        links:
        - name: Helm Chart Source
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb
        - name: Configuration Options
          url: https://github.com/helm/charts/tree/master/stable/cockroachdb#configuration
        - name: CockroachDB Source
          url: https://github.com/cockroachdb/cockroach
        maintainers:
        - email: yandu@cockroachlabs.com
          name: Yandu Oppacher
        maturity: stable
        minKubeVersion: 1.8.0
        provider:
          name: Helm Community
          url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/cockroachdb/cockroach-helm-operator:6.0.0
        - quay.io/openshift-community-operators/cockroachdb@sha256:d3016b1507515fc7712f9c47fd9082baf9ccb070aaab58ed0ef6e5abdedde8ba
        version: 6.0.0
      entries:
      - name: cockroachdb.v6.0.0
        version: 6.0.0
      name: stable-v6.x
    defaultChannel: stable-v6.x
    packageName: cockroachdb
    provider:
      name: Helm Community
      url: https://artifacthub.io/packages/helm/cockroachdb/cockroachdb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netobserv-operator.v1.6.3-community
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:1.6.2-community
          createdAt: "2024-08-22T18:26:56Z"
          description: Network flows collector and monitoring solution
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/1.6.2-community/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |-
          DEPRECATED: this distribution channel is deprecated and doesn't receive new updates. Please read https://github.com/orgs/netobserv/discussions/996 to learn more.

          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline that consists in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - when used in OpenShift, a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Prometheus metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.6.2-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: '[DEPRECATED] NetObserv Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/flowlogs-pipeline:v1.6.2-community
        - quay.io/netobserv/netobserv-ebpf-agent:v1.6.2-community
        - quay.io/netobserv/network-observability-console-plugin:v1.6.2-community
        - quay.io/netobserv/network-observability-operator:1.6.2-community
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/netobserv-operator@sha256:73af4040707ce79baa3115deed1e9148436b248ef3570f1175a6d31fdf9596c4
        version: 1.6.3-community
      entries:
      - name: netobserv-operator.v1.6.3-community
        version: 1.6.3-community
      - name: netobserv-operator.v1.6.2-community
        version: 1.6.2-community
      - name: netobserv-operator.v1.6.1-community
        version: 1.6.1-community
      - name: netobserv-operator.v1.6.0-community
        version: 1.6.0-community
      - name: netobserv-operator.v1.5.0-community
        version: 1.5.0-community
      name: community
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/netobserv/network-observability-operator:0.2.1
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: latest
    - currentCSV: netobserv-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "ipfix": {
                      "cacheActiveTimeout": "20s",
                      "cacheMaxFlows": 400,
                      "clusterNetworkOperator": {
                        "namespace": "openshift-network-operator"
                      },
                      "ovnKubernetes": {
                        "containerName": "ovnkube-node",
                        "daemonSetName": "ovnkube-node",
                        "namespace": "ovn-kubernetes"
                      },
                      "sampling": 400
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "staticLabels": {
                      "app": "netobserv-flowcollector"
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "enableKubeProbes": true,
                    "healthPort": 8080,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: quay.io/netobserv/network-observability-operator:0.2.1
          createdAt: "2022-12-07T09:00:48Z"
          description: Network flows collector and monitoring solution
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
        description: |-
          NetObserv Operator is an OpenShift / Kubernetes operator for network observability. It deploys a monitoring pipeline to collect and enrich network flows. These flows can be produced by the NetObserv eBPF agent, or by any device or CNI able to export flows in IPFIX format, such as OVN-Kubernetes.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When used in OpenShift, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For a normal usage, we recommend two options:

          - Installing the [Loki Operator](https://loki-operator.dev/docs/prologue/quickstart.md/). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          kubectl create namespace netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          kubectl apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionnaly be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/0.2.1/config/samples/flows_v1alpha1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          kubectl edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Agent (`spec.agent.type`) can be `EBPF` (default) or `IPFIX`. eBPF is recommended, as it should work in more situations and offers better performances. If you can't, or don't want to use eBPF, note that the IPFIX option is fully functional only when using [OVN-Kubernetes](https://github.com/ovn-org/ovn-kubernetes/) CNI. Other CNIs are not officially supported, but you may still be able to configure them manually if they allow IPFIX exports.

          - Sampling (`spec.agent.ebpf.sampling` and `spec.agent.ipfix.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50) for eBPF and 400 (1:400) for IPFIX. Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/0.2.1/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub more more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - A [F.A.Q](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: NetObserv Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: mmaciasl@redhat.com
          name: Mario Macias
        - email: ocazade@redhat.com
          name: Olivier Cazade
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/netobserv-operator@sha256:6bc3675bcccd3276a7b1f6627e4409bf61bac2b4a43771431c9b5a0232cbf5ce
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/netobserv/flowlogs-pipeline:v0.1.5
        - quay.io/netobserv/netobserv-ebpf-agent:v0.2.2
        - quay.io/netobserv/network-observability-console-plugin:v0.1.6
        - quay.io/netobserv/network-observability-operator:0.2.1
        version: 0.2.1
      entries:
      - name: netobserv-operator.v0.2.1
        version: 0.2.1
      - name: netobserv-operator.v0.2.0
        version: 0.2.0
      - name: netobserv-operator.v0.1.4
        version: 0.1.4
      - name: netobserv-operator.v0.1.3
        version: 0.1.3
      - name: netobserv-operator.v0.1.2
        version: 0.1.2
      - name: netobserv-operator.v0.1.1
        version: 0.1.1
      - name: netobserv-operator.v0.1.0
        version: 0.1.0
      name: v0.2.x
    defaultChannel: community
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v0.17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/node-maintenance-operator:v0.17.0
          createdAt: "2024-03-13 18:56:47"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=0.12.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/node-maintenance-operator@sha256:80dfaab6f7a5c34cebb45c874ce134bf60e3697e14a95b69f1972dd98ecc172a
        - quay.io/medik8s/node-maintenance-operator:v0.17.0
        version: 0.17.0
      entries:
      - name: node-maintenance-operator.v0.17.0
        version: 0.17.0
      - name: node-maintenance-operator.v0.16.1
        version: 0.16.1
      - name: node-maintenance-operator.v0.16.0
        version: 0.16.0
      - name: node-maintenance-operator.v0.15.0
        version: 0.15.0
      - name: node-maintenance-operator.v0.14.0
        version: 0.14.0
      - name: node-maintenance-operator.v0.13.0
        version: 0.13.0
      - name: node-maintenance-operator.v0.12.1
        version: 0.12.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: gcp-filestore-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure GCP Filestore CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=4.11.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/gcp-filestore-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Operator that installs and configures the CSI driver for GCP Filestore.
        displayName: GCP Filestore CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - filestore
        links:
        - name: Documentation
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/gcp-filestore-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:75eaf7b8b71e1a15e002f6cc5430e5e50039699d7b52c3150f5f71b1595f00c5
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9-operator@sha256:38e31381ede981be8a28316e7aeefc15d77316e54c214dcb69176d01f49c5fa0
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-operator-bundle@sha256:41ece1ea773fd610cf8a57483a4d2846e0bb3ce81db701021a3e0dd91e13b7c8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-gcp-filestore-csi-driver-rhel9@sha256:f7159e880ec55da49ac774db0c454800055e7cd3f37ba94fbcb629d08aac1146
        version: 4.18.0-202509100344
      entries:
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: gcp-filestore-csi-driver-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: gcp-filestore-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Devfile
      provider-url: https://devfile.io
    name: devworkspace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: devworkspace-operator.v0.36.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/internal-objects: '["devworkspaceroutings.controller.devfile.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DevWorkspaceOperatorConfig
            name: devworkspaceoperatorconfigs.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha2
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha1
          - kind: DevWorkspaceTemplate
            name: devworkspacetemplates.workspace.devfile.io
            version: v1alpha2
        description: |
          The DevWorkspace Operator enables cluster-level support for the
          [Devfile 2.0 spec](https://docs.devfile.io), enabling static, reproducible
          configurations for launching cloud-based editors and IDEs in OpenShift and
          Kubernetes clusters.

          Leveraging the language server protocol and the Eclipse Theia web IDE, the
          DevWorkspace operator provides easy configuration of full development
          environments on the cloud with support for a wide variety of languages and
          technologies, including Go, Java, Typescript/Javascript, Python, and more.

          The DevWorkspace Operator is also used in the Web Terminal Operator to
          automatically provision Web Terminal environments.

          ## Installing the operator
          The DevWorkspace Operator can be installed directly from the OperatorHub UI and
          will be available in all namespaces on the cluster. DevWorkspace creation is
          driven by the DevWorkspace custom resource, which can be created in any
          namespace to provision a full development environment. To get started, browse
          the DevWorkspace [spec](https://devfile.io/docs/2.3.0/devfile-schema).

          Once a DevWorkspace is started, it can be accessed via the URL in its
          `.status.mainUrl` field.

          It's recommended to install the DevWorkspace Operator to the
          `openshift-operators` namespace for compatibility.

          ## Uninstalling the operator
          The DevWorkspace Operator utilizes finalizers on resources it creates and
          webhooks to restrict access to development resources. As a result, manual steps
          are required in uninstalling the operator. See the
          [documentation](https://github.com/devfile/devworkspace-operator/blob/main/docs/uninstall.md)
          for details.
        displayName: DevWorkspace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Devworkspace Operator
        - DevWorkspaces
        - Devfile
        links:
        - name: Devworkspace Operator
          url: https://github.com/devfile/devworkspace-operator
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Devfile
          url: https://devfile.io
        relatedImages:
        - registry.redhat.io/ubi9/ubi-minimal@sha256:a50731d3397a4ee28583f1699842183d4d24fadcc565c4688487af9ee4e13a44
        - registry.redhat.io/devworkspace/devworkspace-operator-bundle@sha256:4f8908ba97517826fda14cab3624ed42a6348fb96f0be4abf9cd2a7b68f7d3e9
        - registry.redhat.io/devworkspace/devworkspace-project-clone-rhel9@sha256:312a013201877cea03956e3db2ba19a3d7f842de2c3a65927c65c3e8d22a57fe
        - registry.redhat.io/devworkspace/devworkspace-rhel9-operator@sha256:83b6abf6e57703da1125e037ea9353c224602bddc3146e5b00df47f1b1976e9a
        version: 0.36.0
      entries:
      - name: devworkspace-operator.v0.36.0
        version: 0.36.0
      - name: devworkspace-operator.v0.35.1
        version: 0.35.1
      - name: devworkspace-operator.v0.34.0
        version: 0.34.0
      - name: devworkspace-operator.v0.33.0
        version: 0.33.0
      - name: devworkspace-operator.v0.32.1
        version: 0.32.1
      - name: devworkspace-operator.v0.31.2
        version: 0.31.2
      - name: devworkspace-operator.v0.31.1
        version: 0.31.1
      - name: devworkspace-operator.v0.31.0
        version: 0.31.0
      - name: devworkspace-operator.v0.30.2
        version: 0.30.2
      - name: devworkspace-operator.v0.30.1
        version: 0.30.1
      - name: devworkspace-operator.v0.30.0
        version: 0.30.0
      - name: devworkspace-operator.v0.29.0
        version: 0.29.0
      - name: devworkspace-operator.v0.28.0-0.1719314743.p
        version: 0.28.0+0.1719314743.p
      - name: devworkspace-operator.v0.28.0
        version: 0.28.0
      - name: devworkspace-operator.v0.27.0-0.1714987663.p
        version: 0.27.0+0.1714987663.p
      - name: devworkspace-operator.v0.27.0
        version: 0.27.0
      - name: devworkspace-operator.v0.26.0
        version: 0.26.0
      - name: devworkspace-operator.v0.25.0
        version: 0.25.0
      - name: devworkspace-operator.v0.24.0
        version: 0.24.0
      - name: devworkspace-operator.v0.23.0
        version: 0.23.0
      - name: devworkspace-operator.v0.22.0
        version: 0.22.0
      - name: devworkspace-operator.v0.21.2
        version: 0.21.2
      - name: devworkspace-operator.v0.21.1
        version: 0.21.1
      - name: devworkspace-operator.v0.20.0
        version: 0.20.0
      - name: devworkspace-operator.v0.19.1
        version: 0.19.1
      - name: devworkspace-operator.v0.19.1-0.1682321189.p
        version: 0.19.1+0.1682321189.p
      - name: devworkspace-operator.v0.19.1-0.1679521112.p
        version: 0.19.1+0.1679521112.p
      - name: devworkspace-operator.v0.18.1
        version: 0.18.1
      - name: devworkspace-operator.v0.18.1-0.1675929565.p
        version: 0.18.1+0.1675929565.p
      - name: devworkspace-operator.v0.17.0
        version: 0.17.0
      - name: devworkspace-operator.v0.16.0-0.1666668361.p
        version: 0.16.0+0.1666668361.p
      - name: devworkspace-operator.v0.16.0
        version: 0.16.0
      - name: devworkspace-operator.v0.15.2-0.1661828401.p
        version: 0.15.2+0.1661828401.p
      - name: devworkspace-operator.v0.15.2
        version: 0.15.2
      - name: devworkspace-operator.v0.14.1
        version: 0.14.1
      - name: devworkspace-operator.v0.13.0
        version: 0.13.0
      - name: devworkspace-operator.v0.12.0
        version: 0.12.0
      - name: devworkspace-operator.v0.11.0
        version: 0.11.0
      - name: devworkspace-operator.v0.10.0
        version: 0.10.0
      - name: devworkspace-operator.v0.9.0
        version: 0.9.0
      name: fast
    defaultChannel: fast
    packageName: devworkspace-operator
    provider:
      name: Devfile
      url: https://devfile.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-migration
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-migration.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Migration",
                "metadata": {
                  "name": "migration-sample"
                },
                "spec": {
                  "auth": {
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configuration": {
                    "port": 8080
                  },
                  "database": {
                    "image": {
                      "name": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "esinfo": {
                    "batchSize": 10000,
                    "inputIndex": "wireless_stats"
                  },
                  "image": {
                    "remoteRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:36:21Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Migration
            name: migrations.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Migration Operator
        displayName: Keysight WAP Migration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Migration
          url: https://keysight-wap-migration.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-wap-migration-bundle@sha256:3deff8b48ab1a6f0e5fd0f8315bc1e191da55525cb3cf0f315f2f7afdb47a6c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-migration@sha256:118b466f1c59832a3620fa9b153507cc66ba38ee08b337b3a1a4225b158bebc5
        version: 0.1.21
      entries:
      - name: keysight-wap-migration.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-migration
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco
      provider-url: ""
    name: aci-containers-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aci-containers-operator.v1.0.10
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aci.ctrl/v1alpha1","kind":"AciContainersOperator","metadata":{"name":"acicnioperator","namespace":"aci-containers-system"},"spec":{}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
          createdAt: "2025-04-10T12:00:00Z"
          description: The Cisco Application Centric Infrastructure (ACI) Containers
            Operator brings the ACI CNI Plugin and Policy model to Kubernetes clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://github.com/noironetworks/aci-containers/blob/master/README.md
          repository: https://github.com/noironetworks/aci-containers
          support: acisupport@cisco.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aci Containers Operator
            displayName: AciContainersOperator
            kind: AciContainersOperator
            name: acicontainersoperators.aci.ctrl
            version: v1alpha1
        description: |
          ## ACI CNI Plugin

          The Cisco Application Centric Infrastructure (ACI) CNI plugin brings the ACI Networking and Policy model to Kubernetes clusters that reside on-prem or in the cloud. It is fully open source and relies on the Opflex Protocol to program Open vSwitch instances running on the Kubernetes nodes. It provides IP address management, L2/L3 networking, load balancing, and security functions for container workloads.

          The Cisco ACI CNI plugin for the Red Hat OpenShift Container Platform provides a single, programmable network infrastructure that eliminates bottlenecks between OpenShift workloads and the rest of the infrastructure while providing enterprise-grade security and offers flexible micro segmentation possibilities.

          Unified Networking - By extending the integrated overlay of the ACI fabric to OpenShift workloads, Pod and Service endpoints are managed as native ACI network resources at the same level as Bare Metal or Virtual Machines.

          No Egress Router Required - Applications running in OpenShift Pods that need high-bandwidth, low-latency traffic for data transfers external to the cluster suffer the bottleneck imposed by the egress router implementation. Pods can easily connect with other fabric or non-fabric endpoints leveraging the combined hardware and software distributed routing and switching provided by ACI and the ACI CNI plugin.

          Simplified Operations and Enhanced Visibility - Whether policies and networking are defined by the devops or fabric or security admins, the APIC becomes the single source of truth for operations and security governance. Kubernetes workload visibility is provided to Network Operators within their familiar Cisco tools.

          Hardware-assisted Load Balancing - Ingress connections to LoadBalancer-type services using ACI’s Policy Based Redirect technology moves load balancing into the fabric's hardware.

          Granular Security - Security can be implemented by using native NetworkPolicy or by using ACI EPGs and contracts, or both models complementing each other. Network policy of container workloads becomes a shared responsibility while ensuring global compliance and tenant constraints are enforced.
        displayName: Cisco ACI Containers Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cisco
        - aci
        links:
        - name: Cisco ACI Homepage
          url: https://www.cisco.com/c/en/us/td/docs/switches/datacenter/aci/apic/sw/kb/b_Kubernetes_Integration_with_ACI.html
        maintainers:
        - email: abhishs9@cisco.com
          name: 'Abhishek Sharma '
        - email: sumitnaiksatam@gmail.com
          name: Sumit Naiksatam
        - email: vlella@cisco.com
          name: Vishal Lella
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: Cisco
        relatedImages:
        - quay.io/noiro/acc-provision-operator@sha256:9c8fa6d705602af13d539fb919e72e7ee5064c6998fc9ca95d10108e986c68fe
        - quay.io/noiro/aci-containers-operator@sha256:2b2d685e254c206be5371285852a8f95d0318af69cb85b786427969e035c125c
        - registry.connect.redhat.com/noiro/aci-operator-bundle@sha256:80e25268ab5aa6ea418deae5ccc197fbffe2b9cdf4f27894f5beeb9315f6b123
        version: 1.0.10
      entries:
      - name: aci-containers-operator.v1.0.10
        version: 1.0.10
      name: stable
    defaultChannel: stable
    packageName: aci-containers-operator
    provider:
      name: Cisco
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Datadog
      provider-url: ""
    name: datadog-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datadog-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadog"
                },
                "spec": {
                  "agent": {
                    "apm": {
                      "enabled": false
                    },
                    "log": {
                      "enabled": false,
                      "logsConfigContainerCollectAll": false
                    },
                    "process": {
                      "enabled": true,
                      "processCollection": false
                    },
                    "security": {
                      "compliance": {
                        "enabled": false
                      },
                      "runtime": {
                        "enabled": false
                      }
                    },
                    "systemProbe": {
                      "bpfDebugEnabled": false
                    }
                  },
                  "clusterAgent": {
                    "config": {
                      "admissionController": {
                        "enabled": false
                      },
                      "clusterChecksEnabled": false,
                      "externalMetrics": {
                        "enabled": false
                      }
                    },
                    "replicas": 2
                  },
                  "clusterChecksRunner": {
                    "replicas": 2
                  },
                  "clusterName": "\u003cCLUSTER_NAME\u003e",
                  "credentials": {
                    "apiKey": "\u003cDATADOG_API_KEY\u003e",
                    "appKey": "\u003cDATADOG_APP_KEY\u003e",
                    "token": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e"
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          olm.skipRange: <1.0.4
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/DataDog/datadog-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgent Deployment with Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:48b37460dd7245cf661009267c3353962f0f195efd6a3007c9cc70b51b91a9e3
        - registry.connect.redhat.com/datadog/operator@sha256:e1fb36e82b5438e3efd5768425651beadeb80e8698e0e0ca475d2a2e23386090
        version: 1.0.4
      entries:
      - name: datadog-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: datadog-operator.v1.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogAgentProfile",
                "metadata": {
                  "name": "datadogagentprofile-sample"
                },
                "spec": {
                  "config": {
                    "override": {
                      "nodeAgent": {
                        "containers": {
                          "agent": {
                            "resources": {
                              "requests": {
                                "cpu": "256m"
                              }
                            }
                          }
                        }
                      }
                    }
                  },
                  "profileAffinity": {
                    "profileNodeAffinity": [
                      {
                        "key": "kubernetes.io/os",
                        "operator": "In",
                        "values": [
                          "linux"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogDashboard",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "datadog-operator",
                    "app.kubernetes.io/instance": "datadogdashboard-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datadogdashboard",
                    "app.kubernetes.io/part-of": "datadog-operator"
                  },
                  "name": "datadogdashboard-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogGenericResource",
                "metadata": {
                  "name": "datadoggenericresource-notebook-sample"
                },
                "spec": {
                  "jsonSpec": "{\n  \"data\": {\n    \"attributes\": {\n      \"cells\": [\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"text\": \"## Some test markdown\\n\\n```js\\nvar x, y;\\nx = 5;\\ny = 6;\\n```\",\n              \"type\": \"markdown\"\n            }\n          },\n          \"type\": \"notebook_cells\"\n        },\n        {\n          \"attributes\": {\n            \"definition\": {\n              \"requests\": [\n                {\n                  \"display_type\": \"line\",\n                  \"q\": \"avg:system.load.1{*}\",\n                  \"style\": {\n                    \"line_type\": \"solid\",\n                    \"line_width\": \"normal\",\n                    \"palette\": \"dog_classic\"\n                  }\n                }\n              ],\n              \"show_legend\": true,\n              \"type\": \"timeseries\",\n              \"yaxis\": {\n                \"scale\": \"linear\"\n              }\n            },\n            \"graph_size\": \"m\",\n            \"split_by\": {\n              \"keys\": [],\n              \"tags\": []\n            },\n            \"time\": null\n          },\n          \"type\": \"notebook_cells\"\n        }\n      ],\n      \"name\": \"Example-Notebook\",\n      \"status\": \"published\",\n      \"time\": {\n        \"live_span\": \"1h\"\n      }\n    },\n    \"type\": \"notebooks\"\n  }\n}",
                  "type": "notebook"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMetric",
                "metadata": {
                  "name": "datadogmetric-sample"
                },
                "spec": {
                  "query": "avg:kubernetes.cpu.usage.total{app:foo}.rollup(avg,30)/(avg:kubernetes.cpu.limits{app:foo}.rollup(avg,30)*10000000)"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogMonitor",
                "metadata": {
                  "name": "datadogmonitor-sample"
                },
                "spec": {
                  "message": "Something is wrong and we need to fix it.",
                  "query": "avg(last_15m):avg:foo{env:staging,service:bar} \u003e 1",
                  "tags": [
                    "env:staging",
                    "service:bar"
                  ],
                  "title": "Latency is increasing on staging",
                  "type": "metric alert"
                }
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogPodAutoscaler",
                "metadata": {
                  "name": "datadogpodautoscaler-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "datadoghq.com/v1alpha1",
                "kind": "DatadogSLO",
                "metadata": {
                  "name": "datadogslo-sample"
                },
                "spec": {
                  "description": "This is an example metric SLO from datadog-operator",
                  "name": "datadogslo-sample",
                  "query": {
                    "denominator": "sum:requests.total{service:example,env:prod}.as_count()",
                    "numerator": "sum:requests.success{service:example,env:prod}.as_count()"
                  },
                  "tags": [
                    "service:example",
                    "env:prod"
                  ],
                  "targetThreshold": "99.9",
                  "timeframe": "7d",
                  "type": "metric"
                }
              },
              {
                "apiVersion": "datadoghq.com/v2alpha1",
                "kind": "DatadogAgent",
                "metadata": {
                  "name": "datadogagent-sample"
                },
                "spec": {
                  "features": {
                    "admissionController": {
                      "enabled": false
                    },
                    "apm": {
                      "enabled": false
                    },
                    "clusterChecks": {
                      "enabled": true,
                      "useClusterChecksRunners": true
                    },
                    "liveProcessCollection": {
                      "enabled": false
                    },
                    "logCollection": {
                      "containerCollectAll": true,
                      "enabled": true
                    }
                  },
                  "global": {
                    "clusterAgentToken": "\u003cDATADOG_CLUSTER_AGENT_TOKEN\u003e",
                    "clusterName": "\u003cCLUSTER_NAME\u003e",
                    "credentials": {
                      "apiKey": "\u003cDATADOG_API_KEY\u003e",
                      "appKey": "\u003cDATADOG_APP_KEY\u003e"
                    },
                    "criSocketPath": "/var/run/crio/crio.sock",
                    "kubelet": {
                      "tlsVerify": false
                    }
                  },
                  "override": {
                    "clusterAgent": {
                      "containers": {
                        "cluster-agent": {
                          "securityContext": {
                            "readOnlyRootFilesystem": false
                          }
                        }
                      },
                      "replicas": 2,
                      "serviceAccountName": "datadog-agent-scc"
                    },
                    "clusterChecksRunner": {
                      "replicas": 2
                    },
                    "nodeAgent": {
                      "hostNetwork": true,
                      "securityContext": {
                        "runAsUser": 0,
                        "seLinuxOptions": {
                          "level": "s0",
                          "role": "system_r",
                          "type": "spc_t",
                          "user": "system_u"
                        }
                      },
                      "serviceAccountName": "datadog-agent-scc"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring, Logging & Tracing
          containerImage: registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
          createdAt: "2025-09-08 08:57:37"
          description: Datadog provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with Datadog Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.7.0 <1.18.0'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/DataDog/datadog-operator
          support: Datadog, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DatadogAgentInternal is the Schema for the datadogagentinternals
              API
            displayName: Datadog Agent Internal
            kind: DatadogAgentInternal
            name: datadogagentinternals.datadoghq.com
            version: v1alpha1
          - description: DatadogAgentProfile is the Schema for the datadogagentprofiles
              API
            displayName: Datadog Agent Profile
            kind: DatadogAgentProfile
            name: datadogagentprofiles.datadoghq.com
            version: v1alpha1
          - description: DatadogAgent Deployment with the Datadog Operator.
            displayName: Datadog Agent
            kind: DatadogAgent
            name: datadogagents.datadoghq.com
            version: v2alpha1
          - description: DatadogDashboard is the Schema for the datadogdashboards
              API
            displayName: Datadog Dashboard
            kind: DatadogDashboard
            name: datadogdashboards.datadoghq.com
            version: v1alpha1
          - kind: DatadogGenericResource
            name: datadoggenericresources.datadoghq.com
            version: v1alpha1
          - description: DatadogMetric allows autoscaling on arbitrary Datadog query
            displayName: Datadog Metric
            kind: DatadogMetric
            name: datadogmetrics.datadoghq.com
            version: v1alpha1
          - description: DatadogMonitor allows to define and manage Monitors from
              your Kubernetes Cluster
            displayName: Datadog Monitor
            kind: DatadogMonitor
            name: datadogmonitors.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha1
          - description: DatadogPodAutoscaler is the Schema for the datadogpodautoscalers
              API
            displayName: Datadog Pod Autoscaler
            kind: DatadogPodAutoscaler
            name: datadogpodautoscalers.datadoghq.com
            version: v1alpha2
          - description: DatadogSLO allows a user to define and manage datadog SLOs
              from Kubernetes cluster.
            displayName: Datadog SLO
            kind: DatadogSLO
            name: datadogslos.datadoghq.com
            version: v1alpha1
        description: Datadog provides a modern monitoring and analytics platform.
          Gather metrics, logs and traces for full observability of your Kubernetes
          cluster with Datadog Operator.
        displayName: Datadog Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Datadog
        - Monitoring
        - Logs
        - Tracing
        links:
        - name: Documentation
          url: https://docs.datadoghq.com/agent/kubernetes/
        - name: Kubernetes Monitoring Info
          url: https://www.datadoghq.com/blog/tag/kubernetes/
        maintainers:
        - email: support@datadoghq.com
          name: Datadog Inc.
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Datadog
        relatedImages:
        - registry.connect.redhat.com/datadog/operator-bundle@sha256:9949a581df3960a770117b6cf6b9fee1c63ea12ca09ac2d677f1cbe0a83fbd93
        - registry.connect.redhat.com/datadog/operator@sha256:aef5961c6dcdb4a25e3261f09e704e645684fa26c6057a66f6d822f4eeafc77d
        version: 1.18.0
      entries:
      - name: datadog-operator.v1.18.0
        version: 1.18.0
      - name: datadog-operator.v1.17.0
        version: 1.17.0
      - name: datadog-operator.v1.16.0
        version: 1.16.0
      - name: datadog-operator.v1.15.1
        version: 1.15.1
      - name: datadog-operator.v1.14.0
        version: 1.14.0
      - name: datadog-operator.v1.13.0
        version: 1.13.0
      - name: datadog-operator.v1.12.1
        version: 1.12.1
      - name: datadog-operator.v1.11.1
        version: 1.11.1
      - name: datadog-operator.v1.10.0
        version: 1.10.0
      - name: datadog-operator.v1.9.0
        version: 1.9.0
      - name: datadog-operator.v1.8.0
        version: 1.8.0
      - name: datadog-operator.v1.7.0
        version: 1.7.0
      - name: datadog-operator.v1.6.0
        version: 1.6.0
      - name: datadog-operator.v1.5.0
        version: 1.5.0
      - name: datadog-operator.v1.4.0
        version: 1.4.0
      - name: datadog-operator.v1.3.0
        version: 1.3.0
      - name: datadog-operator.v1.2.1
        version: 1.2.1
      name: stable
    defaultChannel: stable
    packageName: datadog-operator-certified
    provider:
      name: Datadog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: akaris
      provider-url: https://github.com/andreaskaris/sosreport-operator
    name: sosreport-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosreport-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "support.openshift.io/v1alpha1",
                "kind": "Sosreport",
                "metadata": {
                  "name": "sosreport-sample"
                },
                "spec": {
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosreport is the Schema for the sosreports API
            displayName: Sosreport
            kind: Sosreport
            name: sosreports.support.openshift.io
            version: v1alpha1
        description: Operator that facilitates the generation of sosreports
        displayName: sosreport-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosreport
        - debugging
        - troubleshooting
        links:
        - name: Sosreport Operator
          url: https://github.com/andreaskaris/sosreport-operator
        maintainers:
        - email: ak.karis@gmail.com
          name: Andreas Karis
        maturity: alpha
        provider:
          name: akaris
          url: https://github.com/andreaskaris/sosreport-operator
        relatedImages:
        - quay.io/akaris/sosreport-operator:0.0.2
        - quay.io/openshift-community-operators/sosreport-operator@sha256:f0d28a3e2ac2e1cbd0c307797b728ad89ac33375135d72054a7b765e274e4bea
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        version: 0.0.2
      entries:
      - name: sosreport-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sosreport-operator
    provider:
      name: akaris
      url: https://github.com/andreaskaris/sosreport-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cambridge Semantics Inc.
      provider-url: ""
    name: anzounstructured-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzounstructured-operator.v2.0.101
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzounstructured.clusters.cambridgesemantics.com/v1",
                "kind": "AnzoUnstructured",
                "metadata": {
                  "name": "au01"
                },
                "spec": {
                  "auWorker": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-w",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2",
                                "name": "w",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  },
                  "msLeader": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app": "anzounstructured"
                          }
                        },
                        "serviceName": "au-au01-ms",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app": "anzounstructured"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286",
                                "name": "ms",
                                "resources": {
                                  "limits": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "serviceAccountName": "unstructured-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
          createdAt: "2021-08-02"
          description: Kubernetes Operator for Anzo Unstructured
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/anzounstructured-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Cambridge Semantics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoUnstructured is the Schema for the anzounstructureds
              API
            displayName: AnzoUnstructured Operator
            kind: AnzoUnstructured
            name: anzounstructureds.anzounstructured.clusters.cambridgesemantics.com
            version: v1
        description: |-
          The Anzo Unstructured Operator provides the way to install and configure an anzo unstructured setup on Red Hat K8S environment.
          Currently, this is possible only through existing Anzo installation.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/cloud-deployments.htm )

          ### Support

          We offer Support to our customers through [ Cambridge Semantics Customer Center ]( https://customercenter.cambridgesemantics.com/ ).
        displayName: Anzo Unstructured Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Database
        - Big Data
        links:
        - name: Anzo Unstructured Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo-unstructured/blob/v2.0.0/README.md
        maintainers:
        - email: frank@cambridgesemantics.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Cambridge Semantics Inc.
        relatedImages:
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-unstructured-worker@sha256:4db61bdf6dc8b42824a06d2f8d8aeecddde765d123b40a63a7bbf87845719fb2
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/unstructured-operator@sha256:a47af068a7aa55c12be84c80d263feeed5597ecc434986b4f2fff782b568bb04
        - registry.connect.redhat.com/cambridgesemantics/anzo-unstructured-operator-bundle@sha256:20a4f2c56cb186fff4e8d31021d84b690512cc65c3c9629ffe623a0213ca9b0a
        - registry.marketplace.redhat.com/rhm/cambridgesemantics/anzo-microservices-leader@sha256:0579fd97e81e436eaff3b58c5395f13aa924940a3d51f0bdea5163d327620286
        version: 2.0.0
      entries:
      - name: anzounstructured-operator.v2.0.101
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: anzounstructured-operator-rhmp
    provider:
      name: Cambridge Semantics Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: mariadb-operator
      provider-url: https://github.com/mariadb-operator/mariadb-operator
    name: mariadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mariadb-operator.v25.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.mariadb.mmontes.io/v1alpha1",
                "kind": "MariadbOperator",
                "metadata": {
                  "name": "mariadb-operator"
                },
                "spec": {
                  "affinity": {},
                  "certController": {
                    "affinity": {},
                    "caValidity": "35064h",
                    "certValidity": "8766h",
                    "enabled": true,
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "lookaheadValidity": "2160h",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "requeueDuration": "5m",
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  },
                  "clusterName": "cluster.local",
                  "extrArgs": [],
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "fullnameOverride": "",
                  "ha": {
                    "enabled": false,
                    "replicas": 3
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                    "tag": ""
                  },
                  "imagePullSecrets": [],
                  "logLevel": "INFO",
                  "metrics": {
                    "enabled": false,
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "podSecurityContext": {},
                  "rbac": {
                    "enabled": true
                  },
                  "resources": {},
                  "securityContext": {},
                  "serviceAccount": {
                    "annotations": {},
                    "automount": true,
                    "enabled": true,
                    "extraLabels": {},
                    "name": ""
                  },
                  "tolerations": [],
                  "webhook": {
                    "affinity": {},
                    "annotations": {},
                    "cert": {
                      "caPath": "/tmp/k8s-webhook-server/certificate-authority",
                      "certManager": {
                        "duration": "",
                        "enabled": false,
                        "issuerRef": {},
                        "renewBefore": ""
                      },
                      "path": "/tmp/k8s-webhook-server/serving-certs",
                      "secretAnnotations": {}
                    },
                    "extrArgs": [],
                    "extraVolumeMounts": [],
                    "extraVolumes": [],
                    "ha": {
                      "enabled": false,
                      "replicas": 3
                    },
                    "hostNetwork": false,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "ghcr.io/mariadb-operator/mariadb-operator",
                      "tag": ""
                    },
                    "imagePullSecrets": [],
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podSecurityContext": {},
                    "port": 10250,
                    "resources": {},
                    "securityContext": {},
                    "serviceAccount": {
                      "annotations": {},
                      "automount": true,
                      "enabled": true,
                      "extraLabels": {},
                      "name": ""
                    },
                    "serviceMonitor": {
                      "additionalLabels": {},
                      "enabled": true,
                      "interval": "30s",
                      "scrapeTimeout": "25s"
                    },
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup"
                },
                "spec": {
                  "args": [
                    "--single-transaction",
                    "--all-databases"
                  ],
                  "logLevel": "info",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "backups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Connection",
                "metadata": {
                  "name": "connection"
                },
                "spec": {
                  "database": "mariadb",
                  "healthCheck": {
                    "interval": "10s",
                    "retryInterval": "3s"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "params": {
                    "parseTime": "true"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "secretName": "connection",
                  "secretTemplate": {
                    "annotations": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "databaseKey": "database",
                    "hostKey": "host",
                    "key": "dsn",
                    "labels": {
                      "k8s.mariadb.com/connection": "sample"
                    },
                    "passwordKey": "password",
                    "portKey": "port",
                    "usernameKey": "username"
                  },
                  "serviceName": "mariadb",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Database",
                "metadata": {
                  "name": "data-test"
                },
                "spec": {
                  "characterSet": "utf8",
                  "collate": "utf8_general_ci",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "ExternalMariaDB",
                "metadata": {
                  "name": "external-mariadb"
                },
                "spec": {
                  "connection": {
                    "healthCheck": {
                      "interval": "5s"
                    },
                    "secretName": "external-mariadb"
                  },
                  "host": "mariadb.default.svc.cluster.local",
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "port": 3306,
                  "username": "root"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Grant",
                "metadata": {
                  "name": "grant"
                },
                "spec": {
                  "database": "*",
                  "grantOption": true,
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "privileges": [
                    "SELECT",
                    "INSERT",
                    "UPDATE"
                  ],
                  "requeueInterval": "30s",
                  "retryInterval": "5s",
                  "table": "*",
                  "username": "user"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MariaDB",
                "metadata": {
                  "name": "mariadb"
                },
                "spec": {
                  "connection": {
                    "secretName": "mariadb-conn",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "database": "mariadb",
                  "galera": {
                    "enabled": true
                  },
                  "metrics": {
                    "enabled": true,
                    "passwordSecretKeyRef": {
                      "generate": true,
                      "key": "password",
                      "name": "mariadb-metrics"
                    }
                  },
                  "myCnf": "[mariadb]\nbind-address=*\ndefault_storage_engine=InnoDB\nbinlog_format=row\ninnodb_autoinc_lock_mode=2\ninnodb_buffer_pool_size=1024M\nmax_allowed_packet=256M\n",
                  "passwordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-password"
                  },
                  "primaryConnection": {
                    "secretName": "mariadb-conn-primary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "primaryService": {
                    "type": "ClusterIP"
                  },
                  "replicas": 3,
                  "rootPasswordSecretKeyRef": {
                    "generate": true,
                    "key": "password",
                    "name": "mariadb-root"
                  },
                  "secondaryConnection": {
                    "secretName": "mariadb-conn-secondary",
                    "secretTemplate": {
                      "key": "dsn"
                    }
                  },
                  "secondaryService": {
                    "type": "ClusterIP"
                  },
                  "service": {
                    "type": "ClusterIP"
                  },
                  "storage": {
                    "size": "1Gi"
                  },
                  "updateStrategy": {
                    "type": "ReplicasFirstPrimaryLast"
                  },
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "MaxScale",
                "metadata": {
                  "name": "maxscale-galera"
                },
                "spec": {
                  "admin": {
                    "guiEnabled": true,
                    "port": 8989
                  },
                  "auth": {
                    "generate": true
                  },
                  "config": {
                    "sync": {
                      "database": "mysql",
                      "interval": "5s",
                      "timeout": "10s"
                    }
                  },
                  "connection": {
                    "port": 3306,
                    "secretName": "mxs-galera-conn"
                  },
                  "guiKubernetesService": {
                    "metadata": {
                      "annotations": {
                        "metallb.universe.tf/loadBalancerIPs": "172.18.0.231"
                      }
                    },
                    "type": "LoadBalancer"
                  },
                  "kubernetesService": {
                    "annotations": {
                      "metallb.universe.tf/loadBalancerIPs": "172.18.0.224"
                    },
                    "type": "LoadBalancer"
                  },
                  "mariaDbRef": {
                    "name": "mariadb-galera"
                  },
                  "monitor": {
                    "cooperativeMonitoring": "majority_of_all",
                    "interval": "2s",
                    "params": {
                      "available_when_donor": "false",
                      "disable_master_failback": "false",
                      "disable_master_role_setting": "false"
                    }
                  },
                  "replicas": 3,
                  "requeueInterval": "10s",
                  "services": [
                    {
                      "listener": {
                        "params": {
                          "connection_metadata": "tx_isolation=auto"
                        },
                        "port": 3306,
                        "protocol": "MariaDBProtocol"
                      },
                      "name": "rw-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "max_slave_connections": "255",
                        "transaction_replay": "true",
                        "transaction_replay_attempts": "10",
                        "transaction_replay_timeout": "5s"
                      },
                      "router": "readwritesplit"
                    },
                    {
                      "listener": {
                        "port": 3307
                      },
                      "name": "rconn-master-router",
                      "params": {
                        "master_accept_reads": "true",
                        "max_replication_lag": "3s",
                        "router_options": "master"
                      },
                      "router": "readconnroute"
                    },
                    {
                      "listener": {
                        "port": 3308
                      },
                      "name": "rconn-slave-router",
                      "params": {
                        "max_replication_lag": "3s",
                        "router_options": "slave"
                      },
                      "router": "readconnroute"
                    }
                  ]
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "PhysicalBackup",
                "metadata": {
                  "name": "physicalbackup"
                },
                "spec": {
                  "compression": "bzip2",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxRetention": "720h",
                  "podAffinity": true,
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "immediate": true,
                    "suspend": false
                  },
                  "serviceAccountName": "backup",
                  "stagingStorage": {
                    "persistentVolumeClaim": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "1Gi"
                        }
                      }
                    }
                  },
                  "storage": {
                    "s3": {
                      "accessKeyIdSecretKeyRef": {
                        "key": "access-key-id",
                        "name": "minio"
                      },
                      "bucket": "physicalbackups",
                      "endpoint": "minio.minio.svc.cluster.local:9000",
                      "prefix": "mariadb",
                      "region": "us-east-1",
                      "secretAccessKeySecretKeyRef": {
                        "key": "secret-access-key",
                        "name": "minio"
                      },
                      "tls": {
                        "caSecretKeyRef": {
                          "key": "ca.crt",
                          "name": "minio-ca"
                        },
                        "enabled": true
                      }
                    }
                  },
                  "timeout": "1h"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore"
                },
                "spec": {
                  "backupRef": {
                    "name": "backup"
                  },
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "resources": {
                    "limits": {
                      "cpu": "300m",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "targetRecoveryTime": "2023-12-19T09:00:00Z"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "SqlJob",
                "metadata": {
                  "name": "03-stars"
                },
                "spec": {
                  "database": "mariadb",
                  "dependsOn": [
                    {
                      "name": "01-users"
                    },
                    {
                      "name": "02-repos"
                    }
                  ],
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "schedule": {
                    "cron": "*/1 * * * *",
                    "suspend": false
                  },
                  "sql": "CREATE TABLE IF NOT EXISTS stars (\n  id bigint PRIMARY KEY AUTO_INCREMENT,\n  user_id bigint NOT NULL,\n  repo_id bigint NOT NULL,\n  FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE,\n  FOREIGN KEY (repo_id) REFERENCES repos(id) ON DELETE CASCADE,\n  UNIQUE KEY (user_id, repo_id)\n);\nINSERT INTO stars(user_id, repo_id) \n  VALUES((SELECT id FROM users ORDER BY RAND() LIMIT 1), (SELECT id FROM repos ORDER BY RAND() LIMIT 1))\n  ON DUPLICATE KEY UPDATE id=id;\nDELETE FROM stars WHERE id = (SELECT id FROM stars ORDER BY RAND() LIMIT 1);\nSELECT r.name AS repo, COUNT(*) AS stars\nFROM stars s\nJOIN repos r\nON s.repo_id = r.id\nGROUP BY r.id\nORDER BY stars DESC;",
                  "username": "mariadb"
                }
              },
              {
                "apiVersion": "k8s.mariadb.com/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "user"
                },
                "spec": {
                  "host": "%",
                  "mariaDbRef": {
                    "name": "mariadb"
                  },
                  "maxUserConnections": 20,
                  "passwordSecretKeyRef": {
                    "key": "password",
                    "name": "mariadb"
                  },
                  "retryInterval": "5s"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
          createdAt: "2025-09-19T13:24:35Z"
          description: Run and operate MariaDB in a cloud native way
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mariadb-operator/mariadb-operator
          support: mariadb-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configures a backup
            displayName: Backup
            kind: Backup
            name: backups.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a connection
            displayName: Connection
            kind: Connection
            name: connections.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a logical database
            displayName: Database
            kind: Database
            name: databases.k8s.mariadb.com
            version: v1alpha1
          - description: Configures an external MariaDB resources
            displayName: ExternalMariaDB
            kind: ExternalMariaDB
            name: externalmariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Grants permissions to an user in a database
            displayName: Grant
            kind: Grant
            name: grants.k8s.mariadb.com
            version: v1alpha1
          - description: Configures MariaDB helm chart based operator
            displayName: MariadbOperator
            kind: MariadbOperator
            name: mariadboperators.helm.mariadb.mmontes.io
            version: v1alpha1
          - description: Provisions a MariaDB instance
            displayName: MariaDB
            kind: MariaDB
            name: mariadbs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a MaxScale database proxy
            displayName: MaxScale
            kind: MaxScale
            name: maxscales.k8s.mariadb.com
            version: v1alpha1
          - description: Configures a physical backup
            displayName: PhysicalBackup
            kind: PhysicalBackup
            name: physicalbackups.k8s.mariadb.com
            version: v1alpha1
          - description: Restores a backup
            displayName: Restore
            kind: Restore
            name: restores.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a SQL job
            displayName: SqlJob
            kind: SqlJob
            name: sqljobs.k8s.mariadb.com
            version: v1alpha1
          - description: Defines a user
            displayName: User
            kind: User
            name: users.k8s.mariadb.com
            version: v1alpha1
        description: |
          Install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via [OLM](https://olm.operatorframework.io/) using the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator).

          This helm operator provides provides a 1:1 mapping between the official helm chart and the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, allowing to install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) via OLM without having to do any change in the helm chart.

          Normally, you would install [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) providing this `values.yaml` to the helm chart:
          ```yaml
          image:
            repository: ghcr.io/mariadb-operator/mariadb-operator
            pullPolicy: IfNotPresent
          logLevel: INFO
          ha:
            enabled: true
          metrics:
            enabled: true
            serviceMonitor:
              enabled: true
          webhook:
            cert:
              certManager:
                enabled: true
          ```

          This helm chart installation is abstracted in the [`MariadbOperator`](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/config/samples/helm_v1alpha1_mariadboperator.yaml) CRD, which will be reconciled by the helm operator:
          ```yaml
          apiVersion: helm.k8s.mariadb.com/v1alpha1
          kind: MariadbOperator
          metadata:
            name: mariadb-operator
          spec:
            image:
              repository: ghcr.io/mariadb-operator/mariadb-operator
              pullPolicy: IfNotPresent
            logLevel: INFO
            ha:
              enabled: true
            metrics:
              enabled: true
              serviceMonitor:
                enabled: true
            webhook:
              cert:
                certManager:
                  enabled: true
          ```

          Once you have installed the operator, you will able to install a [`MariaDB`](https://github.com/mariadb-operator/mariadb-operator/blob/main/examples/manifests/mariadb_v1alpha1_mariadb.yaml) instance. Refer to the main [`mariadb-operator`](https://github.com/mariadb-operator/mariadb-operator) documentation for getting started with the rest of CRDs.

          ## Documentation
          * [mariadb-operator](https://github.com/mariadb-operator/mariadb-operator/blob/main/README.md)
          * [mariadb-operator-helm](https://github.com/mariadb-operator/mariadb-operator-helm/blob/main/README.md)

          ## Releases
          This operator is automatically published in the following repositories whenever a new version of the [helm chart](https://artifacthub.io/packages/helm/mariadb-operator/mariadb-operator) is released:
          - [k8s-operatorhub/community-operators](https://github.com/k8s-operatorhub/community-operators)
          - [redhat-openshift-ecosystem/community-operators-prod](https://github.com/redhat-openshift-ecosystem/community-operators-prod)

          ## Roadmap
          Take a look at our [roadmap](https://github.com/mariadb-operator/mariadb-operator/blob/main/ROADMAP.md) and feel free to open an issue to suggest new features.

          ## Contributing
          We welcome and encourage contributions to this project! Please check our [contributing](https://github.com/mariadb-operator/mariadb-operator/blob/main/CONTRIBUTING.md) and [development](https://github.com/mariadb-operator/mariadb-operator/blob/main/docs/DEVELOPMENT.md) guides. PRs welcome!

          ## Get in touch
          Join us on Slack: **[MariaDB Community Slack](https://r.mariadb.com/join-community-slack)**.
        displayName: MariaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mariadb
        - mysql
        - operator
        - mariadb-operator
        - database
        - maxscale
        links:
        - name: GitHub
          url: https://github.com/mariadb-operator/mariadb-operator
        maintainers:
        - email: mariadb-operator@proton.me
          name: mmontes11
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: mariadb-operator
          url: https://github.com/mariadb-operator/mariadb-operator
        relatedImages:
        - ghcr.io/mariadb-operator/mariadb-operator-helm:25.8.4
        - quay.io/community-operator-pipeline-prod/mariadb-operator@sha256:3cbbff7ae75fb4f4629c7b8b58ca85ce67b23d5da558241301266d010e8b1188
        version: 25.8.4
      entries:
      - name: mariadb-operator.v25.8.4
        version: 25.8.4
      - name: mariadb-operator.v25.8.3
        version: 25.8.3
      - name: mariadb-operator.v25.8.2
        version: 25.8.2
      - name: mariadb-operator.v25.8.1
        version: 25.8.1
      - name: mariadb-operator.v25.8.0
        version: 25.8.0
      - name: mariadb-operator.v0.38.1
        version: 0.38.1
      - name: mariadb-operator.v0.38.0
        version: 0.38.0
      - name: mariadb-operator.v0.37.1
        version: 0.37.1
      - name: mariadb-operator.v0.37.0
        version: 0.37.0
      - name: mariadb-operator.v0.36.0
        version: 0.36.0
      - name: mariadb-operator.v0.35.1
        version: 0.35.1
      - name: mariadb-operator.v0.35.0
        version: 0.35.0
      - name: mariadb-operator.v0.34.0
        version: 0.34.0
      - name: mariadb-operator.v0.33.0
        version: 0.33.0
      - name: mariadb-operator.v0.32.0
        version: 0.32.0
      - name: mariadb-operator.v0.31.0
        version: 0.31.0
      - name: mariadb-operator.v0.30.0
        version: 0.30.0
      - name: mariadb-operator.v0.29.0
        version: 0.29.0
      - name: mariadb-operator.v0.28.1
        version: 0.28.1
      - name: mariadb-operator.v0.27.0
        version: 0.27.0
      - name: mariadb-operator.v0.26.0
        version: 0.26.0
      - name: mariadb-operator.v0.24.0
        version: 0.24.0
      - name: mariadb-operator.v0.23.1
        version: 0.23.1
      - name: mariadb-operator.v0.23.0
        version: 0.23.0
      - name: mariadb-operator.v0.22.0
        version: 0.22.0
      - name: mariadb-operator.v0.21.0
        version: 0.21.0
      - name: mariadb-operator.v0.20.0
        version: 0.20.0
      - name: mariadb-operator.v0.19.0
        version: 0.19.0
      - name: mariadb-operator.v0.18.0
        version: 0.18.0
      - name: mariadb-operator.v0.17.0
        version: 0.17.0
      - name: mariadb-operator.v0.16.0
        version: 0.16.0
      - name: mariadb-operator.v0.15.1
        version: 0.15.1
      - name: mariadb-operator.v0.15.0
        version: 0.15.0
      - name: mariadb-operator.v0.14.0
        version: 0.14.0
      - name: mariadb-operator.v0.13.0
        version: 0.13.0
      - name: mariadb-operator.v0.11.0
        version: 0.11.0
      - name: mariadb-operator.v0.10.0
        version: 0.10.0
      - name: mariadb-operator.v0.9.1
        version: 0.9.1
      - name: mariadb-operator.v0.9.0
        version: 0.9.0
      - name: mariadb-operator.v0.8.0
        version: 0.8.0
      - name: mariadb-operator.v0.7.1
        version: 0.7.1
      - name: mariadb-operator.v0.7.0
        version: 0.7.0
      - name: mariadb-operator.v0.6.7
        version: 0.6.7
      - name: mariadb-operator.v0.6.6
        version: 0.6.6
      - name: mariadb-operator.v0.6.5
        version: 0.6.5
      - name: mariadb-operator.v0.6.4
        version: 0.6.4
      name: alpha
    defaultChannel: alpha
    packageName: mariadb-operator
    provider:
      name: mariadb-operator
      url: https://github.com/mariadb-operator/mariadb-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: k8gb.io
      provider-url: https://github.com/k8gb-io/k8gb
    name: k8gb
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8gb.v0.11.4
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"k8gb.absa.oss/v1beta1","kind":"Gslb","metadata":{"name":"test-gslb-failover","namespace":"test-gslb"},"spec":{"ingress":{"rules":[{"host":"failover.test.k8gb.io","http":{"paths":[{"backend":{"service":{"name":"frontend-podinfo","port":{"name":"http"}}},"path":"/"}]}}]},"strategy":{"primaryGeoTag":"eu-west-1","type":"failover"}}}]'
          capabilities: Seamless Upgrades
          categories: Networking
          certified: "false"
          containerImage: docker.io/absaoss/k8gb:v0.11.4
          createdAt: "2021-09-24 12:00:00"
          description: A cloud native Kubernetes Global Balancer
          operatorframework.io/suggested-namespace: k8gb
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: k8gb
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/k8gb-io/k8gb
          support: cncf-k8gb-maintainers@lists.cncf.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gslb resource for global load balancing strategy configuration
            displayName: Gslb
            kind: Gslb
            name: gslbs.k8gb.absa.oss
            version: v1beta1
          - description: Using ExternalDNS it synchronizes exposed Kubernetes Services
              and Ingresses with DNS providers
            displayName: DNSEndpoint
            kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
        description: |
          **A cloud native Kubernetes Global Balancer**

          A Global Service Load Balancing solution with a focus on having cloud native qualities and work natively in a Kubernetes context.

          Key Differentiators:
           - Load balancing is based on timeproof DNS protocol which is perfect for global scope and extremely reliable
           - No dedicated management cluster and no single point of failure
           - Kubernetes native application health checks utilizing status of Liveness and Readiness probes for load balancing decisions
           - Configuration with a single Kubernetes CRD of `Gslb` kind

          Operator needs to be configured by setting couple of environment variables. This can be done by modifying
          the subscription yaml in the web ui or via kubectl. Here is an example of changing the `EDGE_DNS_ZONE`:

          ```yaml
          kind: Subscription
            ...
          spec:
            ...
            config:
              env:
              - name: CLUSTER_GEO_TAG
                value: us
              - name: EDGE_DNS_ZONE
                value: mycloud.example.com
                ...
           ```
           For more deployment variables see [helm code](https://github.com/k8gb-io/k8gb/blob/v0.8.3/chart/k8gb/templates/operator.yaml#L53:L68).

           For more information see [k8gb.io](https://k8gb.io).
        displayName: k8gb
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gslb
        - dns-lb
        - kuberneters-global-balancer
        - kubernetes-operator
        - balancer
        - multi-cluster
        links:
        - name: Home
          url: https://www.k8gb.io/
        - name: Source
          url: https://github.com/k8gb-io/k8gb
        maintainers:
        - email: dinar.valeev@absa.africa
          name: Dinar Valeev
        - email: jiri.kremser@gmail.com
          name: Jiri Kremser
        - email: kuritka@gmail.com
          name: Michal Kuritka
        - email: timofey.ilinykh@absa.africa
          name: Timofey Ilinykh
        - email: yury.tsarev@absa.africa
          name: Yury Tsarev
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: k8gb.io
          url: https://github.com/k8gb-io/k8gb
        relatedImages:
        - docker.io/absaoss/k8gb:v0.11.4
        - quay.io/openshift-community-operators/k8gb@sha256:9743f957b00e5ad344fed0c062d616d8222990139ac4afbf14871fd9813f0ef0
        - absaoss/k8s_crd:v0.0.11
        version: 0.11.4
      entries:
      - name: k8gb.v0.11.4
        version: 0.11.4
      name: alpha
    defaultChannel: alpha
    packageName: k8gb
    provider:
      name: k8gb.io
      url: https://github.com/k8gb-io/k8gb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Cisco Networking - Compute Solutions
      provider-url: ""
    name: ucs-ci-solutions-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ucs-ci-solutions-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intersight.cisco.com/v1alpha1",
                "kind": "UCSCISolutions",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ucs-ci-solutions-operator",
                    "app.kubernetes.io/instance": "ucscisolutions-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ucscisolutions",
                    "app.kubernetes.io/part-of": "ucs-ci-solutions-operator"
                  },
                  "name": "ucscisolutions-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: docker.io/pkoppa/ucs-ci-solutions-operator:latest
          createdAt: "2023-07-15T15:19:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: UCSCISolutions
            name: ucscisolutions.intersight.cisco.com
            version: v1alpha1
        description: The operator is used to automate the creation of  Pools, Policies
          and Server Profile Templates required to deploy Cisco UCS Converged Infrastructure
          Solution in Intersight.
        displayName: UCS-CI-Solutions-Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ucscisolutions
        - intersight
        - ucs
        - ci
        - flexpod
        - flashstack
        links:
        - name: Ucs Ci Solutions Operator
          url: https://ucs-ci-solutions-operator.domain
        maintainers:
        - email: pkoppa@cisco.com
          name: Paniraj Koppa
        - email: chrobrie@cisco.com
          name: Chris O'Brien
        maturity: alpha
        provider:
          name: Cisco Networking - Compute Solutions
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/ucs-ci-solutions-operator@sha256:e284253263183eb2cd313f6fb009bb830b9a79a4d4e2481468821c30a9cdce03
        - docker.io/pkoppa/ucs-ci-solutions-operator:latest
        version: 0.0.1
      entries:
      - name: ucs-ci-solutions-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ucs-ci-solutions-operator
    provider:
      name: Cisco Networking - Compute Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: secrets-store-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secrets-store-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
          createdAt: "2023-06-12T00:00:00Z"
          description: Install and configure Secrets Store CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.13.0-0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secrets-store-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecretProviderClass is the Schema for the secretproviderclasses
              API
            displayName: Secret Provider Class
            kind: SecretProviderClass
            name: secretproviderclasses.secrets-store.csi.x-k8s.io
            version: v1
          - description: SecretProviderClassPodStatus is the Schema for the secretproviderclassespodstatus
              API
            displayName: Secret Provider Class Pod Status
            kind: SecretProviderClassPodStatus
            name: secretproviderclasspodstatuses.secrets-store.csi.x-k8s.io
            version: v1
        description: |
          Operator that installs and configures the CSI driver for Secrets Store.
        displayName: Secrets Store CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: Documentation
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/secrets-store-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-operator-bundle@sha256:67723df19ab7d7871a6b639da3975c230d95dd250fa2995a9f44bdbb4b141454
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9-operator@sha256:d783af179a7f19cb9be613c26208094429206705cf6e341649b7d6d3849e214a
        - registry.redhat.io/openshift4/ose-secrets-store-csi-driver-rhel9@sha256:439aeb806c3264948a15e325f7b1e71037099a085c8d971646188ace77fcf5d3
        version: 4.18.0-202509100344
      entries:
      - name: secrets-store-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: secrets-store-csi-driver-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: secrets-store-csi-driver-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: secrets-store-csi-driver-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: secrets-store-csi-driver-operator.v4.18.0-202506241202
        version: 4.18.0-202506241202
      - name: secrets-store-csi-driver-operator.v4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: secrets-store-csi-driver-operator.v4.18.0-202505271635
        version: 4.18.0-202505271635
      - name: secrets-store-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: secrets-store-csi-driver-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: secrets-store-csi-driver-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: secrets-store-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: secrets-store-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: secrets-store-csi-driver-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: secrets-store-csi-driver-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: secrets-store-csi-driver-operator.v4.18.0-202502280802
        version: 4.18.0-202502280802
      - name: secrets-store-csi-driver-operator.v4.18.0-202502260503
        version: 4.18.0-202502260503
      - name: secrets-store-csi-driver-operator.v4.18.0-202502040032
        version: 4.18.0-202502040032
      name: stable
    defaultChannel: stable
    packageName: secrets-store-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-multicluster-orchestrator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-multicluster-orchestrator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-09T14:17:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:c5bd4f10120088f285a3b65e85bba825f9f55964758901f3b5827ac31516675b
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:26935de5f21a7865aba484599136df6a251c009e8fa9781e1040da3b19e7dd33
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:e05074d46e062268461d2689f7048c1a5dc4ed89a039d93e382f5ebc0726a269
        version: 4.17.13-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-multicluster-orchestrator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-multicluster-orchestrator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-multicluster-orchestrator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-multicluster-orchestrator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-multicluster-orchestrator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-multicluster-orchestrator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-multicluster-orchestrator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-multicluster-orchestrator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-multicluster-orchestrator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-multicluster-orchestrator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-multicluster-orchestrator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-multicluster-orchestrator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-multicluster-orchestrator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-multicluster-orchestrator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multicluster.odf.openshift.io/v1alpha1",
                "kind": "MirrorPeer",
                "metadata": {
                  "name": "mirrorpeer-sample"
                },
                "spec": {
                  "items": [
                    {
                      "clusterName": "local-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    },
                    {
                      "clusterName": "spoke-cluster",
                      "storageClusterRef": {
                        "name": "ocs-storagecluster",
                        "namespace": "openshift-storage"
                      }
                    }
                  ],
                  "manageS3": false,
                  "schedulingIntervals": [
                    "5m"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          console.openshift.io/plugins: '["odf-multicluster-console"]'
          createdAt: "2025-09-11T15:26:27Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MirrorPeer is the Schema for the mirrorpeers API
            displayName: Mirror Peer
            kind: MirrorPeer
            name: mirrorpeers.multicluster.odf.openshift.io
            version: v1alpha1
        description: |
          Orchestrator for OpenShift Data Foundation clusters running across multiple OpenShift clusters.
          It uses Red Hat Advanced Cluster Management for Kubernetes as the multicluster control plane.
        displayName: ODF Multicluster Orchestrator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - multicluster
        - orchestrator
        - storage
        - acm
        links:
        - name: ODF Multicluster Orchestrator
          url: https://github.com/red-hat-storage/odf-multicluster-orchestrator
        maintainers:
        - email: uchapaga@redhat.com
          name: Umanga Chapagain
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-multicluster-rhel9-operator@sha256:4b78c8df4f4aa4be805a442b059b5e2822464bb029826f2f0090cf8b6b747768
        - registry.redhat.io/odf4/odf-multicluster-operator-bundle@sha256:47c754d0e2782dd8d885758e68803173c32f1fe6319ac09c786dd24e5124df25
        - registry.redhat.io/odf4/odf-multicluster-console-rhel9@sha256:957805157e7690deedaae63856c9ce71e5c299610bca401524397a40e2820537
        version: 4.18.11-rhodf
      entries:
      - name: odf-multicluster-orchestrator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-multicluster-orchestrator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-multicluster-orchestrator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-multicluster-orchestrator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-multicluster-orchestrator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-multicluster-orchestrator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-multicluster-orchestrator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-multicluster-orchestrator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-multicluster-orchestrator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-multicluster-orchestrator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-multicluster-orchestrator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-multicluster-orchestrator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-multicluster-orchestrator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Entando, Inc
      provider-url: ""
    name: entando-k8s-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: alpha
    - currentCSV: entando-k8s-operator.v6.3.2-pr3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: beta
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - docker.io/entando/entando-redhat-sso@sha256:3692adb3694b3bf278bed7a678c9d0eaa0aba8e3e3ed4117ca04156525bcac72
        - docker.io/entando/entando-k8s-app-plugin-link-controller@sha256:aa7ba0861705213434e610530c5bf5da1bdb33f1834d3c346174f5a28145c0bd
        - docker.io/entando/entando-de-app-eap@sha256:1f0e35d7f7ee8123721cf8e80ac96ac2f594f21a2b68cf9e720917d090c8095e
        - docker.io/entando/entando-k8s-app-controller@sha256:c68d3b129fc625f5d1ee10ed2a107e28c145049d4b513dcdb2c736555025e339
        - docker.io/entando/entando-k8s-database-service-controller@sha256:58ca4ad2fca83a715aba9b68609f9056f633d556eb9845b3aff5c05769369539
        - docker.io/entando/app-builder@sha256:1ca90c09302b9cadf8950278d92ad4d7f3a14f02dea8704320794aa77ada8631
        - docker.io/entando/entando-component-manager@sha256:668191fc86b090452baaba18ec7bbc3d07df5a99750e0c5b98759599d565cfbe
        - docker.io/entando/entando-k8s-keycloak-controller@sha256:f1ed42001260802353197433cea7df4c2d0e5fd9bcd926bb5ae7c654dfad2ffe
        - registry.redhat.io/rhel8/postgresql-12@sha256:f4e5c728b644bf1888ec8086424852ed74b5596a511be29e636fb10218fc9b6f
        - docker.io/entando/entando-k8s-service@sha256:775391f0005e64764288bfe007d9ea576f13618c02ea400d030a7502cfb551bd
        - registry.redhat.io/rhel8/mysql-80@sha256:570bce7dc649ac7ac85e7ecb196aabc281c77a94590aa7df7d9746869ce757ea
        - docker.io/entando/entando-k8s-composite-app-controller@sha256:4bc6797bb41b3e407b64530c3cff7a78f3fc2c22863caad9c606ccd650310b61
        - docker.io/entando/entando-k8s-dbjob@sha256:8ae47dbf93201965106663299f06daad544982fa244d4a0f365a1b717727e5ed
        - registry.connect.redhat.com/entando/entando-operator@sha256:bbb6a9d744c57fd56c372447efd4f625cddca29fe2bd539d045c9a53a17b9308
        - docker.io/entando/entando-k8s-cluster-infrastructure-controller@sha256:86dbea6696349d7b2393e12bacef4d1acec2dd3f8551e397fa26c92706a18dd3
        - entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        - docker.io/entando/entando-k8s-plugin-controller@sha256:a20c753dda86e0f54f210e446b404733045ba426fc7186dc3936daed8224fbcb
        version: 6.3.2-pr3
      entries:
      - name: entando-k8s-operator.v6.3.2-pr3
        version: 6.3.2-pr3
      name: beta
    - currentCSV: entando-k8s-operator.v6.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {"apiVersion":"entando.org/v1","kind":"EntandoDeBundle","metadata":{"name":"my-bundle","namespace":"my-namespace"},"spec":{"details":{"name":"my-bundle"}}},
              {"apiVersion":"entando.org/v1","kind":"EntandoDatabaseService","metadata":{"name":"my-entando-database-service","namespace":"my-namespace"},"spec":{"dbms":"postgresql","databaseName":"my_database","createDeployment":true}},
              {"apiVersion":"entando.org/v1","kind":"EntandoKeycloakServer","metadata":{"name":"my-keycloak","namespace":"my-namespace"},"spec":{"standardImage":"redhat-sso", "dbms":"postgresql","replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoClusterInfrastructure","metadata":{"name":"my-entando-cluster-infrastructure","namespace":"my-namespace"},"spec":{"replicas":1,"environmentVariables":[]}},
              {"apiVersion":"entando.org/v1","kind":"EntandoPlugin","metadata":{"name":"my-entando-plugin","namespace":"my-namespace"},"spec": {"dbms": "postgresql","healthCheckPath": "/management/health","image":
            "entando/entando-avatar-plugin:6.0.5","ingressHostName": "my-plugin.apps-crc.testing","ingressPath": "/avatarPlugin","replicas": 1,"roles": [{"code": "admin","name": "user"},{"code": "user","name": "user"}],"securityLevel": "strict"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoApp","metadata":{"name":"my-app","namespace":"my-namespace"},"spec":{"dbms":"postgresql","environmentVariables":[],"ingressHostName":"my-app.apps-crc.testing","replicas":1,"standardServerImage":"eap"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoAppPluginLink","metadata":{"name":"my-link","namespace":"my-namespace"},"spec":{"entandoAppName":"my-app","entandoPluginName":"my-plugin"}},
              {"apiVersion":"entando.org/v1","kind":"EntandoCompositeApp","metadata":{"name":"my-entando-composite-app","namespace":"my-namespace"},"spec":{"dbmsOverride":"postgresql","ingressHostNameOverride":"entando.apps-crc.testing","components":[{"kind":"EntandoDatabaseService","metadata":{"name":"inline-entando-database-service"},"spec":{"createDeployment":true}},{"kind":"EntandoKeycloakServer","metadata":{"name":"inline-keycloak"},"spec":{"standardImage":"redhat-sso"}},{"kind":"EntandoClusterInfrastructure","metadata":{"name":"inline-entando-cluster-infrastructure"},"spec":{}},{"kind":"EntandoApp","metadata":{"name":"inline-app"},"spec":{"standardServerImage":"eap"}},{"kind":"EntandoPlugin","metadata":{"name":"inline-plugin"},"spec":{"healthCheckPath":"/management/health","image":"entando/entando-avatar-plugin:6.0.5","ingressPath":"/avatarPlugin","roles":[{"code":"admin","name":"admin"},{"code":"user","name":"user"}]}},{"kind":"EntandoAppPluginLink","metadata":{"name":"inline-link"},"spec":{"entandoAppName":"inline-app","entandoPluginName":"inline-plugin"}}]}}
            ]
          capabilities: Basic Install
          categories: Application Runtime, Development Tools
          certified: "false"
          containerImage: registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
          createdAt: ""
          description: Processes EntandoKeycloakServer, EntandoApp and EntandoPlugin
            custom resources and deploys the relevant containers in the Kubernetes
            cluster.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/entando-k8s-operator-rhmp/support?utm_source=openshift_console
          repository: ""
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              An Entando Composite Application is the easiest way to get started with Entando. It allows the user to combine Entando Custom Resources and deploy them in the sequence specified. This is useful when one Entando Custom Resource (e.g. an EntandoPlugin) is dependent on another Entando Custom Resource. Simply order the components in the list of components, deploy the EntandoCompositeApp, and wait for the whole application to be ready.
            displayName: Entando Composite Application
            kind: EntandoCompositeApp
            name: entandocompositeapps.entando.org
            version: v1
          - description: |
              The EntandoDatabaseService is used to share a database instance across multiple deployments. This custom resource can either point to an external database service, or it can actually deploy a container that hosts the database service. Any subsequent deployments in the same namespace that need  the same database specified in the EntandoDatabaseService will be configured to point to this database service.
            displayName: Entando Database Service
            kind: EntandoDatabaseService
            name: entandodatabaseservices.entando.org
            version: v1
          - description: |
              The EntandoKeycloakServer custom resource instructs the Entando Operator to deploy either the Keycloak Community OIDC provider, or its equivalent commercial product, Red Hat SSO. Once a Keycloak server has been deployed, subsequent EntandoApp and EntandoPlugin deployments can use it as an OIDC provider. The Entand Operator will automatically create OIDC Clients for these deployments in the resulting Keycloak Server.
            displayName: Entando Keycloak Server
            kind: EntandoKeycloakServer
            name: entandokeycloakservers.entando.org
            version: v1
          - description: |
              The EntandoClusterInfrastructure resource allows the Entando Operator to deploy its abstraction layer on Kubernetes, the Entando Kubernetes Service.
            displayName: Entando Cluster Infrastructure
            kind: EntandoClusterInfrastructure
            name: entandoclusterinfrastructures.entando.org
            version: v1
          - description: |
              The EntandoPlugin allows customers to extend the functionality of an Entando application by deploying a microservice based on the Entando JHipster Blueprint. EntandoPlugins can also be included in, and deployed via, an Entando Bundle
            displayName: Entando Plugin Microservice
            kind: EntandoPlugin
            name: entandoplugins.entando.org
            version: v1
          - description: An Entando App deploys the components required to host an
              Entando App. The server side components that are deployed include the
              Entando App Engine, the Entando Component Manager, the Entando App Builder,
              and the user facing application.
            displayName: EntandoApp
            kind: EntandoApp
            name: entandoapps.entando.org
            version: v1
          - description: The EntandoAppPluginLink custom resource is used to associate
              an Entando Application and a plugin microservice so that the microservice
              APIs can be accessed using the applications ingress.
            displayName: EntandoAppPluginLink
            kind: EntandoAppPluginLink
            name: entandoapppluginlinks.entando.org
            version: v1
          - description: The EntandoDeBundle custom resource defines a bundle that
              can be deployed to the Entando Component Repository. See [here](https://dev.entando.org/v6.3/tutorials/ecr/publish-simple-bundle.html#build-and-publish-a-simple-bundle)
              for instructions on creating and deploying bundles.
            displayName: EntandoDeBundle
            kind: EntandoDeBundle
            name: entandodebundles.entando.org
            version: v1
        description: |
          ## Entando

          The Entando platform accelerates the development and lifecycle  management of fully modularized applications on Kubernetes. It  provides tools to help developers create and manage applications  using modular frontend and backend components.
          The Entando Operator automates the installation, provisioning, and  configuration management of the components that make up an Entando  application. Specifically, the operator manages the following custom  resources:
          **EntandoKeycloakServers** for centralized authentication of frontend  and backend components. The operator can deploy Keycloak, or in certified  environments, Red Hat SSO servers that can then be used by subsequent  deployments as an OpenID Connect provider.
          **EntandoApps** for hosting an Entando application. EntandoApps are hosted on Wildfly or JBoss EAP containers, and can also be used to deploy custom  EntandoApp containers.
          **EntandoPlugins** for deploying microservices to customize or enhance your EntandoApp. Entando microservice plugins are deployed to your cluster, and  then linked to one or more EntandoApps.
          ## Using the Operator
          The Entando Operator can be deployed using the default settings without any  configuration. Once deployed, the operator can be customized by editing  the *configmap* and secrets.
          ### ConfigMap: entando-operator-config
          The 'entando-operator-config' ConfigMap can be added after deployment and any changes to it will be picked up by the operator on subsequent event processing. It supports the following keys:

              entando.k8s.operator.gc.controller.pods: set this to "false" to keep controller pods after completion.
              entando.k8s.operator.compliance.mode: set this to "community" if there is no requirement for Red Hat compliance.
              entando.k8s.operator.image.pull.secrets: a comma separated list containing the names of pull secrets that will be linked to all service accounts.
              entando.k8s.operator.disable.pvc.garbage.collection: set this to "false" if you want Persistent Volume Claims to be deleted with the custom resources they are associated with.
              entando.k8s.operator.impose.default.limits: set this to "false" if there is no need to limit the resource consumption of pods on your cluster.
              entando.k8s.operator.request.to.limit.ratio: a decimal number that default limits will be multiplied by to calculate default requests for resources.
              entando.k8s.operator.force.db.password.reset: set this to "true" if you plan to delete Secrets from your namespace but you want to retain the Database they point to.
              entando.k8s.operator.pull.policy.override: specify your preferred pullPolicy for images. The default is Always.
              entando.tls.secret.name: The name of a standard TLS secret to use for HTTPS Ingresses. See the section entando-tls-secret.
              entando.ca.secret.name: The name of a secret containing CA certificates. See the section entando-ca-cert-secret.
              entando.assume.external.https.provider: Set this to "true" if your cloud provider handles HTTPS for you.
              entando.use.auto.cert.generation: Set this to "true" to have Openshift use its internal CA to generate certificates for your Routes.
              entando.default.routing.suffix: The domain name that can be suffixed to deployment names when the ingressHostName is omitted. Needs to be preconfigured on your DNS provider.
              entando.pod.completion.timeout.seconds: The time it will take before Entando fails a run-to-completion Pod.
              entando.pod.readiness.timeout.seconds: The time it will take before Entando fails a Service Pod.
              entando.pod.shutdown.timeout.seconds: The time Entando will give a Pod to shutdown gracefully.


          ### entando-pull-secret
          The secret to be used to pull Entando images from the Red Hat container registry. The name of this secret is hard coded as it is required for the ClusterServiceVersion of this Operator
          ### entando-tls-secret
          A standard Kubernetes TLS secret that will be used on all deployments where no custom TLS secret name is specified.
          ### entando-ca-cert-secret
          This is an opaque secret in the Entando Operator's namespace that contains the certificates of all trusted certificate authorities in your environment. This is generally used mainly for self signed certificates. As is generally the case for opaque secrets, there are no constraints on the keys in this secret. However, limit the files inside the secret to X509 certificates or certificate chains. The Entando Operator will load all of these files into a Java keystore that it then configures as the trust store for each container that uses Java.
        displayName: Entando Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MFE
        links:
        - name: Documentation
          url: https://dev.entando.org/v6.3/docs/
        - name: Support
          url: https://www.entando.com/
        maintainers:
        - email: dev@entando.com
          name: support
        - email: a.barnard@entando.com
          name: author
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Entando, Inc
        relatedImages:
        - registry.connect.redhat.com/entando/entando-operator@sha256:3b683338f5849b887e1bc81ea078a9b799d272f824abd4247ff87e70ca7acbf2
        - registry.hub.docker.com/entando/entando-k8s-controller-coordinator@sha256:c8a93ebd69af3cf822dd7d4fcce5d4241cd8dca229c4e25c771bcfecca916839
        version: 6.3.2
      entries:
      - name: entando-k8s-operator.v6.3.2
        version: 6.3.2
      name: stable
    defaultChannel: stable
    packageName: entando-k8s-operator-rhmp
    provider:
      name: Entando, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sumo Logic
      provider-url: https://www.sumologic.com/
    name: sumologic-kubernetes-collection-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "helm-operator.sumologic.com/v1alpha1",
                    "kind": "SumologicCollection",
                    "metadata": {
                        "name": "collection"
                    },
                    "spec": {
                        "sumologic": {
                            "accessId": "",
                            "accessKey": "",
                            "clusterName": "kubernetes",
                            "scc": {
                                "create": true
                            }
                        },
                        "kube-prometheus-stack": {
                            "prometheus-node-exporter": {
                                "service": {
                                    "port": 9200,
                                    "targetPort": 9200
                                }
                            }
                        },
                        "otellogs": {
                            "daemonset": {
                                "containers": {
                                    "otelcol": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                },
                                "initContainers": {
                                    "changeowner": {
                                        "securityContext": {
                                            "privileged": true
                                        }
                                    }
                                }
                            }
                        }
                    }
                }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing,Monitoring
          containerImage: registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
          createdAt: "2024-07-19T10:48:00Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator.git
          support: Sumo Logic
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Configuration for Sumo Logic Kubernetes Collection
            displayName: SumologicCollection
            kind: SumologicCollection
            name: sumologiccollections.helm-operator.sumologic.com
            version: v1alpha1
        description: Helm Operator for the Sumo Logic Kubernetes Collection Helm Chart
          v4.13.0-2
        displayName: Sumo Logic Kubernetes Collection Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - collection
        - logging
        - metrics
        - monitoring
        links:
        - name: Sumo Logic Kubernetes Collection Helm Operator
          url: https://github.com/SumoLogic/sumologic-kubernetes-collection-helm-operator
        maintainers:
        - email: collection@sumologic.com
          name: Sumo Logic
        maturity: alpha
        provider:
          name: Sumo Logic
          url: https://www.sumologic.com/
        relatedImages:
        - registry.connect.redhat.com/sumologic/thanos@sha256:16faf7191668e86d8b576bba055dd0cbdeb009635976b9e07bd00f926032f11c
        - registry.connect.redhat.com/sumologic/kubernetes-setup@sha256:f07eaf68fa11ffc708e4485a83bd3763fc5712f7d73d29edfe5b9c9e08f05df6
        - registry.connect.redhat.com/sumologic/opentelemetry-operator@sha256:5587fe5b7ebfa57ad320ad2521103f908317388d1741c40e5f7be5ed9e5f9ad7
        - registry.connect.redhat.com/sumologic/kube-state-metrics@sha256:3551ededa327a35db798ec8286cd843580e3ae395d2231e041e8c4b7621127ee
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator-bundle@sha256:333a6f476d03adcbfaa4ab38802859719a7dd8857783f8487c99b008d06815c5
        - registry.connect.redhat.com/sumologic/autoinstrumentation-java@sha256:242958b7d8905f4ac02f6ad38897ba58b9badf2a6213a389e67dcc71fee3630c
        - registry.connect.redhat.com/sumologic/nginx-unprivileged@sha256:7260555d16c94c4fe1e016626b4cb8a5c0de6b700e2eea28744331fb298f71ae
        - registry.connect.redhat.com/sumologic/telegraf@sha256:75efd2e509c6017ea0fc5c9103ce068cbc89087b16506a4ef48cb5ff00eab84e
        - registry.connect.redhat.com/sumologic/busybox@sha256:209bc8e0b1d958fc699f3bb75b1248544e372ad35cdc46c991bd1698f641e1f9
        - registry.connect.redhat.com/sumologic/prometheus-config-reloader@sha256:0b605d36782afd8f50fd631d0e9c7b19d39ce962af89b7a737b2f09a8ceff291
        - registry.connect.redhat.com/sumologic/tailing-sidecar-operator@sha256:c16b8b18e9b263fc3eb24059bba512bcfc3201b32105af6e5518392826ddd66c
        - registry.connect.redhat.com/sumologic/prometheus@sha256:0ef6eaf570a40499dd5de92a79593e950bc771b014e0ee6ff7419b4b82a36fcc
        - registry.connect.redhat.com/sumologic/autoinstrumentation-nodejs@sha256:c206eef4fdecaa58b4f01e5e7f0dbd74e9606eab2b4debf47461d36191e248d1
        - registry.connect.redhat.com/sumologic/node-exporter@sha256:11e9e604432a66d9b388cb1bf70eac077c7f0372d1264342ef927d76b8ef0272
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:e1f49fcf5b7e86d9eb49c16dbbf9484cb486f55f999fd51d7c738c3b1f2666cf
        - registry.connect.redhat.com/sumologic/telegraf-operator@sha256:ff958222be01c0c3966474463e43bbd5bcf348790dd34bb761718cb1cea2d793
        - registry.connect.redhat.com/sumologic/autoinstrumentation-python@sha256:f0fa45fa45e64c7df88f21181ae18ee96fe83cd3988258663d0806ea6f39e6bb
        - registry.connect.redhat.com/sumologic/sumologic-kubernetes-collection-helm-operator@sha256:92f7cf8f9613b1bb4d1c32dd8592f1dbd463de3ced6a6febd5373a7d23d556c4
        - registry.connect.redhat.com/sumologic/autoinstrumentation-dotnet@sha256:f4f385c4489c74b5c43f2fe6b6620a9eb8d9ecab470567236b3c45f636106640
        - registry.connect.redhat.com/sumologic/metrics-server@sha256:65d2a3ddc4b4183f438fa18b655079c93c905cb9e3b8c54977336205dc1fff07
        - registry.connect.redhat.com/sumologic/prometheus-operator@sha256:58a329e6582456b32abebb470086191f8366b927af4703741613993465f116b8
        - registry.connect.redhat.com/sumologic/kubernetes-tools-kubectl@sha256:d3155b864cc74f327adedf7d013a8f0fcb11c2e73a5655b7cafa848824e80172
        - registry.connect.redhat.com/sumologic/sumologic-otel-collector@sha256:4554d43dba204e3ec8e91b36442a438f10234fa0cb4db7fb7880b257a6485112
        - registry.connect.redhat.com/sumologic/tailing-sidecar@sha256:738d653a78dc415243ab3627db9135768175fb7f51d0803566f5d9e824c8744b
        version: 4.13.0-2
      entries:
      - name: sumologic-kubernetes-collection-helm-operator.v4.13.0-2
        version: 4.13.0-2
      name: alpha
    defaultChannel: alpha
    packageName: sumologic-kubernetes-collection-helm-operator
    provider:
      name: Sumo Logic
      url: https://www.sumologic.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: candidate
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: fast
    - currentCSV: stackgres.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - registry.connect.redhat.com/ongres/stackgres-certified-1@sha256:8ad367e032815b75d95b76bc47167b94538bdaca26ab2268f731f50713122be7
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres.v1.17.2
        version: 1.17.2
      - name: stackgres.v1.17.1
        version: 1.17.1
      - name: stackgres.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-certified
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acm-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acm.services.k8s.aws/v1alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
          createdAt: "2025-09-20T01:19:56Z"
          description: AWS ACM controller is a service controller for managing ACM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Certificate represents the state of an AWS acm Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acm.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage AWS Certificate Manager (ACM) resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM**

          AWS Certificate Manager (ACM) handles the complexity of creating, storing, and renewing public and private SSL/TLS X.509 certificates and keys that protect your AWS websites and applications. You can provide certificates for your [integrated AWS services](https://docs.aws.amazon.com/acm/latest/userguide/acm-services.html) either by issuing them directly with ACM or by [importing](https://docs.aws.amazon.com/acm/latest/userguide/import-certificate.html) third-party certificates into the ACM management system. ACM certificates can secure singular domain names, multiple specific domain names, wildcard domains, or combinations of these. ACM wildcard certificates can protect an unlimited number of subdomains. You can also [export](https://docs.aws.amazon.com/acm/latest/userguide/export-private.html) ACM certificates signed by AWS Private CA for use anywhere in your internal PKI.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon ACM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM Developer Resources
          url: https://aws.amazon.com/certificate-manager/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acm maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acm-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acm-controller@sha256:4d63542b39f41c458dd1eec28e44edce3a919f4dd9137ff64a9452cac1a1f954
        version: 1.1.1
      entries:
      - name: ack-acm-controller.v1.1.1
        version: 1.1.1
      - name: ack-acm-controller.v1.1.0
        version: 1.1.0
      - name: ack-acm-controller.v1.0.13
        version: 1.0.13
      - name: ack-acm-controller.v1.0.12
        version: 1.0.12
      - name: ack-acm-controller.v1.0.11
        version: 1.0.11
      - name: ack-acm-controller.v1.0.10
        version: 1.0.10
      - name: ack-acm-controller.v1.0.9
        version: 1.0.9
      - name: ack-acm-controller.v1.0.8
        version: 1.0.8
      - name: ack-acm-controller.v1.0.7
        version: 1.0.7
      - name: ack-acm-controller.v1.0.6
        version: 1.0.6
      - name: ack-acm-controller.v1.0.5
        version: 1.0.5
      - name: ack-acm-controller.v1.0.4
        version: 1.0.4
      - name: ack-acm-controller.v1.0.1
        version: 1.0.1
      - name: ack-acm-controller.v1.0.0
        version: 1.0.0
      - name: ack-acm-controller.v0.0.20
        version: 0.0.20
      - name: ack-acm-controller.v0.0.19
        version: 0.0.19
      - name: ack-acm-controller.v0.0.18
        version: 0.0.18
      - name: ack-acm-controller.v0.0.17
        version: 0.0.17
      - name: ack-acm-controller.v0.0.16
        version: 0.0.16
      - name: ack-acm-controller.v0.0.12
        version: 0.0.12
      - name: ack-acm-controller.v0.0.10
        version: 0.0.10
      - name: ack-acm-controller.v0.0.9
        version: 0.0.9
      - name: ack-acm-controller.v0.0.6
        version: 0.0.6
      - name: ack-acm-controller.v0.0.5
        version: 0.0.5
      - name: ack-acm-controller.v0.0.4
        version: 0.0.4
      - name: ack-acm-controller.v0.0.2
        version: 0.0.2
      - name: ack-acm-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acm-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Debezium Authors
      provider-url: https://debezium.io/
    name: debezium-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: debezium-operator.v2.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.4.2.Final
          createdAt: "2023-12-13T12:20:56Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1a156d67fd5beb9ffddecab724490acb7143f8141ff3b392c6f519e76c3ad374
        - quay.io/debezium/operator:2.4.2.Final
        version: 2.4.2
      entries:
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.4.x
    - currentCSV: debezium-operator.v2.5.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.5.3.Final
          createdAt: "2024-03-19T08:17:17Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.5.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:645fcb18684884aef8ca4d5299f0d053d496c3a5d3f37cadd66c51d38514a680
        version: 2.5.3-final
      entries:
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.5.x
    - currentCSV: debezium-operator.v2.6.1-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.6.1.Final
          createdAt: "2024-04-12T17:18:16Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.5.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:5a73ce3a8cc8422aeda37fd9ef97f5aecdbf8c69fb6ff488df26342e989b5c57
        - quay.io/debezium/operator:2.6.1.Final
        version: 2.6.1-final
      entries:
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.6.x
    - currentCSV: debezium-operator.v2.7.3-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:2.7.3.Final
          createdAt: "2024-09-20T10:31:16Z"
          description: An Operator for installing and managing Debezium
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:2.7.3.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:1dfe9180a9fca4f773c1282dfd0006daa0bb48dfbac87e352661e051107f9033
        version: 2.7.3-final
      entries:
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-2.7.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-3.0.x
    - currentCSV: debezium-operator.v3.0.0-final
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
                {
                  "apiVersion": "debezium.io/v1alpha1",
                  "kind": "DebeziumServer",
                  "metadata": {
                    "name": "debezium-test"
                  },
                  "spec": {
                    "quarkus": {
                      "config": {
                        "log.console.json": false
                      }
                    },
                    "sink": {
                      "type": "kafka",
                      "config": {
                        "producer.bootstrap.servers": "dbz-kafka-kafka-bootstrap.debezium:9092",
                        "producer.key.serializer": "org.apache.kafka.common.serialization.StringSerializer",
                        "producer.value.serializer": "org.apache.kafka.common.serialization.StringSerializer"
                      }
                    },
                    "source": {
                      "class": "io.debezium.connector.mongodb.MongoDbConnector",
                      "config": {
                        "topic.prefix": "dbserver1",
                        "offset.storage.file.filename": "/debezium/data/offsets.dat",
                        "database.history": "io.debezium.relational.history.FileDatabaseHistory",
                        "mongodb.connection.string": "mongodb://debezium:dbz@mongo.debezium:27017/?replicaSet=rs0"
                      }
                    }
                  }
                }
            ]
          capabilities: Basic Install
          categories: Big Data, Database, Integration & Delivery, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/debezium/operator:3.0.0.Final
          createdAt: "2024-10-02T09:55:38Z"
          description: An Operator for installing and managing Debezium
          olm.skipRange: '>2.7.3-final'
          repository: https://github.com/debezium/debezium-operator/debezium-operator-dist
          support: Debezium Authors
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Debezium Server
            displayName: DebeziumServer
            kind: DebeziumServer
            name: debeziumservers.debezium.io
            version: v1alpha1
        description: Debezium is an open source distributed platform for change data
          capture. Start it up, point it at your databases, and your apps can start
          responding to all of the inserts, updates, and deletes that other apps commit
          to your databases
        displayName: Debezium Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: true
          type: MultiNamespace
        keywords:
        - Debezium
        - CDC
        - Data
        - Streaming
        links:
        - name: Debezium
          url: https://debezium.io/
        - name: Documentation
          url: https://debezium.io/documentation/reference/stable/
        - name: Debezium Zulip Chat
          url: https://debezium.zulipchat.com
        maintainers:
        - email: debezium@googlegroups.com
          name: Debezium Authors
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        provider:
          name: Debezium Authors
          url: https://debezium.io/
        relatedImages:
        - quay.io/debezium/operator:3.0.0.Final
        - quay.io/community-operator-pipeline-prod/debezium-operator@sha256:10671f1f6a7bab7b8bf7c8d986321484a7972b727f08947894ba7a6ccef98cd8
        version: 3.0.0-final
      entries:
      - name: debezium-operator.v3.0.0-final
        version: 3.0.0-final
      - name: debezium-operator.v2.7.3-final
        version: 2.7.3-final
      - name: debezium-operator.v2.7.0-final
        version: 2.7.0-final
      - name: debezium-operator.v2.6.1-final
        version: 2.6.1-final
      - name: debezium-operator.v2.6.0-final
        version: 2.6.0-final
      - name: debezium-operator.v2.5.3-final
        version: 2.5.3-final
      - name: debezium-operator.v2.5.2-final
        version: 2.5.2-final
      - name: debezium-operator.v2.5.1
        version: 2.5.1
      - name: debezium-operator.v2.5.0
        version: 2.5.0
      - name: debezium-operator.v2.4.2
        version: 2.4.2
      - name: debezium-operator.v2.4.0
        version: 2.4.0
      name: debezium-latest
    defaultChannel: debezium-latest
    packageName: debezium-operator
    provider:
      name: Debezium Authors
      url: https://debezium.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigatewayv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigatewayv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Deployment",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "API",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Authorizer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Route",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigatewayv2.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
          createdAt: "2025-09-19T21:58:07Z"
          description: AWS API Gateway v2 controller is a service controller for managing
            API Gateway v2 resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIMapping represents the state of an AWS apigatewayv2 APIMapping
              resource.
            displayName: APIMapping
            kind: APIMapping
            name: apimappings.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: API represents the state of an AWS apigatewayv2 API resource.
            displayName: API
            kind: API
            name: apis.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigatewayv2 Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigatewayv2 Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: DomainName represents the state of an AWS apigatewayv2 DomainName
              resource.
            displayName: DomainName
            kind: DomainName
            name: domainnames.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigatewayv2 Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Route represents the state of an AWS apigatewayv2 Route resource.
            displayName: Route
            kind: Route
            name: routes.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigatewayv2 Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigatewayv2.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigatewayv2 VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigatewayv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway v2 resources in AWS from within your Kubernetes cluster. Currently only HTTP APIs are supported. WebSocket API support will be added soon. For HTTP APIs, DomainName and APIMapping resource support will be added soon.

          **About Amazon API Gateway v2**

          Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.

          API Gateway handles all the tasks involved in accepting and processing up to hundreds of thousands of concurrent API calls, including traffic management, CORS support, authorization and access control, throttling, monitoring, and API version management. API Gateway has no minimum fees or startup costs. You pay for the API calls you receive and the amount of data transferred out and, with the API Gateway tiered pricing model, you can reduce your cost as your API usage scales.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway v2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigatewayv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway v2 Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: api gateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/apigatewayv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-apigatewayv2-controller@sha256:1cf12140fcd67b0f3e8c5791a91c5b705bb3af8b2e03ea4ba2d4ce6f9f7a4809
        version: 1.1.1
      entries:
      - name: ack-apigatewayv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-apigatewayv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigatewayv2-controller.v1.0.30
        version: 1.0.30
      - name: ack-apigatewayv2-controller.v1.0.29
        version: 1.0.29
      - name: ack-apigatewayv2-controller.v1.0.28
        version: 1.0.28
      - name: ack-apigatewayv2-controller.v1.0.26
        version: 1.0.26
      - name: ack-apigatewayv2-controller.v1.0.25
        version: 1.0.25
      - name: ack-apigatewayv2-controller.v1.0.23
        version: 1.0.23
      - name: ack-apigatewayv2-controller.v1.0.21
        version: 1.0.21
      - name: ack-apigatewayv2-controller.v1.0.17
        version: 1.0.17
      - name: ack-apigatewayv2-controller.v1.0.16
        version: 1.0.16
      - name: ack-apigatewayv2-controller.v1.0.15
        version: 1.0.15
      - name: ack-apigatewayv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-apigatewayv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-apigatewayv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-apigatewayv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigatewayv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigatewayv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigatewayv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-apigatewayv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigatewayv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigatewayv2-controller.v0.1.6
        version: 0.1.6
      - name: ack-apigatewayv2-controller.v0.1.4
        version: 0.1.4
      - name: ack-apigatewayv2-controller.v0.1.3
        version: 0.1.3
      - name: ack-apigatewayv2-controller.v0.1.2
        version: 0.1.2
      - name: ack-apigatewayv2-controller.v0.1.1
        version: 0.1.1
      - name: ack-apigatewayv2-controller.v0.1.0
        version: 0.1.0
      - name: ack-apigatewayv2-controller.v0.0.22
        version: 0.0.22
      - name: ack-apigatewayv2-controller.v0.0.21
        version: 0.0.21
      - name: ack-apigatewayv2-controller.v0.0.20
        version: 0.0.20
      - name: ack-apigatewayv2-controller.v0.0.19
        version: 0.0.19
      - name: ack-apigatewayv2-controller.v0.0.18
        version: 0.0.18
      - name: ack-apigatewayv2-controller.v0.0.17
        version: 0.0.17
      - name: ack-apigatewayv2-controller.v0.0.16
        version: 0.0.16
      - name: ack-apigatewayv2-controller.v0.0.15
        version: 0.0.15
      - name: ack-apigatewayv2-controller.v0.0.14
        version: 0.0.14
      - name: ack-apigatewayv2-controller.v0.0.13
        version: 0.0.13
      - name: ack-apigatewayv2-controller.v0.0.12
        version: 0.0.12
      - name: ack-apigatewayv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-apigatewayv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-apigatewayv2-controller.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigatewayv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "certmanager": {
                        "install": false
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-26T07:16:35Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/gitlab-operator-kubernetes@sha256:a5357c10307eb26380b9a7263d64d19aa57bf6da6f0112be20803b34dc7e9230
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3281fe12aaae428fd7102332e7c0d7033515042d7830eba3953891ab0c3e3a29
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      - name: gitlab-operator-kubernetes.v2.1.1
        version: 2.1.1
      - name: gitlab-operator-kubernetes.v2.1.0
        version: 2.1.0
      - name: gitlab-operator-kubernetes.v2.0.2
        version: 2.0.2
      - name: gitlab-operator-kubernetes.v2.0.1
        version: 2.0.1
      - name: gitlab-operator-kubernetes.v2.0.0
        version: 2.0.0
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-netobs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2
    - currentCSV: skupper-netobs-operator.v2.1.1-rh-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability.skupper.io/v2alpha1",
                "kind": "NetworkObserver",
                "metadata": {
                  "name": "networkobserver-sample"
                },
                "spec": {
                  "auth": {
                    "basic": {
                      "create": true,
                      "secretName": ""
                    },
                    "openshift": {
                      "cookieSecretName": "",
                      "createCookieSecret": true,
                      "serviceAccount": {
                        "create": true,
                        "nameOverride": ""
                      }
                    },
                    "strategy": "openshift"
                  },
                  "extraArgs": null,
                  "fullnameOverride": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "registry.redhat.io/service-interconnect/skupper-network-observer-rhel9",
                    "tag": "2.1.1"
                  },
                  "ingress": {
                    "annotations": {},
                    "className": "",
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": [
                          {
                            "path": "/",
                            "pathType": "ImplementationSpecific"
                          }
                        ]
                      }
                    ],
                    "tls": []
                  },
                  "nameOverride": "",
                  "nginx": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/ubi9/nginx-124",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": 9.6
                  },
                  "openshiftOauthProxy": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-oauth-proxy-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.19.0"
                  },
                  "podAnnotations": {},
                  "podLabels": {},
                  "podSecurityContext": {
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "prometheus": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.redhat.io/openshift4/ose-prometheus-rhel9",
                    "securityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      }
                    },
                    "tag": "v4.16.0"
                  },
                  "resources": {},
                  "route": {
                    "enabled": true
                  },
                  "router": {
                    "certificate": {
                      "create": true,
                      "nameOverride": ""
                    },
                    "endpoint": "amqps://skupper-router-local"
                  },
                  "securityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    }
                  },
                  "service": {
                    "port": 443,
                    "type": "ClusterIP"
                  },
                  "skipManagementLabels": false,
                  "tls": {
                    "openshiftIssued": true,
                    "secretName": "",
                    "skupperIssued": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
          createdAt: "2025-08-27T20:21:18Z"
          description: Red Hat Service Interconnect Network Observer Operator provides
            the ability to monitor and observe a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkObserver
            name: networkobservers.observability.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect Network Observer
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-network-observer-operator-bundle@sha256:e2f2038278f8c8256a18dbc3ce6bcb424ba6bb2bf165a048391a9ee677d006ac
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9-operator@sha256:361eabb3fced000f7b304af5ea4de73219516839aba27076c09e0826ad26e53c
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        - registry.redhat.io/ubi9/nginx-124@sha256:cbcca4743015b4ef167383a080f4711a6049e275e0b66f80a74f8b087b0dfc7f
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:d3056b35d9a205b9f2c48d924f199c5ac23904eb18d526e4bff229e7c7181415
        version: 2.1.1-rh-8
      entries:
      - name: skupper-netobs-operator.v2.1.1-rh-8
        version: 2.1.1-rh-8
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-netobs-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-enterprise
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-enterprise.v1.33.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
          createdAt: "2025-04-15T11:32:47Z"
          description: The MongoDB Enterprise Kubernetes Operator enables easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-enterprise-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          The MongoDB Enterprise Kubernetes Operator enables easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8f8735d377d214f64961381b5b1ed7cd3c306e6354dbc6b45c51134a25444b93
        - quay.io/mongodb/mongodb-agent-ubi@sha256:0f8d102780a9d2ce54b7453f9575d0e24b15f014cd9b74816b1e6a4090b979b2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:13186ad45a2ba3ab55ed5c2a008eca494ee3937c8519475379476cc8e6f1d055
        - quay.io/mongodb/mongodb-agent-ubi@sha256:442f5d99d02674099525abaa69daa42cec758ce5554198f978927aa5c49006c8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c54eeb910855594ed4bf5dcbbbcc14aa6f733e97313896a8381fb4be6aeac583
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1bae6375f846af6fca7faeeff4aedf9fab41a9d45b515ba74e9fd7336408617
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7aa00dca44559318356627e88440765576dd6706187044edd1f21d0d0dcb0bf6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5421e00c0935b7a9ced51a11e935773ef5df4d259f3e1b1930fbf2f1b9d46586
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:110a560f5a0ebe1fde1b2c20ae5ff5638682229d1c8656b993bdaf7b71eb94bc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:36c84115c6c645fd7d65d1004d2507ef3592aa1f07176645b1ee8e8b2a4f5f73
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6ebe232ed0ea87045792e578ecb11dd46411333ad0ea61383c247704b4a0efe9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b175972792aa1be268d5fb3cf3ad0156f4aecffa323e1b473360fc2a3a8d8891
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f64d99a77d6a6b9ed10de9a28b1b8b8155414eca2e561f3fdb04bd66914fb3c5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:766e6cbeec27f995211aa509ff7f118b6945f3e728141f28f9c4ea0dd2297af5
        - quay.io/mongodb/mongodb-agent-ubi@sha256:97bb6bb5ecd5a802e8a91e69aa10d8bde0f648b106dc5591d49ee94209822daa
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d876f57ea45582e84748b02459748e440af443a7eb0882585638973409c08f6a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:db63f6d93e03a3e644ef6ef4846ed7d24576c688ffb584226c78e1d1fba38878
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c03e4a7663206f54c5ef39974502816f66d8d64936c19849f0337ea4547e614e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0b2d01d543d38feddcc9d351e711df33a0a5187b22d95cdb926f6015ee1093bd
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:dfec259209995bb040a1eff1fcafb1d6c3fdd145129f12c04a56ec74834b80a6
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:584255582db31d0bb1f9593feb3b675d934d8750766fc98d3a1b3340feae4995
        - quay.io/mongodb/mongodb-agent-ubi@sha256:3b57a03dc17a301fda86c8afed1917c10d9b98947680116399c975ece944f5de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - registry.connect.redhat.com/mongodb/enterprise-operator-bundle@sha256:23c3c8387eb9377b52327d7193f59eeea5831f4050248b3cbb54cbcd9437fc71
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c07f20f8e7a5479e43cac7c74fc9f18cb4d4a1bd8b6b46f7dcc6f7ec7525fe28
        - quay.io/mongodb/mongodb-enterprise-database-ubi@sha256:e2ca129f04ad6bf4e0dbb988badb083187725f28bac01bb7080cc80cad9ac316
        - quay.io/mongodb/mongodb-agent-ubi@sha256:5119a1ceeae6440dbe88b89550a7ae18de8e85c9a2059f467ad66adf7ec5d75f
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2395fe9d38a1a9160f7cb2f400b776ea3fc2e72187d23b6dee7fd1e4fb9ca721
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-agent-ubi@sha256:03763de5018fd5fa14843444d24adfc5e387605970b531801ef2d0916233a010
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:cd0cdf1b2a58f4c3b514e602cc1d96543410f631825045bbf65bd3fb84113f1f
        - quay.io/mongodb/mongodb-enterprise-init-ops-manager-ubi@sha256:6ccc8148c266a9037ea744e8d19875146b1c231909f29a8f01c7ce415ca15936
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:83af86f8098e3c81b6e0879eabd582592ef6cbb1d6592f5e175c0794b9df2a0e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:411f85772e7a4b2c76e21d26d320b7743612080bf0cb4c7900d9a626cfeea33b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        - quay.io/mongodb/mongodb-enterprise-init-database-ubi@sha256:ab12a330e6b40735ee516168073887a4f37b58694bd10026a451eb9972e55500
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:605dcca06a45e5ea127edb9ff262a1c3ddafdab2f30b6d3a18f73bfb8a6acdb0
        - quay.io/mongodb/mongodb-agent-ubi@sha256:38175af108178d89e4d3359ee2754534e5e9b24e843f12711fc14fe1dead516a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:c96ad084edbdc6a6d907f57795fa38f39199b730328dbc64997d56ea50d55dfc
        - quay.io/mongodb/mongodb-agent-ubi@sha256:f5f3b14ee6533b21d89d9ed6d4af47dbdee9e071f909dc686dfcdf6742179e46
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b2bd8eb7f8b96adb874468ff1b0ebae2db4ebe490b6e1da4c9a6c5108f813ff5
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-agent-ubi@sha256:23d962d038888c6aba5540c8f5d81d7e86c4ccfaa07b9e159825d98676609f8e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d3248506b6bd6b01c5973ab7bdac51ee41efa0ec8dd6211be4cc0b76a53a40c1
        - quay.io/mongodb/mongodb-agent-ubi@sha256:7144ac5e5038e92a6d142bf39eba46554c1b5fafeb8ea59f0390896a0d35ed17
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2be0c6c89898b18618932b6a6303aeefdde60f48ec37d5c1fca5b8d04a00fd78
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ca2aaf57d70a0b9d743a6533c4e1b88324056197453de540d3352e679a3c9c92
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:7d93aee462c60cacc4e5e66729f5131c5c0fa434e6e6400c8e2ccac4bd5e20ba
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent-ubi@sha256:695074fbbfcadf454230b80fbf1094260e631f6108cefb2050473edaee2a217a
        - quay.io/mongodb/mongodb-agent-ubi@sha256:2dbaaece4dd48087370e43d9bcf6b0a0312f86cb5dfe916e8725323bd5401992
        - quay.io/mongodb/mongodb-agent-ubi@sha256:6d007271faed24bdf06eb64f2b54e5cdc3e11ce71840a7f586dabe2ee02c640e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:b23cfa639c5a4d20381245772dfe396ec915af3a3046df5d9b10016b900431f5
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:2127d7f2684ae36b123422f55237055aad02fdb71c4c21a5b8f0b0f6d8baedfc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0ee80032abe873b330537d4b64511ee76c5efb512af95287f28ddeffab13bb6d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-enterprise-operator-ubi@sha256:6d394a5247cc346c4421bd9b37948c0378970dd3f4fc5d6eaac681f4c9655957
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:83db493ac6158d654d3ce55619f312d53cfbac258e0faf22e9110ca004e174bd
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-agent-ubi@sha256:66b49d9dce414dcd657c3550b26f71964b82edfb67a41a3ce8d2386b04ef07e8
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8467b1c2924a19929fc101ef2ff8d7e53cb0115a0cdc2e371dc0733bd0262395
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:985f770667e6c4c083e0c4bb79a948b83efa0cdc4ed3012b1dbf72fe48b38986
        - quay.io/mongodb/mongodb-enterprise-server@sha256:35b263861d6ea66b541abc190ce1aaf4d4e36f128e5d8ba2a5e9fa8e25d84446
        - quay.io/mongodb/mongodb-enterprise-init-appdb-ubi@sha256:be4295843d92ce0f325b8daaeb3a5490c11772d0348cbd9ea35a6126d038b7fb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:31cfefd2660da3efe1b6d63e39090464c28ae006d6ed1dffb2189a09183b9c61
        - quay.io/mongodb/mongodb-agent-ubi@sha256:8c11348eac180f5faa113db2804817302ec78847472e1891b063744adba4aa60
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd868284d9dfcb7c5a20e1160218d0172358384229b556c37e0789af0030eb24
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:c9a211a2c31b97eed3a25a814198547f2aa3470df93d661ca3700b828100d8de
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-agent-ubi@sha256:bd7edb8d0931dd952524f1f0ef99f61b0f69385d9e61efd034a4f45fd315a40b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-agent-ubi@sha256:575a43be4c667f3249f46f1fdcc06dd2e68cacc9eed989da0f6edd7e214d51b7
        - quay.io/mongodb/mongodb-agent-ubi@sha256:85f883076bc8edc516800704e2290c54b1bee14a9a533df84928c7ce1ae807d2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-agent-ubi@sha256:9d1f21df008b14ca721c9ed0ff905af23e1a56b64cd4eddab4df1e8e9b63df8c
        - quay.io/mongodb/mongodb-agent-ubi@sha256:d927afa9d3d569d943518f0a684e19afd4adb0875f55df7c904b9dd019589e5e
        - quay.io/mongodb/mongodb-agent-ubi@sha256:ee150e969c0bb786969c7e0a45fb1812766b752aee539072e28648817826433b
        version: 1.33.0
      entries:
      - name: mongodb-enterprise.v1.33.0
        version: 1.33.0
      - name: mongodb-enterprise.v1.32.0
        version: 1.32.0
      - name: mongodb-enterprise.v1.31.0
        version: 1.31.0
      - name: mongodb-enterprise.v1.30.0
        version: 1.30.0
      - name: mongodb-enterprise.v1.29.0
        version: 1.29.0
      - name: mongodb-enterprise.v1.28.0
        version: 1.28.0
      - name: mongodb-enterprise.v1.27.0
        version: 1.27.0
      - name: mongodb-enterprise.v1.26.0
        version: 1.26.0
      - name: mongodb-enterprise.v1.25.0
        version: 1.25.0
      - name: mongodb-enterprise.v1.24.0
        version: 1.24.0
      - name: mongodb-enterprise.v1.23.0
        version: 1.23.0
      - name: mongodb-enterprise.v1.22.0
        version: 1.22.0
      - name: mongodb-enterprise.v1.21.0
        version: 1.21.0
      - name: mongodb-enterprise.v1.20.1
        version: 1.20.1
      - name: mongodb-enterprise.v1.19.1
        version: 1.19.1
      - name: mongodb-enterprise.v1.18.0
        version: 1.18.0
      - name: mongodb-enterprise.v1.17.2
        version: 1.17.2
      - name: mongodb-enterprise.v1.17.1
        version: 1.17.1
      - name: mongodb-enterprise.v1.17.0
        version: 1.17.0
      - name: mongodb-enterprise.v1.16.4
        version: 1.16.4
      - name: mongodb-enterprise.v1.16.3
        version: 1.16.3
      - name: mongodb-enterprise.v1.16.2
        version: 1.16.2
      - name: mongodb-enterprise.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: mongodb-enterprise
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-websocket-service
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-websocket-service.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "websocket.loadcore.keysight.com/v1alpha1",
                "kind": "Websocket",
                "metadata": {
                  "name": "websocket-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "websocket": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:26:39Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Websocket
            name: websockets.websocket.loadcore.keysight.com
            version: v1alpha1
        description: Keysight Websocket Operator
        displayName: Keysight Websocket Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Websocket Service
          url: https://keysight-websocket-service.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.connect.redhat.com/loadcore/keysight-websocket-service-bundle@sha256:da100dd9c5b542ba46de7166e783e28a0f7e675f9ac896f211405e5a9e2082ce
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-websocket-service@sha256:c91da3557ddcb8f052c9caa514d2dbf6c22b4963c080117a9d1859f1458fc404
        version: 0.1.21
      entries:
      - name: keysight-websocket-service.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-websocket-service
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/network-operator/
    name: nvidia-network-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-network-operator.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "doca3.1.0-25.07-0.9.7.0-0"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "nvcr.io/nvidia/mellanox",
                    "version": "sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
          createdAt: "2025-08-21T19:12:20Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:d41eea65b102dc5bb5792871ae212b59a65dae492ca109c52138d5626eba1ffd
        - nvcr.io/nvidia/mellanox/nic-feature-discovery@sha256:1330e3a7ea2491bc310bfc4766769389ceb56d050831bf9c145b7bfb6aba36c4
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:51270965284cceeda659fb1098e70a888b39664ceade7e9d455fb95ac85914b7
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:30354799264480e1aee057397eef2c52b277bbc6872184e2b7b811f4152b3a67
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:f9d4b8358c68ba28f849ca710bf769c4327a639733c0ff30282ca33a744078f6
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:13c2a2d055e501c03c6451413b85476ceafbc2fbffc7443952294c196f3f89f3
        - nvcr.io/nvidia/mellanox/sriov-network-device-plugin@sha256:7508a160f3f124e0b78d2b4d5a52deca6b9f56e226fee957c4d156e425eb101a
        - nvcr.io/nvidia/mellanox/nic-configuration-operator-daemon@sha256:b78c15904cecd343fa48374757b3619b5cbdc22059d5ee9731da87e35b94a2e2
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2cc940711539edd93a23355dcb06365821d140346e1df248386a77667c0f47ea
        - nvcr.io/nvidia/mellanox/ipoib-cni@sha256:7564ad1142c863e61f1dd84f5714deba9a7c6f6d85efb383ac3a2566eb4f4b71
        - nvcr.io/nvidia/mellanox/nic-configuration-operator@sha256:03f34f7fde3f65ca7b214b66c6fdbc1250d4f3863f9c02b0fba78e8aa84efca3
        - nvcr.io/nvidia/mellanox/ib-kubernetes@sha256:1fa251979b5f5201af50bc7aa6936829103d742f42bfbf752f20852f5fe44dfb
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:5ad37cd5ae96153e6f87670f437ff5f6e44febe7bb6eaeb1a0f219e28284e53d
        - nvcr.io/nvidia/mellanox/k8s-rdma-shared-dev-plugin@sha256:a87096761d155eeb6f470e042d2d167bb466d57e63b4aba957f57d745e15a9b2
        - nvcr.io/nvidia/mellanox/network-operator-init-container@sha256:02ee60ac6ee163acc97706bd4229d268f608a95bcbcb6fccee90ceb1d8a6c62b
        - nvcr.io/nvidia/mellanox/nvidia-k8s-ipam@sha256:3f692273c1598525b9c698c9d9cd463e5ddc55b638c6c031639f38786fc67475
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:423aea08b4a0f01e2e19b6a709e14347bd08f8274f40002e0224090dfa4904fc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9a831bfdf85f313b1f5749b7c9b2673bb8fff18b4ff768c9242dabaa4468e449
        version: 25.7.0
      entries:
      - name: nvidia-network-operator.v25.7.0
        version: 25.7.0
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: stable
    - currentCSV: nvidia-network-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "HostDeviceNetwork",
                "metadata": {
                  "name": "example-hostdevice-network"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.3.225/28\",\n  \"exclude\": [\n   \"192.168.3.229/30\",\n   \"192.168.3.236/32\"\n  ]\n}\n",
                  "networkNamespace": "default",
                  "resourceName": "hostdev"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "IPoIBNetwork",
                "metadata": {
                  "name": "example-ipoibnetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.6.225/28\",\n  \"exclude\": [\n   \"192.168.6.229/30\",\n   \"192.168.6.236/32\"\n  ]\n}\n",
                  "master": "ibs3f1",
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "MacvlanNetwork",
                "metadata": {
                  "name": "example-macvlannetwork"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"whereabouts\",\n  \"range\": \"192.168.2.225/24\",\n  \"exclude\": [\n   \"192.168.2.229/30\",\n   \"192.168.2.236/32\"\n  ]\n}\n",
                  "master": "ens2f0",
                  "mode": "bridge",
                  "mtu": 1500,
                  "networkNamespace": "default"
                }
              },
              {
                "apiVersion": "mellanox.com/v1alpha1",
                "kind": "NicClusterPolicy",
                "metadata": {
                  "name": "nic-cluster-policy"
                },
                "spec": {
                  "ofedDriver": {
                    "image": "doca-driver",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 30
                    },
                    "readinessProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 30
                    },
                    "repository": "nvcr.io/nvidia/mellanox",
                    "startupProbe": {
                      "initialDelaySeconds": 10,
                      "periodSeconds": 20
                    },
                    "terminationGracePeriodSeconds": 300,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": true,
                        "enable": true,
                        "force": true,
                        "podSelector": "",
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1
                    },
                    "version": "25.04-0.6.1.0-2"
                  },
                  "rdmaSharedDevicePlugin": {
                    "config": "{\n  \"configList\": [\n    {\n      \"resourceName\": \"rdma_shared_device_a\",\n      \"rdmaHcaMax\": 63,\n      \"selectors\": {\n        \"vendors\": [\"15b3\"]\n      }\n    }\n  ]\n}\n",
                    "image": "k8s-rdma-shared-dev-plugin",
                    "repository": "ghcr.io/mellanox",
                    "version": "sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
          createdAt: "2025-05-28T19:40:27Z"
          description: Deploy and manage NVIDIA networking resources in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-network-operator
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: https://github.com/Mellanox/network-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: HostDeviceNetwork
            name: hostdevicenetworks.mellanox.com
            version: v1alpha1
          - kind: IPoIBNetwork
            name: ipoibnetworks.mellanox.com
            version: v1alpha1
          - kind: MacvlanNetwork
            name: macvlannetworks.mellanox.com
            version: v1alpha1
          - kind: NicClusterPolicy
            name: nicclusterpolicies.mellanox.com
            version: v1alpha1
        description: |
          ## NVIDIA Network Operator
          The NVIDIA Network Operator simplifies the provisioning and management of NVIDIA networking resources  in a Kubernetes cluster. The operator automatically installs the required host networking software - bringing together all the needed components to provide high-speed network connectivity. These components include the NVIDIA networking driver, Kubernetes device plugin, CNI plugins, IP address management (IPAM) plugin and others.
          The NVIDIA Network Operator works in conjunction with the NVIDIA GPU Operator to deliver high-throughput, low-latency networking for scale-out, GPU computing clusters.

          The Network Operator uses Node Feature Discovery (NFD) labels in order to properly schedule resources.
          Nodes can be labelled manually or using the NFD Operator. An example of `NodeFeatureDiscovery`
          configuration is available in the documentation.
          The following NFD labels are used:
          `feature.node.kubernetes.io/pci-15b3.present` for nodes containing NVIDIA Networking hardware.
          `feature.node.kubernetes.io/pci-10de.present` for nodes containing NVIDIA GPU hardware.

          The Network Operator is open-source. For more information on contributions and release artifacts, see the [GitHub repo](https://github.com/Mellanox/network-operator/)
        displayName: NVIDIA Network Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - sriov
        - networking
        links:
        - name: Network Operator Documentation
          url: https://docs.nvidia.com/networking/software/cloud-orchestration/index.html
        - name: GitHub repository
          url: https://github.com/Mellanox/network-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/network-operator/
        relatedImages:
        - ghcr.io/mellanox/network-operator-init-container@sha256:67e93ccf3ecb61f17597567faf0f72e1b8ddcf73c5d7440baeadcc1cb6bb811b
        - ghcr.io/mellanox/ipoib-cni@sha256:513d88d639a01ef98b65365f711cee7fbaa166137a56245b7ab32818f745418e
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:4a620bd39bc70928cd6e249855176ced0f4feb585c8c7624c0b9f5a3e57664c9
        - ghcr.io/k8snetworkplumbingwg/sriov-network-device-plugin@sha256:cabce074d10a0f1d62135e2cc5442d65b49094b95b8297fdd024a1a5f461319f
        - ghcr.io/mellanox/nic-configuration-operator-daemon@sha256:0e8478898e0b30aec13030b11081b205308666c2a9c5c5bf2c00121079901b41
        - ghcr.io/mellanox/nvidia-k8s-ipam@sha256:1b20b78f889339834ed74e0da621fc5da582719b2537b36d8967ddc6a04679b8
        - nvcr.io/nvidia/cloud-native/network-operator@sha256:93c2063f4da6a93810784be95df0083f07618486699d93245c97a29a63861dee
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:76529c24ea17e68b89a2c0d4823d66504fa6f53759760a449308fda396574097
        - nvcr.io/nvidia/doca/doca_telemetry@sha256:866971717cb2683f2742d265c9793f93242f55b6e8bd9a0a88a558869f9b9cd7
        - ghcr.io/mellanox/k8s-rdma-shared-dev-plugin@sha256:1eb5ab7f4b6f0cf89c954cf7ac8eb2e6fb2290a4f5f57417136a5302f20b12c8
        - ghcr.io/mellanox/ib-kubernetes@sha256:9774f774df6ede9ddf780b9193ca9d919618056e682d985e6b41e477b13c3432
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-network-operator@sha256:2418015d00846dd0d7a8aca11927f1e89b4d8d525e6ae936360e3e3b3bd9e22f
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:56a41010b4dbee0dc31e8f9c1467a739241a2d248ed8b5054d4b418f7f79f04d
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:35ff3b173b2694198e57b61720545be67e944c1da18ac004007a5b4e68562adc
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:69e6ab6a0249be918f60c1e04fc53a07f280b33563f8edc953806538a9cc3c30
        - ghcr.io/mellanox/nic-feature-discovery@sha256:35b1820fe94a833a43613ef5c9db1f58b293ea0dbcf96ec0742b0987f23f133c
        - ghcr.io/mellanox/nic-configuration-operator@sha256:353ab27f5f5fba08197cbe8968222e1c115431cc1b9b8d8445f1facb3c3c4d7b
        - nvcr.io/nvidia/mellanox/doca-driver@sha256:9bc09051aca763f83cde99ee8909cf989b1d6c5d6ab683fbb1ca441d47f61fdf
        version: 25.4.0
      entries:
      - name: nvidia-network-operator.v25.4.0
        version: 25.4.0
      name: v25.4
    defaultChannel: stable
    packageName: nvidia-network-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/network-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Mantisnet
      provider-url: https://www.mantisnet.com
    name: tawon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tawon-operator.v2.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "Directive",
                "metadata": {
                  "name": "amf-h2"
                },
                "spec": {
                  "condition": {
                    "equal": {
                      "field": "process.name",
                      "value": "amf"
                    }
                  },
                  "container": {
                    "resources": {
                      "limits": {
                        "cpu": "500m"
                      }
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  },
                  "tasks": [
                    {
                      "config": {
                        "filter": "port 8080"
                      },
                      "task": "payload"
                    },
                    {
                      "task": "h2"
                    },
                    {
                      "config": {
                        "addr": "nats://nats:4222",
                        "topic": "hello"
                      },
                      "task": "publish"
                    }
                  ]
                }
              },
              {
                "apiVersion": "tawon.mantisnet.com/v1alpha1",
                "kind": "TopologyAggregator",
                "metadata": {
                  "name": "topology-aggregator"
                },
                "spec": {
                  "messaging": {
                    "nats": {
                      "url": "nats://nats:4222"
                    }
                  },
                  "pod": {
                    "nodeSelector": {
                      "myNodeType": "nodeType1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-11-13T01:16:29Z"
          olm.skipRange: '>=2.23.0 <2.29.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Directive allows to define a Tawon task pipeline that will
              be applied to a Kubernetes cluster.
            displayName: Tawon Directive
            kind: Directive
            name: directives.tawon.mantisnet.com
            version: v1alpha1
          - description: TopologyAggregator is the Schema for the topologyaggregators
              API
            displayName: Tawon Topology Aggregator
            kind: TopologyAggregator
            name: topologyaggregators.tawon.mantisnet.com
            version: v1alpha1
        description: Tawon is a cloud-native observability fabric for deeply instrumenting
          microservice-based applications.
        displayName: Tawon
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - networking
        - metrics
        - ebpf
        links:
        - name: Tawon Docs
          url: https://docs.tawon.xmantis.net
        maintainers:
        - email: elliott@mantisnet.com
          name: Elliott Starin
        - email: nicolas@mantisnet.com
          name: Nicolas Paton
        - email: thiago@mantisnet.com
          name: Thiago Navarro
        maturity: alpha
        provider:
          name: Mantisnet
          url: https://www.mantisnet.com
        relatedImages:
        - quay.io/mantisnet/nats@sha256:6afc7ced92181a7c12d709e9693740847297dc5d60ab78cbac7b9b6c859ecdd1
        - quay.io/mantisnet/tawon-operator@sha256:d5669dde46bb30411964baa86844d97721655886769b96809626246a4123caad
        - quay.io/mantisnet/tawon@sha256:83575079fe00ae6e53bb13005b94ef0047cc1974d95015c19e64dd53e13bcb93
        - quay.io/mantisnet/tawonctl@sha256:d436ef902bb08d830f64f34fdb86b20d2bd18d86233c8240014f0722c60e5e94
        - registry.connect.redhat.com/a5f1f0cc7ac3db90370970f9b/tawon@sha256:3691c0948c8eb266ada348de504622227af6516f3b3d331eb09fd258dd0609a5
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/mantisnet/kernel-headers@sha256:242e837af831b6eed98d522b6bae7416928b9fbe451fb6e02ffa678686771068
        - quay.io/mantisnet/kernel-headers@sha256:2aa3bc7871fe0fc0a79ac812d941e55cd0d25a5f0d1d9b3e2435aab6f3d70c67
        version: 2.29.1
      entries:
      - name: tawon-operator.v2.29.1
        version: 2.29.1
      name: alpha
    defaultChannel: alpha
    packageName: tawon-operator
    provider:
      name: Mantisnet
      url: https://www.mantisnet.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: dapr.io
      provider-url: https://dapr.io
    name: dapr-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dapr-kubernetes-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprControlPlane",
                "metadata": {
                  "name": "dapr-control-plane",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprCruiseControl",
                "metadata": {
                  "name": "dapr-cruise-control",
                  "namespace": "dapr-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.dapr.io/v1alpha1",
                "kind": "DaprInstance",
                "metadata": {
                  "name": "dapr-instance",
                  "namespace": "dapr-system"
                },
                "spec": {
                  "values": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
          createdAt: "2025-09-19T13:48:41Z"
          description: Dapr Control Plane Operator
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/dapr/kubernetes-operator
          support: contact@dapr.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DaprControlPlane is the Schema for the Dapr ControlPlane
              API.
            displayName: DaprControlPlane
            kind: DaprControlPlane
            name: daprcontrolplanes.operator.dapr.io
            version: v1alpha1
          - kind: DaprCruiseControl
            name: daprcruiscontrols.operator.dapr.io
            version: v1alpha1
          - description: DaprInstance is the Schema for the Dapr Instance API.
            displayName: DaprInstance
            kind: DaprInstance
            name: daprinstances.operator.dapr.io
            version: v1alpha1
        description: Dapr Control Plane Operator
        displayName: Dapr Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dapr
        links:
        - name: Dapr
          url: https://dapr.io
        - name: Dapr Kubernetes Operator
          url: https://github.com/dapr/kubernetes-operator
        maintainers:
        - email: lburgazzoli@gmail.com
          name: Luca Burgazzoli
        - email: salaboy@gmail.com
          name: Mauricio Salatino
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: dapr.io
          url: https://dapr.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/dapr-kubernetes-operator@sha256:8f2087df1987437f1c5e62aa9453fea220a41012077fec67b6ba8368ec267421
        - docker.io/daprio/kubernetes-operator@sha256:e1205581a07d1dd6e23367c33663014f44412eb70641115a9db8cb88890a8a57
        version: 0.0.11
      entries:
      - name: dapr-kubernetes-operator.v0.0.11
        version: 0.0.11
      - name: dapr-kubernetes-operator.v0.0.10
        version: 0.0.10
      - name: dapr-kubernetes-operator.v0.0.9
        version: 0.0.9
      - name: dapr-kubernetes-operator.v0.0.8
        version: 0.0.8
      - name: dapr-kubernetes-operator.v0.0.6
        version: 0.0.6
      - name: dapr-kubernetes-operator.v0.0.5
        version: 0.0.5
      - name: dapr-kubernetes-operator.v0.0.4
        version: 0.0.4
      - name: dapr-kubernetes-operator.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: dapr-kubernetes-operator
    provider:
      name: dapr.io
      url: https://dapr.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The cert-manager maintainers
      provider-url: https://cert-manager.io/
    name: cert-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.16.0-beta.0
        version: 1.16.0-beta.0
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.14.2-rc1
        version: 1.14.2-rc1
      - name: cert-manager.v1.14.1-rc1
        version: 1.14.1-rc1
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.3-rc1
        version: 1.13.3-rc1
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.12.2-rc1
        version: 1.12.2-rc1
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.4-rc1
        version: 1.11.4-rc1
      - name: cert-manager.v1.11.3-rc1
        version: 1.11.3-rc1
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.11.0-rc1
        version: 1.11.0-rc1
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.2-rc1
        version: 1.10.2-rc1
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.1-rc1
        version: 1.10.1-rc1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.10.0-rc1
        version: 1.10.0-rc1
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.9.1-rc1
        version: 1.9.1-rc1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.2-rc1
        version: 1.8.2-rc1
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.8.0-rc1
        version: 1.8.0-rc1
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.7.1-rc1
        version: 1.7.1-rc1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.2-rc1
        version: 1.6.2-rc1
      name: candidate
    - currentCSV: cert-manager.v1.16.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Issuer",
                "metadata": {
                  "name": "example-issuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "ClusterIssuer",
                "metadata": {
                  "name": "example-clusterissuer"
                },
                "spec": {
                  "selfSigned": {}
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "Certificate",
                "metadata": {
                  "name": "example-certificate"
                },
                "spec": {
                  "dnsNames": [
                    "example.com"
                  ],
                  "issuerRef": {
                    "name": "example-issuer"
                  },
                  "secretName": "example-certificate-tls"
                }
              },
              {
                "apiVersion": "cert-manager.io/v1",
                "kind": "CertificateRequest",
                "metadata": {
                  "name": "example-certificaterequest"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Order",
                "metadata": {
                  "name": "example-order"
                },
                "spec": {}
              },
              {
                "apiVersion": "acme.cert-manager.io/v1",
                "kind": "Challenge",
                "metadata": {
                  "name": "example-challenge"
                },
                "spec": {}
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: quay.io/jetstack/cert-manager-controller:v1.16.5
          createdAt: 2025-05-02T09:11:31
          olm.skipRange: '>=1.16.0 <1.16.5'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/internal-objects: |-
            [
              "challenges.acme.cert-manager.io",
              "orders.acme.cert-manager.io"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/cert-manager/cert-manager
          support: The cert-manager maintainers
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              A CertificateRequest is used to request a signed certificate from one of the
              configured issuers.

              All fields within the CertificateRequest's `spec` are immutable after creation.
              A CertificateRequest will either succeed or fail, as denoted by its `Ready` status
              condition and its `status.failureTime` field.

              A CertificateRequest is a one-shot resource, meaning it represents a single
              point in time request for a certificate and cannot be re-used.
            displayName: CertificateRequest
            kind: CertificateRequest
            name: certificaterequests.cert-manager.io
            version: v1
          - description: |-
              A Certificate resource should be created to ensure an up to date and signed
              X.509 certificate is stored in the Kubernetes Secret resource named in `spec.secretName`.

              The stored certificate will be renewed before it expires (as configured by `spec.renewBefore`).
            displayName: Certificate
            kind: Certificate
            name: certificates.cert-manager.io
            version: v1
          - description: Challenge is a type to represent a Challenge request with
              an ACME server
            displayName: (Internal) Challenge
            kind: Challenge
            name: challenges.acme.cert-manager.io
            version: v1
          - description: |-
              A ClusterIssuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is similar to an Issuer, however it is cluster-scoped and therefore can
              be referenced by resources that exist in *any* namespace, not just the same
              namespace as the referent.
            displayName: ClusterIssuer
            kind: ClusterIssuer
            name: clusterissuers.cert-manager.io
            version: v1
          - description: |-
              An Issuer represents a certificate issuing authority which can be
              referenced as part of `issuerRef` fields.
              It is scoped to a single namespace and can therefore only be referenced by
              resources within the same namespace.
            displayName: Issuer
            kind: Issuer
            name: issuers.cert-manager.io
            version: v1
          - description: Order is a type to represent an Order with an ACME server
            displayName: (Internal) Order
            kind: Order
            name: orders.acme.cert-manager.io
            version: v1
        description: |
          cert-manager is a Kubernetes add-on to automate the management and issuance of TLS certificates from various issuing sources.
          It will ensure certificates are valid and up to date periodically, and attempt to renew certificates at an appropriate time before expiry.

          cert-manager has the following [built in issuers][]:

          * ACME: sign certificates using Let's Encrypt (or any other [ACME][] compliant signer).
          * Vault: sign certificates using HashiCorp Vault.
          * Venafi: sign certificates using Venafi Cloud or Venafi Trust Protection Platform.
          * SelfSigned: the private key of the certificate will be used to sign the certificate.
          * CA: a private key stored as a Kubernetes Secret will be used to sign the certificate.

          [built in issuers]: https://cert-manager.io/docs/configuration/#supported-issuer-types
          [ACME]: https://datatracker.ietf.org/doc/html/rfc8555

          cert-manager also supports [external issuers][],
          which are out-of-tree components that provide support for other certificate authorities, such as:
          [Google Cloud Certificate Authority](https://github.com/jetstack/google-cas-issuer),
          [AWS Private Certificate Authority](https://github.com/cert-manager/aws-privateca-issuer), and
          [Cloudflare Origin CA](https://github.com/cloudflare/origin-ca-issuer)

          [external issuers]: https://cert-manager.io/docs/configuration/external/

          In order to configure cert-manager to issue certificates, you must first create an Issuer or ClusterIssuer.
          These resources represent a particular signing authority and tell cert-manager how to handle your certificate requests.
          The [Issuer Documentation][] has more detail.

          [configure cert-manager]: https://cert-manager.io/docs/configuration/
          [Issuer Documentation]: https://cert-manager.io/docs/concepts/issuer/

          Once an Issuer has been configured, you can create a Certificate resource,
          which is a definition of a certificate request that is to be honored by an issuer
          and which is to be kept up-to-date.
          This is the usual way that you will interact with cert-manager to request signed certificates.
          The [Certificate Documentation][] has more detail.

          [Certificate Documentation]: https://cert-manager.io/docs/usage/certificate/
        displayName: cert-manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Documentation
          url: https://cert-manager.io/
        - name: Release Notes
          url: https://cert-manager.io/docs/release-notes
        maintainers:
        - email: cert-manager-maintainers@googlegroups.com
          name: The cert-manager maintainers
        maturity: stable
        minKubeVersion: 1.19.0-0
        provider:
          name: The cert-manager maintainers
          url: https://cert-manager.io/
        relatedImages:
        - quay.io/jetstack/cert-manager-webhook:v1.16.5
        - quay.io/community-operator-pipeline-prod/cert-manager@sha256:d8037cd7653b257737cec92308042cf71d559c6c2886d07b1879b5e432eeebcd
        - quay.io/jetstack/cert-manager-cainjector:v1.16.5
        - quay.io/jetstack/cert-manager-controller:v1.16.5
        version: 1.16.5
      entries:
      - name: cert-manager.v1.16.5
        version: 1.16.5
      - name: cert-manager.v1.16.1
        version: 1.16.1
      - name: cert-manager.v1.15.2
        version: 1.15.2
      - name: cert-manager.v1.15.0
        version: 1.15.0
      - name: cert-manager.v1.14.2
        version: 1.14.2
      - name: cert-manager.v1.13.3
        version: 1.13.3
      - name: cert-manager.v1.13.1
        version: 1.13.1
      - name: cert-manager.v1.13.1-rc1
        version: 1.13.1-rc1
      - name: cert-manager.v1.12.2
        version: 1.12.2
      - name: cert-manager.v1.11.4
        version: 1.11.4
      - name: cert-manager.v1.11.0
        version: 1.11.0
      - name: cert-manager.v1.10.2
        version: 1.10.2
      - name: cert-manager.v1.10.1
        version: 1.10.1
      - name: cert-manager.v1.10.0
        version: 1.10.0
      - name: cert-manager.v1.9.1
        version: 1.9.1
      - name: cert-manager.v1.8.2
        version: 1.8.2
      - name: cert-manager.v1.8.0
        version: 1.8.0
      - name: cert-manager.v1.7.2
        version: 1.7.2
      - name: cert-manager.v1.7.1
        version: 1.7.1
      - name: cert-manager.v1.6.2
        version: 1.6.2
      - name: cert-manager.v1.6.1
        version: 1.6.1
      - name: cert-manager.v1.6.0
        version: 1.6.0
      - name: cert-manager.v1.5.4
        version: 1.5.4
      - name: cert-manager.v1.5.3
        version: 1.5.3
      - name: cert-manager.v1.4.4
        version: 1.4.4
      - name: cert-manager.v1.4.3
        version: 1.4.3
      - name: cert-manager.v1.4.2
        version: 1.4.2
      - name: cert-manager.v1.4.1
        version: 1.4.1
      - name: cert-manager.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: cert-manager
    provider:
      name: The cert-manager maintainers
      url: https://cert-manager.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ODH
      provider-url: ""
    name: opendatahub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opendatahub-operator.v2.34.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Managed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "feastoperator": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "nim": {
                        "managementState": "Managed"
                      },
                      "rawDeploymentServiceConfig": "Headed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "OpenshiftDefaultIngress"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Managed"
                    },
                    "llamastackoperator": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Managed",
                      "registriesNamespace": "odh-model-registries"
                    },
                    "ray": {
                      "managementState": "Managed"
                    },
                    "trainingoperator": {
                      "managementState": "Managed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "metrics": {},
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.34.0
          createdAt: "2025-09-03T09:59:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["featuretrackers.features.opendatahub.io",
            "codeflares.components.platform.opendatahub.io", "dashboards.components.platform.opendatahub.io",
            "datasciencepipelines.components.platform.opendatahub.io", "kserves.components.platform.opendatahub.io",
            "kueues.components.platform.opendatahub.io", "modelmeshservings.components.platform.opendatahub.io",
            "modelregistries.components.platform.opendatahub.io", "rays.components.platform.opendatahub.io",
            "trainingoperators.components.platform.opendatahub.io", "trustyais.components.platform.opendatahub.io",
            "workbenches.components.platform.opendatahub.io", "monitorings.services.platform.opendatahub.io",
            "servicemeshes.services.platform.opendatahub.io", "modelcontrollers.components.platform.opendatahub.io",
            "feastoperators.components.platform.opendatahub.io", "llamastackoperators.components.platform.opendatahub.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auth is the Schema for the auths API
            displayName: Auth
            kind: Auth
            name: auths.services.platform.opendatahub.io
            version: v1alpha1
          - description: CodeFlare is the Schema for the codeflares API
            displayName: Code Flare
            kind: CodeFlare
            name: codeflares.components.platform.opendatahub.io
            version: v1alpha1
          - description: Dashboard is the Schema for the dashboards API
            displayName: Dashboard
            kind: Dashboard
            name: dashboards.components.platform.opendatahub.io
            version: v1alpha1
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DataSciencePipelines is the Schema for the datasciencepipelines
              API
            displayName: Data Science Pipelines
            kind: DataSciencePipelines
            name: datasciencepipelines.components.platform.opendatahub.io
            version: v1alpha1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSCInitialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - description: FeastOperator is the Schema for the FeastOperator API
            displayName: Feast Operator
            kind: FeastOperator
            name: feastoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
          - description: HardwareProfile is the Schema for the hardwareprofiles API.
            displayName: Hardware Profile
            kind: HardwareProfile
            name: hardwareprofiles.infrastructure.opendatahub.io
            version: v1alpha1
          - description: Kserve is the Schema for the kserves API
            displayName: Kserve
            kind: Kserve
            name: kserves.components.platform.opendatahub.io
            version: v1alpha1
          - description: Kueue is the Schema for the kueues API
            displayName: Kueue
            kind: Kueue
            name: kueues.components.platform.opendatahub.io
            version: v1alpha1
          - description: LlamaStackOperator is the Schema for the LlamaStackOperator
              API
            displayName: Llama Stack Operator
            kind: LlamaStackOperator
            name: llamastackoperators.components.platform.opendatahub.io
            version: v1alpha1
          - kind: ModelController
            name: modelcontrollers.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelMeshServing is the Schema for the modelmeshservings
              API
            displayName: Model Mesh Serving
            kind: ModelMeshServing
            name: modelmeshservings.components.platform.opendatahub.io
            version: v1alpha1
          - description: ModelRegistry is the Schema for the modelregistries API
            displayName: Model Registry
            kind: ModelRegistry
            name: modelregistries.components.platform.opendatahub.io
            version: v1alpha1
          - description: Monitoring is the Schema for the monitorings API
            displayName: Monitoring
            kind: Monitoring
            name: monitorings.services.platform.opendatahub.io
            version: v1alpha1
          - description: Ray is the Schema for the rays API
            displayName: Ray
            kind: Ray
            name: rays.components.platform.opendatahub.io
            version: v1alpha1
          - description: ServiceMesh is the Schema for the servicemesh API
            displayName: Service Mesh
            kind: ServiceMesh
            name: servicemeshes.services.platform.opendatahub.io
            version: v1alpha1
          - description: TrainingOperator is the Schema for the trainingoperators
              API
            displayName: Training Operator
            kind: TrainingOperator
            name: trainingoperators.components.platform.opendatahub.io
            version: v1alpha1
          - description: TrustyAI is the Schema for the trustyais API
            displayName: Trusty AI
            kind: TrustyAI
            name: trustyais.components.platform.opendatahub.io
            version: v1alpha1
          - description: Workbenches is the Schema for the workbenches API
            displayName: Workbenches
            kind: Workbenches
            name: workbenches.components.platform.opendatahub.io
            version: v1alpha1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Argo
          Workflows\n* Model Mesh - ModelMesh Serving is the Controller for managing
          ModelMesh, a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks\n* Feast - Feature
          Store is the fastest path to manage existing infrastructure to productionize
          analytic data for model training and online inference.\n* Llama Stack –
          unified open‑source APIs for inference, RAG, agents, safety, evals & telemetry"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed workloads
        - trustyai
        - modelregistry
        - feast
        - featurestore
        - llamastack
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opendatahub-operator:2.34.0
        - quay.io/opendatahub/opendatahub-operator:v2.34.0
        version: 2.34.0
      entries:
      - name: opendatahub-operator.v2.34.0
        version: 2.34.0
      - name: opendatahub-operator.v2.33.0
        version: 2.33.0
      - name: opendatahub-operator.v2.32.0
        version: 2.32.0
      - name: opendatahub-operator.v2.31.0
        version: 2.31.0
      - name: opendatahub-operator.v2.30.0
        version: 2.30.0
      - name: opendatahub-operator.v2.29.0
        version: 2.29.0
      - name: opendatahub-operator.v2.28.0
        version: 2.28.0
      - name: opendatahub-operator.v2.27.0
        version: 2.27.0
      - name: opendatahub-operator.v2.26.0
        version: 2.26.0
      - name: opendatahub-operator.v2.25.0
        version: 2.25.0
      - name: opendatahub-operator.v2.24.0
        version: 2.24.0
      - name: opendatahub-operator.v2.23.0
        version: 2.23.0
      - name: opendatahub-operator.v2.22.0
        version: 2.22.0
      - name: opendatahub-operator.v2.21.0
        version: 2.21.0
      - name: opendatahub-operator.v2.20.0
        version: 2.20.0
      - name: opendatahub-operator.v2.19.0
        version: 2.19.0
      - name: opendatahub-operator.v2.18.2
        version: 2.18.2
      - name: opendatahub-operator.v2.18.1
        version: 2.18.1
      - name: opendatahub-operator.v2.18.0
        version: 2.18.0
      - name: opendatahub-operator.v2.17.0
        version: 2.17.0
      - name: opendatahub-operator.v2.16.0
        version: 2.16.0
      - name: opendatahub-operator.v2.15.0
        version: 2.15.0
      - name: opendatahub-operator.v2.14.0
        version: 2.14.0
      - name: opendatahub-operator.v2.13.0
        version: 2.13.0
      - name: opendatahub-operator.v2.12.1
        version: 2.12.1
      - name: opendatahub-operator.v2.12.0
        version: 2.12.0
      - name: opendatahub-operator.v2.11.1
        version: 2.11.1
      - name: opendatahub-operator.v2.11.0
        version: 2.11.0
      - name: opendatahub-operator.v2.10.0
        version: 2.10.0
      name: fast
    - currentCSV: opendatahub-operator.v2.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datasciencecluster.opendatahub.io/v1",
                "kind": "DataScienceCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "datasciencecluster",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsc"
                },
                "spec": {
                  "components": {
                    "codeflare": {
                      "managementState": "Removed"
                    },
                    "dashboard": {
                      "managementState": "Managed"
                    },
                    "datasciencepipelines": {
                      "managementState": "Managed"
                    },
                    "kserve": {
                      "managementState": "Managed",
                      "serving": {
                        "ingressGateway": {
                          "certificate": {
                            "type": "SelfSigned"
                          }
                        },
                        "managementState": "Managed",
                        "name": "knative-serving"
                      }
                    },
                    "kueue": {
                      "managementState": "Removed"
                    },
                    "modelmeshserving": {
                      "managementState": "Managed"
                    },
                    "modelregistry": {
                      "managementState": "Removed"
                    },
                    "ray": {
                      "managementState": "Removed"
                    },
                    "trustyai": {
                      "managementState": "Managed"
                    },
                    "workbenches": {
                      "managementState": "Managed"
                    }
                  }
                }
              },
              {
                "apiVersion": "dscinitialization.opendatahub.io/v1",
                "kind": "DSCInitialization",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dscinitialization",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-dsci"
                },
                "spec": {
                  "applicationsNamespace": "opendatahub",
                  "monitoring": {
                    "managementState": "Managed",
                    "namespace": "opendatahub"
                  },
                  "serviceMesh": {
                    "controlPlane": {
                      "metricsCollection": "Istio",
                      "name": "data-science-smcp",
                      "namespace": "istio-system"
                    },
                    "managementState": "Managed"
                  },
                  "trustedCABundle": {
                    "customCABundle": "",
                    "managementState": "Managed"
                  }
                }
              },
              {
                "apiVersion": "features.opendatahub.io/v1",
                "kind": "FeatureTracker",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "opendatahub-operator",
                    "app.kubernetes.io/instance": "default",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "default-feature",
                    "app.kubernetes.io/part-of": "opendatahub-operator"
                  },
                  "name": "default-feature"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v2.8.0
          createdAt: "2024-2-20T00:00:00Z"
          operatorframework.io/initialization-resource: |-
            {
            "apiVersion": "dscinitialization.opendatahub.io/v1",
            "kind": "DSCInitialization",
            "metadata": {
            "labels": {
             "app.kubernetes.io/created-by": "opendatahub-operator",
             "app.kubernetes.io/instance": "default",
             "app.kubernetes.io/managed-by": "kustomize",
             "app.kubernetes.io/name": "dscinitialization",
             "app.kubernetes.io/part-of": "opendatahub-operator"
            },
            "name": "default-dsci"
            },
            "spec": {
            "applicationsNamespace": "opendatahub",
            "monitoring": {
             "managementState": "Managed",
             "namespace": "opendatahub"
            },
            "serviceMesh": {
             "controlPlane": {
               "metricsCollection": "Istio",
               "name": "data-science-smcp",
               "namespace": "istio-system"
             },
             "managementState": "Managed"
            },
            "trustedCABundle": {
             "customCABundle": "",
             "managementState": "Managed"
            }
            }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/internal-objects: '[dscinitialization.opendatahub.io]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DataScienceCluster is the Schema for the datascienceclusters
              API.
            displayName: Data Science Cluster
            kind: DataScienceCluster
            name: datascienceclusters.datasciencecluster.opendatahub.io
            version: v1
          - description: DSCInitialization is the Schema for the dscinitializations
              API.
            displayName: DSC Initialization
            kind: DSCInitialization
            name: dscinitializations.dscinitialization.opendatahub.io
            version: v1
          - kind: FeatureTracker
            name: featuretrackers.features.opendatahub.io
            version: v1
        description: "The Open Data Hub is a machine-learning-as-a-service platform
          built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open
          Data Hub integrates multiple AI/ML open source components into one operator
          that can easily be downloaded and installed by OpenShift users.\n\nOpen
          Data Hub operator allows users to install and manage components of the Open
          Data Hub. Users can mix and match tools from each project to fulfill the
          needs of their use case. Each of the projects share some components, but
          can be mostly seen as an extension of each other to provide a complete solution
          for both novice and skilled enterprise users.\n    \n### Components\n\n*
          Open Data Hub Dashboard - A web dashboard that displays installed Open Data
          Hub components with easy access to component UIs and documentation\n* ODH
          Notebook Controller - Secure management of Jupyter Notebook in Kubernetes
          environments built on top of Kubeflow Notebook Controller with support for
          OAuth\n* Jupyter Notebooks - JupyterLab notebook that provide Python support
          for GPU workloads\n* Data Science Pipelines - Pipeline solution for end
          to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton\n*
          Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh,
          a general-purpose model serving management/routing layer\n* Distributed
          Workloads(Incubation) - Stack built to make managing distributed compute
          infrastructure in the cloud easy and intuitive for Data Scientists. This
          stack consists of three components \n                                      Codeflare
          , KubeRay and Kueue.\n* Kserve - Kserve is the Controller for for serving
          machine learning (ML) models on arbitrary frameworks"
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odh
        - notebooks
        - serving
        - training
        - pipelines
        - modelmesh
        - workbenches
        - dashboard
        - kserve
        - distributed-workloads
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: ODH
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v2.8.0
        - quay.io/community-operator-pipeline-prod/opendatahub-operator@sha256:d3e3d5ae02b834522481fa9046bf2ad3ea06d706c5cb06a32d32e61e229e4a6c
        version: 2.8.1
      entries:
      - name: opendatahub-operator.v2.8.1
        version: 2.8.1
      name: odh-2.8.z
    - currentCSV: opendatahub-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator/"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "parameters": [
                          {
                            "name": "monitoring-namespace",
                            "value": "opendatahub"
                          }
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "file:///opt/manifests/odh-manifests.tar.gz"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.11.0
          createdAt: "2023-10-31T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AcceleratorProfile is the Schema for the acceleratorprofiles
            displayName: AcceleratorProfile
            kind: AcceleratorProfile
            name: acceleratorprofiles.dashboard.opendatahub.io
            version: v1alpha
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.
          * TrustyAI - TrustyAI is an open source AI Explainability (XAI) Toolkit

          To install one or multiple of these components use the default KfDef provided with the operator.
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io/opendatahub-community
        maintainers:
        - email: opendatahub@redhat.com
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - quay.io/opendatahub/opendatahub-operator:v1.11.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:2992841660e8b1664fa2ce06da894aa828d4ea1be3eb5b27476efdd73731486a
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.11.0
      entries:
      - name: opendatahub-operator.v1.11.0
        version: 1.11.0
      - name: opendatahub-operator.v1.10.1
        version: 1.10.1
      - name: opendatahub-operator.v1.10.0
        version: 1.10.0
      - name: opendatahub-operator.v1.9.0
        version: 1.9.0
      - name: opendatahub-operator.v1.8.0
        version: 1.8.0
      - name: opendatahub-operator.v1.7.0
        version: 1.7.0
      - name: opendatahub-operator.v1.6.0
        version: 1.6.0
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: rolling
    - currentCSV: opendatahub-operator.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kfdef.apps.kubeflow.org/v1",
                "kind": "KfDef",
                "metadata": {
                  "name": "opendatahub",
                  "namespace": "opendatahub"
                },
                "spec": {
                  "applications": [
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-common"
                        }
                      },
                      "name": "odh-common"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-dashboard"
                        }
                      },
                      "name": "odh-dashboard"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/cluster"
                        }
                      },
                      "name": "prometheus-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "prometheus/operator"
                        }
                      },
                      "name": "prometheus-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/cluster"
                        }
                      },
                      "name": "grafana-cluster"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "grafana/grafana"
                        }
                      },
                      "name": "grafana-instance"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "odh-notebook-controller"
                        }
                      },
                      "name": "odh-notebook-controller"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "notebook-images"
                        }
                      },
                      "name": "notebook-images"
                    },
                    {
                      "kustomizeConfig": {
                        "overlays": [
                          "odh-model-controller"
                        ],
                        "repoRef": {
                          "name": "manifests",
                          "path": "model-mesh"
                        }
                      },
                      "name": "model-mesh"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "data-science-pipelines-operator"
                        }
                      },
                      "name": "data-science-pipelines-operator"
                    },
                    {
                      "kustomizeConfig": {
                        "repoRef": {
                          "name": "manifests",
                          "path": "trustyai-service"
                        }
                      },
                      "name": "trustyai"
                    }
                  ],
                  "repos": [
                    {
                      "name": "manifests",
                      "uri": "https://github.com/opendatahub-io/odh-manifests/tarball/v1.5"
                    }
                  ]
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhApplication",
                "metadata": {
                  "name": "example-odhapplication",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started,Model development,Model training,Notebook environments,Package management,Data Management"
                  }
                },
                "spec": {
                  "displayName": "ODHAPPLICATION DISPLAY NAME",
                  "provider": "PROVIDER_NAME",
                  "description": "Description of your ODHApplication",
                  "kfdefApplications": [
                    "<NAME OF THE ODH APPLICATION IN KFDEF>"
                  ],
                  "route": "<APPLICATION_OCP_ROUTE_OBJECT_NAME>",
                  "img": "<svg xmlns=\"http://www.w3.org/2000/svg\" role=\"img\" viewBox=\"-1.13 -1.13 58.50 48.94\"><title>Open Data Hub logo</title><g><g><g fill=\"#4d5258\" stroke-width=\".006\"><path fill=\"#bbb\" d=\"M112.30231 22.797083c-.2034 0-.35737.143517-.35737.337407 0 .192939.15397.339307.35737.339307.20339 0 .35736-.147318.35736-.339307 0-.19294-.15397-.337407-.35736-.337407zm.002.131161c.1112 0 .20149.08934.20149.206246 0 .116904-.0903.208146-.20149.208146-.1112 0-.2053-.09124-.2053-.208146 0-.116904.0931-.206246.2053-.206246zM113.04141 22.801836h-.28228v.665308h.15017v-.190088h.13211c.16918 0 .26612-.08934.26612-.244263 0-.147319-.0969-.230957-.26612-.230957zm-.008.349762h-.12451v-.225255h.12451c.0836 0 .13116.03707.13116.111202 0 .07603-.0475.114053-.13116.114053zM113.40328 22.801836v.665308h.51703v-.124508h-.36687v-.147318h.32125V23.07081h-.32125v-.144467h.35642v-.124507zM114.03814 22.801836v.665308h.14352v-.426748l.3222.426748h.13686v-.665308h-.14257v.427698L114.175 22.801836z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/><path d=\"M115.29685 22.801836h-.28228v.665308h.27658c.21004 0 .35546-.136863.35546-.332654 0-.196741-.14351-.332654-.34976-.332654zm.004.538899h-.13591v-.412491h.12926c.11405 0 .19959.08459.19959.207196 0 .121657-.0817.205295-.19294.205295zM116.21878 23.467144h.16062l-.28418-.665308h-.15492l-.28894.665308h.15493l.0523-.12926h.30794zm-.3108-.249966l.1055-.25852.10455.25852zM116.31525 22.801836v.127359h.20244v.537949h.15017v-.537949h.2034v-.127359zM117.37527 23.467144h.16063l-.28419-.665308h-.15492l-.28893.665308h.15492l.0523-.12926h.30795zm-.31079-.249966l.1055-.25852.10454.25852zM117.83433 22.801836v.665308h.15017v-.262322h.30129v.262322h.15017v-.665308h-.15017v.277528h-.30129v-.277528zM118.72025 23.188665v-.386829h-.15017v.386829c0 .175831.11785.285132.30224.285132.18343 0 .29844-.109301.29844-.285132v-.386829h-.15017v.386829c0 .09599-.0589.153971-.14827.153971-.0893 0-.15207-.05798-.15207-.153971zM119.61059 22.801836h-.30604v.665308h.31649c.15492 0 .25092-.07033.25092-.184385 0-.08744-.0532-.152071-.13876-.168228.0713-.01806.115-.07413.115-.146368 0-.102648-.0912-.166327-.23761-.166327zm-.008.268024h-.14826v-.145418h.14826c.0561 0 .0913.02566.0913.07128 0 .04657-.0352.07413-.0913.07413zm0 .274677h-.14826v-.158723h.14826c.0713 0 .11406.02946.11406.07794 0 .05132-.0428.08079-.11406.08079z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g><g fill=\"#58595b\" stroke-width=\".01\"><path d=\"M112.19192 24.15027h.0381l-.13717-.306759h-.0364l-.1376.306759h.0372l.0342-.07844h.16696zm-.18712-.110871l.0688-.157763.0697.157763zM112.27858 23.843511v.306759h.0351v-.306759zM112.64205 23.843511h-.11525v.306759h.0351v-.09729h.0802c.0758 0 .11964-.039.11964-.10649 0-.0653-.0438-.102983-.11964-.102983zm-.001.177044h-.0789V23.87594h.0789c.0552 0 .0872.02454.0872.07143 0 .0482-.032.07318-.0872.07318zM112.81775 23.825105v.325165h.0342v-.325165zM113.1137 24.15027l-.00044-.152503c-.00044-.04908-.032-.07932-.089-.07932-.0394 0-.0644.01139-.0942.03111l.0145.0241c.0241-.01709.0478-.02673.0732-.02673.0408 0 .0618.0206.0618.05434v.01446h-.0758c-.0543.000439-.0863.02586-.0863.06617 0 .03944.0307.06968.0784.06968.0368 0 .0662-.01227.0846-.03725v.03593zm-.11131-.02717c-.0324 0-.053-.01709-.053-.04251 0-.02585.0202-.03769.0587-.03769h.0719v.02805c-.009.03287-.0386.05215-.0776.05215zM113.29915 24.108201c-.0127.0074-.025.01139-.0368.01183-.0197 0-.0307-.01139-.0307-.03944v-.127086h.0732v-.02673h-.0732v-.06048h-.0337v.06048h-.0333v.02673h.0333v.132783c0 .04558.025.06573.0587.0653.0184 0 .0359-.0061.0521-.01709zM113.44228 23.854466c.0118 0 .0245.0044.0351.01183l.014-.02629c-.014-.01052-.032-.01622-.0495-.01622-.0386-.000438-.0741.02717-.0741.07713v.02586h-.0298v.02673h.0298v.196764h.0337v-.196764h.0662v-.02673h-.0662v-.03024c0-.02717.0184-.04251.0408-.04207zM113.59236 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM113.80323 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM114.22747 23.918448c-.0443.000438-.0767.01972-.0911.06004-.01-.03813-.0394-.06004-.0828-.06004-.0416.000438-.0727.01753-.0881.05346v-.05215h-.0337v.230508h.0337v-.113501c0-.04996.0285-.08458.0758-.08589.0403 0 .064.02454.064.06617v.133221h.0346v-.113501c0-.04996.0276-.08458.0754-.08589.0399 0 .064.02454.064.06617v.133221h.0342V24.00828c0-.05609-.032-.08984-.0859-.08984zM114.63474 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0487.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11088-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM114.89934 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .117-.04777.117-.117007 0-.06836-.0486-.11613-.117-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM115.02844 23.919762l.0863.230508h.0355l.0675-.188876.0679.188876h.0355l.0855-.230508h-.0355l-.0675.195888-.0679-.195888h-.0359l-.067.195888-.0684-.195888zM115.53009 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM115.73075 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM115.95034 23.918448c-.067.000438-.11438.04777-.11438.116568 0 .0688.0469.116569.11482.116569.0386 0 .071-.01402.0929-.03813l-.0188-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18843c.004-.07801-.0333-.129277-.1078-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM116.29474 23.825105v.144615c-.0171-.03287-.0491-.05171-.0903-.05171-.0653 0-.11087.04777-.11087.116569 0 .0688.0456.117445.11131.117445.0412 0 .0727-.01884.0898-.05127v.04952h.0342v-.325165zm-.0833.297557c-.0491 0-.0837-.03637-.0837-.08721 0-.05083.0346-.08765.0837-.08765.0486 0 .0833.03681.0833.08765 0 .05083-.0346.08721-.0833.08721zM116.65984 23.918448c-.0403 0-.0719.01884-.089.05083v-.144177h-.0342v.325165h.0342v-.04952c.0171.03199.0486.05083.0898.05083.0653 0 .11-.04777.11-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM116.89657 24.183138l.10605-.263376h-.0351l-.0732.191944-.0767-.191944h-.0359l.0955.228756-.01.02191c-.009.02191-.0241.03331-.0421.03331-.0114 0-.0219-.0035-.0324-.01095l-.0145.02717c.014.01139.0294.01665.0486.01665.0307 0 .0548-.01841.0692-.05346zM117.29947 23.842634c-.0898 0-.16083.06793-.16083.154257 0 .08633.071.154694.16083.154694.0894 0 .16039-.06836.16039-.154694 0-.08633-.071-.154257-.16039-.154257zm0 .03243c.0684 0 .12446.0539.12446.121828 0 .06792-.0561.122703-.12446.122703-.0684 0-.12533-.05478-.12533-.122703 0-.06793.057-.121828.12533-.121828zM117.64414 23.918448c-.0403 0-.0719.01884-.089.05083v-.04952h-.0342v.315525h.0342v-.134536c.0171.03199.0486.05083.0898.05083.0653 0 .10999-.04777.10999-.11613 0-.0688-.0456-.117007-.11087-.117007zm-.006.202899c-.0486 0-.0828-.03593-.0828-.08677 0-.0504.0342-.08677.0828-.08677.0482 0 .0824.03681.0824.08677 0 .0504-.0342.08677-.0824.08677zM117.9061 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362zM118.19616 23.918448c-.0421.000438-.0732.01709-.0894.05215v-.05084h-.0342v.230508h.0342v-.12358c.004-.0447.032-.07494.0776-.07537.0408 0 .0653.02454.0653.06573v.133221h.0342v-.141986c0-.05609-.0324-.08984-.0876-.08984zM118.65836 23.901795l.0149-.03111c-.0254-.01709-.0618-.02848-.0973-.02848-.0635 0-.10605.03068-.10605.07888 0 .109995.17266.06661.17222.148559 0 .0298-.0276.04777-.0719.04777-.0346 0-.0736-.01578-.1021-.04295l-.0158.03024c.0289.02761.0727.04689.11745.04689.0657 0 .11087-.03287.11087-.0837.00044-.11131-.17223-.06968-.17223-.149436 0-.02629.0259-.04207.0653-.04207.0245 0 .057.0075.0846.02542zM118.84168 23.918448c-.0684 0-.11701.04777-.11701.11613 0 .06924.0486.117007.11701.117007.0684 0 .11701-.04777.11701-.117007 0-.06836-.0487-.11613-.11701-.11613zm0 .03024c.0482 0 .0828.0355.0828.08633 0 .05171-.0346.08677-.0828.08677-.0482 0-.0828-.03506-.0828-.08677 0-.05083.0346-.08633.0828-.08633zM119.18446 23.919762v.113063c0 .04996-.0281.08502-.0758.08589-.0399 0-.064-.02454-.064-.06573v-.133222h-.0342v.141986c0 .05565.0324.08984.0863.08984.0416-.000438.0723-.01753.0876-.05346v.05215h.0342v-.230508zM119.33586 23.971035v-.05127h-.0342v.230508h.0342V24.02757c.004-.04645.0342-.07669.0828-.07581v-.0333c-.039.000438-.0675.01884-.0828.05259zM119.62557 23.976294l.0188-.02366c-.021-.02147-.0517-.03418-.089-.03418-.0666 0-.11438.0482-.11438.116568 0 .0688.0473.116569.11438.116569.0399 0 .0719-.0149.0925-.03988l-.0197-.02016c-.0162.01928-.0412.0298-.0714.0298-.0478 0-.082-.03506-.082-.08633 0-.05083.0342-.08589.082-.08589.028 0 .0513.01008.0688.02717zM119.78997 23.918448c-.067.000438-.11437.04777-.11437.116568 0 .0688.0469.116569.11481.116569.0386 0 .071-.01402.0929-.03813l-.0189-.01972c-.018.01884-.0429.02936-.0714.02936-.0451 0-.078-.0298-.0837-.07538h.18844c.004-.07801-.0333-.129277-.10781-.129277zm-.0806.102107c.005-.0447.0364-.07362.0806-.07362.0451 0 .0749.02892.0771.07362z\" transform=\"translate(-789.851 -124.305) scale(7.05736)\"/></g></g><g><g><path fill=\"#a586f3\" d=\"m 7808.8125,2097.1875 c -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 v 329.875 c -0.8781,10.2195 7.1804,19 17.4375,19 10.2571,0 19.0227,-8.7805 18.1446,-19 v -77.7324 l 0.4179,-6.1426 c 0,-55.1597 44.2153,-99.5625 99.375,-99.5625 55.1596,0 99.5625,44.4028 99.5625,99.5625 0.01,0.1837 0.015,0.3674 0.027,0.5508 -0.026,0.4409 -0.035,0.8826 -0.027,1.3242 v 110 c 0,9.6822 7.849,17.5312 17.5312,17.5312 9.6823,0 17.5313,-7.849 17.5313,-17.5312 v -110 c 0,-0.3245 -0.012,-0.6488 -0.031,-0.9727 0.018,-0.3004 0.029,-0.6013 0.031,-0.9023 1e-4,-74.075 -60.55,-134.4374 -134.625,-134.4375 -39.9177,1e-4 -75.8465,17.5354 -100.5,45.293 V 2115 c 0.1749,-9.8356 -7.7881,-17.8833 -17.625,-17.8125 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><path fill=\"#5ed3ee\" d=\"m 7765.8027,2098.123 c -0.038,5e-4 -0.077,0 -0.1152,0 -9.6898,0.1378 -17.423,8.1233 -17.25,17.8125 V 2361 c 0,41.9333 -26.0162,79.2671 -65.4375,93.5625 -39.4213,14.2954 -83.5594,2.3741 -110.4375,-29.8125 -26.878,-32.1865 -30.6579,-77.5719 -9.5625,-113.8125 21.0955,-36.2406 62.4262,-55.2897 103.6875,-47.8125 9.5173,1.7397 18.6463,-4.5547 20.4043,-14.0686 1.758,-9.514 -4.5187,-18.6551 -14.0293,-20.4314 -55.6354,-10.082 -111.8054,15.8219 -140.25,64.6875 -28.4446,48.8656 -23.3039,110.5381 12.9375,153.9375 36.2415,43.3993 95.908,59.4005 149.0625,40.125 44.7667,-16.2337 77.0286,-54.658 86.0977,-100.1562 0,0 1.9812,-4.7686 2.5898,-9.1563 0.6086,-4.3877 0,-17.0625 0,-17.0625 v -245.0625 c 0.1769,-9.865 -7.8313,-17.9263 -17.6973,-17.8145 z\" color=\"#000\" transform=\"translate(-449.265 -129.165) scale(.06159)\" style=\"text-indent:0;text-align:start;text-decoration-line:none;text-decoration-style:solid;text-decoration-color:#000;text-transform:none;text-orientation:mixed;shape-padding:0;isolation:auto;mix-blend-mode:normal;solid-color:#000;solid-opacity:1\" overflow=\"visible\"/><g fill=\"none\" stroke-width=\".365\" transform=\"translate(-449.265 -83.342) scale(5.91262)\"><circle cx=\"79.677\" cy=\"16.823\" r=\".645\" stroke=\"#c6ea70\"/><path stroke=\"#dff3b1\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 80.808656,18.234303 a 1.8088735,1.8088735 0 0 1 -2.309183,-0.03848 1.8088735,1.8088735 0 0 1 -0.389241,-2.276466 1.8088735,1.8088735 0 0 1 2.165211,-0.803532\"/><path stroke=\"#f2fadf\" stroke-linecap=\"round\" stroke-linejoin=\"round\" d=\"m 82.216272,17.248644 a 2.455857,2.4425189 0 0 1 -2.468226,2.016391 2.455857,2.4425189 0 0 1 -2.383743,-2.114547 2.455857,2.4425189 0 0 1 1.733108,-2.668494\"/></g></g></g></g></svg>",
                  "category": "Open Data Hub",
                  "support": "open data hub",
                  "docsLink": "http://REPLACE.THIS.URL",
                  "quickStart": "example-odhquickstart",
                  "getStartedLink": "http://REPLACE.THIS.URL",
                  "getStartedMarkDown": "# Header for Getting Started MarkDown section\nHere is where you add information about your ODHApplication in markdown format\n\n1. List item\n1. List item\n1. List item"
                }
              },
              {
                "apiVersion": "dashboard.opendatahub.io/v1",
                "kind": "OdhDocument",
                "metadata": {
                  "name": "example-odhdocument",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHDocument",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "type": "how-to",
                  "description": "This is an example of an ODHDocument custom resource",
                  "url": "http://REPLACE.THIS.URL",
                  "durationMinutes": 5
                }
              },
              {
                "apiVersion": "console.openshift.io/v1",
                "kind": "OdhQuickStart",
                "metadata": {
                  "name": "example-odhquickstart",
                  "annotations": {
                    "opendatahub.io/categories": "Getting started"
                  }
                },
                "spec": {
                  "displayName": "Example ODHQuickstart Name",
                  "appName": "<EXAMPLE-ODHAPPLICATION-CUSTOM-RESOURCE-NAME>",
                  "durationMinutes": 5,
                  "description": "Description of an example quickstart",
                  "introduction": "### Introduction Header\nThis is the body of the introduction for an ODHQuickstart",
                  "tasks": [
                    {
                      "title": "Title of the first task",
                      "description": "### Description header for the first task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the first task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    },
                    {
                      "title": "Title of the second task",
                      "description": "### Description header for the second task\n1. First Step\n2. Second Step\n3. Third Step",
                      "review": {
                        "instructions": "#### Header for the review instructions\nThis is the body of the review instructions for the second task",
                        "failedTaskHelp": "This task is not verified yet. Try the task again."
                      },
                      "summary": {
                        "success": "Success Message",
                        "failed": "Failure Message"
                      }
                    }
                  ],
                  "conclusion": "Message for the conclusion of the ODHQuickstart",
                  "nextQuickStart": null
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, Big Data
          certified: "False"
          containerImage: quay.io/opendatahub/opendatahub-operator:v1.5.0
          createdAt: "2022-10-18T00:00:00Z"
          description: Open Data Hub Operator for deployment and management of Open
            Data Hub
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/opendatahub-io/opendatahub-operator
          support: Open Data Hub
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KfDef is the Schema for the kfdefs API
            displayName: Kf Def
            kind: KfDef
            name: kfdefs.kfdef.apps.kubeflow.org
            version: v1
          - description: OdhApplication is the Schema for the odhapplications
            displayName: Open Data Hub Dashboard Application
            kind: OdhApplication
            name: odhapplications.dashboard.opendatahub.io
            version: v1
          - description: OdhDashboardConfig is the Schema for the odhdashboardconfigs
            displayName: Open Data Hub Dashboard Configuration
            kind: OdhDashboardConfig
            name: odhdashboardconfigs.opendatahub.io
            version: v1alpha
          - description: OdhDocument is the Schema for the odhdocuments
            displayName: Open Data Hub Dashboard Documentation
            kind: OdhDocument
            name: odhdocuments.dashboard.opendatahub.io
            version: v1
          - description: Extension for guiding user through various workflows in the
              Open Data Hub dashboard.
            displayName: Open Data Hub Dashboard QuickStart
            kind: OdhQuickStart
            name: odhquickstarts.console.openshift.io
            version: v1
        description: |
          The Open Data Hub is a machine-learning-as-a-service platform built on Red Hat's Kubernetes-based OpenShift® Container Platform. Open Data Hub integrates multiple AI/ML open source components into one operator that can easily be downloaded and installed by OpenShift users.

          Open Data Hub operator allows users to install and manage components of the Open Data Hub. Users can mix and match tools from each project to fulfill the needs of their use case. Each of the
          projects share some components, but can be mostly seen as an extension of each other to provide a complete solution for both novice and skilled enterprise users.

          ### Components
          * Open Data Hub Dashboard - A web dashboard that displays installed Open Data Hub components with easy access to component UIs and documentation
          * ODH Notebook Controller - Secure management of Jupyter Notebook in Kubernetes environments built on top of Kubeflow Notebook Controller with support for OAuth
          * Jupyter Notebooks - JupyterLab notebook that provide Python support for GPU workloads
          * Data Science Pipelines - Pipeline solution for end to end MLOps workflows that support the Kubeflow Pipelines SDK and Tekton
          * Model Mesh - ModelMesh Serving is the Controller for managing ModelMesh, a general-purpose model serving management/routing layer.

          To install one or multiple of these components use the default KfDef provided with the operator.

          ### Available Channels

          #### Stable

          Channel `stable` offers the major releases of Open Data Hub operator and ODH manifests. It is based on Kubeflow Operator and Kustomize for deployment configuration. This version is compatible with [Open Data Hub manifests](https://github.com/opendatahub-io/odh-manifests/) as well as [Kubeflow manifests](https://github.com/opendatahub-io/manifests). This channel will offer new components and features that have been thoroughly tested by consumers of the `rolling` channel

          NOTE: The `stable` channel will be deprecated starting with the ODH 1.6 release and we will only be publishing updates on the `rolling` channel for all future releases

          #### Rolling

          Channel `rolling` will offer the latest release of Open Data Hub operator and manifests. Subscribing to this branch will provide faster updates and access to components that will be in developement and testing for the next stable release
        displayName: Open Data Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Operator
        - OpenShift
        - Open Data Hub
        - opendatahub
        links:
        - name: Open Data Hub
          url: https://opendatahub.io
        - name: Open Data Hub Community
          url: https://github.com/opendatahub-io
        maintainers:
        - email: maintainers@lists.opendatahub.io
          name: Open Data Hub Maintainers
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Open Data Hub
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opendatahub/opendatahub-operator:v1.5.0
        - quay.io/openshift-community-operators/opendatahub-operator@sha256:bf85a9da8a100a90deb3745497bc1505852910ef23dd12c624fa17de51bd8377
        version: 1.5.0
      entries:
      - name: opendatahub-operator.v1.5.0
        version: 1.5.0
      - name: opendatahub-operator.v1.4.2
        version: 1.4.2
      - name: opendatahub-operator.v1.4.1
        version: 1.4.1
      - name: opendatahub-operator.v1.4.0
        version: 1.4.0
      - name: opendatahub-operator.v1.3.0
        version: 1.3.0
      - name: opendatahub-operator.v1.2.0
        version: 1.2.0
      - name: opendatahub-operator.v1.1.2
        version: 1.1.2
      - name: opendatahub-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: fast
    packageName: opendatahub-operator
    provider:
      name: ODH
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: amq-streams
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amqstreams.v2.2.2-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.2.3",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.2"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.2.3",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.2.3",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
          createdAt: "2022-07-01 10:33:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.1.0-0 <2.2.2-2'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat AMQ", "Red Hat Application Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### CRD Upgrades
          **!!! IMPORTANT !!!**  This release supports only the API version `v1beta2` and CRD version `apiextensions.k8s.io/v1`.  If upgrading from AMQ Streams 1.7, migration to `v1beta2` needs to be completed for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! If upgrading from AMQ Streams version earlier than 1.7, you need to first install the CRDs from AMQ Streams 1.7 and complete the migration to `v1beta2` for all AMQ Streams CRDs and CRs before the upgrade to 2.2 is done! For more details about the CRD upgrades, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.2/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: Red Hat Integration - AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:826354c9c06a06f998e522a88d6a81239e003e1d6a2633e9b7b27b8a61d5fe38
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:5d49bd23da40590e74422424b9457398bad89da8276718ff6a8a088e901c8345
        - registry.redhat.io/amq7/amq-streams-kafka-31-rhel8@sha256:d0607285fae95365b9d34b9474466824fb5ad012871204e8b13e52e438b66bf5
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:e3bbfff505f12c0360e175f4640c2dd9a7adcf3dce1ae2f586f2bde805ed8b1f
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:f908d848b5e67de4b162597b190ef18d449ffd4b6540457b24b491445d5c5ea3
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        version: 2.2.2-2
      entries:
      - name: amqstreams.v2.2.2-2
        version: 2.2.2-2
      - name: amqstreams.v2.2.2-1
        version: 2.2.2-1
      - name: amqstreams.v2.2.2-0
        version: 2.2.2-0
      - name: amqstreams.v2.2.1-7
        version: 2.2.1-7
      - name: amqstreams.v2.2.1-6
        version: 2.2.1-6
      - name: amqstreams.v2.2.1-5
        version: 2.2.1-5
      - name: amqstreams.v2.2.1-4
        version: 2.2.1-4
      - name: amqstreams.v2.2.1-3
        version: 2.2.1-3
      - name: amqstreams.v2.2.1-2
        version: 2.2.1-2
      - name: amqstreams.v2.2.1-1
        version: 2.2.1-1
      - name: amqstreams.v2.2.1-0
        version: 2.2.1-0
      - name: amqstreams.v2.2.0-5
        version: 2.2.0-5
      name: amq-streams-2.2.x
    - currentCSV: amqstreams.v2.3.0-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.3.1",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.3.1",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.3.1",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.2.0-0 <2.3.0-4'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.3.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 2.2 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and_upgrading_amq_streams_on_openshift/assembly-upgrade-str#assembly-upgrade-resources-str

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.3/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq7/amq-streams-kafka-33-rhel8@sha256:046c2539fef32bd928fc1f4f4c41e63f9a143b371fcc201d78c66f48fc34ea32
        - registry.redhat.io/amq7/amq-streams-operator-bundle@sha256:a59435a22c2f470b149ff721240f14144d2e56200b141e59a4f1011267a6c925
        - registry.redhat.io/amq7/amq-streams-rhel8-operator@sha256:2e46a25144d697761f0119deaf528c6f1c2ccd04e2acd95c661b0df4f32e1a9f
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq7/amq-streams-bridge-rhel8@sha256:c822b3d87f7dea233fc523238d8d0516e42f9a3d252f50f4d427c934995f9b22
        - registry.redhat.io/amq7/amq-streams-kafka-32-rhel8@sha256:0152a85f2d0bc85881dacaee62c6e8b0b8b04f57a542b9b82dc93ad5303d6bf9
        version: 2.3.0-4
      entries:
      - name: amqstreams.v2.3.0-4
        version: 2.3.0-4
      name: amq-streams-2.3.x
    - currentCSV: amqstreams.v2.4.0-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.4.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.3"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.4.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operation":"Read",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Write",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Create",
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operation":"Describe",
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "my-cluster-target",
                  "clusters": [
                    {
                      "alias": "my-cluster-source",
                      "bootstrapServers": "my-cluster-source-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "my-cluster-target",
                      "bootstrapServers": "my-cluster-target-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "my-cluster-source",
                      "targetCluster": "my-cluster-target",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": 1,
                          "offset-syncs.topic.replication.factor": 1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "heartbeatConnector": {
                        "config": {
                          "heartbeats.topic.replication.factor": 1
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": 1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
          createdAt: "2022-10-31 17:05:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          olm.skipRange: '>=2.3.0-0 <2.4.0-3'
          operators.openshift.io/infrastructure-features: '["Disconnected", "Proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.4.0 release, AMQ Streams supports only OCP 4.8 to 4.12!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more details, see the [documentation](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and_managing_amq_streams_on_openshift/assembly-upgrade-str)

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. See [the upgrade guide](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/2.4/html/deploying_and upgrading_amq_streams_on_openshift/assembly-upgrade-str) for more information.
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:cd1fefcd199400a504b4b9942c24ba6becafcecc952a9e39fa2729228be79e77
        - registry.access.redhat.com/ubi8/openjdk-11@sha256:d9739b266415be7c915daf3c61f8f1cc4e14d0b237a0ab664693bc2e0d3577e4
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:8d467fc746b78440df06f6d88d266f6ac9b5a4574a2b17fd870c173a107f8d65
        - registry.redhat.io/amq-streams/kafka-33-rhel8@sha256:aa780a250444c2426be79e95366050a33f4a4c9d68c08b58743413ea67a2d6d7
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:50b3be789993cfeedc4923a8e760c1cf52498b2178be3abfa3719fc2d15ce723
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:1346dcee82f2120bd2af22be9f041e057bdc32f9ad1c3ef717d599ba6519faeb
        version: 2.4.0-3
      entries:
      - name: amqstreams.v2.4.0-3
        version: 2.4.0-3
      - name: amqstreams.v2.4.0-2
        version: 2.4.0-2
      - name: amqstreams.v2.4.0-1
        version: 2.4.0-1
      name: amq-streams-2.4.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.5.x
    - currentCSV: amqstreams.v2.6.0-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.6.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.6"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.6.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.6.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
          createdAt: "2023-10-14 18:18:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.6.0-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: "**Red Hat AMQ Streams** is a massively scalable, distributed,
          and high performance data streaming platform based on the Apache Kafka®
          project. AMQ Streams provides an event streaming backbone that allows microservices
          and other application components to exchange data with extremely high throughput
          and low latency.\n**From the 2.6.0 release, AMQ Streams supports only OCP
          4.11 to 4.14! \n**The core capabilities include:** * A pub/sub messaging
          model, similar to a traditional enterprise messaging system, in which application
          components publish and consume events to/from an ordered stream\n* The long
          term, fault-tolerant storage of events\n* The ability for a consumer to
          replay streams of events\n* The ability to partition topics for horizontal
          scalability\n### Upgrades\n**!!! IMPORTANT !!!** **Direct upgrade from AMQ
          Stream 1.7 or earlier is not supported!** You have to upgrade first to one
          of the previous versions of AMQ Streams. You will also need to convert the
          CRD resources. For more information, see the [documentation for AMQ Streams
          on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Supported Features\n* **Manages the Kafka Cluster** - Deploys and manages
          all of the components of this complex application, including dependencies
          like Apache ZooKeeper® that are traditionally hard to administer.\n* **Includes
          Kafka Connect** - Allows for configuration of common data sources and sinks
          to move data into and out of the Kafka cluster.\n* **Topic Management**
          - Creates and manages Kafka Topics within the cluster.\n* **User Management**
          - Creates and manages Kafka Users within the cluster.\n* **Connector Management**
          - Creates and manages Kafka Connect connectors.\n* **Includes Kafka Mirror
          Maker 1 and 2** - Allows for mirroring data between different Apache Kafka®
          clusters.\n* **Includes HTTP Kafka Bridge** - Allows clients to send and
          receive messages through an Apache Kafka® cluster via HTTP protocol.\n*
          **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing
          partition replicas according to specified goals in order to achieve the
          best cluster performance.\n* **Monitoring** - Built-in support for monitoring
          using Prometheus and provided Grafana dashboards\n### Upgrading your Clusters\nThe
          AMQ Streams Operator understands how to run and upgrade between a set of
          Kafka versions. When specifying a new version in your config, check to make
          sure you aren't using any features that may have been removed. For more
          information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).\n###
          Storage\nAn efficient data storage infrastructure is essential to the optimal
          performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires
          block storage. The use of file storage (for example, NFS) is not recommended.\nThe
          AMQ Streams Operator supports three types of data storage:\n* Ephemeral
          (Recommended for development only!)\n* Persistent\n* JBOD (Just a Bunch
          of Disks, suitable for Kafka only. Not supported in Zookeeper.)\nAMQ Streams
          also supports advanced operations such as adding or removing disks in Apache
          Kafka® brokers or resizing the persistent volumes (where supported by the
          infrastructure).\n### Documentation\nDocumentation for the current release
          can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).\n"
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:44a5c4cc3086e4eac42dc78a72b7492d13df0098017288278d9006e767a480a1
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:be7466ffd7f624f769eedca2e4106c597d2bf1d95cc9679f235e6dcdec66cfa9
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:fa8361e5089ab4449261f5a0ad57bd0bd059dc254b94bd9ab68340faae386947
        version: 2.6.0-2
      entries:
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.6.x
    - currentCSV: amqstreams.v2.7.0-7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.7.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.7"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.7.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.7.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
          createdAt: "2024-03-09 21:27:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.6.0-0 <2.7.0-7'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.6.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:b9dbbce235b7a4f45f42184cf0b908e91e17094e971bc29a8710f35127a6b06f
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:671c1a2fcb0856cb208c0ead539fff9571fa2281a1e3d45b3df9ebd3df8c787e
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:be9b82af13cbc120bb415abf4fcc97420e543172d6450b836c47344e4dcacd00
        - registry.redhat.io/amq-streams/kafka-36-rhel9@sha256:8890c3217343d9aa3dd2ab9c7fa1742f960b5a26e09906581ce6075e61dc2888
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:66362c800b78003d26df9471b7f21f5c9ad7fca8e8fd2511428dfc549cdf46c2
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:a7620559f3e5a0a4d5ecee44538619f698158847d9d0a06c7168404ed6f7cd36
        version: 2.7.0-7
      entries:
      - name: amqstreams.v2.7.0-7
        version: 2.7.0-7
      - name: amqstreams.v2.7.0-6
        version: 2.7.0-6
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.7.x
    - currentCSV: amqstreams.v2.8.0-0-0.1738265624.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.8.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.8"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.8.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.8.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.7.0-0 <2.8.0-0'
          olm.substitutesFor: amqstreams.v2.8.0-0
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.8.0 release, Streams supports only OCP 4.12 to 4.14!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:b46457a0a7f63518bc12b2789f75afb0e44496e62de918b9b584129c7078de86
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:026418bae8b2d425d45b2b3481f65c241d89e25f455e3fbfcf2f79536d978bca
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:4df906cf1490922a4587d96250f3be6ff605fe2caf87dfba7b812c144c5c5774
        - registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:17b2d255439f38c6ef54c0a56929ae636743b12c688d42223ac3fb58729e51ad
        version: 2.8.0-0+0.1738265624.p
      entries:
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.8.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.9.x
    - currentCSV: amqstreams.v2.9.2-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.9.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.9"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "pattern":"*.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.9.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.9.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
          createdAt: "2024-08-22 15:19:00"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.5.0-0 <2.9.2-1'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.9.0 release, Streams supports only OCP 4.14 to 4.18!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream * The long term, fault-tolerant storage of events * The ability for a consumer to replay streams of events * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!**
          **Direct upgrade from Streams 1.7 or earlier is not supported anymore!** You have to upgrade first to one of the previous versions of Streams. You will also need to convert the CRD resources. For more information, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Supports KRaft** - You can run your Apache Kafka clusters without Apache ZooKeeper.
          * **Tiered storage** - Offloads older, less critical data to a lower-cost, lower-performance storage tier, such as object storage.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Auto-rebalancing when scaling** - Automatically rebalance the Kafka cluster after a scale-up or before a scale-down.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for Streams for Apache Kafka on OpenShift](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:86ea2e823cd83c476521f2a1e9da5619c4810661c60d28295d0f5482c8aa519d
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:67647bff1a76adb5b51f43c99d5f88a8aa38055c1e87190add8972bb621ddc15
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:fb1bb746eabaf71a0b132903394d5a15c43a78c824ba596e8e67783a677fc447
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:96f6eb4ee13ccf8f6236e372955f97fc26a265b6bfe0ec14a36b031f2d665063
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:f5985592b7f05c9b3c64caa0ac90f34f26b9b4277b342b6bd4a6415d2c563f9a
        - registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:ddfb50ec2f1cf45ccccc632182bf0ff549c14969da35b7248802eab67d9c2e9d
        version: 2.9.2-1
      entries:
      - name: amqstreams.v2.9.2-1
        version: 2.9.2-1
      - name: amqstreams.v2.9.2-0
        version: 2.9.2-0
      - name: amqstreams.v2.9.1-2
        version: 2.9.1-2
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-2.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.0.x
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-3.x
    - currentCSV: amqstreams.v2.5.2-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"Kafka",
                "metadata":{
                  "name":"my-cluster"
                },
                "spec":{
                  "kafka":{
                      "version":"3.5.0",
                      "replicas":3,
                      "listeners": [
                        {
                          "name": "plain",
                          "port": 9092,
                          "type": "internal",
                          "tls": false
                        },
                        {
                          "name": "tls",
                          "port": 9093,
                          "type": "internal",
                          "tls": true
                        }
                      ],
                      "config":{
                        "offsets.topic.replication.factor":3,
                        "transaction.state.log.replication.factor":3,
                        "transaction.state.log.min.isr":2,
                        "default.replication.factor":3,
                        "min.insync.replicas":2,
                        "inter.broker.protocol.version":"3.5"
                      },
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "zookeeper":{
                      "replicas":3,
                      "storage":{
                        "type":"ephemeral"
                      }
                  },
                  "entityOperator":{
                      "topicOperator":{
                        },
                      "userOperator":{
                        }
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaConnect",
                "metadata":{
                  "name":"my-connect-cluster"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9093",
                  "tls":{
                      "trustedCertificates":[
                        {
                            "secretName":"my-cluster-cluster-ca-cert",
                            "certificate":"ca.crt"
                        }
                      ]
                  },
                  "config": {
                      "group.id": "connect-cluster",
                      "offset.storage.topic": "connect-cluster-offsets",
                      "config.storage.topic": "connect-cluster-configs",
                      "status.storage.topic": "connect-cluster-status",
                      "config.storage.replication.factor": -1,
                      "offset.storage.replication.factor": -1,
                      "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaMirrorMaker",
                "metadata":{
                  "name":"my-mirror-maker"
                },
                "spec":{
                  "version":"3.5.0",
                  "replicas":1,
                  "consumer":{
                      "bootstrapServers":"my-source-cluster-kafka-bootstrap:9092",
                      "groupId":"my-source-group-id"
                  },
                  "producer":{
                      "bootstrapServers":"my-target-cluster-kafka-bootstrap:9092"
                  },
                  "include":".*"
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaBridge",
                "metadata":{
                  "name":"my-bridge"
                },
                "spec":{
                  "replicas":1,
                  "bootstrapServers":"my-cluster-kafka-bootstrap:9092",
                  "http":{
                      "port":8080
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaTopic",
                "metadata":{
                  "name":"my-topic",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "partitions":10,
                  "replicas":3,
                  "config":{
                      "retention.ms":604800000,
                      "segment.bytes":1073741824
                  }
                }
              },
              {
                "apiVersion":"kafka.strimzi.io/v1beta2",
                "kind":"KafkaUser",
                "metadata":{
                  "name":"my-user",
                  "labels":{
                      "strimzi.io/cluster":"my-cluster"
                  }
                },
                "spec":{
                  "authentication":{
                      "type":"tls"
                  },
                  "authorization":{
                      "type":"simple",
                      "acls":[
                        {
                            "resource":{
                              "type":"topic",
                              "name":"my-topic",
                              "patternType":"literal"
                            },
                            "operations":["Read", "Describe", "Write", "Create"],
                            "host":"*"
                        },
                        {
                            "resource":{
                              "type":"group",
                              "name":"my-group",
                              "patternType":"literal"
                            },
                            "operations":["Read"],
                            "host":"*"
                        }
                      ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version":"3.4.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "strimzi-pod-set-example"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "strimzi-pod-set-example"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "strimzi-pod-set-example"
                    }
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mirror-maker": {
                "description": "Example Kafka Mirror Maker 1 deployment"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example AMQ Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
          createdAt: "2023-07-10 21:47:00"
          description: Red Hat AMQ Streams is a massively scalable, distributed, and
            high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.2.0-0 <2.5.2-8'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Red Hat AMQ Streams** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project. AMQ Streams provides an event streaming backbone that allows microservices and other application components to exchange data with extremely high throughput and low latency.
          **From the 2.5.2 release, AMQ Streams supports only OCP 4.12 to 4.16!
          **The core capabilities include:** * A pub/sub messaging model, similar to a traditional enterprise messaging system, in which application components publish and consume events to/from an ordered stream
          * The long term, fault-tolerant storage of events
          * The ability for a consumer to replay streams of events
          * The ability to partition topics for horizontal scalability
          ### Upgrades
          **!!! IMPORTANT !!!** **Direct upgrade from AMQ Stream 1.7 or earlier is not supported!** You have to upgrade first to one of the previous versions of AMQ Streams. You will also need to convert the CRD resources. For more information, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).

          ### Supported Features
          * **Manages the Kafka Cluster** - Deploys and manages all of the components of this complex application, including dependencies like Apache ZooKeeper® that are traditionally hard to administer.
          * **Includes Kafka Connect** - Allows for configuration of common data sources and sinks to move data into and out of the Kafka cluster.
          * **Topic Management** - Creates and manages Kafka Topics within the cluster.
          * **User Management** - Creates and manages Kafka Users within the cluster.
          * **Connector Management** - Creates and manages Kafka Connect connectors.
          * **Includes Kafka Mirror Maker 1 and 2** - Allows for mirroring data between different Apache Kafka® clusters.
          * **Includes HTTP Kafka Bridge** - Allows clients to send and receive messages through an Apache Kafka® cluster via HTTP protocol.
          * **Cluster Rebalancing** - Uses built-in Cruise Control for redistributing partition replicas according to specified goals in order to achieve the best cluster performance.
          * **Monitoring** - Built-in support for monitoring using Prometheus and provided Grafana dashboards
          ### Upgrading your Clusters
          The AMQ Streams Operator understands how to run and upgrade between a set of Kafka versions. When specifying a new version in your config, check to make sure you aren't using any features that may have been removed. For more information on upgrading, see the [documentation for AMQ Streams on OpenShift](https://access.redhat.com/documentation/en-us/red_hat_amq_streams/).
          ### Storage
          An efficient data storage infrastructure is essential to the optimal performance of Apache Kafka®. Apache Kafka® deployed via AMQ Streams requires block storage. The use of file storage (for example, NFS) is not recommended.
          The AMQ Streams Operator supports three types of data storage:
          * Ephemeral (Recommended for development only!)
          * Persistent
          * JBOD (Just a Bunch of Disks, suitable for Kafka only. Not supported in Zookeeper.)
          AMQ Streams also supports advanced operations such as adding or removing disks in Apache Kafka® brokers or resizing the persistent volumes (where supported by the infrastructure).
          ### Documentation
          Documentation for the current release can be found on the Red Hat [Customer Portal](https://access.redhat.com/documentation/en-us/red_hat_amq_streams).
        displayName: AMQ Streams
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_amq_streams
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/kafka-34-rhel8@sha256:95bc35951eb545b918114af1e92a7139e2e3323b7d054dcb2a060843c7326d28
        - registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:8c7e8ccb6c2d5da32b70d9d0c3919c303ad73a25447f6993d964f7b19cca1072
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:3d5bcf889f3d534402a8bbfa51fc4a9499b7d6dfaac0d7b11e91363b379be192
        - registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:67ba7b0280fbddd4e7281d07d5844dfd37d69aa81d71dc42fddbf91b59cd479a
        - registry.redhat.io/ubi8/openjdk-17@sha256:8796da64c626c5f63f31a4e3ac0ac3b179333a7c6d5472a195b9b3f61ccd1099
        - registry.redhat.io/amq-streams/bridge-rhel8@sha256:4ede5eeed159bddacfe830a768bab4996003d4c57a496fc8ec9b41df916d4acc
        version: 2.5.2-8
      entries:
      - name: amqstreams.v2.5.2-8
        version: 2.5.2-8
      - name: amqstreams.v2.5.2-7
        version: 2.5.2-7
      - name: amqstreams.v2.5.2-6
        version: 2.5.2-6
      - name: amqstreams.v2.5.2-5
        version: 2.5.2-5
      - name: amqstreams.v2.5.2-4
        version: 2.5.2-4
      - name: amqstreams.v2.5.2-3
        version: 2.5.2-3
      - name: amqstreams.v2.5.2-2
        version: 2.5.2-2
      - name: amqstreams.v2.5.2-1
        version: 2.5.2-1
      - name: amqstreams.v2.5.2-0
        version: 2.5.2-0
      - name: amqstreams.v2.5.1-5
        version: 2.5.1-5
      - name: amqstreams.v2.5.1-4
        version: 2.5.1-4
      - name: amqstreams.v2.5.1-3
        version: 2.5.1-3
      - name: amqstreams.v2.5.1-2
        version: 2.5.1-2
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: amq-streams-lts
    - currentCSV: amqstreams.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster",
                  "annotations": {
                    "strimzi.io/node-pools": "enabled",
                    "strimzi.io/kraft": "enabled"
                  }
                },
                "spec": {
                  "kafka": {
                    "version": "4.0.0",
                    "metadataVersion": "4.0-IV3",
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaMirrorMaker2",
                "metadata": {
                  "name": "my-mm2-cluster"
                },
                "spec": {
                  "version": "4.0.0",
                  "replicas": 1,
                  "connectCluster": "cluster-b",
                  "clusters": [
                    {
                      "alias": "cluster-a",
                      "bootstrapServers": "cluster-a-kafka-bootstrap:9092"
                    },
                    {
                      "alias": "cluster-b",
                      "bootstrapServers": "cluster-b-kafka-bootstrap:9092",
                      "config": {
                        "config.storage.replication.factor": -1,
                        "offset.storage.replication.factor": -1,
                        "status.storage.replication.factor": -1
                      }
                    }
                  ],
                  "mirrors": [
                    {
                      "sourceCluster": "cluster-a",
                      "targetCluster": "cluster-b",
                      "sourceConnector": {
                        "config": {
                          "replication.factor": -1,
                          "offset-syncs.topic.replication.factor": -1,
                          "sync.topic.acls.enabled": "false"
                        }
                      },
                      "checkpointConnector": {
                        "config": {
                          "checkpoints.topic.replication.factor": -1
                        }
                      },
                      "topicsPattern": ".*",
                      "groupsPattern": ".*"
                    }
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaBridge",
                "metadata": {
                  "name": "my-bridge"
                },
                "spec": {
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9092",
                  "http": {
                    "port": 8080
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "core.strimzi.io/v1beta2",
                "kind": "StrimziPodSet",
                "metadata": {
                  "name": "my-podset"
                },
                "spec": {
                  "pods": [
                    {
                      "apiVersion": "v1",
                      "kind": "Pod",
                      "metadata": {
                        "labels": {
                          "app": "my-podset"
                        },
                        "name": "example-pod"
                      },
                      "spec": {
                        "containers": [
                          {
                            "command": [
                              "sleep",
                              3600
                            ],
                            "image": "busybox:latest",
                            "imagePullPolicy": "IfNotPresent",
                            "name": "busybox"
                          }
                        ]
                      }
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app": "my-podset"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster (to be used together with KafkaNodePool)"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-mm2-cluster": {
                "description": "Example Kafka Mirror Maker 2 deployment"
              },
              "my-bridge": {
                "description": "Example Streams Kafka HTTP Bridge deployment"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              },
              "my-podset": {
                "description": "Example StrimziPodSet custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
          createdAt: "2025-09-24T12:21:07Z"
          description: Streams for Apache Kafka is a massively scalable, distributed,
            and high performance data streaming platform based on Apache Kafka®.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          repository: https://github.com/strimzi/strimzi-kafka-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |
          **Streams for Apache Kafka** is a massively scalable, distributed, and high performance data streaming platform based on the Apache Kafka® project.

          It provides an event streaming backbone that enables microservices and other application components to exchange data with extremely high throughput and low latency.

          > NOTE: For supported OpenShift versions, see the latest [Streams for Apache Kafka on OpenShift Release Notes](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/).

          ***

          ### Core Capabilities

          * **Pub/sub messaging** – Delivers high-throughput, low-latency event streams.
          * **Durable storage** – Replicates data for fault tolerance and long-term retention.
          * **Replay support** – Replays past messages from persistent storage.
          * **Horizontal scaling** – Distributes processing using partitioned topics.

          ***

          ### Key Features

          * **Manage Kafka clusters** – Deploy, scale, and upgrade Kafka and its components.
          * **Manage topics, users, and connectors** – Create and configure Kafka resources.
          * **Integrate with Kafka Connect** – Move data between Kafka and external systems.
          * **Mirror with MirrorMaker 2** – Replicate data across Kafka clusters.
          * **Bridge with HTTP Kafka Bridge** – Send and receive Kafka messages over HTTP.
          * **Offload data with tiered storage** – Store older data in low-cost object storage.
          * **Optimize with Cruise Control** – Rebalance clusters based on performance goals.
          * **Auto-rebalance on scale** – Trigger rebalancing during scale-up or scale-down.
          * **Monitor with built-in tools** – Export metrics to Prometheus and visualize with Grafana.

          ***

          ### Upgrades

          Kafka is KRaft-only starting with version 4.0.
          As a result, **Streams for Apache Kafka 2.9** was the last version that supported Kafka clusters using ZooKeeper.

          > **Important:** To upgrade from versions earlier than **2.7**, first upgrade to a **2.7–2.9** release and complete the migration to KRaft. For migration steps, refer to the [Streams for Apache Kafka on OpenShift documentation](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/) for version **2.9**.

          **Upgrade Options**

          * **Incremental upgrade** – Move between consecutive minor versions.
          * **Multi-version upgrade** – Skip one or more minor versions between [Long-Term Support (LTS) releases](https://access.redhat.com/articles/6975608) by switching to a version-specific channel and manually upgrading Kafka using the `Kafka` resource.

          > **Note:** Multi-version upgrades are **not supported in the deprecated LTS channel (`amq-streams-lts`)**. No new LTS versions are being added. Use version-specific channels for future LTS-to-LTS upgrades.

          **Recommendation:** Always test upgrades in a controlled environment before applying them in production.

          ***

          ### Storage

          Kafka requires **block storage** for performance.
          File storage, such as NFS, is **not recommended**.

          Supported storage types:

          * **Ephemeral** (for development)
          * **Persistent**
          * **JBOD**

          Streams for Apache Kafka supports advanced storage operations:

          * Adding/removing broker disks
          * Resizing persistent volumes (if supported by infrastructure)

          ***

          ### Documentation

          Documentation for the current release can be found on the [Red Hat Documentation site](https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka).
        displayName: Streams for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/amq-streams/strimzi-operator-bundle@sha256:a423992ca43b6264635bc9460e8037dcd36dbb4bafd614d95a3245cca2e54265
        - registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:42733933be49e2f08ffd06c4c8f31dbfc6eac0327779a54da79ec78f487d19b1
        - registry.redhat.io/amq-streams/bridge-rhel9@sha256:a42a58b07e6c0c6bfd3c908095232d5c82e789a365372a512e5c1cb5ebe23d09
        - registry.redhat.io/amq-streams/kafka-39-rhel9@sha256:6014828970f83e9546f2e670a9a97d916dd063ff4f9b13910f91eee0068f96fa
        - registry.redhat.io/amq-streams/kafka-40-rhel9@sha256:38354c92b8883144a2c2cc37699046323338fd42c8d852eed0c4e711ed7e5cbe
        - registry.redhat.io/amq-streams/maven-builder-rhel9@sha256:efa03b0d29d2b5ac6cab8f5f7e455b7342e30401f83b45f2563e3a1b928d2d2a
        version: 3.0.1-2
      entries:
      - name: amqstreams.v3.0.1-2
        version: 3.0.1-2
      - name: amqstreams.v3.0.1-1
        version: 3.0.1-1
      - name: amqstreams.v3.0.0-13
        version: 3.0.0-13
      - name: amq-streams.v3.0.0-9
        version: 3.0.0-9
      - name: amqstreams.v2.9.1-1
        version: 2.9.1-1
      - name: amqstreams.v2.9.1-0
        version: 2.9.1-0
      - name: amqstreams.v2.8.0-0-0.1738265624.p
        version: 2.8.0-0+0.1738265624.p
      - name: amqstreams.v2.8.0-0
        version: 2.8.0-0
      - name: amqstreams.v2.7.0-5
        version: 2.7.0-5
      - name: amqstreams.v2.7.0-4
        version: 2.7.0-4
      - name: amqstreams.v2.7.0-3
        version: 2.7.0-3
      - name: amqstreams.v2.7.0-2
        version: 2.7.0-2
      - name: amqstreams.v2.7.0-1
        version: 2.7.0-1
      - name: amqstreams.v2.7.0-0
        version: 2.7.0-0
      - name: amqstreams.v2.6.0-2
        version: 2.6.0-2
      - name: amqstreams.v2.6.0-1
        version: 2.6.0-1
      - name: amqstreams.v2.6.0-0
        version: 2.6.0-0
      - name: amqstreams.v2.5.1-1
        version: 2.5.1-1
      - name: amqstreams.v2.5.1-0
        version: 2.5.1-0
      - name: amqstreams.v2.5.0-1
        version: 2.5.0-1
      - name: amqstreams.v2.5.0-0
        version: 2.5.0-0
      name: stable
    defaultChannel: stable
    packageName: amq-streams
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC Corporation
      provider-url: https://jpn.nec.com/webotx/
    name: webotx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: webotx-operator.v11.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "webotx.nec.com/v3",
                "kind": "ApplicationServer",
                "metadata": {
                  "name": "webotx-express"
                },
                "spec": {
                  "deployment": {
                    "image": "registry.connect.redhat.com/nec/webotx-express:11.11.00.00@sha256:73a5dc48b4e3340951dbe207d744f5ac68474701dd134fa7aa05e65e3510c8ad",
                    "replicas": 1
                  },
                  "service": {
                    "externalTrafficPolicy": "Cluster",
                    "ports": [
                      {
                        "name": "http",
                        "nodePort": 30080,
                        "port": 8080,
                        "protocol": "TCP",
                        "targetPort": 8080
                      },
                      {
                        "name": "https",
                        "nodePort": 30443,
                        "port": 8443,
                        "protocol": "TCP",
                        "targetPort": 8443
                      }
                    ],
                    "sessionAffinity": "ClientIP",
                    "sessionAffinityConfig": {
                      "clientIP": {
                        "timeoutSeconds": 10800
                      }
                    },
                    "type": "NodePort"
                  },
                  "webotx": {
                    "metering": {
                      "enabled": false,
                      "port": 8443,
                      "scheme": "https"
                    },
                    "sessionReplication": {
                      "enabled": false,
                      "useAllProcessGroups": false,
                      "useJmxAgent": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApplicationServer is the Schema for the applicationservers
              API
            displayName: Application Server
            kind: ApplicationServer
            name: applicationservers.webotx.nec.com
            version: v3
        description: WebOTX Operator is the Operator for WebOTX Application Server.
          WebOTX Operator enables to deploy and administrate WebOTX Application Server
          automatically.
        displayName: WebOTX Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - webotx
        - WebOTX
        - nec
        - NEC
        links:
        - name: WebOTX Operator for Kubernetes
          url: https://github.com/webotx/operator
        maintainers:
        - email: info-webotx@isd.jp.nec.com
          name: NEC Corporation
        maturity: stable
        provider:
          name: NEC Corporation
          url: https://jpn.nec.com/webotx/
        relatedImages:
        - registry.connect.redhat.com/nec/webotx-operator:11.11.00.00@sha256:df0d0315ed88be875ca7a6474500f43f17855f763c5c7d0f4d3ec58a2e46f6fb
        - registry.connect.redhat.com/nec/webotx-operator-bundle@sha256:7e655933a7fab3ac3aa88eb5e4ca9ac3a71b98875ef1ef5bc49e87284f160ff8
        version: 11.11.0
      entries:
      - name: webotx-operator.v11.11.0
        version: 11.11.0
      name: stable
    defaultChannel: stable
    packageName: webotx-operator
    provider:
      name: NEC Corporation
      url: https://jpn.nec.com/webotx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-kcos-licensing
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-kcos-licensing.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kcos.loadcore.keysight.com/v1alpha1",
                "kind": "Licensing",
                "metadata": {
                  "name": "licensing-sample"
                },
                "spec": {
                  "environment": {
                    "deploymentType": "openshift"
                  },
                  "hostid": {
                    "enabled": true,
                    "environment": {},
                    "hostMounts": {},
                    "image": "prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "image": {
                    "dbimage": "prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b",
                    "initOwnership": "prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf",
                    "repository": "prod.harbor.keysight.digital/loadcore"
                  },
                  "ingress": {
                    "annotations": {
                      "nginx.ingress.kubernetes.io/auth-method": "POST",
                      "nginx.ingress.kubernetes.io/auth-response-headers": "X-AUTH-USER",
                      "nginx.ingress.kubernetes.io/auth-url": "https://127.0.0.1/token/validate"
                    },
                    "enabled": true
                  },
                  "nodeSelector": {
                    "node-role.kubernetes.io/kcos-licensing": "kcos-licensing"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:29:04Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Licensing
            name: licensings.kcos.loadcore.keysight.com
            version: v1alpha1
        description: Keysight KCOS Licensing Operator
        displayName: Keysight KCOS Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Kcos Licensing
          url: https://keysight-kcos-licensing.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/postgresql-ubi-image@sha256:17f6a8f775d06650939d2c73e82b45d73d50190cc735ca70695d082f7790834b
        - prod.harbor.keysight.digital/loadcore/wap-rhel8-toolbox@sha256:58fa302bba072951a559d749801cd39266af06207c79c92994d4f08ebf6eefbf
        - registry.connect.redhat.com/loadcore/keysight-kcos-licensing-bundle@sha256:2d3abd0263183bc92718316581675b92073f37c3786784fae24b3587a9441def
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/kcos-licensing-hostid-vmuuid@sha256:076a2c6aeec875470414af7905613a25b56bcb2aca5edcb1c4c53037cd01b332
        - prod.harbor.keysight.digital/loadcore/keysight-kcos-licensing@sha256:a09291923ad8d7533b33e40e1a3c4e735c2a4ae3b2b87be7c9d1a1a9244b3dad
        version: 0.1.21
      entries:
      - name: keysight-kcos-licensing.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-kcos-licensing
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent
      provider-url: ""
    name: cilium
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cilium.v1.10.15-x084926b
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm:970b1d74dbe2b87c28e8dc1c68936de33123a4df-v1.10.15@sha256:afa3ac140fae24b151ab6ff5a8cdef11566a58f7234d702a6a45a398822e6d34
        - quay.io/cilium/hubble-relay@sha256:67a9183fd1febd9e51f4d02bd35092b81c8612a1716015988f69fffba79d999a
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - quay.io/cilium/cilium@sha256:adbb60f9d19d641a5496fcfc4769ebb55c1d4cd7699fef752f7c94754194cf83
        - quay.io/cilium/operator-generic@sha256:7f2f7b59b541e4f660d65ed0a21d25ab69e49022172529771dd1134315e95c82
        - quay.io/cilium/hubble-ui-backend@sha256:a07e13df3e9d7ec0dcc3b158f243cda5a08f6eca38a25a5aa3171f6b3ec491ed
        - quay.io/cilium/hubble-ui@sha256:dec6e7e541cac91ab859245dd2a7fa02de5363b75caa3509a79cf76d4f622550
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        - quay.io/cilium/clustermesh-apiserver@sha256:e73151c3295589de00dccae75819aabdff8fd907260817020f9e8fe3738e78b2
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:329d6eaeff787bcde5db0ba4f9403eee533a601b45dc2d5d349b318705452591
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        version: 1.10.15+x084926b
      entries:
      - name: cilium.v1.10.15-x084926b
        version: 1.10.15+x084926b
      - name: cilium.v1.10.13-xa8bb1e5
        version: 1.10.13+xa8bb1e5
      - name: cilium.v1.10.12-x362ea11
        version: 1.10.12+x362ea11
      - name: cilium.v1.10.10-x6adc6ff
        version: 1.10.10+x6adc6ff
      name: "1.10"
    - currentCSV: cilium.v1.11.20-xdb5ed32
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/cilium@sha256:60df3cb7155886e0b62060c7a4a31e457933c6e35af79febad5fd6e43bab2a99
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3677840d57e1d74e3ac14db0dfceb83088e6ebd70775d775855cdd3bfed68059
        - quay.io/cilium/hubble-relay@sha256:e2f38b901fd8bd5adc9a765a5e68836364ebd1e7dfb85c2bcd8a5488b23c3470
        - quay.io/cilium/clustermesh-apiserver@sha256:46760182f8c98227cfac27627275616987b71509227775350573d834133a6d49
        - quay.io/cilium/hubble-ui@sha256:bbf8b041152041d3ef000d6074a8dd5d40ed910b5c0067f365b00e52e449d467
        - quay.io/cilium/operator-generic@sha256:1439954acf620f048ef663524ae70b4a25693c58527a2f2cee51124496e29f90
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/hubble-ui-backend@sha256:46070270b8ffe7ac3f1794721be0b5a09f906dff3492f8a199a5f995db362d2a
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:2cac5fdb55664c08928a7d924eacb18cc4c03c609261e3b6d812590e4d166375
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        version: 1.11.20+xdb5ed32
      entries:
      - name: cilium.v1.11.20-xdb5ed32
        version: 1.11.20+xdb5ed32
      name: "1.11"
    - currentCSV: cilium.v1.12.19-xef55de8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.12.0 <1.12.19+xef55de8'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:b35da8aa4e6746952e5cfdbbac72950a48f7ad32a3d762a470ce62118c78643a
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/operator-generic@sha256:da4d72facecde862c88d08bf99c3f476845beb407052a0cc4c06b68cc93b7570
        - quay.io/cilium/clustermesh-apiserver@sha256:61f8e871760dceaa0c0459bbf44a3049dd0f4ab15e57e063d9b3284434cb3dd8
        - quay.io/cilium/hubble-ui-backend@sha256:f88a73a120521eeafbcdbd908f517117f6557bf61e115847853fac371f0d774c
        - quay.io/cilium/cilium@sha256:b997994ab759443d9fc9807eb39b4784bf03e13730a8132ffe5df3edef22a312
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:d35994cf5cf319b88ff5a7985cc972edd23df5a282360f9ba5516d610390ebf8
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/hubble-relay@sha256:19f2dee371523888985e6e34b30d07ef3f84f2a9206ad79bdfc8dede96e25d5c
        - quay.io/cilium/hubble-ui@sha256:e15af59a2ded739e420be82de6cbdd0ce22f8d3f00e3a10b3d2d2734e074a394
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        version: 1.12.19+xef55de8
      entries:
      - name: cilium.v1.12.19-xef55de8
        version: 1.12.19+xef55de8
      - name: cilium.v1.12.18-x63f4f4c
        version: 1.12.18+x63f4f4c
      - name: cilium.v1.12.17-x6cc0c18
        version: 1.12.17+x6cc0c18
      - name: cilium.v1.12.16-x14b1d1d
        version: 1.12.16+x14b1d1d
      - name: cilium.v1.12.15-x7d84617
        version: 1.12.15+x7d84617
      - name: cilium.v1.12.14-x17ed60a
        version: 1.12.14+x17ed60a
      - name: cilium.v1.12.13-xb414403
        version: 1.12.13+xb414403
      - name: cilium.v1.12.12-xd280cc3
        version: 1.12.12+xd280cc3
      - name: cilium.v1.12.11-xc8fb7eb
        version: 1.12.11+xc8fb7eb
      - name: cilium.v1.12.10-xe372776
        version: 1.12.10+xe372776
      - name: cilium.v1.12.9-x2ccb2b6
        version: 1.12.9+x2ccb2b6
      - name: cilium.v1.12.8-xa53bac1
        version: 1.12.8+xa53bac1
      - name: cilium.v1.12.7-xba75145
        version: 1.12.7+xba75145
      - name: cilium.v1.12.0-xb42b669
        version: 1.12.0+xb42b669
      name: "1.12"
    - currentCSV: cilium.v1.13.12-x6e56d47
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.13.0 <1.13.12+x6e56d47'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium@sha256:d99204aa7b3b7bd2c9ab47fd398cc9f40290799bc0c7a4386c8dc5c1780cd3d3
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/clustermesh-apiserver@sha256:f1b1d0a85bab65e7d6adc90d000513a56ac58bdb071aa391a8580d73d20e6b6a
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:f83734bbe270f961d545c7929152785507ce04a05d818ebc9776941723736d02
        - quay.io/cilium/hubble-relay@sha256:01b23ea40bcd81145dde6bfcbfc4d542749d08c2a1c6348954c85123a8d2b1fc
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:e5a0cb8af7879bf13d940df278ccdb8a5c400e5d7108f23b4abef94c964e40d6
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b0d22d4376b4c2b2b3b988955227f1ed2ad49a363c5367e53f64edae2cec0ff5
        version: 1.13.12+x6e56d47
      entries:
      - name: cilium.v1.13.12-x6e56d47
        version: 1.13.12+x6e56d47
      - name: cilium.v1.13.11-xba1ab0b
        version: 1.13.11+xba1ab0b
      - name: cilium.v1.13.10-x412ae64
        version: 1.13.10+x412ae64
      - name: cilium.v1.13.9-x3ae7099
        version: 1.13.9+x3ae7099
      - name: cilium.v1.13.8-xe0355c7
        version: 1.13.8+xe0355c7
      - name: cilium.v1.13.7-x37f9269
        version: 1.13.7+x37f9269
      - name: cilium.v1.13.6-xdb5afcd
        version: 1.13.6+xdb5afcd
      - name: cilium.v1.13.5-xaf5caa6
        version: 1.13.5+xaf5caa6
      - name: cilium.v1.13.4-x129f2a7
        version: 1.13.4+x129f2a7
      - name: cilium.v1.13.3-x5854d64
        version: 1.13.3+x5854d64
      - name: cilium.v1.13.2-x82c79c4
        version: 1.13.2+x82c79c4
      - name: cilium.v1.13.1-x10a5de9
        version: 1.13.1+x10a5de9
      - name: cilium.v1.13.0-x32540df
        version: 1.13.0+x32540df
      name: "1.13"
    - currentCSV: cilium.v1.14.7-x3522df7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.14.0 <1.14.7+x3522df7'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:291a05c58de8d7daefeb0689c23cadd92c9368aea09751de70eaa817f7fcc358
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/coreos/etcd@sha256:a67fb152d4c53223e96e818420c37f11d05c2d92cf62c05ca5604066c37295e9
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/operator-generic@sha256:37ef0bd85c27c765c637cd58c3ff4a559f8734ae39f9d1839a3ac7803de7b952
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:91aae0500cd3cfc9a4a536241feb4fe50288e2a68f69ab5c602c02169e5f5b2e
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/clustermesh-apiserver@sha256:28f3ffe53365ca79831af600f09a95c0b3e9959f5f891b416dab8cedd90c263d
        - quay.io/cilium/cilium@sha256:45ce2b87696082ecf7d53ba1c64ceeb4217578033e5ef28ac479ec049a48bc32
        - quay.io/cilium/hubble-relay@sha256:46762393daf4a0aaef76b106614c2615942f98f10aeacd435ea3fb1a0bdf69e4
        version: 1.14.7+x3522df7
      entries:
      - name: cilium.v1.14.7-x3522df7
        version: 1.14.7+x3522df7
      - name: cilium.v1.14.6-x11464e1
        version: 1.14.6+x11464e1
      - name: cilium.v1.14.5-x3583390
        version: 1.14.5+x3583390
      - name: cilium.v1.14.4-xb89471d
        version: 1.14.4+xb89471d
      - name: cilium.v1.14.3-x041dc10
        version: 1.14.3+x041dc10
      - name: cilium.v1.14.2-xedd5ce2
        version: 1.14.2+xedd5ce2
      - name: cilium.v1.14.1-xdd859b7
        version: 1.14.1+xdd859b7
      name: "1.14"
    - currentCSV: cilium.v1.15.1-x7095b76
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Networking,Security
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.15.0 <1.15.1+x7095b76'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:40c7d896355dfe67ff3616f1e1df86548bfcd0524f7e56d4d72ab4f5f1739f6a
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:9ab6be29447125e886300e9258b9a06bedf0a9d87405832aa8b6565ed1ba4215
        - quay.io/cilium/certgen@sha256:f09fccb919d157fc0a83de20011738192a606250c0ee3238e3610b6cb06c0981
        - quay.io/cilium/cilium@sha256:351d6685dc6f6ffbcd5451043167cfa8842c6decf80d8c8e426a417c73fb56d4
        - quay.io/cilium/hubble-relay@sha256:3254aaf85064bc1567e8ce01ad634b6dd269e91858c83be99e47e685d4bb8012
        - quay.io/cilium/hubble-ui-backend@sha256:6a396a3674b7d90ff8c408a2e13bc70b7871431bddd63da57afcdeea1d77d27c
        - quay.io/cilium/clustermesh-apiserver@sha256:b353badd255c2ce47eaa8f394ee4cbf70666773d7294bd887693e0c33503dc37
        - quay.io/cilium/hubble-ui@sha256:cc0d4f6f610409707566087895062ac40960d667dd79e4f33a4f0f393758fc1e
        - quay.io/cilium/startup-script@sha256:a1454ca1f93b69ecd2c43482c8e13dc418ae15e28a46009f5934300a20afbdba
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/operator-generic@sha256:819c7281f5a4f25ee1ce2ec4c76b6fbc69a660c68b7825e9580b1813833fa743
        version: 1.15.1+x7095b76
      entries:
      - name: cilium.v1.15.1-x7095b76
        version: 1.15.1+x7095b76
      name: "1.15"
    - currentCSV: cilium.v1.9.14-x70dc065
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:3a9ca9a4c2f78109f265c4758f5f693d01865f2d0e31bc5f1ecbc065e878628b
        - quay.io/cilium/hubble-relay@sha256:fd6ab1aea260abc5f64eca26c1b1e7009983e4aaa8e5d098e8d442f7659603fb
        - quay.io/cilium/hubble-ui-backend@sha256:2cf1c7b1fde6161f575f3147c2ef885739e60b8c59a1ed444fe74edbdf22c7c1
        - quay.io/cilium/operator-generic@sha256:bdcfd2eade99933f2fda55ef79ea697ddfad3512f65b15bcd0ba7702518c1ba3
        - quay.io/cilium/certgen@sha256:0c2b71bb3469990e7990e7e26243617aa344b5a69a4ce465740b8577f9d48ab9
        - quay.io/cilium/cilium-etcd-operator@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc
        - quay.io/cilium/cilium@sha256:2c6ce93fa7e625979043a387eb998c17ad57df8768d89facb9b715da42a4c51c
        - quay.io/cilium/clustermesh-apiserver@sha256:a0da5edf0372899647da51de1b277f0bab8e676d694aee7f939cddfdd3172010
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:3f69a124f9c31ebc0c78f249872a69b1289622bf4f5aad4c739f6e8caf9cb85b
        - docker.io/envoyproxy/envoy@sha256:55a141c0b8e39497b32074cb758b4e5dc2384bd6d9926144969a4452eaa611fd
        - quay.io/cilium/hubble-ui@sha256:c267219946b84f79a5fd2849f5cc71f39453e29d93039e57b9cef2b5802e9c50
        - quay.io/cilium/startup-script@sha256:1daf817f34000399fcb5da9a94cb299e2810d2c7a52e51de22ba0d4783b6ce84
        version: 1.9.14+x70dc065
      entries:
      - name: cilium.v1.9.14-x70dc065
        version: 1.9.14+x70dc065
      name: "1.9"
    - currentCSV: cilium.v1.10.4-x5bfd7b3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"cilium.io/v1alpha1","kind":"CiliumConfig","metadata":{"name":"cilium-openshift-default","namespace":"placeholder"},"spec":{"nativeRoutingCIDR":"10.128.0.0/14","endpointRoutes":{"enabled":true},"kubeProxyReplacement":"probe","cni":{"binPath":"/var/lib/cni/bin","confPath":"/var/run/multus/cni/net.d"},"ipam":{"operator":{"clusterPoolIPv4PodCIDR":"10.128.0.0/14","clusterPoolIPv4MaskSize":"23"},"mode":"cluster-pool"},"prometheus":{"serviceMonitor":{"enabled":false}},"hubble":{"tls":{"enabled":false}}}}]'
          alm-examples-metadata: '{"cilium-openshift-default":{"description":"Default
            CiliumConfig CR for OpenShift"}}'
          capabilities: Basic Install
          categories: Networking,Security
          repository: http://github.com/cilium/cilium
          support: support@isovalent.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Cilium - eBPF-based Networking, Security, and Observability
        displayName: Cilium
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - networking
        - security
        - observability
        - eBPF
        links:
        - name: Cilium Homepage
          url: https://cilium.io/
        maintainers:
        - email: maintainer@cilium.io
          name: Cilium
        maturity: stable
        provider:
          name: Isovalent
        relatedImages:
        - registry.connect.redhat.com/isovalent/cilium-olm-metadata@sha256:33a2e21a069ff24af4bf0ffc51e83ddf398df5d9735d0123120768e9dc2c27f9
        - registry.connect.redhat.com/isovalent/cilium-olm@sha256:b45ff6a6552d71cbc26e358e0b25759d33818fd92301f1656453c5bcf826ff13
        version: 1.10.4+x5bfd7b3
      entries:
      - name: cilium.v1.10.4-x5bfd7b3
        version: 1.10.4+x5bfd7b3
      name: stable
    defaultChannel: "1.15"
    packageName: cilium
    provider:
      name: Isovalent
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhbk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22
    - currentCSV: rhbk-operator.v22.0.13-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
          createdAt: "2024-09-17T11:20:22Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:884afe0ca1698d85a02aa47b6d60b8222e50d70ad280db6f98e1a4bbdbe13c40
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:c342da2a8183de278abebf98ef23332ce382544233a8b1128bf1ff74126e9c14
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:21e808c676cf3e231f0d5592a48ec668c069ba00cbc7c5aaac08f61d52323ad8
        version: 22.0.13-opr.1
      entries:
      - name: rhbk-operator.v22.0.13-opr.1
        version: 22.0.13-opr.1
      - name: rhbk-operator.v22.0.12-opr.1
        version: 22.0.12-opr.1
      - name: rhbk-operator.v22.0.11-opr.2
        version: 22.0.11-opr.2
      - name: rhbk-operator.v22.0.11-opr.1
        version: 22.0.11-opr.1
      - name: rhbk-operator.v22.0.10-opr.1
        version: 22.0.10-opr.1
      - name: rhbk-operator.v22.0.9-opr.1
        version: 22.0.9-opr.1
      - name: rhbk-operator.v22.0.8-opr.2
        version: 22.0.8-opr.2
      - name: rhbk-operator.v22.0.8-opr.1
        version: 22.0.8-opr.1
      - name: rhbk-operator.v22.0.7-opr.1
        version: 22.0.7-opr.1
      - name: rhbk-operator.v22.0.6-opr.1
        version: 22.0.6-opr.1
      name: stable-v22.0
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24
    - currentCSV: rhbk-operator.v24.0.11-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
          createdAt: "2025-03-26T02:27:51Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:df74d1b1479c6cf371e45b4775bc6c905a9855cf182fb6ff8a752c6e846f55a6
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:ee815bf36c438e81b9553f0359a7725f0b3f633a5143f04cc822ea865940801a
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:d339ecf788d1681ba9945a2ea63ca675caec863ec75f497ef916103f4c898003
        version: 24.0.11-opr.1
      entries:
      - name: rhbk-operator.v24.0.11-opr.1
        version: 24.0.11-opr.1
      - name: rhbk-operator.v24.0.10-opr.1
        version: 24.0.10-opr.1
      - name: rhbk-operator.v24.0.9-opr.1
        version: 24.0.9-opr.1
      - name: rhbk-operator.v24.0.8-opr.1
        version: 24.0.8-opr.1
      - name: rhbk-operator.v24.0.7-opr.1
        version: 24.0.7-opr.1
      - name: rhbk-operator.v24.0.6-opr.1
        version: 24.0.6-opr.1
      - name: rhbk-operator.v24.0.5-opr.1
        version: 24.0.5-opr.1
      - name: rhbk-operator.v24.0.4-opr.1
        version: 24.0.4-opr.1
      - name: rhbk-operator.v24.0.3-opr.1
        version: 24.0.3-opr.1
      name: stable-v24.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26
    - currentCSV: rhbk-operator.v26.0.15-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
          createdAt: "2025-08-28T19:34:30Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:fc5fc8cfb2c100c5b3a33fc11a78dfca569f5f34f45d88439340392d2c60afb2
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:4bb4d3070fe118bdd7de24e8b68133b8b6e158cea48c9f1abad2b4297be0dfdb
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:74775c7851ac98ba2903435c71b9d02cc504ca8ee98b65ad7282d671392756a6
        version: 26.0.15-opr.1
      entries:
      - name: rhbk-operator.v26.0.15-opr.1
        version: 26.0.15-opr.1
      - name: rhbk-operator.v26.0.14-opr.1
        version: 26.0.14-opr.1
      - name: rhbk-operator.v26.0.13-opr.1
        version: 26.0.13-opr.1
      - name: rhbk-operator.v26.0.12-opr.1
        version: 26.0.12-opr.1
      - name: rhbk-operator.v26.0.11-opr.1
        version: 26.0.11-opr.1
      - name: rhbk-operator.v26.0.10-opr.1
        version: 26.0.10-opr.1
      - name: rhbk-operator.v26.0.9-opr.1
        version: 26.0.9-opr.1
      - name: rhbk-operator.v26.0.8-opr.1
        version: 26.0.8-opr.1
      - name: rhbk-operator.v26.0.7-opr.1
        version: 26.0.7-opr.1
      - name: rhbk-operator.v26.0.6-opr.1
        version: 26.0.6-opr.1
      - name: rhbk-operator.v26.0.5-opr.1
        version: 26.0.5-opr.1
      name: stable-v26.0
    - currentCSV: rhbk-operator.v26.2.9-opr.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
          createdAt: "2025-09-18T15:03:55Z"
          description: An Operator for installing and managing Keycloak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Red Hat build of Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhbk/keycloak-operator-bundle@sha256:a588a2eed385d29ccaf1f81bd719ca4b4ee85ab7359081706b2e810fd5e2db79
        - registry.redhat.io/rhbk/keycloak-rhel9-operator@sha256:9d5d8332fab50df73589b1961272eb5fd6a606d0115d3389c37e68881f4843cd
        - registry.redhat.io/rhbk/keycloak-rhel9@sha256:50a62b46af44ce0ff7e6c3a15e1bfb2a9611ba4a8b65b8d0ab65817064692c55
        version: 26.2.9-opr.1
      entries:
      - name: rhbk-operator.v26.2.9-opr.1
        version: 26.2.9-opr.1
      - name: rhbk-operator.v26.2.8-opr.1
        version: 26.2.8-opr.1
      - name: rhbk-operator.v26.2.7-opr.1
        version: 26.2.7-opr.1
      - name: rhbk-operator.v26.2.6-opr.1
        version: 26.2.6-opr.1
      - name: rhbk-operator.v26.2.5-opr.1
        version: 26.2.5-opr.1
      - name: rhbk-operator.v26.2.4-opr.2
        version: 26.2.4-opr.2
      - name: rhbk-operator.v26.2.4-opr.1
        version: 26.2.4-opr.1
      name: stable-v26.2
    defaultChannel: stable-v26.2
    packageName: rhbk-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: www.ibm.com
    name: ibm-elasticsearch-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-elasticsearch-operator.v1.1.2667
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
          createdAt: "2025-01-30"
          description: Elasticsearch operator for use within Cloudpak
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-elasticsearch-operator
          olm.skipRange: <0.0.0 >1.1.0
          operators.openshift.io/valid-subscription: '["IBM Elasticsearch"]'
          operators.operatorframework.io/internal-objects: '["elasticsearchclusters.elasticsearch.opencontent.ibm.com"]'
          operators.operatorframework.io/operator-type: non-standalone
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ElasticsearchCluster
            name: elasticsearchclusters.elasticsearch.opencontent.ibm.com
            version: v1
        description: |-
          # Introduction
          An Elasticsearch operator for use with IBM Cloudpak services.
          # Details
          ibm-elasticsearch-operator supports installing Elasticsearch major version 6 and 7. This operator supports two CRDs Elasticsearch and Elasticsearchdetails. Currently Elasticsearch is an abstraction of Elasticsearchdetails, but there is no difference between both the CRDs. Creating either of them provisions elasticsearch instances. This abstraction layer has been created to support any changes to the operator with out breaking the CRD/API for adopting products.

          ## Prerequisites
          - Openshift 3.11+
          ## Resources Required
          In short the answer is it depends. Especially if you are running Elasticsearch 6 you need 3 or a greater odd number of replicas to make a cluster and not have split brain. Each replica will have a sidecar that needs 0.1 cpu and 2gb of ram. You should size Elasticsearch based on your use-case. The default is a full cpu and 4 cpu and 64 gb ram. To assist sizing requirements we have included a copy of [esrally](https://esrally.readthedocs.io/en/stable/) on all the docker images.
          ## Installing
          ElasticsearchCluster is the strategic Kind moving forward. There are two flavors of CRD supported that will soon be deprecated. A simple Elasticseach "All in one" crd and a more detailed CRD for teams who need greater control. Upon deploy the elasticsearch is available from the service (not the headless) created by the operator. Only https will be available on port 443. the service will also be in the ElasticsearchDetails status.

          ### Case install
          To install the component using case you must either use the native install or catalog install. Upon installing the catalog you can then use either OLM or ODLM to install but must create the OperatorGroup & Subscription for either method to avoid conflict.

          # Configuration
          Architecture:
            - x86
            - s390x
            - ppc64le
          Storage: RWO Encrypted Block Storage with 120+ IOPS/GB
          Clouds:
          - IBM CLoud (ibm-block-gold, Gen2 custom iops storage class)
          - AWS (EBS)
          - Azure (For Azure please ensure you are using a higher tier of AzureDisk storage)
          - OCS (RBD not Ceph)
          - Portworx (Non-shared replacated)
          - Please do not use with NFS

          ### Limitations
          This product is not deployable standalone and it is part of the LICCR already accepted
          ## SecurityContextConstraints Requirements
          This component requires a restricted SCC.
        displayName: IBM OpenContent Elasticsearch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - internal
        - Elasticsearch
        maintainers:
        - email: noreply@ibm.com
          name: IBM CloudpakOpen Community
        maturity: alpha
        provider:
          name: IBM
          url: www.ibm.com
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.connect.redhat.com/cpopen/ibm-elasticsearch-operator@sha256:0bedcf95b31e28d41b124780f3b97c68a87db5626ecf1d5e81bb06b8b3fcd91a
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        version: 1.1.2667
      entries:
      - name: ibm-elasticsearch-operator.v1.1.2667
        version: 1.1.2667
      name: v1.1
    defaultChannel: v1.1
    packageName: ibm-elasticsearch-operator
    provider:
      name: IBM
      url: www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ProphetStor Data Services, Inc.
      provider-url: ""
    name: federatorai-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: federatorai.v4.7.1-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "federatorai.containers.ai/v1alpha1",
                "kind": "AlamedaService",
                "metadata": {
                  "name": "my-alamedaservice"
                },
                "spec": {
                  "alamedaAi": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "alamedaInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "100Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "fedemeterInfluxdb": {
                    "storages": [
                      {
                        "accessModes": ["ReadWriteOnce"],
                        "size": "10Gi",
                        "type": "pvc",
                        "usage": "data"
                      }
                    ]
                  },
                  "imageLocation": "registry.connect.redhat.com/prophetstor",
                  "storages": [
                    {
                      "accessMode": ["ReadWriteOnce"],
                      "size": "2Gi",
                      "type": "pvc",
                      "usage": "log"
                    }
                  ],
                  "version": "v4.7.1-b1760"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
          createdAt: "2021-10-12T01:14:00Z"
          description: Federator.ai Operator provides easy configuration and management
            of AI-based Kubernetes resource orchestrator
          operatorframework.io/suggested-namespace: federatorai
          repository: https://access.redhat.com/containers/#/registry.connect.redhat.com/prophetstor/federatorai-operator
          support: ProphetStor Data Services, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An instance of Federator.ai.
            displayName: AlamedaService
            kind: AlamedaService
            name: alamedaservices.federatorai.containers.ai
            version: v1alpha1
        description: |-
          Federator.ai helps enterprises optimize cloud resources, maximize application performance, and save significant cost without excessive over-provisioning or under-provisioning of resources, meeting the service-level requirements of their applications.

          Enterprises often lack understanding of the resources needed to support their applications. This leads to either excessive over-provisioning or under-provisioning of resources (CPU, memory, storage). Using machine learning, Federator.ai determines the optimal cloud resources needed to support any workload on OpenShift and helps users find the best-cost instances from cloud providers for their applications.


          **Multi-layer workload prediction**

          Using machine learning and math-based algorithms, Federator.ai predicts containerized application and cluster node resource usage as the basis for resource recommendations at application level as well as at cluster node level. Federator.ai supports prediction for both physical/virtual CPUs and memories.


          **Auto-scaling via resource recommendation**

          Federator.ai utilizes the predicted resource usage to recommend the right number and size of pods for applications. Integrated with Datadog's WPA, applications are automatically scaled to meet the predicted resource usage.


          **Application-aware recommendation execution**

          Optimizing the resource usage and performance goals, Federator.ai uses application specific metrics for workload prediction and pod capacity estimation to auto-scale the right number of pods for best performance without overprovisioning.


          **Multi-cloud Cost Analysis**

          With resource usage prediction, Federator.ai analyzes potential cost of a cluster on different public cloud providers. It also recommend appropriate cluster nodes and instance types based on resource usage.


          **Custom Datadog/Sysdig Dashboards**

          Predefined custom Datadog/Sysdig Dashboards for workload prediction/recommendation visualization for cluster nodes and applications.


          **Additional resources**

          Want more product information? Explore detailed information about using this product and where to find additional help.

          * [Federator.ai Datasheet](https://prophetstor.com/wp-content/uploads/datasheets/Federator.ai.pdf)
          * [Quick Start Guide](https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf)
          * [Installation Guide](https://prophetstor.com/wp-content/uploads/2021/08/ProphetStor-Federator.ai-v4.7.0-Installation-Guide-v1.0.pdf)
          * [User Guide](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-User-Guide.pdf)
          * [Release Notes](https://prophetstor.com/wp-content/uploads/2021/08/Federator.ai-4.7.0-Release-Notes.pdf)
          * [Company Information](https://www.prophetstor.com/)


          ### Prerequisite
          *  The [Openshift](https://www.openshift.com) version 4.4.x or later if using OpenShift platform.
          *  The [Kubernetes](https://kubernetes.io/) version 1.16 or later if using Kubernetes environment.
        displayName: Federator.ai
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - AI
        - Resource Orchestration
        - NoOps
        links:
        - name: Website
          url: https://prophetstor.com/federator-ai-2/
        - name: Quickstart guide
          url: https://prophetstor.com/wp-content/uploads/documentation/Federator.ai/Latest%20Version/ProphetStor%20Federator.ai%20Quick%20Installation%20Guide.pdf
        maintainers:
        - email: support@prophetstor.com
          name: ProphetStor Data Services, Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: ProphetStor Data Services, Inc.
        relatedImages:
        - registry.connect.redhat.com/prophetstor/alameda-executor-ubi@sha256:4512e3be7a89d0ca9f120b68f6ab9766d000bf2b01b279cd55b2dbc433aacfb6
        - registry.connect.redhat.com/prophetstor/federatorai-operator-bundle@sha256:d0ac978d749654ddf4cc20823805ba7d33ee62e9a7ec98fdd53d9926b697c347
        - registry.connect.redhat.com/prophetstor/federatorai-data-adapter@sha256:adf5d655fa030dc05133516d45642d7282fdd66e0aa8e7a7331ed614a41d3fe0
        - registry.connect.redhat.com/prophetstor/federatorai-operator@sha256:bdd87f6b7d7ba9d12ed8ddb9c5151e5a89781e88d7a02b73814b16a766968047
        - registry.connect.redhat.com/prophetstor/fedemeter-api-ubi@sha256:451456207fe8d16b9824b3f64d00645766dcccbc8d685b6942e6c00d0d0c627e
        - registry.connect.redhat.com/prophetstor/alameda-recommender-ubi@sha256:13e941df9252cea3024ecb5e2247a696b64928015c1a3a4a04ee4e362388b1e4
        - registry.connect.redhat.com/prophetstor/alameda-operator-ubi@sha256:9a863394cb1c89228223b26addb0d12bb91a97233fe2d5d19ecd5998e0cfe0d6
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-backend@sha256:022dbf6696edb8eaf389fdbd1dff14c40a51c237c01f8a02a7ea65c60cd59f38
        - registry.connect.redhat.com/prophetstor/fedemeter-influxdb@sha256:36904f003e8e29d4a9fd738655f401fbe840d60158d7d820cda79a09858dc2c2
        - registry.connect.redhat.com/prophetstor/federatorai-agent-ubi@sha256:cfadc84f31f3776ded15d4d8dda92e00b7d9b1bce0ff5440277176292f9304d3
        - registry.connect.redhat.com/prophetstor/alameda-datahub-ubi@sha256:3819d6acce8d376dbe3b12469be2528d0ce9ac95c2a86507aeb76b6e3a5f5f38
        - registry.connect.redhat.com/prophetstor/alameda-notifier-ubi@sha256:1f75034fb78d62924a87cb4a5435661b4afa990d6960431e784029742fbae0da
        - registry.connect.redhat.com/prophetstor/alameda-rabbitmq@sha256:b396d9aa0470e593a36f5799c04fc3c9e59b5723f703f4b019f7ef92528e6741
        - registry.connect.redhat.com/prophetstor/federatorai-dashboard-frontend@sha256:1c1147425ca6302103363a1a3bcf27baeef13b10ec6780ef266b7f52fe27ffeb
        - registry.connect.redhat.com/prophetstor/alameda-ai-dispatcher@sha256:6e0dc1136b0f8e5971d92b95ce0d6dfac475044c073bada50f23b315f4478ae4
        - registry.connect.redhat.com/prophetstor/federatorai-rest-ubi@sha256:3e92528615543ea900bf38387c4eb8439b60759be5be2ac5311a9dbaafc7026b
        - registry.connect.redhat.com/prophetstor/alameda-analyzer-ubi@sha256:6117ac4692793d49bc0204f13091dec119bdbc7b169e9deb86c243aca00ded27
        - registry.connect.redhat.com/prophetstor/federatorai-agent-preloader@sha256:66daea6193e399f8363df5ba42e1fcd643db6df3a9075dbcd8c3464fdc91e776
        - registry.connect.redhat.com/prophetstor/alameda-ai@sha256:9293298bb2a72d0b0764050b6f6d8da707b46a83363a3b6aa7be60f512e2ec61
        - registry.connect.redhat.com/prophetstor/alameda-influxdb@sha256:485c2119a6a56807e28ad03b2087129950118b55531126a66bb414063052ae76
        version: 4.7.1-1
      entries:
      - name: federatorai.v4.7.1-1
        version: 4.7.1-1
      - name: federatorai.v4.6.1-4
        version: 4.6.1-4
      name: stable
    defaultChannel: stable
    packageName: federatorai-certified
    provider:
      name: ProphetStor Data Services, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: cert-utils-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cert-utils-operator.v1.3.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
          createdAt: "2024-01-23T12:29:04Z"
          description: Set of utilities for TLS certificates
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: cert-utils-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/cert-utils-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Cert utils operator is a set of functionalities around certificates packaged in a [Kubernetes operator](https://github.com/operator-framework/operator-sdk).

          Certificates are assumed to be available in a [secret](https://kubernetes.io/docs/concepts/configuration/secret/) of type `kubernetes.io/tls` (other types of secrets are *ignored* by this operator).
          By convention this type of secrets have three optional entries:

          1. `tls.key`: the private key of the certificate.
          2. `tls.crt`: the actual certificate.
          3. `ca.crt`: the CA bundle that validates the certificate.

          The functionalities are the following:

          1. [Ability to populate route certificates](https://github.com/redhat-cop/cert-utils-operator#Populating-route-certificates)
          2. [Ability to create java keystore and truststore from the certificates](https://github.com/redhat-cop/cert-utils-operator#Creating-java-keystore-and-truststore)
          3. [Ability to show info regarding the certificates](https://github.com/redhat-cop/cert-utils-operator#Showing-info-on-the-certificates)
          4. [Ability to alert when a certificate is about to expire](https://github.com/redhat-cop/cert-utils-operator#Alerting-when-a-certificate-is-about-to-expire)
          5. [Ability to inject ca bundles in ValidatingWebhookConfiguration, MutatingWebhookConfiguration, CustomResourceDefinition object](https://github.com/redhat-cop/cert-utils-operator#CA-injection)

          All these feature are activated via opt-in annotations.
        displayName: Cert Utils Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TLS
        - secrets
        - security
        - certificates
        links:
        - name: Cert Utils Operator
          url: https://github.com/redhat-cop/cert-utils-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/cert-utils-operator
        maintainers:
        - email: rspazzoli@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/cert-utils-operator@sha256:0290e7b2800a5aa8eb57b0268908f2a20a93b16403c76df4c61021048161d6ff
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/community-operator-pipeline-prod/cert-utils-operator@sha256:92a0c0b893b56330593e0a255b8df61031e92a61d172aecc58be5805d847798a
        version: 1.3.12
      entries:
      - name: cert-utils-operator.v1.3.12
        version: 1.3.12
      - name: cert-utils-operator.v1.3.11
        version: 1.3.11
      - name: cert-utils-operator.v1.3.10
        version: 1.3.10
      - name: cert-utils-operator.v1.3.9
        version: 1.3.9
      - name: cert-utils-operator.v1.3.8
        version: 1.3.8
      - name: cert-utils-operator.v1.3.7
        version: 1.3.7
      - name: cert-utils-operator.v1.3.6
        version: 1.3.6
      - name: cert-utils-operator.v1.3.5
        version: 1.3.5
      - name: cert-utils-operator.v1.3.4
        version: 1.3.4
      - name: cert-utils-operator.v1.3.3
        version: 1.3.3
      - name: cert-utils-operator.v1.3.2
        version: 1.3.2
      - name: cert-utils-operator.v1.3.1
        version: 1.3.1
      - name: cert-utils-operator.v1.3.0
        version: 1.3.0
      - name: cert-utils-operator.v1.2.0
        version: 1.2.0
      - name: cert-utils-operator.v1.1.0
        version: 1.1.0
      - name: cert-utils-operator.v1.0.7
        version: 1.0.7
      - name: cert-utils-operator.v1.0.6
        version: 1.0.6
      - name: cert-utils-operator.v1.0.5
        version: 1.0.5
      - name: cert-utils-operator.v1.0.4
        version: 1.0.4
      - name: cert-utils-operator.v1.0.3
        version: 1.0.3
      - name: cert-utils-operator.v1.0.2
        version: 1.0.2
      - name: cert-utils-operator.v1.0.1
        version: 1.0.1
      - name: cert-utils-operator.v1.0.0
        version: 1.0.0
      - name: cert-utils-operator.v0.2.2
        version: 0.2.2
      - name: cert-utils-operator.v0.2.1
        version: 0.2.1
      - name: cert-utils-operator.v0.2.0
        version: 0.2.0
      - name: cert-utils-operator.v0.1.1
        version: 0.1.1
      - name: cert-utils-operator.v0.1.0
        version: 0.1.0
      - name: cert-utils-operator.v0.0.3
        version: 0.0.3
      - name: cert-utils-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cert-utils-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: gitwebhook-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitwebhook-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitWebhook",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "gitwebhook-operator",
                    "app.kubernetes.io/instance": "gitwebhook-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gitwebhook",
                    "app.kubernetes.io/part-of": "gitwebhook-operator"
                  },
                  "name": "gitwebhook-sample"
                },
                "spec": null
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
          createdAt: "2024-05-10T18:17:13Z"
          description: An operator to create git webhooks declaratively
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: gitwebhook-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/gitwebhook-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitWebhook is the Schema for the gitwebhooks API
            displayName: Git Webhook
            kind: GitWebhook
            name: gitwebhooks.redhatcop.redhat.io
            version: v1alpha1
        description: "This operator automates the setting up of git webhooks. Its
          main purpose is to stream line the configuration of operators such as Tekton
          or ArgoCD which rely on git webhooks for some of their functionalities.\n\n##
          The GitWebhook CRD\n \nA CRD is provided by this operator to create a git
          web hook, here is an example for github\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GitWebhook\nmetadata:\n  name: gitwebhook-github\nspec:\n  gitHub:\n    gitServerCredentials:\n
          \     name: github-pat\n  repositoryOwner: ${repo_owner}\n  ownerType: user\n
          \ repositoryName: ${repo_name}\n  webhookURL: https://hellowebhook.com\n
          \ insecureSSL: false\n  webhookSecret:\n    name: webhook-secret\n  events:\n
          \   - push  \n  contentType: json\n  active: true  \n```    \n\nhere is
          an explanation of each field:\n\n- `gihub` specifies how to connect to the
          git api server. It also requires a local reference to a secret (in the same
          namespace) containing a key `token` with a valid github token to be used
          to authenticate. A similar `gitLab` section exists when connecting to gitlab.
          Only one of `gitLab` or `gitHub` can be defined. \n- `repositoryOwner` and
          `repositoryName` identify the repository for which we want to receive events.\n-
          `ownerType` can have two values: `user` and `organization` and identifies
          the kind of owner.\n- `webhookURL` is the URL for to be called.\n - `insecureSSL`
          defines whether the target URL certificate should be validated (default
          `false`). \n- `webhookSecret` defines a local reference to a secret containing
          the `secret` key. The value is a shared secret between the webhook caller
          and the received for farther validation or identification of the caller.\n-
          `events` is the list of the repo-level events that the webhook should generate.
          The list of valid events for github can be found [here](https://docs.github.com/en/developers/webhooks-and-events/webhooks/webhook-events-and-payloads).
          The list of valid events for gitlab can be found [here](https://docs.gitlab.com/ee/user/project/integrations/webhook_events.html).\n
          - `contentType` defines the format of the webhook payload (default `json`)
          (github).\n - `active` whether the webhook should be turned on (default
          `true`) (github).\n - `pushEventBranchFilter` a regular expression to filter
          from which branches push events should be generated (gitlab only).\n\n##
          Security Considerations\n\nThis operator does not own credentials for the
          git server, but instead always allocate a new connection based on the credentials
          referenced in the CR and every reconcile cycle. As a result there is no
          risk of security escalation or credential leaking between tenants of a cluster
          using this operator. On the other hand it is the responsibility of the namespace
          owners or the platform owner to ensure that valid git credentials are always
          available in the namespace where the GitWebhook CRs need to defined.\n\n##
          Current support\n\nCurrently this operator support creating repo-level webhooks
          for github and gitlab. Potentially this operator could be extended to support
          org-level webhook or other git systems. Contributions are welcome.  "
        displayName: GitWebhook operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - webhook
        - github
        - gitlab
        links:
        - name: GitWbehook Operator
          url: https://github.com/redhat-cop/gitwebhook-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/gitwebhook-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - quay.io/community-operator-pipeline-prod/gitwebhook-operator@sha256:2d33e4f4f4b607ee2006481efc7ad9341049085ef988777d23e586f7406a5109
        - quay.io/redhat-cop/gitwebhook-operator@sha256:2810d75eebb24069dbbedf6a82285136c6f16b4cc58b59f1b0fbdc38560cf8a8
        version: 0.1.2
      entries:
      - name: gitwebhook-operator.v0.1.2
        version: 0.1.2
      - name: gitwebhook-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: gitwebhook-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: The bpfman Community
      provider-url: https://bpfman.io/
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:v0.5.6"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:v0.5.6"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:v0.5.6"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:v0.5.6"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:v0.5.6"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:v0.5.6"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:v0.5.6
          createdAt: "2025-02-18T13:44:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The bpfman Operator is a Kubernetes Operator for deploying [bpfman](https://bpfman.netlify.app/),
          a system daemon\nfor managing eBPF programs. It deploys bpfman itself along
          with CRDs to make deploying\neBPF programs in Kubernetes much easier.\n\n##
          Quick Start\n\nTo get bpfman up and running quickly simply click 'install'
          to deploy the bpfman-operator in the bpfman namespace via operator-hub.\n##
          Configuration\n\nThe `bpfman-config` configmap is automatically created
          in the `bpfman` namespace and used to configure the bpfman deployment.\n\nTo
          edit the config simply run\n\n```bash\nkubectl edit cm bpfman-config\n```\n\nThe
          following fields are adjustable\n\n- `bpfman.agent.image`: The image used
          for the bpfman-agent`\n- `bpfman.image`: The image used for bpfman`\n- `bpfman.log.level`:
          the log level for bpfman, currently supports `debug`, `info`, `warn`, `error`,
          and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`: the log level
          for the bpfman-agent currently supports `info`, `debug`, and `trace` \n\nThe
          bpfman operator deploys eBPF programs via CRDs. The following CRDs are currently
          available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n - KprobeProgram\n-
          TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n- XdpProgram\n
          - BpfNsApplication\n- TcNsProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n- BpfProgram\n - BpfNsProgram\n\n ## More information\n\nPlease
          checkout the [bpfman community website](https://bpfman.io/) for more information."
        displayName: Bpfman Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: The bpfman Community
          url: https://bpfman.io/
        relatedImages:
        - quay.io/bpfman/bpfman-operator:v0.5.6
        - quay.io/community-operator-pipeline-prod/bpfman-operator@sha256:6b3603c7c1917c67b5aa57202d1d2fdc0d9ac6b3a9373afa7172f322b3b3fa9b
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      - name: bpfman-operator.v0.5.5
        version: 0.5.5
      - name: bpfman-operator.v0.5.4
        version: 0.5.4
      - name: bpfman-operator.v0.5.2
        version: 0.5.2
      - name: bpfman-operator.v0.5.1
        version: 0.5.1
      - name: bpfman-operator.v0.5.0
        version: 0.5.0
      name: alpha
    defaultChannel: alpha
    packageName: bpfman-operator
    provider:
      name: The bpfman Community
      url: https://bpfman.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/crunchy-postgres-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle-marketplace@sha256:19c9d44cb9a0c7f52501fcb589750f3e85e270d066423bcadac593dd79e48ff4
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator-rhmp
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: boxboat-github-practice
      provider-url: https://github.com/boxboat-github-practice
    name: github-arc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: github-arc-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "github-practice.boxboat.com/v1alpha1",
                "kind": "ActionsRunnerController",
                "metadata": {
                  "name": "arc-sample"
                },
                "spec": {
                  "authSecret": {
                    "name": "ghauth"
                  },
                  "createRunnerNamespaces": true,
                  "openshift": true,
                  "runnerNamespaces": [
                    "test1",
                    "test2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery, OpenShift Optional
          containerImage: ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
          createdAt: "2023-03-03T20:49:16Z"
          description: Operator for deploying the github actions runner controller
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/boxboat-github-practice/github-arc-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActionsRunnerController is the schema for the ActionsRunnerController
              API
            displayName: Actions Runner Controller
            kind: ActionsRunnerController
            name: actionsrunnercontrollers.github-practice.boxboat.com
            version: v1alpha1
        description: helm based operator for github actions runner controller
        displayName: github-arc-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - github
        - actions
        - boxboat
        - ibm
        - self-hosted
        - runners
        links:
        - name: Github Arc Operator
          url: https://github.com/boxboat-github-practice/github-arc-operator
        - name: Actions Runner Controller
          url: https://github.com/actions/actions-runner-controller
        - name: Boxboat
          url: https://boxboat.com/
        maintainers:
        - email: Michael.McMullen@ibm.com
          name: mbmcmullen27
        maturity: alpha
        provider:
          name: boxboat-github-practice
          url: https://github.com/boxboat-github-practice
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - ghcr.io/boxboat-github-practice/github-arc-operator:1.1.0
        - quay.io/openshift-community-operators/github-arc-operator@sha256:d8d97d19cea466c4c240097b79bfb7de90312d937ad4db63f9e90901686e54da
        version: 1.1.0
      entries:
      - name: github-arc-operator.v1.1.0
        version: 1.1.0
      - name: github-arc-operator.v1.0.4
        version: 1.0.4
      name: alpha
    defaultChannel: alpha
    packageName: github-arc-operator
    provider:
      name: boxboat-github-practice
      url: https://github.com/boxboat-github-practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: power-monitoring-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: power-monitoring-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "Kepler",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "kepler",
                    "app.kubernetes.io/name": "kepler",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "kepler"
                },
                "spec": {
                  "exporter": {
                    "deployment": {
                      "port": 9103
                    },
                    "redfish": null
                  }
                }
              },
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
          createdAt: "2025-07-18T06:13:30Z"
          description: Deploys and Manages Kepler on OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "keplerinternals.kepler.system.sustainable.computing.io",
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift-power-monitoring/power-monitoring-operator
          support: Team Power Monitoring (team-powermon@redhat.com)
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KeplerInternal is the Schema for internal/unsupported API
            displayName: KeplerInternal
            kind: KeplerInternal
            name: keplerinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: Kepler is deprecated and will be removed in a future release.
              Please use PowerMonitor instead.
            displayName: Kepler
            kind: Kepler
            name: keplers.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |2

          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Deprecation Notice
          **Important:** The Kepler CR (Custom Resource) is deprecated and will be removed in a future release. Please use the new PowerMonitor instead.

          ### Prerequisites
          To view the metrics exposed by Kepler,
          **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          Documentation and installation guide can be found below:
            * [Installation Guide](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring)
        displayName: Power monitoring for Red Hat OpenShift
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://docs.redhat.com/documentation/openshift_container_platform/latest/html/power_monitoring/installing-power-monitoring
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: kaliu@redhat.com
          name: KaiYiLiu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-operator-bundle@sha256:5ebbc06088e99b937327cd457324461cafdda9649b2dd2db1bcb892be8556eed
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:141b069ff9925390607315b0cf675003abbdac9555b29700e3b999fadd6c89f2
        - registry.redhat.io/openshift-power-monitoring/power-monitoring-rhel9-operator@sha256:96fc903f6602c671a8a62b7d7638c7c99c46ed53d827302bbe6e12c9a84293bc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:971174692a1aae68ac07ffa132254eb3c018a52507eb32621b6dc286fd25ca0d
        version: 0.5.0
      entries:
      - name: power-monitoring-operator.v0.5.0
        version: 0.5.0
      - name: power-monitoring-operator.v0.4.0
        version: 0.4.0
      - name: power-monitoring-operator.v0.3.0
        version: 0.3.0
      - name: power-monitoring-operator.v0.2.0
        version: 0.2.0
      - name: power-monitoring-operator.v0.1.0
        version: 0.1.0
      name: tech-preview
    defaultChannel: tech-preview
    packageName: power-monitoring-operator
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cryostat-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cryostat-operator.v4.0.2-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.cryostat.io/v1beta2",
                "kind": "Cryostat",
                "metadata": {
                  "name": "cryostat-sample"
                },
                "spec": {
                  "enableCertManager": true,
                  "eventTemplates": [],
                  "networkPolicies": {},
                  "reportOptions": {
                    "replicas": 0
                  },
                  "trustedCertSecrets": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Developer Tools
          containerImage: registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
          createdAt: "2025-09-11T18:04:40Z"
          description: JVM monitoring and profiling tool
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <4.0.2'
          olm.substitutesFor: cryostat-operator.v4.0.2
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "operator.cryostat.io/v1beta2",
              "kind": "Cryostat",
              "metadata": {
                "name": "cryostat-sample"
              },
              "spec": {
                "enableCertManager": true,
                "reportOptions": {
                  "replicas": 0
                }
              }
            }
          operators.openshift.io/valid-subscription: '["Red Hat build of OpenJDK"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/cryostatio/cryostat-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cryostat allows you to install Cryostat for a single namespace,
              or multiple namespaces. It contains configuration options for controlling
              the Deployment of the Cryostat application and its related components.
              A Cryostat instance must be created to instruct the operator to deploy
              the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta2
          - description: Cryostat allows you to install Cryostat for a single namespace.
              It contains configuration options for controlling the Deployment of
              the Cryostat application and its related components. A Cryostat instance
              must be created to instruct the operator to deploy the Cryostat application.
            displayName: Cryostat
            kind: Cryostat
            name: cryostats.operator.cryostat.io
            version: v1beta1
        description: |
          The Red Hat build of Cryostat provides a cloud-based solution for interacting with the JDK Flight Recorder already present in OpenJDK 11+ JVMs. With Cryostat, users can remotely start, stop, retrieve, and even analyze JFR event data, providing the capability to easily take advantage of Flight Recorder's extremely low runtime cost and overhead and the flexibility to monitor applications and analyze recording data without transferring data outside of the cluster the application runs within.
          ##Prerequisites
          Cryostat requires [cert-manager](https://cert-manager.io/) to run. If not already installed in your cluster, please [install](https://cert-manager.io/docs/installation/) it using your preferred method.
          When running on Kubernetes, the operator also requires [Ingress configurations](https://github.com/cryostatio/cryostat-operator/blob/v4.0.0/docs/config.md#network-options) for each of its services to make them available outside of the cluster. The user is responsible for providing the hostnames for each Ingress.
        displayName: Red Hat build of Cryostat
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flightrecorder
        - java
        - jdk
        - jfr
        - jmc
        - missioncontrol
        - monitoring
        - profiling
        - diagnostic
        links:
        - name: Upstream Project
          url: https://github.com/cryostatio/cryostat
        - name: Website
          url: https://cryostat.io/
        maintainers:
        - email: cryostat-development@googlegroups.com
          name: The Cryostat Authors
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/cryostat/cryostat-db-rhel9@sha256:b6a6e401042daf1780ae6be19aedc551ed176d1a13bf1f0f48359552c638a156
        - registry.redhat.io/cryostat/cryostat-grafana-dashboard-rhel9@sha256:8fd30d86f5468227899560071a28e92197c5790ea820e918c47559f0695cfb8a
        - registry.redhat.io/cryostat/cryostat-openshift-console-plugin-rhel9@sha256:d2788c8732fec62c7a27e7bd1497fea1ab3ff605e7e913bbe17d8ef97567329b
        - registry.redhat.io/cryostat/cryostat-rhel9-operator@sha256:011a7297c3fbe6f10cb4388d025005bbe20bf966111671a7dbd487ac3b2de10a
        - registry.redhat.io/cryostat/jfr-datasource-rhel9@sha256:550de94fb20462c5cc5c52166820319544e774b8d45fc7beed7a095f5c1300aa
        - registry.redhat.io/cryostat/cryostat-reports-rhel9@sha256:bbee66a44c2e3a5525c2fa81db72e8528bccf40583d0bd665841644473308310
        - registry.redhat.io/ubi9/nginx-124@sha256:126903359331eb9f2c87950f952442db456750190956b9301d00acfcadbc19ed
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/cryostat/cryostat-rhel9@sha256:67e313e50c6ef100d6938dc1a734728661f328ab1f9757667c7a06de11689967
        - registry.redhat.io/cryostat/cryostat-agent-init-rhel9@sha256:d9720fefa3e34d05a5536fe1a66e23b30fe0845ae2b96b9c9be4ee705be9633b
        - registry.redhat.io/cryostat/cryostat-operator-bundle@sha256:455ecde240a152987fcfe0013abee992014510de9e89c014559a5d0665a56512
        - registry.redhat.io/cryostat/cryostat-ose-oauth-proxy-rhel9@sha256:9ba156d802c9c3eafc948afef88e0eaca6b5e73bf32ba80ea7b0651ba7ce923d
        - registry.redhat.io/cryostat/cryostat-storage-rhel9@sha256:2a5e1d2d39a7d9d2bc74316110e1a65e9e18f3a9fb7f0a498c7457dd9e6684ee
        version: 4.0.2+4
      entries:
      - name: cryostat-operator.v4.0.2-4
        version: 4.0.2+4
      - name: cryostat-operator.v4.0.2
        version: 4.0.2
      - name: cryostat-operator.v4.0.1
        version: 4.0.1
      - name: cryostat-operator.v4.0.1-5
        version: 4.0.1+5
      - name: cryostat-operator.v4.0.1-4
        version: 4.0.1+4
      - name: cryostat-operator.v4.0.1-3
        version: 4.0.1+3
      - name: cryostat-operator.v4.0.1-2
        version: 4.0.1+2
      - name: cryostat-operator.v4.0.0-10
        version: 4.0.0+10
      - name: cryostat-operator.v4.0.0
        version: 4.0.0
      - name: cryostat-operator.v4.0.0-11
        version: 4.0.0+11
      - name: cryostat-operator.v4.0.0-9
        version: 4.0.0+9
      - name: cryostat-operator.v3.0.1-7
        version: 3.0.1+7
      - name: cryostat-operator.v3.0.1-9
        version: 3.0.1+9
      - name: cryostat-operator.v3.0.1-8
        version: 3.0.1+8
      - name: cryostat-operator.v3.0.1
        version: 3.0.1
      - name: cryostat-operator.v3.0.1-6
        version: 3.0.1+6
      - name: cryostat-operator.v3.0.1-5
        version: 3.0.1+5
      - name: cryostat-operator.v3.0.1-3
        version: 3.0.1+3
      - name: cryostat-operator.v3.0.1-13
        version: 3.0.1+13
      - name: cryostat-operator.v3.0.1-12
        version: 3.0.1+12
      - name: cryostat-operator.v3.0.1-11
        version: 3.0.1+11
      - name: cryostat-operator.v3.0.1-10
        version: 3.0.1+10
      - name: cryostat-operator.v3.0.0-12
        version: 3.0.0+12
      - name: cryostat-operator.v3.0.0
        version: 3.0.0
      - name: cryostat-operator.v3.0.0-6
        version: 3.0.0+6
      - name: cryostat-operator.v3.0.0-8
        version: 3.0.0+8
      - name: cryostat-operator.v2.4.0
        version: 2.4.0
      - name: cryostat-operator.v2.4.0-8
        version: 2.4.0+8
      - name: cryostat-operator.v2.4.0-7
        version: 2.4.0+7
      - name: cryostat-operator.v2.4.0-6
        version: 2.4.0+6
      - name: cryostat-operator.v2.4.0-4
        version: 2.4.0+4
      - name: cryostat-operator.v2.4.0-3
        version: 2.4.0+3
      - name: cryostat-operator.v2.3.1-6
        version: 2.3.1+6
      - name: cryostat-operator.v2.3.1-8
        version: 2.3.1+8
      - name: cryostat-operator.v2.3.1-9
        version: 2.3.1+9
      - name: cryostat-operator.v2.3.1
        version: 2.3.1
      - name: cryostat-operator.v2.3.0
        version: 2.3.0
      - name: cryostat-operator.v2.3.0-6
        version: 2.3.0+6
      - name: cryostat-operator.v2.2.1-14
        version: 2.2.1+14
      - name: cryostat-operator.v2.2.1-8
        version: 2.2.1+8
      - name: cryostat-operator.v2.2.1-5
        version: 2.2.1+5
      - name: cryostat-operator.v2.2.1-13
        version: 2.2.1+13
      - name: cryostat-operator.v2.2.1-12
        version: 2.2.1+12
      - name: cryostat-operator.v2.2.1-10
        version: 2.2.1+10
      - name: cryostat-operator.v2.2.1
        version: 2.2.1
      - name: cryostat-operator.v2.2.0
        version: 2.2.0
      - name: cryostat-operator.v2.1.1-2.1659469835.p
        version: 2.1.1+2.1659469835.p
      - name: cryostat-operator.v2.1.1-2.1655334614.p
        version: 2.1.1+2.1655334614.p
      - name: cryostat-operator.v2.1.1
        version: 2.1.1
      - name: cryostat-operator.v2.1.0
        version: 2.1.0
      name: stable
    defaultChannel: stable
    packageName: cryostat-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Cloudera
      provider-url: https://cloudera.com
    name: cloudera-streams-messaging-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: 1.2.x
    - currentCSV: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "Kafka",
                "metadata": {
                  "name": "my-cluster"
                },
                "spec": {
                  "kafka": {
                    "version": "3.9.0.1.4",
                    "replicas": 3,
                    "listeners": [
                      {
                        "name": "plain",
                        "port": 9092,
                        "type": "internal",
                        "tls": false
                      },
                      {
                        "name": "tls",
                        "port": 9093,
                        "type": "internal",
                        "tls": true
                      }
                    ],
                    "config": {
                      "offsets.topic.replication.factor": 3,
                      "transaction.state.log.replication.factor": 3,
                      "transaction.state.log.min.isr": 2,
                      "default.replication.factor": 3,
                      "min.insync.replicas": 2,
                      "inter.broker.protocol.version": "3.9"
                    },
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "zookeeper": {
                    "replicas": 3,
                    "storage": {
                      "type": "ephemeral"
                    }
                  },
                  "entityOperator": {
                    "topicOperator": {},
                    "userOperator": {}
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnect",
                "metadata": {
                  "name": "my-connect-cluster"
                },
                "spec": {
                  "version": "3.9.0.1.4",
                  "replicas": 1,
                  "bootstrapServers": "my-cluster-kafka-bootstrap:9093",
                  "tls": {
                    "trustedCertificates": [
                      {
                        "secretName": "my-cluster-cluster-ca-cert",
                        "pattern": "*.crt"
                      }
                    ]
                  },
                  "config": {
                    "group.id": "connect-cluster",
                    "offset.storage.topic": "connect-cluster-offsets",
                    "config.storage.topic": "connect-cluster-configs",
                    "status.storage.topic": "connect-cluster-status",
                    "config.storage.replication.factor": -1,
                    "offset.storage.replication.factor": -1,
                    "status.storage.replication.factor": -1
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "my-topic",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "partitions": 10,
                  "replicas": 3,
                  "config": {
                    "retention.ms": 604800000,
                    "segment.bytes": 1073741824
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaUser",
                "metadata": {
                  "name": "my-user",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "authentication": {
                    "type": "tls"
                  },
                  "authorization": {
                    "type": "simple",
                    "acls": [
                      {
                        "resource": {
                          "type": "topic",
                          "name": "my-topic",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read",
                          "Describe",
                          "Write",
                          "Create"
                        ],
                        "host": "*"
                      },
                      {
                        "resource": {
                          "type": "group",
                          "name": "my-group",
                          "patternType": "literal"
                        },
                        "operations": [
                          "Read"
                        ],
                        "host": "*"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaConnector",
                "metadata": {
                  "name": "my-source-connector",
                  "labels": {
                    "strimzi.io/cluster": "my-connect-cluster"
                  }
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSourceConnector",
                  "tasksMax": 1,
                  "config": {
                    "file": "/opt/kafka/LICENSE",
                    "topic": "my-topic"
                  }
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaRebalance",
                "metadata": {
                  "name": "my-rebalance",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "goals": [
                    "CpuCapacityGoal",
                    "NetworkInboundCapacityGoal",
                    "DiskCapacityGoal",
                    "RackAwareGoal",
                    "MinTopicLeadersPerBrokerGoal",
                    "NetworkOutboundCapacityGoal",
                    "ReplicaCapacityGoal"
                  ]
                }
              },
              {
                "apiVersion": "kafka.strimzi.io/v1beta2",
                "kind": "KafkaNodePool",
                "metadata": {
                  "name": "my-pool",
                  "labels": {
                    "strimzi.io/cluster": "my-cluster"
                  }
                },
                "spec": {
                  "replicas": 3,
                  "roles": [
                    "broker"
                  ],
                  "storage": {
                    "type": "jbod",
                    "volumes": [
                      {
                        "id": 0,
                        "type": "persistent-claim",
                        "size": "100Gi",
                        "deleteClaim": false
                      }
                    ]
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "my-cluster": {
                "description": "Example Kafka cluster with Ephemeral storage"
              },
              "my-connect-cluster": {
                "description": "Example Kafka Connect cluster"
              },
              "my-topic": {
                "description": "Example KafkaTopic custom resource"
              },
              "my-user": {
                "description": "Example KafkaUser custom resource"
              },
              "my-source-connector": {
                "description": "Example Kafka Connect source connector"
              },
              "my-rebalance": {
                "description": "Example Cruise Control rebalance request"
              },
              "my-pool": {
                "description": "Example KafkaNodePool custom resource"
              }
            }
          capabilities: Deep Insights
          categories: Streaming & Messaging
          certified: "false"
          containerImage: container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
          createdAt: "2025-07-14 10:09:41"
          description: Run Apache Kafka with Cloudera Streams Messaging - Kubernetes
            Operator
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy",
            "proxy-aware", "fips"]'
          operators.openshift.io/valid-subscription: A valid license and current subscription
            for Cloudera Streaming - Kubernetes Operator is required. Contact Cloudera
            sales to obtain a license. https://www.cloudera.com/contact-sales.html
          operators.operatorframework.io/internal-objects: '["strimzipodsets.core.strimzi.io"]'
          support: Cloudera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Kafka cluster
            displayName: Kafka
            kind: Kafka
            name: kafkas.kafka.strimzi.io
            version: v1beta2
          - description: Represents a pool of Kafka brokers
            displayName: Kafka Node Pool
            kind: KafkaNodePool
            name: kafkanodepools.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Connect cluster
            displayName: Kafka Connect
            kind: KafkaConnect
            name: kafkaconnects.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker cluster
            displayName: Kafka Mirror Maker
            kind: KafkaMirrorMaker
            name: kafkamirrormakers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka Bridge cluster
            displayName: Kafka Bridge
            kind: KafkaBridge
            name: kafkabridges.kafka.strimzi.io
            version: v1beta2
          - description: Represents a topic inside a Kafka cluster
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.kafka.strimzi.io
            version: v1beta2
          - description: Represents a user inside a Kafka cluster
            displayName: Kafka User
            kind: KafkaUser
            name: kafkausers.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Connector inside a Kafka Connect cluster
            displayName: Kafka Connector
            kind: KafkaConnector
            name: kafkaconnectors.kafka.strimzi.io
            version: v1beta2
          - description: Represents a Kafka MirrorMaker 2 cluster
            displayName: Kafka MirrorMaker 2
            kind: KafkaMirrorMaker2
            name: kafkamirrormaker2s.kafka.strimzi.io
            version: v1beta2
          - description: Triggeres rebalance of replicas in a Kafka cluster
            displayName: Kafka Rebalance
            kind: KafkaRebalance
            name: kafkarebalances.kafka.strimzi.io
            version: v1beta2
          - description: Internal custom resource used to manage Strimzi pods
            displayName: Strimzi Pod Set
            kind: StrimziPodSet
            name: strimzipodsets.core.strimzi.io
            version: v1beta2
        description: |-
          Cloudera Streams Messaging - Kubernetes Operator brings enterprise-grade Kafka deployments to your existing Kubernetes infrastructure.

          Cloudera packages extensively tested and stable releases of Kafka for enterprise deployments. To learn more about Cloudera Streaming, [read this deep dive blog post](https://blog.cloudera.com/accelerating-deployments-of-streaming-pipelines-announcing-data-in-motion-on-kubernetes/).

          Cloudera Streams Messaging - Kubernetes Operator consists of and ships multiple components including Apache Kafka, Apache ZooKeeper, Cruise Control, Strimzi, and others.

          ## Installation prerequisites

          Installing Cloudera Streams Messaging - Kubernetes Operator requires a valid license as well as Cloudera credentials (username and password) to be available in the installation namespace as `Secrets`. The license is required for operator applications to function properly. The credentials provide access to the Cloudera Docker registry (`container.repository.cloudera.com`) where installation artifacts are pulled from.

          ### Obtaining a license

          Contact [Cloudera Sales](https://www.cloudera.com/contact-sales.html) to obtain a license.

          ### Creating Secrets

          Before installation, create  two `Secrets`. One containing your Cloudera credentials, and one containing your license. Cloudera credentials are provided to you together with your license.

          Create a `Secret` containing your Cloudera credentials.
          ```
          oc create secret docker-registry cloudera-container-repository-credentials --docker-username=[***USERNAME***] --docker-password=[***PASSWORD***] --docker-server=container.repository.cloudera.com
          ```
          Create a `Secret` containing your license.
          ```
          oc create secret generic csm-op-license --from-file=license.txt=[***PATH TO LICENSE FILE***]
          ```

          Note the following:
          - The name of both `Secrets` is fixed.
          - The `Secret` containing your Cloudera credentials (`cloudera-container-repository-credentials`) must be available in all namespaces where you plan to deploy Kafka or Kafka Connect clusters. Cloudera recommends that you create the `Secret` in all required namespaces before installation if you know what namespaces you will be using to deploy Kafka or Kafka Connect.

          ## Learn more
          To learn more, visit the [Cloudera Streams Messaging - Kubernetes Operator documentation library](https://docs.cloudera.com/csm-operator).

          Additionally, to learn more about the various open-source components included in Cloudera Streams Messaging - Kubernetes Operator, visit the following resources:

          * [Strimzi](https://strimzi.io/)
          * [Apache Kafka](https://kafka.apache.org/)
          * [Apache ZooKeeper](https://zookeeper.apache.org/)
          * [Cruise Control](https://github.com/linkedin/cruise-control)
        displayName: Cloudera Streams Messaging - Kubernetes Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - messaging
        - kafka-connect
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Documentation
          url: https://docs.cloudera.com/csm-operator
        - name: Cloudera Community Forum
          url: https://community.cloudera.com/
        - name: Cloudera Stream Processing
          url: https://www.cloudera.com/products/stream-processing.html
        - name: Contact Cloudera
          url: https://www.cloudera.com/contact-sales.html
        maintainers:
        - email: support@cloudera.com
          name: Cloudera
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Cloudera
          url: https://cloudera.com
        relatedImages:
        - container.repository.cloudera.com/cloudera/maven-builder@sha256:9f18203b175193e5079fdc0c2b71c14887641d997f92b4a1c918546f777166df
        - registry.connect.redhat.com/cloudera/cloudera-streams-messaging-kubernetes-operator-bundle@sha256:e7c79d2e6b21ce423d0199316012ccde1ccf16da412da66e4a36c54539aac81d
        - container.repository.cloudera.com/cloudera/kafka-operator@sha256:d6812fed7ddf139dab4d12cab6a0ae292e2d0619303b0ad72f2eef642e5a183a
        - container.repository.cloudera.com/cloudera/kafka@sha256:aa64e161b9672153a6f6559f9ae9ef88d0e40d6040a80e21bd3dc8ee39df6523
        - container.repository.cloudera.com/cloudera/kafka@sha256:b675e7426a558c4681e53d8e03f278264d7ad555f7c02422017683a29708ac2b
        version: 1.4.0-b199
      entries:
      - name: cloudera-streams-messaging-kubernetes-operator.1.4.0-b199
        version: 1.4.0-b199
      name: stable
    defaultChannel: stable
    packageName: cloudera-streams-messaging-kubernetes-operator
    provider:
      name: Cloudera
      url: https://cloudera.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cognitoidentityprovider-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cognitoidentityprovider-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
          createdAt: "2025-09-19T20:58:49Z"
          description: AWS Cognito Identity Provider controller is a service controller
            for managing Cognito Identity Provider resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UserPool represents the state of an AWS cognitoidentityprovider
              UserPool resource.
            displayName: UserPool
            kind: UserPool
            name: userpools.cognitoidentityprovider.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Cognito Identity Provider resources in AWS from within your Kubernetes cluster.

          **About Amazon Cognito Identity Provider**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Cognito Identity Provider
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cognitoidentityprovider
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Cognito Identity Provider Developer Resources
          url: https://aws.amazon.com/Cognito Identity Provider/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cognitoidentityprovider maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cognitoidentityprovider-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cognitoidentityprovider-controller@sha256:093670fcaee6ef20e471c0df45c9ea7f40e5f0c43208c9447623324fdb39dbe0
        version: 1.1.1
      entries:
      - name: ack-cognitoidentityprovider-controller.v1.1.1
        version: 1.1.1
      - name: ack-cognitoidentityprovider-controller.v1.1.0
        version: 1.1.0
      - name: ack-cognitoidentityprovider-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-cognitoidentityprovider-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/self-node-remediation-operator:v0.9.0
          createdAt: "2024-07-25 14:40:52"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/self-node-remediation-operator:v0.9.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/self-node-remediation@sha256:549585ca18d331cfbd5911fa3e79912428a0be3308588be2438ece3247746bd8
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      - name: self-node-remediation.v0.7.0
        version: 0.7.0
      - name: self-node-remediation.v0.6.1
        version: 0.6.1
      - name: self-node-remediation.v0.6.0
        version: 0.6.0
      - name: self-node-remediation.v0.5.0
        version: 0.5.0
      - name: self-node-remediation.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: deployment-validation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: deployment-validation-operator.v0.7.12
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Application Runtime, Monitoring, Security
          certified: "false"
          containerImage: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
          createdAt: "2024-11-27T00:00:00Z"
          description: The deployment validation operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          ignore-check.kube-linter.io/minimum-three-replicas: This deployment uses
            1 pod as currently replicating does not replicate metric data causing
            installation issues
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/app-sre/deployment-validation-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Deployment Validation Operator (DVO) checks deployments and other resources against a curated collection of best practices.
          These best practices focus mainly on ensuring that the applications are fault-tolerant. DVO reports failed validations via Prometheus metrics.
          If the best-practice check has failed, the metrics will report `1`.
        displayName: Deployment Validation Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: AllNamespaces
        - supported: false
          type: MultiNamespace
        keywords:
        - dvo
        links:
        - name: repository
          url: https://github.com/app-sre/deployment-validation-operator
        - name: containerImage
          url: registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        maintainers:
        - email: dvo-owners@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/dvo/deployment-validation-rhel8-operator@sha256:cea717212b60d860eeabb63d96e8b9c47faceaeb4c2f0fcb879d486a19c37b3d
        - registry.redhat.io/dvo/deployment-validation-operator-bundle@sha256:c04bae05ee0c24fe5bf94257b0dd2cb020f5c9b5f173ceffffdd041b6dff4ee9
        version: 0.7.12
      entries:
      - name: deployment-validation-operator.v0.7.12
        version: 0.7.12
      - name: deployment-validation-operator.v0.7.11
        version: 0.7.11
      - name: deployment-validation-operator.0.7.9
        version: 0.7.9
      - name: deployment-validation-operator.0.7.8
        version: 0.7.8
      - name: deployment-validation-operator.0.7.5
        version: 0.7.5
      name: alpha
    defaultChannel: alpha
    packageName: deployment-validation-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CephCSI Community
      provider-url: ""
    name: cephcsi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cephcsi-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:12:33Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:b023d95637934b16ec8342994edd4e3198579517b68817ce6b4cc14958746b9b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:10cfe59ee2d625ed67f3804ba7e471f10fa1e12bc4f96de5e2f453e5cdefcf89
        version: 4.17.13-rhodf
      entries:
      - name: cephcsi-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: cephcsi-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: cephcsi-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: cephcsi-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: cephcsi-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      name: stable-4.17
    - currentCSV: cephcsi-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:20:52Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CephConnection
            name: cephconnections.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfileMapping
            name: clientprofilemappings.csi.ceph.io
            version: v1alpha1
          - kind: ClientProfile
            name: clientprofiles.csi.ceph.io
            version: v1alpha1
          - kind: Driver
            name: drivers.csi.ceph.io
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.csi.ceph.io
            version: v1alpha1
        description: Operator to deploy ceph csi drivers.
        displayName: CephCSI operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: Github Repo
          url: https://github.com/ceph/ceph-csi-operator
        maintainers:
        - email: ceph-csi-team@redhat.com
          name: CephCSI Community
        maturity: alpha
        provider:
          name: CephCSI Community
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/cephcsi-operator-bundle@sha256:903e47fa88119b5507a0d773d58318236bbce53201b5e6a18a16311f704b1120
        - registry.redhat.io/odf4/cephcsi-rhel9-operator@sha256:562703a8b197f03b29d0564df329ecfda192a94b7818c6fd5f919329a5cc9bd7
        version: 4.18.11-rhodf
      entries:
      - name: cephcsi-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: cephcsi-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: cephcsi-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: cephcsi-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: cephcsi-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: cephcsi-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: cephcsi-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: cephcsi-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: cephcsi-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: cephcsi-operator
    provider:
      name: CephCSI Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pulp Community
      provider-url: https://github.com/pulp/pulp-operator
    name: pulp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pulp-operator.v1.0.0-alpha.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache_enabled": true,
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "ingress_type": "Route",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "storage_type": "File",
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true,
                    "redis_storage_class": "standard"
                  },
                  "content": {
                    "replicas": 1
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "pulp",
                  "file_storage_access_mode": "ReadWriteOnce",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "image_version": "nightly",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/",
                    "telemetry": false
                  },
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "nightly",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "nightly",
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/api/galaxy/pulp/",
                    "telemetry": false
                  },
                  "signing_scripts_configmap": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "storage_type": "File",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "instance_name": "pulp",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1alpha1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-alpha.1
          createdAt: "2021-03-26 16:57:40"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1alpha1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1alpha1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1alpha1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/pulp/pulp-operator:v1.0.0-alpha.1
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/openshift-community-operators/pulp-operator@sha256:1926e96b185443ff0c7a2861cefc2e3aded0c0493e0d7f7d78e45447e06750d4
        - quay.io/pulp/pulp-minimal:stable
        version: 1.0.0-alpha.1
      entries:
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: alpha
    - currentCSV: pulp-operator.v1.0.0-beta.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "ingress_type": "nodeport",
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "api_root": "/pulp/"
                  },
                  "worker": {
                    "replicas": 2,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "Pulp",
                "metadata": {
                  "name": "galaxy-example"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "api": {
                    "replicas": 1
                  },
                  "content": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "deployment_type": "galaxy",
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "10Gi",
                  "file_storage_storage_class": "standard",
                  "image": "quay.io/pulp/galaxy-minimal",
                  "image_version": "4.6.3",
                  "image_web": "quay.io/pulp/galaxy-web",
                  "image_web_version": "4.6.3",
                  "ingress_type": "nodeport",
                  "nodeport_port": 30000,
                  "pulp_settings": {
                    "allowed_export_paths": [
                      "/tmp"
                    ],
                    "allowed_import_paths": [
                      "/tmp"
                    ],
                    "analytics": false,
                    "ansible_api_hostname": "http://nodeport.local:30000",
                    "api_root": "/api/galaxy/pulp/",
                    "content_origin": "http://nodeport.local:30000",
                    "galaxy_feature_flags": {
                      "ai_deny_index": true,
                      "execution_environments": true
                    },
                    "token_server": "http://nodeport.local:30000/token/"
                  },
                  "signing_scripts": "signing-scripts",
                  "signing_secret": "signing-galaxy",
                  "web": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "256Mi"
                      }
                    }
                  },
                  "worker": {
                    "replicas": 1,
                    "resource_requirements": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "256Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "galaxy-example",
                  "deployment_type": "galaxy",
                  "postgres_configuration_secret": "galaxy-example-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1beta2",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "galaxy-example"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.0.0-beta.4
          createdAt: "2024-01-18T20:33:37Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1beta2
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1beta2
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1beta2
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://docs.pulpproject.org/](https://docs.pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          Note that Pulp operator works with three different types of service containers (the operator itself, the main service and the web service):

          |           | Operator | Main | Web |
          | --------- | -------- | ---- | --- |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[pulp-minimal](https://quay.io/repository/pulp/pulp-minimal?tab=tags) | [pulp-web](https://quay.io/repository/pulp/pulp-web?tab=tags) |
          | **Image** | [pulp-operator](https://quay.io/repository/pulp/pulp-operator?tab=tags) |[galaxy-minimal](https://quay.io/repository/pulp/galaxy-minimal?tab=tags) | [galaxy-web](https://quay.io/repository/pulp/galaxy-web?tab=tags) |

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://docs.pulpproject.org/pulp_operator/container/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://docs.pulpproject.org/pulp_operator/](https://docs.pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://docs.pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: beta
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:1e75649538f13cf3bb2e6131160f038a70689d6f92a110bdcec348e865040c4a
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.0.0-beta.4
        - quay.io/pulp/pulp-web:stable
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.4
        version: 1.0.0-beta.4
      entries:
      - name: pulp-operator.v1.0.0-beta.4
        version: 1.0.0-beta.4
      - name: pulp-operator.v1.0.0-beta.3
        version: 1.0.0-beta.3
      - name: pulp-operator.v1.0.0-beta.2
        version: 1.0.0-beta.2
      - name: pulp-operator.v1.0.0-beta.1
        version: 1.0.0-beta.1
      - name: pulp-operator.v1.0.0-alpha.9
        version: 1.0.0-alpha.9
      - name: pulp-operator.v1.0.0-alpha.8
        version: 1.0.0-alpha.8
      - name: pulp-operator.v1.0.0-alpha.7
        version: 1.0.0-alpha.7
      - name: pulp-operator.v1.0.0-alpha.6
        version: 1.0.0-alpha.6
      - name: pulp-operator.v1.0.0-alpha.5
        version: 1.0.0-alpha.5
      - name: pulp-operator.v1.0.0-alpha.4
        version: 1.0.0-alpha.4
      - name: pulp-operator.v1.0.0-alpha.3
        version: 1.0.0-alpha.3
      - name: pulp-operator.v1.0.0-alpha.2
        version: 1.0.0-alpha.2
      - name: pulp-operator.v1.0.0-alpha.1
        version: 1.0.0-alpha.1
      name: beta
    - currentCSV: pulp-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "Pulp",
                "metadata": {
                  "name": "example-pulp"
                },
                "spec": {
                  "api": {
                    "replicas": 1
                  },
                  "cache": {
                    "enabled": true
                  },
                  "content": {
                    "replicas": 1
                  },
                  "custom_pulp_settings": "settings",
                  "database": {
                    "postgres_storage_class": "standard"
                  },
                  "file_storage_access_mode": "ReadWriteMany",
                  "file_storage_size": "2Gi",
                  "file_storage_storage_class": "standard",
                  "ingress_type": "nodeport",
                  "web": {
                    "replicas": 1
                  },
                  "worker": {
                    "replicas": 1
                  }
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpBackup",
                "metadata": {
                  "name": "pulpbackup-sample"
                },
                "spec": {
                  "admin_password_secret": "example-pulp-admin-password",
                  "backup_storage_class": "standard",
                  "deployment_name": "example-pulp",
                  "postgres_configuration_secret": "example-pulp-postgres-configuration"
                }
              },
              {
                "apiVersion": "repo-manager.pulpproject.org/v1",
                "kind": "PulpRestore",
                "metadata": {
                  "name": "ci-pulprestore"
                },
                "spec": {
                  "backup_name": "pulpbackup-sample",
                  "deployment_name": "example-pulp"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: quay.io/pulp/pulp-operator:v1.1.0
          createdAt: "2025-09-18T18:00:01Z"
          description: Pulp is a platform for managing repositories of software packages
            and making them available to a large number of consumers.
          metadata.annotations.support: Community
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pulp/pulp-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PulpBackup is the Schema for the pulpbackups API
            displayName: Pulp Backup
            kind: PulpBackup
            name: pulpbackups.repo-manager.pulpproject.org
            version: v1
          - description: PulpRestore is the Schema for the pulprestores API
            displayName: Pulp Restore
            kind: PulpRestore
            name: pulprestores.repo-manager.pulpproject.org
            version: v1
          - description: Pulp is the Schema for the pulps API
            displayName: Pulp
            kind: Pulp
            name: pulps.repo-manager.pulpproject.org
            version: v1
        description: |-
          [Pulp](https://pulpproject.org/) is a platform for managing repositories of content, such as software packages, and making them available to a large number of consumers.

          With Pulp you can:

          * Locally mirror all or part of a repository
          * Host your own content in a new repository
          * Manage content from multiple sources in one place
          * Promote content through different repos in an organized way

          If you have dozens, hundreds, or thousands of software packages and need a better way to manage them, Pulp can help.

          Pulp is completely free and open-source!

          * License: GPLv2+
          * Documentation: [https://pulpproject.org/](https://pulpproject.org/)
          * Source: [https://github.com/pulp/pulpcore/](https://github.com/pulp/pulpcore/)

          For more information, check out the project website: [https://pulpproject.org](https://pulpproject.org)

          If you want to evaluate Pulp quickly, try [Pulp in One Container](https://pulpproject.org/pulp-in-one-container/)

          ## Pulp Operator

          Pulp Operator is under active development, with the goal to provide a scalable and robust cluster for Pulp 3.

          <br>Pulp operator is manually built and [hosted on quay.io](https://quay.io/repository/pulp/pulp-operator). Read more about the container images [here](https://pulpproject.org/pulp-oci-images/docs/admin/reference/available-images/single-process-images/).

          ## Custom Resource Definitions
          Pulp Operator currently provides three different kinds of [Custom Resources](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#custom-resources): Pulp, Pulp Backup and Pulp Restore.
          ### Pulp
          Manages the Pulp application and its deployments, services, etc.
          ### Pulp Backup
          Manages pulp backup.

          ### Pulp Restore
          Manages the restoration of a pulp backup.

          ## Get Help

          Documentation: [https://pulpproject.org/pulp_operator/](https://pulpproject.org/pulp_operator/)

          Issue Tracker: [https://github.com/pulp/pulp-operator/issues](https://github.com/pulp/pulp-operator/issues)

          Forum: [https://discourse.pulpproject.org/](https://discourse.pulpproject.org/)

          Join [**#pulp** on Matrix](https://matrix.to/#/#pulp:matrix.org)

          Join [**#pulp-dev** on Matrix](https://matrix.to/#/#pulp-dev:matrix.org) for Developer discussion.
        displayName: Pulp Project
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - pulp
        - pulpcore
        - content
        - repositories
        - lifecycle
        links:
        - name: Pulp Operator
          url: https://github.com/pulp/pulp-operator
        - name: Pulp Project
          url: https://pulpproject.org/
        - name: Docs
          url: https://pulpproject.org/pulp_operator/
        maintainers:
        - email: pulp-list@redhat.com
          name: Pulp Community
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: Pulp Community
          url: https://github.com/pulp/pulp-operator
        relatedImages:
        - docker.io/library/postgres:13
        - docker.io/library/redis:latest
        - quay.io/community-operator-pipeline-prod/pulp-operator@sha256:92a3e95cb3b1e50100933ea08cb44074fc505f6c0e7fa24d028ec46fa9e95590
        - quay.io/pulp/pulp-minimal:stable
        - quay.io/pulp/pulp-operator:v1.1.0
        - quay.io/pulp/pulp-web:stable
        version: 1.1.0
      entries:
      - name: pulp-operator.v1.1.0
        version: 1.1.0
      - name: pulp-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pulp-operator
    provider:
      name: Pulp Community
      url: https://github.com/pulp/pulp-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: clusterresourceoverride
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterresourceoverride-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.autoscaling.openshift.io/v1",
                "kind": "ClusterResourceOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "podResourceOverride": {
                    "spec": {
                      "memoryRequestToLimitPercent": 50,
                      "cpuRequestToLimitPercent": 25,
                      "limitCPUToMemoryPercent": 200
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certifiedLevel: Primed
          containerImage: registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
          createdAt: 2019/11/15
          description: An operator to manage the OpenShift ClusterResourceOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          healthIndex: B
          olm.skipRange: '>=4.3.0 <4.18.0-202509100344'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-resource-override-admission-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of ClusterResourceOverride Admission
              Webhook
            displayName: ClusterResourceOverride
            kind: ClusterResourceOverride
            name: clusterresourceoverrides.operator.autoscaling.openshift.io
            version: v1
        description: "ClusterResourceOverride\n==============\n\nContainers can specify
          compute resource requests and limits. Requests are used for scheduling your
          container and provide a minimum service guarantee. Limits constrain the
          amount of compute resource that may be consumed on your node.\n\nThe scheduler
          attempts to optimize the compute resource use across all nodes in your cluster.
          It places pods onto specific nodes, taking the pods' compute resource requests
          and nodes' available capacity into consideration.\n\nRequests and limits
          enable administrators to allow and manage the overcommitment of resources
          on a node, which may be desirable in development environments where a trade
          off of guaranteed performance for capacity is acceptable.\n\n### Requests
          and Limits\n\nFor each compute resource, a container may specify a resource
          request and limit. Scheduling decisions are made based on the request to
          ensure that a node has enough capacity available to meet the requested value.
          If a container specifies limits, but omits requests, the requests are defaulted
          to the limits. A container is not able to exceed the specified limit on
          the node.\n\nThe enforcement of limits is dependent upon the compute resource
          type. If a container makes no request or limit, the container is scheduled
          to a node with no resource guarantees. In practice, the container is able
          to consume as much of the specified resource as is available with the lowest
          local priority. In low resource situations, containers that specify no resource
          requests are given the lowest quality of service.\n\n### Compute Resources\n\nThe
          node-enforced behavior for compute resources is specific to the resource
          type.\n\n#### CPU\n\nA container is guaranteed the amount of CPU it requests
          and is additionally able to consume excess CPU available on the node, up
          to any limit specified by the container. If multiple containers are attempting
          to use excess CPU, CPU time is distributed based on the amount of CPU requested
          by each container.\n\nFor example, if one container requested 500m of CPU
          time and another container requested 250m of CPU time, then any extra CPU
          time available on the node is distributed among the containers in a 2:1
          ratio. If a container specified a limit, it will be throttled not to use
          more CPU than the specified limit.\n\nCPU requests are enforced using the
          CFS shares support in the Linux kernel. By default, CPU limits are enforced
          using the CFS quota support in the Linux kernel over a 100ms measuring interval,
          though this can be disabled.\n\n#### Memory\n\nA container is guaranteed
          the amount of memory it requests. A container may use more memory than requested,
          but once it exceeds its requested amount, it could be killed in a low memory
          situation on the node.\n\nIf a container uses less memory than requested,
          it will not be killed unless system tasks or daemons need more memory than
          was accounted for in the node’s resource reservation. If a container specifies
          a limit on memory, it is immediately killed if it exceeds the limit amount.\n\n###
          Configuring the Cluster for Overcommitment\n\nScheduling is based on resources
          requested, while quota and hard limits refer to resource limits, which can
          be set higher than requested resources. The difference between request and
          limit determines the level of overcommit; for instance, if a container is
          given a memory request of 1Gi and a memory limit of 2Gi, it is scheduled
          based on the 1Gi request being available on the node, but could use up to
          2Gi; so it is 200% overcommitted. If OpenShift Container Platform administrators
          would like to control the level of overcommit and manage container density
          on nodes, ClusterResourceOverride Admission Webhook can be configured to
          override the ratio between request and limit set on developer containers.
          In conjunction with a per-project LimitRange specifying limits and defaults,
          this adjusts the container limit and request to achieve the desired level
          of overcommit.\n\nThis requires creating a custom resource of `ClusterResourceOverride`
          type as in the following example:\n\n    \n    - apiVersion: operator.autoscaling.openshift.io/v1\n
          \   - kind: ClusterResourceOverride\n    - metadata:\n    -   name: cluster\n
          \   - spec:\n    -   podResourceOverride:\n    -     spec:\n    -       memoryRequestToLimitPercent:
          25\n    -       cpuRequestToLimitPercent: 25\n    -       limitCPUToMemoryPercent:
          200\n    \n **memoryRequestToLimitPercent**: (optional, 1-100) If a container
          memory limit has been specified or defaulted, the memory request is overridden
          to this percentage of the limit.\n\n **cpuRequestToLimitPercent**: (optional,
          1-100) If a container CPU limit has been specified or defaulted, the CPU
          request is overridden to this percentage of the limit.\n\n **limitCPUToMemoryPercent**:
          (optional, positive integer) If a container memory limit has been specified
          or defaulted, the CPU limit is overridden to a percentage of the memory
          limit, with a 100 percentage scaling 1Gi of RAM to equal 1 CPU core. This
          is processed prior to overriding CPU request (if configured).\n\n Note that
          these overrides have no effect if no limits have been set on containers.
          [Create a LimitRange object] (https://docs.openshift.com/container-platform/3.3/admin_guide/limits.html#admin-guide-limits)
          with default limits (per individual project, or in the [project template](https://docs.openshift.com/container-platform/3.3/admin_guide/managing_projects.html#modifying-the-template-for-new-projects))
          in order to ensure that the overrides apply.\n\nWhen configured, overrides
          can be enabled per-project by applying the following label.\n    ```\n    clusterresourceoverrides.admission.autoscaling.openshift.io/enabled:
          \"true\"\n    ```\n\n"
        displayName: ClusterResourceOverride Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - deschedule
        - scale
        - binpack
        - efficiency
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9@sha256:fbf847c41e1736fb6efab7a6acdc03acc562fa7e667306773880b277883bc4e5
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-operator-bundle@sha256:7c808f6e4ea469d88703300bdcc6bdce160634833cbd10bd110dc69d05be588c
        - registry.redhat.io/openshift4/ose-clusterresourceoverride-rhel9-operator@sha256:8ed208ccfb5cf300cdb63743ca3ca652672723a815ce22a9e27440397ea65265
        version: 4.18.0-202509100344
      entries:
      - name: clusterresourceoverride-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: clusterresourceoverride-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: clusterresourceoverride-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: clusterresourceoverride-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: clusterresourceoverride-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: clusterresourceoverride-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: clusterresourceoverride-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: clusterresourceoverride-operator.v4.18.0-202505070803
        version: 4.18.0-202505070803
      - name: clusterresourceoverride-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: clusterresourceoverride-operator.v4.18.0-202504251302
        version: 4.18.0-202504251302
      - name: clusterresourceoverride-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: clusterresourceoverride-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: clusterresourceoverride-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: clusterresourceoverride-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: clusterresourceoverride-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: clusterresourceoverride-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: clusterresourceoverride-operator.v4.18.0-202501231202
        version: 4.18.0-202501231202
      name: stable
    defaultChannel: stable
    packageName: clusterresourceoverride
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: service-registry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-registry-operator.v2.0.7-0.1662549961.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
          createdAt: "2022-06-24"
          description: Deploy and manage Red Hat Integration service registry on OpenShift.
          olm.substitutesFor: service-registry-operator.v2.0.7-0.1661784420.p
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat Integration Service
              Registry instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Service Registry
          Service Registry stores and retrieves API designs and event schemas, and gives you control of their evolution.
          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf - Provides a REST API and web UI to manage the artifacts and artifact meta-data - Includes Serializers and Deserializers for Kafka client integration - Configurable rules to control schema validity and evolution (compatibility) - Storage options: Kafka, PostgreSQL - Compatible with Confluent and IBM APIs - Runs on a lightweight Quarkus platform - Includes Maven plugin to integrate with Maven based builds
          ## Service Registry Operator
          Provides a quick and easy way to deploy and manage service registry on OpenShift.
          **Features**
          - Supports installation and configuration of the Registry for all storage options - Easily perform a rolling upgrade of the Registry
          ## Prerequisites
          This operator does not deploy storage for the Registry. Therefore, some storage options require that the chosen persistence service is already set up.
          ## License
          Red Hat Service Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - Service Registry Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        links:
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/service-registry-rhel8-operator-metadata@sha256:a2f0513cd8b96d71855464384cad7f1cd8c39f8d445be1ff1c1b91f22008d16d
        - registry.redhat.io/integration/service-registry-rhel8-operator@sha256:6dfd539951a739ec06fb6620b79c42c6159d4ee148c0d9d336547e786cd4780c
        - registry.redhat.io/integration/service-registry-sql-rhel8@sha256:6c97fa8ead1f6f0323811f8f35a80ae33f0e160f6c58f2bdb31713dbeaa3f014
        - registry.redhat.io/integration/service-registry-kafkasql-rhel8@sha256:fc97a2a08b2f77d620ea86ec77bb4f750bfc0109ede3ad1080c2cb694019f8b8
        version: 2.0.7+0.1662549961.p
      entries:
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.0.x
    - currentCSV: service-registry-operator.v2.6.13-r2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-kafkasql"
                },
                "spec": {
                  "configuration": {
                    "kafkasql": {
                      "bootstrapServers": "\u003cservice name\u003e.\u003cnamespace\u003e.svc:9092"
                    },
                    "persistence": "kafkasql"
                  }
                }
              },
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry",
                "metadata": {
                  "name": "example-apicurioregistry-sql"
                },
                "spec": {
                  "configuration": {
                    "persistence": "sql",
                    "sql": {
                      "dataSource": {
                        "password": "\u003cpassword\u003e",
                        "url": "jdbc:postgresql://\u003cservice name\u003e.\u003cnamespace\u003e.svc:5432/\u003cdatabase name\u003e",
                        "userName": "postgres"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
          createdAt: "2025-09-15"
          description: Deploy and manage Red Hat build of Apicurio Registry on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Apicurio/apicurio-registry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry represents a Red Hat build of Apicurio Registry
              instance
            displayName: Apicurio Registry
            kind: ApicurioRegistry
            name: apicurioregistries.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry

          Red Hat build of Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Provides a REST API and web UI to manage the artifacts and artifact meta-data
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL
          - Compatible with Confluent and IBM APIs
          - Runs on a lightweight Quarkus platform
          - Includes Maven plugin to integrate with Maven based builds

          ## Red Hat build of Apicurio Registry Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry on OpenShift.

          **Features**
          - Supports installation and configuration of the Registry for all storage options
          - Easily perform a rolling upgrade of Red Hat build of Apicurio Registry

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry Operator is licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry-operator/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - service
        - service-registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        - name: Twitter
          url: https://twitter.com/Apicurio
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel8-operator@sha256:61f00395b51819312dda7abcdaa47190a11a83b1e2b523101be0ba776b1c5b7c
        - registry.redhat.io/apicurio/apicurio-registry-sql-rhel8@sha256:a55425a9af1a16959ab5e528a5740d47ea06f6935593cc925e7727cded4e7d13
        - registry.redhat.io/apicurio/apicurio-registry-kafkasql-rhel8@sha256:f2da7cd92e126187d91043e2ae40521f8574586f8b3975be27d8370fe31b6732
        - registry.redhat.io/apicurio/apicurio-registry-operator-bundle@sha256:c2229c2cfad9bbbca9179cb6617bce0033f5be40731641a69b41f9a90e034c8b
        version: 2.6.13-r2
      entries:
      - name: service-registry-operator.v2.6.13-r2
        version: 2.6.13-r2
      - name: service-registry-operator.v2.6.13-r1
        version: 2.6.13-r1
      - name: service-registry-operator.v2.6.12-r1
        version: 2.6.12-r1
      - name: service-registry-operator.v2.6.11-r1
        version: 2.6.11-r1
      - name: service-registry-operator.v2.6.10
        version: 2.6.10
      - name: service-registry-operator.v2.6.9
        version: 2.6.9
      - name: service-registry-operator.v2.6.8-0.1740477005.p
        version: 2.6.8+0.1740477005.p
      - name: service-registry-operator.v2.6.8-0.1739457262.p
        version: 2.6.8+0.1739457262.p
      - name: service-registry-operator.v2.6.8
        version: 2.6.8
      - name: service-registry-operator.v2.6.6-0.1736942038.p
        version: 2.6.6+0.1736942038.p
      - name: service-registry-operator.v2.6.6
        version: 2.6.6
      - name: service-registry-operator.v2.6.5-0.1731946870.p
        version: 2.6.5+0.1731946870.p
      - name: service-registry-operator.v2.6.5
        version: 2.6.5
      - name: service-registry-operator.v2.6.5-0.1729511936.p
        version: 2.6.5+0.1729511936.p
      - name: service-registry-operator.v2.6.5-0.1730982111.p
        version: 2.6.5+0.1730982111.p
      - name: service-registry-operator.v2.6.3
        version: 2.6.3
      - name: service-registry-operator.v2.6.1-0.1724927626.p
        version: 2.6.1+0.1724927626.p
      - name: service-registry-operator.v2.6.1
        version: 2.6.1
      - name: service-registry-operator.v2.3.5
        version: 2.3.5
      - name: service-registry-operator.v2.3.4-0.1716987426.p
        version: 2.3.4+0.1716987426.p
      - name: service-registry-operator.v2.3.4
        version: 2.3.4
      - name: service-registry-operator.v2.3.3
        version: 2.3.3
      - name: service-registry-operator.v2.3.2
        version: 2.3.2
      - name: service-registry-operator.v2.3.1-0.1706181988.p
        version: 2.3.1+0.1706181988.p
      - name: service-registry-operator.v2.3.1
        version: 2.3.1
      - name: service-registry-operator.v2.3.0
        version: 2.3.0
      - name: service-registry-operator.v2.2.3
        version: 2.2.3
      - name: service-registry-operator.v2.2.3-0.1698369839.p
        version: 2.2.3+0.1698369839.p
      - name: service-registry-operator.v2.2.2
        version: 2.2.2
      - name: service-registry-operator.v2.2.2-0.1697116886.p
        version: 2.2.2+0.1697116886.p
      - name: service-registry-operator.v2.2.1
        version: 2.2.1
      - name: service-registry-operator.v2.2.0
        version: 2.2.0
      - name: service-registry-operator.v2.1.5
        version: 2.1.5
      - name: service-registry-operator.v2.1.4
        version: 2.1.4
      - name: service-registry-operator.v2.1.3
        version: 2.1.3
      - name: service-registry-operator.v2.1.2-0.1676389596.p
        version: 2.1.2+0.1676389596.p
      - name: service-registry-operator.v2.1.2
        version: 2.1.2
      - name: service-registry-operator.v2.1.1
        version: 2.1.1
      - name: service-registry-operator.v2.1.0
        version: 2.1.0
      - name: service-registry-operator.v2.0.7
        version: 2.0.7
      - name: service-registry-operator.v2.0.7-0.1662549961.p
        version: 2.0.7+0.1662549961.p
      - name: service-registry-operator.v2.0.7-0.1661784420.p
        version: 2.0.7+0.1661784420.p
      - name: service-registry-operator.v2.0.6
        version: 2.0.6
      - name: service-registry-operator.v2.0.5-0.1648813064.p
        version: 2.0.5+0.1648813064.p
      - name: service-registry-operator.v2.0.5-0.1646309885.p
        version: 2.0.5+0.1646309885.p
      - name: service-registry-operator.v2.0.5-0.1654503709.p
        version: 2.0.5+0.1654503709.p
      - name: service-registry-operator.v2.0.5-0.1652887585.p
        version: 2.0.5+0.1652887585.p
      - name: service-registry-operator.v2.0.5-0.1652429755.p
        version: 2.0.5+0.1652429755.p
      - name: service-registry-operator.v2.0.5-0.1651047781.p
        version: 2.0.5+0.1651047781.p
      - name: service-registry-operator.v2.0.5-0.1655396591.p
        version: 2.0.5+0.1655396591.p
      - name: service-registry-operator.v2.0.5-0.1647938417.p
        version: 2.0.5+0.1647938417.p
      - name: service-registry-operator.v2.0.5
        version: 2.0.5
      - name: service-registry-operator.v2.0.4
        version: 2.0.4
      - name: service-registry-operator.v2.0.3
        version: 2.0.3
      - name: service-registry-operator.v2.0.2
        version: 2.0.2
      - name: service-registry-operator.v2.0.1
        version: 2.0.1
      - name: service-registry-operator.v2.0.0
        version: 2.0.0
      name: 2.x
    defaultChannel: 2.x
    packageName: service-registry-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hazelcast, Inc
      provider-url: ""
    name: hazelcast-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hazelcast-platform-operator.v5.15.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Cache",
                "metadata": {
                  "name": "cache-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "CronHotBackup",
                "metadata": {
                  "name": "cronhotbackup-sample"
                },
                "spec": {
                  "hotBackupTemplate": {
                    "metadata": {
                      "labels": {
                        "cron-hotbackup": "true"
                      }
                    },
                    "spec": {
                      "hazelcastResourceName": "hazelcast"
                    }
                  },
                  "schedule": "*/30 * * * *"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Flow",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "flow"
                },
                "spec": {
                  "database": {
                    "host": "flow-postgresql",
                    "secretName": "flow-db-secret"
                  },
                  "env": [
                    {
                      "name": "OPTIONS",
                      "value": "--flow.analytics.persistRemoteCallResponses=true --flow.analytics.persistResults=false --flow.config.custom.managementCenterUrl=http://34.140.160.158/mc\n"
                    }
                  ],
                  "externalConnectivity": {
                    "ingress": {
                      "annotations": {
                        "kubernetes.io/ingress.class": "nginx",
                        "nginx.ingress.kubernetes.io/enable-opentelemetry": "true",
                        "nginx.ingress.kubernetes.io/ssl-redirect": "false"
                      },
                      "hostname": "",
                      "ingressClassName": "nginx",
                      "path": "/"
                    }
                  },
                  "imagePullSecrets": [
                    {
                      "name": "hazelcast-cloud-hazelcast-connect-demo-pull-secret"
                    }
                  ],
                  "licenseKeySecretName": "flow-mc-license-key",
                  "repository": "quay.io/hazelcast_cloud/hazelcast-flow-internal",
                  "size": 3,
                  "version": "next"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast"
                },
                "spec": {
                  "clusterSize": 3,
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Hazelcast",
                "metadata": {
                  "name": "hazelcast-sample"
                },
                "spec": {
                  "clusterSize": 3,
                  "jet": {
                    "bucketConfig": {
                      "bucketURI": "gs://operator-user-code/jetJobs",
                      "secretName": "br-secret-gcp"
                    },
                    "enabled": true,
                    "resourceUploadEnabled": true
                  },
                  "licenseKeySecretName": "hazelcast-license-key",
                  "repository": "docker.io/hazelcast/hazelcast-enterprise"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HazelcastEndpoint",
                "metadata": {
                  "name": "hazelcastendpoint-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "port": 5701,
                  "type": "Discovery"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "HotBackup",
                "metadata": {
                  "name": "hot-backup"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-longrun"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-longrun-2.0.0.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJob",
                "metadata": {
                  "name": "jet-job-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "jarName": "jet-pipeline-1.0.2.jar",
                  "state": "Running"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "JetJobSnapshot",
                "metadata": {
                  "name": "jetjobsnapshot-sample"
                },
                "spec": {
                  "cancelJob": false,
                  "jetJobResourceName": "jet-job"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ManagementCenter",
                "metadata": {
                  "name": "managementcenter"
                },
                "spec": {
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "hazelcastClusters": [
                    {
                      "address": "hazelcast",
                      "name": "dev"
                    }
                  ],
                  "licenseKeySecretName": "hazelcast-license-key",
                  "persistence": {
                    "enabled": true,
                    "size": "10Gi"
                  },
                  "repository": "hazelcast/management-center"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Map",
                "metadata": {
                  "name": "map"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "MultiMap",
                "metadata": {
                  "name": "multimap-sample"
                },
                "spec": {
                  "backupCount": 3,
                  "binary": true,
                  "collectionType": "LIST",
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "queue-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "ReplicatedMap",
                "metadata": {
                  "name": "replicatedmap-sample"
                },
                "spec": {
                  "asyncFillup": true,
                  "hazelcastResourceName": "hazelcast",
                  "inMemoryFormat": "OBJECT"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "topic-sample"
                },
                "spec": {
                  "globalOrderingEnabled": true,
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "UserCodeNamespace",
                "metadata": {
                  "name": "ucn-sample"
                },
                "spec": {
                  "bucketConfig": {
                    "bucketURI": "gs://operator-user-code/userCodeNamespace",
                    "secretName": "br-secret-gcp"
                  },
                  "hazelcastResourceName": "hazelcast"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "VectorCollection",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hazelcast-platform-operator"
                  },
                  "name": "vectorcollection-sample"
                },
                "spec": {
                  "hazelcastResourceName": "hazelcast",
                  "indexes": [
                    {
                      "dimension": 10,
                      "efConstruction": 256,
                      "maxDegree": 32,
                      "metric": "Dot",
                      "useDeduplication": false
                    }
                  ]
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanReplication",
                "metadata": {
                  "name": "wanreplication-sample"
                },
                "spec": {
                  "endpoints": "35.192.33.252",
                  "resources": [
                    {
                      "kind": "Map",
                      "name": "map-sample"
                    },
                    {
                      "kind": "Hazelcast",
                      "name": "hz-sample"
                    }
                  ],
                  "targetClusterName": "dev"
                }
              },
              {
                "apiVersion": "hazelcast.com/v1alpha1",
                "kind": "WanSync",
                "metadata": {
                  "name": "wansync-sample"
                },
                "spec": {
                  "wanReplicationResourceName": "wan-replication-sample"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
          createdAt: "2025-03-13T11:59:59Z"
          description: Install Hazelcast clusters in Kubernetes environments.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hazelcast/hazelcast-platform-operator/
          support: Hazelcast, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cache is the Schema for the caches API
            displayName: Cache
            kind: Cache
            name: caches.hazelcast.com
            version: v1alpha1
          - description: CronHotBackup is the Schema for the cronhotbackups API
            displayName: Cron Hot Backup
            kind: CronHotBackup
            name: cronhotbackups.hazelcast.com
            version: v1alpha1
          - description: Flow is the Schema for the flows API
            displayName: Flow
            kind: Flow
            name: flows.hazelcast.com
            version: v1alpha1
          - description: HazelcastEndpoint is the Schema for the hazelcastendpoints
              API
            displayName: Hazelcast Endpoint
            kind: HazelcastEndpoint
            name: hazelcastendpoints.hazelcast.com
            version: v1alpha1
          - description: Hazelcast is the Schema for the hazelcasts API
            displayName: Hazelcast
            kind: Hazelcast
            name: hazelcasts.hazelcast.com
            version: v1alpha1
          - description: HotBackup is the Schema for the hot backup API
            displayName: Hot Backup
            kind: HotBackup
            name: hotbackups.hazelcast.com
            version: v1alpha1
          - description: JetJob is the Schema for the jetjobs API
            displayName: Jet Job
            kind: JetJob
            name: jetjobs.hazelcast.com
            version: v1alpha1
          - description: JetJobSnapshot is the Schema for the jetjobsnapshots API
            displayName: Jet Job Snapshot
            kind: JetJobSnapshot
            name: jetjobsnapshots.hazelcast.com
            version: v1alpha1
          - description: ManagementCenter is the Schema for the managementcenters
              API
            displayName: Management Center
            kind: ManagementCenter
            name: managementcenters.hazelcast.com
            version: v1alpha1
          - description: Map is the Schema for the maps API
            displayName: Map
            kind: Map
            name: maps.hazelcast.com
            version: v1alpha1
          - description: MultiMap is the Schema for the multimaps API
            displayName: Multi Map
            kind: MultiMap
            name: multimaps.hazelcast.com
            version: v1alpha1
          - description: Queue is the Schema for the queues API
            displayName: Queue
            kind: Queue
            name: queues.hazelcast.com
            version: v1alpha1
          - description: ReplicatedMap is the Schema for the replicatedmaps API
            displayName: Replicated Map
            kind: ReplicatedMap
            name: replicatedmaps.hazelcast.com
            version: v1alpha1
          - description: Topic is the Schema for the topics API
            displayName: Topic
            kind: Topic
            name: topics.hazelcast.com
            version: v1alpha1
          - description: UserCodeNamespace is the Schema for the usercodenamespaces
              API
            displayName: User Code Namespace
            kind: UserCodeNamespace
            name: usercodenamespaces.hazelcast.com
            version: v1alpha1
          - description: VectorCollection is the Schema for the vectorcollections
              API
            displayName: Vector Collection
            kind: VectorCollection
            name: vectorcollections.hazelcast.com
            version: v1alpha1
          - description: WanReplication is the Schema for the wanreplications API
            displayName: Wan Replication
            kind: WanReplication
            name: wanreplications.hazelcast.com
            version: v1alpha1
          - description: WanSync is the Schema for the wansyncs API
            displayName: Wan Sync
            kind: WanSync
            name: wansyncs.hazelcast.com
            version: v1alpha1
        description: |
          Easily deploy Hazelcast clusters and Management Center into Kubernetes environments and manage their lifecycles.

          ## Before You Start

          If you are planning to create Hazelcast Platform Enterprise clusters, you need to [create a secret](https://docs.hazelcast.com/operator/latest/get-started#step-2-start-the-hazelcast-cluster) for the license. You can request a trial license key from [here](https://trialrequest.hazelcast.com).

          For Hazelcast Platform clusters, you can simply continue.

          ## Documentation

          1. [Get started](https://docs.hazelcast.com/operator/latest/get-started) with the Operator
          2. [Connect to the cluster from outside Kubernetes](https://guides.hazelcast.org/hazelcast-platform-operator-expose-externally)
          3. [Restore a Cluster from Cloud Storage with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-external-backup-restore)
          4. [Replicate Data between Two Hazelcast Clusters with Hazelcast Platform Operator](https://docs.hazelcast.com/tutorials/hazelcast-platform-operator-wan-replication)

          ## Features

          Hazelcast Platform Operator supports the features below:

          * Custom resource for Hazelcast Platform (Enterprise) and Management Center
          * Observe status of Hazelcast clusters and Management Center instances
          * High Availability Mode configuration to create clusters that are resilient to node and zone failures
          * Support for TLS connections between members using self-signed certificates
          * Scale up and down Hazelcast clusters
          * Expose Hazelcast cluster to external clients ([Smart & Unisocket](https://docs.hazelcast.com/hazelcast/latest/clients/java#java-client-operation-modes))
          * Backup Hazelcast persistence data to cloud storage with the possibility of scheduling it and restoring the data accordingly
          * WAN Replication feature when you need to synchronize multiple Hazelcast clusters, which are connected by WANs
          * Full and Delta WAN Sync support
          * Tiered Storage support for Hazelcast and Map CRs
          * CP Subsystem configuration support for Hazelcast CR
          * User Code Deployment feature, which allows you to deploy custom and domain classes from cloud storages and URLs to Hazelcast members
          * The User Code Namespaces feature allows you to deploy custom and domain classes at runtime without a members restart
          * Support the configuration of advanced networking options
          * Support Multi-namespace configuration
          * ExecutorService and EntryProcessor support
          * Support several data structures like Map, Topic, MultiMap, ReplicatedMap, Queue and Cache which can be created dynamically via specific Custom Resources
          * MapStore, Near Cache and off-heap memory (HD memory and native memory) support for the Map CR
          * Native Memory support for the Cache CR
          * Support Jet configuration and Jet Job submission using the JetJob CR
          * Support for exporting the snapshots of JetJob CRs using JetJobSnapshot CR
          * Support for custom configurations using ConfigMap
        displayName: Hazelcast Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hazelcast
        - keyvalue
        - in-memory
        - database
        - caching
        links:
        - name: Documentation
          url: https://docs.hazelcast.com/operator/latest/
        - name: Source Code
          url: https://github.com/hazelcast/hazelcast-platform-operator
        maintainers:
        - email: cloudnative@hazelcast.com
          name: Hazelcast Cloud Native Team
        maturity: alpha
        minKubeVersion: 1.22.7
        provider:
          name: Hazelcast, Inc
        relatedImages:
        - docker.io/hazelcast/hazelcast-platform-operator@sha256:59328a4ea7a4e2110e0750364db792cf266f1e7231e3bb321e3de09d535dab23
        - registry.connect.redhat.com/hazelcast/hazelcast-platform-operator-bundle@sha256:0aabd3e7f5f993c1288450678d55e8f7b36c8b30894e5d844521a474aa6b2714
        version: 5.15.0
      entries:
      - name: hazelcast-platform-operator.v5.15.0
        version: 5.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hazelcast-platform-operator
    provider:
      name: Hazelcast, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudtrail-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudtrail-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "EventDataStore",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudtrail.services.k8s.aws/v1alpha1",
                "kind": "Trail",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
          createdAt: "2025-09-19T21:16:25Z"
          description: AWS CloudTrail controller is a service controller for managing
            CloudTrail resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventDataStore represents the state of an AWS cloudtrail
              EventDataStore resource.
            displayName: EventDataStore
            kind: EventDataStore
            name: eventdatastores.cloudtrail.services.k8s.aws
            version: v1alpha1
          - description: Trail represents the state of an AWS cloudtrail Trail resource.
            displayName: Trail
            kind: Trail
            name: trails.cloudtrail.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (CloudTrail) resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudTrail**

          AWS CloudTrail is an AWS service that helps you enable operational and risk auditing, governance, and compliance of your AWS account. Actions taken by a user, role, or an AWS service are recorded as events in CloudTrail. Events include actions taken in the AWS Management Console, AWS Command Line Interface, and AWS SDKs and APIs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudTrail
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudtrail
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudTrail Developer Resources
          url: https://aws.amazon.com/cloudtrail/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudtrail maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudtrail-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-cloudtrail-controller@sha256:20fe3c3ba7688906de40d7fd36c00552be280595708935ff800ad5228829f781
        version: 1.1.1
      entries:
      - name: ack-cloudtrail-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudtrail-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudtrail-controller.v1.0.28
        version: 1.0.28
      - name: ack-cloudtrail-controller.v1.0.26
        version: 1.0.26
      - name: ack-cloudtrail-controller.v1.0.25
        version: 1.0.25
      - name: ack-cloudtrail-controller.v1.0.24
        version: 1.0.24
      - name: ack-cloudtrail-controller.v1.0.23
        version: 1.0.23
      - name: ack-cloudtrail-controller.v1.0.22
        version: 1.0.22
      - name: ack-cloudtrail-controller.v1.0.21
        version: 1.0.21
      - name: ack-cloudtrail-controller.v1.0.20
        version: 1.0.20
      - name: ack-cloudtrail-controller.v1.0.19
        version: 1.0.19
      - name: ack-cloudtrail-controller.v1.0.16
        version: 1.0.16
      - name: ack-cloudtrail-controller.v1.0.15
        version: 1.0.15
      - name: ack-cloudtrail-controller.v1.0.14
        version: 1.0.14
      - name: ack-cloudtrail-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudtrail-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudtrail-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudtrail-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudtrail-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudtrail-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudtrail-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudtrail-controller.v1.0.3
        version: 1.0.3
      - name: ack-cloudtrail-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudtrail-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudtrail-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudtrail-controller.v0.1.3
        version: 0.1.3
      - name: ack-cloudtrail-controller.v0.1.2
        version: 0.1.2
      - name: ack-cloudtrail-controller.v0.1.1
        version: 0.1.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudtrail-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: smart-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: smart-gateway-operator.v5.0.1757650606
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "smartgateway.infra.watch/v2",
                "kind": "SmartGateway",
                "metadata": {
                  "name": "demo"
                },
                "spec": {
                  "applications": [
                    {
                      "config": "host: localhost\nport: 8081\nwithtimestamp: false\n",
                      "name": "prometheus"
                    }
                  ],
                  "blockEventBus": false,
                  "bridge": {
                    "amqpBlock": false,
                    "amqpUrl": "amqp://amq-interconnect:5672/collectd",
                    "enabled": true,
                    "ringBufferCount": 15000,
                    "ringBufferSize": 2048,
                    "socketBlock": true,
                    "statsPeriod": 60,
                    "stopCount": 0,
                    "unixSocketPath": "/tmp",
                    "verbose": true
                  },
                  "handleErrors": false,
                  "logLevel": "info",
                  "pluginDir": "/usr/lib64/sg-core",
                  "services": [
                    {
                      "name": "prometheus",
                      "ports": [
                        {
                          "name": "prom-http",
                          "port": 8081,
                          "protocol": "TCP",
                          "targetPort": 8081
                        }
                      ]
                    }
                  ],
                  "size": 1,
                  "transports": [
                    {
                      "config": "path: /tmp/smartgateway\n",
                      "handlers": [
                        {
                          "config": "",
                          "name": "collectd-metrics"
                        }
                      ],
                      "name": "socket"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
          createdAt: "2025-09-12T04:17:59Z"
          description: Operator for managing the Smart Gateway Custom Resources, resulting
            in deployments of the Smart Gateway.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenStack Platform", "Cloud
            Infrastructure", "Cloud Suite"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/infrawatch/smart-gateway-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of the Smart Gateway
            displayName: SmartGateway
            kind: SmartGateway
            name: smartgateways.smartgateway.infra.watch
            version: v2
        description: |-
          # Smart Gateway for Service Telemetry Framework

          The Smart Gateway is middleware for OpenShift that takes metrics and events
          data streams from an AMQP 1.x message bus, and provides a method to having
          that data stream stored within supported data stores such as Prometheus and
          ElasticSearch.

          ## Core capabilities

          The Smart Gateway provides two modes:

          * metrics: provides an HTTP scrape endpoint for Prometheus
          * events: writes events directly to an ElasticSearch endpoint

          ## Operator features

          * **Two Service Types** - Supports the `metrics` and `events` service types which defines the type of Smart Gateway to be deployed

          * **Configuration** - Configuration of the Smart Gateway to allow for it to be connected to an AMQP 1.x bus and pointed at the appropriate storage mechanism based on service type.

          ## Before getting started

          1. Ensure an AMQP 1.x bus has been setup (such as AMQ Interconnect)
          2. If the Smart Gateway is running in events mode, be sure the ElasticSearch Operator and ElasticSearch instance have been pre-deployed.
        displayName: Smart Gateway Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - servicetelemetry
        - monitoring
        - telemetry
        - notifications
        - smartgateway
        links:
        - name: Source Code
          url: https://github.com/infrawatch/smart-gateway-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/stf/sg-bridge-rhel9@sha256:7fae342cf6cecc676eaa0a8fb64ee380ad101d667d6e8f0d89c943938bb6baf4
        - registry.redhat.io/stf/sg-core-rhel9@sha256:0e12903da0b15c64421102b5b914ca4c82d107295485c9b76ed19e91a45a4420
        - registry.redhat.io/stf/smart-gateway-rhel9-operator@sha256:3363abcb91319f8642212b9b80ca49bb9f90cae69255b1f75f03929af3d3a8c3
        - registry.redhat.io/stf/smart-gateway-operator-bundle@sha256:f12b2f1dd34c73b9c8a5ad461e771c0ce56c9ea4eb88f03225be1462a42c5a78
        - registry.redhat.io/openshift4/ose-oauth-proxy:latest@sha256:dc24aa382e02c4ac058d4b824a21f0e90a6259d6d817d0341b2a41809b5daa78
        version: 5.0.1757650606
      entries:
      - name: smart-gateway-operator.v5.0.1757650606
        version: 5.0.1757650606
      - name: smart-gateway-operator.v5.0.1743671602
        version: 5.0.1743671602
      name: stable-1.5
    defaultChannel: stable-1.5
    packageName: smart-gateway-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: aiu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aiu-operator.v2.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.1"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "featureDiscovery": {
                    "image": "aiu-feature-discovery",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "imagePullPolicy": "Always",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_03",
                    "version": "2.3.2"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "labels": {
                    "operator-version": "2.3.3"
                  },
                  "name": "aiunodestate"
                },
                "spec": {
                  "nodeName": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
          createdAt: "2025-04-03T01:19:38Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu/release_2025_03/aiu-operator@sha256:2ea8268e94d44581b3490cd4bbbd5525359073129bae42c7f007c05727afe1c3
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:0cda4be8f74da7524a72c961c8762927c66e145d82e59af92073a54179bc9936
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        version: 2.3.3
      entries:
      - name: aiu-operator.v2.3.3
        version: 2.3.3
      name: stable-v2.3
    - currentCSV: aiu-operator.v2.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuClusterPolicy",
                "metadata": {
                  "name": "aiuclusterpolicy"
                },
                "spec": {
                  "cardManagement": {
                    "cardmgmtClaimName": "cardmgmt",
                    "config": {
                      "aiuFilter": ".",
                      "pfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453",
                      "vfRunnerImage": "icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453"
                    },
                    "enabled": false,
                    "image": "aiucardmgmt",
                    "repository": "icr.io/ibmaiu_internal/x86_64",
                    "version": "release_2025_05"
                  },
                  "devicePlugin": {
                    "configName": "senlib_config.json",
                    "configPath": "/etc/aiu",
                    "image": "aiu-device-plugin",
                    "initContainer": {
                      "executePolicy": "IfNotPresent",
                      "image": "e2e_stable",
                      "repository": "icr.io/ibmaiu_internal/x86_64/dd2",
                      "version": "release_2025_05"
                    },
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "experimentalMode": [
                    "perDeviceAllocation",
                    "topologyAwareAllocation",
                    "externalDeviceReservation"
                  ],
                  "metricsExporter": {
                    "enabled": false,
                    "image": "aiu-exporter",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "podValidator": {
                    "enabled": true,
                    "image": "aiu-webhook-validator",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  },
                  "scheduler": {
                    "image": "aiu-kube-scheduler-ocp-4.16",
                    "repository": "icr.io/ibmaiu/release_2025_05",
                    "version": "2.4.0"
                  }
                }
              },
              {
                "apiVersion": "aiu.ibm.com/v2",
                "kind": "AiuNodeState",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by operator.\nDo not create/modify.\n"
                  },
                  "name": "example-worker-node"
                },
                "spec": {
                  "nodeName": "worker-1"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
          createdAt: "2025-05-28T05:53:45Z"
          description: operator for AIU devices
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: use with IBM AIU
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["aiunodestates.aiu.ibm.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AiuClusterPolicy is the Schema for the AIU API
            displayName: AIU Cluster Policy
            kind: AiuClusterPolicy
            name: aiuclusterpolicies.aiu.ibm.com
            version: v2
          - description: AiuNodeState is the Schema for the AiuNodeState API
            displayName: AIU Node State
            kind: AiuNodeState
            name: aiunodestates.aiu.ibm.com
            version: v2
        description: operator for AIU devices
        displayName: AIU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - app
        - operator
        links:
        - name: Aiu Operator
          url: https://aiu-operator.domain
        maintainers:
        - email: tmishina@jp.ibm.com
          name: Takuya Mishina
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: laurentiu.bradin@ibm.com
          name: Laurentiu Bradin
        - email: htchang@us.ibm.com
          name: Ted Chang
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: IBM
        relatedImages:
        - icr.io/ibmaiu_internal/x86_64/dd2/e2e_stable@sha256:beeb2677e8c7fc81dfd952942842d5dd8c39ed3fd00dc30448a0cbc070757453
        - registry.connect.redhat.com/aiu/aiu-operator-bundle@sha256:4bab095c5dd6f8c95a17a8ff2bf70ff3f53faea48b56eea8baa9cf9a42a5227a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:fcb3b8ab93dfb5ef2b290e39ea5899dbb5e0c6d430370b8d281e59e74d94d749
        - icr.io/ibmaiu/release_2025_05/aiu-operator@sha256:9796ab1c39d1dd55f8aa79c9f585be0f988b5026745184bdc890dfd1846909df
        version: 2.4.0
      entries:
      - name: aiu-operator.v2.4.0
        version: 2.4.0
      name: stable-v2.4
    defaultChannel: stable-v2.4
    packageName: aiu-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keycloak-operator.v19.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3-legacy
          createdAt: "2022-10-07 17:00:00"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Keycloak User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
          * Install Extensions
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/keycloak-operator@sha256:a4da3bcbb4e99d60b8217d9e62742eeb83eef163dfefb2f7efda131f312e46e3
        - quay.io/keycloak/keycloak-operator:19.0.3-legacy
        version: 19.0.3
      entries:
      - name: keycloak-operator.v19.0.3
        version: 19.0.3
      - name: keycloak-operator.v19.0.2
        version: 19.0.2
      - name: keycloak-operator.v19.0.1
        version: 19.0.1
      - name: keycloak-operator.v19.0.0
        version: 19.0.0
      - name: keycloak-operator.v18.0.2
        version: 18.0.2
      - name: keycloak-operator.v18.0.1
        version: 18.0.1
      - name: keycloak-operator.v18.0.0
        version: 18.0.0
      - name: keycloak-operator.v17.0.1
        version: 17.0.1
      - name: keycloak-operator.v17.0.0
        version: 17.0.0
      - name: keycloak-operator.v16.1.0
        version: 16.1.0
      - name: keycloak-operator.v16.0.0
        version: 16.0.0
      - name: keycloak-operator.v15.1.1
        version: 15.1.1
      - name: keycloak-operator.v15.1.0
        version: 15.1.0
      - name: keycloak-operator.v15.0.2
        version: 15.0.2
      - name: keycloak-operator.v15.0.1
        version: 15.0.1
      - name: keycloak-operator.v15.0.0
        version: 15.0.0
      - name: keycloak-operator.v14.0.0
        version: 14.0.0
      name: alpha
    - currentCSV: keycloak-operator.v20.0.0-alpha.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": "example.org",
                  "tlsSecret": "my-tls-secret"
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "False"
          containerImage: quay.io/keycloak/keycloak-operator:19.0.3
          createdAt: 10/07/22 17:00:00
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/docs/latest/server_installation/index.html#_operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@lists.jboss.org
          name: Keycloak DEV mailing list
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/keycloak/keycloak-operator:19.0.3
        - quay.io/openshift-community-operators/keycloak-operator@sha256:65fe00f310289b70538a6dde897ef663731d0da44a7bb1d0f55a3001a50e53fa
        version: 20.0.0-alpha.7
      entries:
      - name: keycloak-operator.v20.0.0-alpha.7
        version: 20.0.0-alpha.7
      - name: keycloak-operator.v20.0.0-alpha.6
        version: 20.0.0-alpha.6
      - name: keycloak-operator.v20.0.0-alpha.5
        version: 20.0.0-alpha.5
      - name: keycloak-operator.v20.0.0-alpha.4
        version: 20.0.0-alpha.4
      - name: keycloak-operator.v20.0.0-alpha.3
        version: 20.0.0-alpha.3
      - name: keycloak-operator.v20.0.0-alpha.2
        version: 20.0.0-alpha.2
      - name: keycloak-operator.v20.0.0-alpha.1
        version: 20.0.0-alpha.1
      name: candidate
    - currentCSV: keycloak-operator.v26.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "hostname": {
                    "hostname": "example.org"
                  },
                  "http": {
                    "tlsSecret": "my-tls-secret"
                  }
                }
              },
              {
                "apiVersion": "k8s.keycloak.org/v2alpha1",
                "kind": "KeycloakRealmImport",
                "metadata": {
                  "name": "example-keycloak-realm-import",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "keycloakCRName": "example-keycloak",
                  "realm": {}
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          certified: "false"
          containerImage: quay.io/keycloak/keycloak-operator:26.4.0
          createdAt: "2025-09-30T11:58:06Z"
          description: An Operator for installing and managing Keycloak
          repository: https://github.com/keycloak/keycloak
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Keycloak Realm Import
            displayName: KeycloakRealmImport
            kind: KeycloakRealmImport
            name: keycloakrealmimports.k8s.keycloak.org
            version: v2alpha1
          - description: Represents a Keycloak Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.k8s.keycloak.org
            version: v2alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Keycloak.

          Keycloak lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
        displayName: Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://www.keycloak.org/guides#operator
        - name: Keycloak
          url: https://www.keycloak.org/
        - name: Keycloak Discourse
          url: https://keycloak.discourse.group/
        maintainers:
        - email: keycloak-dev@googlegroups.com
          name: Keycloak DEV mailing list
        maturity: stable
        nativeApis:
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: batch
          kind: Job
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        - group: networking.k8s.io
          kind: NetworkPolicy
          version: v1
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/keycloak-operator@sha256:70bbd44f6ecf86aa8e9710175a96dd3c1d813aec40667a09800241ce30afd93f
        - quay.io/keycloak/keycloak-operator:26.4.0
        version: 26.4.0
      entries:
      - name: keycloak-operator.v26.4.0
        version: 26.4.0
      - name: keycloak-operator.v26.3.5
        version: 26.3.5
      - name: keycloak-operator.v26.3.4
        version: 26.3.4
      - name: keycloak-operator.v26.3.3
        version: 26.3.3
      - name: keycloak-operator.v26.3.2
        version: 26.3.2
      - name: keycloak-operator.v26.3.1
        version: 26.3.1
      - name: keycloak-operator.v26.3.0
        version: 26.3.0
      - name: keycloak-operator.v26.2.5
        version: 26.2.5
      - name: keycloak-operator.v26.2.4
        version: 26.2.4
      - name: keycloak-operator.v26.2.3
        version: 26.2.3
      - name: keycloak-operator.v26.2.2
        version: 26.2.2
      - name: keycloak-operator.v26.2.1
        version: 26.2.1
      - name: keycloak-operator.v26.2.0
        version: 26.2.0
      - name: keycloak-operator.v26.1.5
        version: 26.1.5
      - name: keycloak-operator.v26.1.4
        version: 26.1.4
      - name: keycloak-operator.v26.1.3
        version: 26.1.3
      - name: keycloak-operator.v26.1.2
        version: 26.1.2
      - name: keycloak-operator.v26.1.1
        version: 26.1.1
      - name: keycloak-operator.v26.1.0
        version: 26.1.0
      - name: keycloak-operator.v26.0.8
        version: 26.0.8
      - name: keycloak-operator.v26.0.7
        version: 26.0.7
      - name: keycloak-operator.v26.0.6
        version: 26.0.6
      - name: keycloak-operator.v26.0.5
        version: 26.0.5
      - name: keycloak-operator.v26.0.4
        version: 26.0.4
      - name: keycloak-operator.v26.0.2
        version: 26.0.2
      - name: keycloak-operator.v26.0.1
        version: 26.0.1
      - name: keycloak-operator.v26.0.0
        version: 26.0.0
      - name: keycloak-operator.v25.0.6
        version: 25.0.6
      - name: keycloak-operator.v25.0.5
        version: 25.0.5
      - name: keycloak-operator.v25.0.4
        version: 25.0.4
      - name: keycloak-operator.v25.0.2
        version: 25.0.2
      - name: keycloak-operator.v25.0.1
        version: 25.0.1
      - name: keycloak-operator.v25.0.0
        version: 25.0.0
      - name: keycloak-operator.v24.0.5
        version: 24.0.5
      - name: keycloak-operator.v24.0.4
        version: 24.0.4
      - name: keycloak-operator.v24.0.3
        version: 24.0.3
      - name: keycloak-operator.v24.0.2
        version: 24.0.2
      - name: keycloak-operator.v24.0.1
        version: 24.0.1
      - name: keycloak-operator.v24.0.0
        version: 24.0.0
      - name: keycloak-operator.v23.0.7
        version: 23.0.7
      - name: keycloak-operator.v23.0.6
        version: 23.0.6
      - name: keycloak-operator.v23.0.5
        version: 23.0.5
      - name: keycloak-operator.v23.0.4
        version: 23.0.4
      - name: keycloak-operator.v23.0.3
        version: 23.0.3
      - name: keycloak-operator.v23.0.2
        version: 23.0.2
      - name: keycloak-operator.v23.0.1
        version: 23.0.1
      - name: keycloak-operator.v23.0.0
        version: 23.0.0
      - name: keycloak-operator.v22.0.5
        version: 22.0.5
      - name: keycloak-operator.v22.0.4
        version: 22.0.4
      - name: keycloak-operator.v22.0.3
        version: 22.0.3
      - name: keycloak-operator.v22.0.2
        version: 22.0.2
      - name: keycloak-operator.v22.0.1
        version: 22.0.1
      - name: keycloak-operator.v22.0.0
        version: 22.0.0
      - name: keycloak-operator.v21.1.2
        version: 21.1.2
      - name: keycloak-operator.v21.1.1
        version: 21.1.1
      - name: keycloak-operator.v21.1.0
        version: 21.1.0
      - name: keycloak-operator.v21.0.2
        version: 21.0.2
      - name: keycloak-operator.v21.0.1
        version: 21.0.1
      - name: keycloak-operator.v21.0.0
        version: 21.0.0
      - name: keycloak-operator.v20.0.3
        version: 20.0.3
      - name: keycloak-operator.v20.0.2
        version: 20.0.2
      - name: keycloak-operator.v20.0.1
        version: 20.0.1
      - name: keycloak-operator.v20.0.0
        version: 20.0.0
      name: fast
    defaultChannel: fast
    packageName: keycloak-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NEC
      provider-url: ""
    name: nspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nspc-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.nec.com/v1",
                "kind": "NSPC",
                "metadata": {
                  "name": "nspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
          createdAt: "2025-07-24T00:00:00Z"
          description: An operator for managing NEC Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: NEC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NSPC is the Schema for the nspcs API
            displayName: NSPC
            kind: NSPC
            name: nspcs.csi.nec.com
            version: v1
        description: |-
          ## About
          NEC Storage Plug-in for Containers (NSPC) is a plugin that integrates NEC Storage V Series into Kubernetes based clusters.
          NSPC provides persistent volume provisioning capabilities from NEC Storage V Series arrays.

          For full documentation, refer to the reference guide.

          ## Requirements

          ### Supported Driver Version

          * NSPC v3.16.0
        displayName: NEC Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - NEC
        - Storage
        - iStorage
        - V Series
        - Container Storage Interface
        - CSI
        maturity: stable
        provider:
          name: NEC
        relatedImages:
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.connect.redhat.com/nec/nspc-operator-bundle@sha256:21b69b5eb896a37691a8e086f3a51c0f68086a0ff9370fd66667acc6c935a850
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/nec/nspc-operator@sha256:50fe10f05d8aa7db1b97280e5842bb0b67c6f9bfb01c0d17c5be3cd11b78e853
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/nec/nspc-csi-driver@sha256:e50d60e025b3c8ae522a8c5ae78a07f8a624a1210ed7a078c8084efd76ccc646
        version: 1.16.0
      entries:
      - name: nspc-operator.v1.16.0
        version: 1.16.0
      name: stable
    defaultChannel: stable
    packageName: nspc-operator
    provider:
      name: NEC
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: unsupported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: mta-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mta-operator.v6.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
          createdAt: "2023-04-26T14:04:20Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          olm.skipRange: '>=0.0.0 <6.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:a4200fc0a39df67eae31c53ee9ae0c05769fb410d0312ebd0bbf8bf1567b2a48
        - registry.redhat.io/rhel8/postgresql-12@sha256:3a69efa85af6a6f95665eca69085b007fa1fcb645a37f356895aeb4c9527a621
        - registry.redhat.io/mta/mta-admin-addon-rhel8@sha256:812b8dc621a06bb99cf2c59abb8257da3ceeb165c563e14d258e69803a557cc8
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:c75e1d9d3307d95be392a53440d9123601a6ada3693e3f2573e03022565e4339
        - registry.redhat.io/mta/mta-operator-bundle@sha256:65c0932bfda32b1e62be9edd0f96bb4d6e41e302a2e8080fc6cc794ee5c52276
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:a66dd030eb99207b6aaacc6b7c7cb5182901da836ebf50373c14f8e5ecb7b642
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:f5d10cd2528bebf5094c7744e47d9de6fe827fd09024c5ffc0e1b74f1a6cc3b4
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:de6cb95ce3663922ecc725b186876b043a9c7794e95510459ffbcba7d668fc26
        version: 6.0.3
      entries:
      - name: mta-operator.v6.0.3
        version: 6.0.3
      - name: mta-operator.v6.0.2
        version: 6.0.2
      - name: mta-operator.v6.0.1
        version: 6.0.1
      - name: mta-operator.v6.0.0
        version: 6.0.0
      name: stable-v6.0
    - currentCSV: mta-operator.v6.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
          createdAt: "2023-10-26T15:29:26Z"
          description: MTA contains tools that support the modernization and migration
            of applications to Kubernetes and OpenShift
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.1.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/tackle2-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is a collection of tools that supports large-scale application modernization and migration projects to Kubernetes and OpenShift.


          Migration Toolkit for Applications allows users to maintain their portfolio of applications with a full set of metadata and to assess their suitability for modernization leveraging a questionnaire based approach.


          ### Install

          By default, MTA requires a RWX capable storage class for caching purposes, this can be disabled via the rwx_supported parameter in Tackle CR.

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-operator-bundle@sha256:f3d585e24d65905b2e9f594dcdb67721aa95057ab7587ae531c3487aa2f94e53
        - registry.redhat.io/mta/mta-pathfinder-rhel8@sha256:1df881da1e20c2501caf53e1cf59b6b366281d842f0948ca5f91694ec1db829d
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6da71d4caecfbe049e63576d4325d4919002196da2e875f79703d8ff0d5a1705
        - registry.redhat.io/mta/mta-ui-rhel8@sha256:3845d6d246003ebeb49bcf1d95d6be2eb9cd05a280a5bf8320b201e878a41427
        - registry.redhat.io/mta/mta-windup-addon-rhel8@sha256:6bcaf8e8b63e169652d6568d77e91a289084e621a586105a2d429b5726569058
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/mta/mta-hub-rhel8@sha256:32624f2bddfe8b6b9c6c1c3f2be541c8c61be29680d13b6d446e8874bdf8e219
        version: 6.1.4
      entries:
      - name: mta-operator.v6.1.4
        version: 6.1.4
      - name: mta-operator.v6.1.3
        version: 6.1.3
      - name: mta-operator.v6.1.2
        version: 6.1.2
      - name: mta-operator.v6.1.1
        version: 6.1.1
      - name: mta-operator.v6.1.0
        version: 6.1.0
      name: stable-v6.1
    - currentCSV: mta-operator.v6.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
          createdAt: "2024-06-10T22:12:58Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0 <6.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:e59725832547e80cab58077a1b6e42c313a4057621defa841f35c0a7627f149d
        - registry.redhat.io/mta/mta-operator-bundle@sha256:a13643117c2867351718a872f7f1b2350c67855ca73a727a1dc140754ffe6589
        - registry.redhat.io/mta/mta-pathfinder-rhel9@sha256:a2a8b6d204d99494a6a005e8cb1fbd92afdc84dc01e4c1de4ad7d8798bd83741
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:6938c3fc6dd8fb032e1765fc11d9586dbfd1aaba4b423b0d87da4120ee0bb60d
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:c35f3ed429a7d351cb8c2d07600b4ad35615a5ba08e5cc9f8b9bd504aea3f7ee
        - registry.redhat.io/mta/mta-windup-addon-rhel9@sha256:d1f4939b92dd13c15cb7344ec47a571669f6d108dc2e7d795ce29350f2cb3838
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda
        - registry.redhat.io/rhel8/postgresql-12@sha256:c4716bee5e2cf2e43698bc99e5aa42f16a0f2d68c1fbdcd3602696bf88e408dc
        version: 6.2.3
      entries:
      - name: mta-operator.v6.2.3
        version: 6.2.3
      - name: mta-operator.v6.2.2
        version: 6.2.2
      - name: mta-operator.v6.2.1
        version: 6.2.1
      - name: mta-operator.v6.2.0
        version: 6.2.0
      name: stable-v6.2
    - currentCSV: mta-operator.v7.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "addon-example",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "image": "addon-fqin"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
          createdAt: "2024-05-10T19:37:30Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.0.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Tackle Instance
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Represents a Tackle Addon Instance
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:da619f6617d99790b4fe00fdb23ab4aea0cdb0592590162d51b3ef94aca01a7b
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:18c611415589a093e16e9c2f2fba9f3c83d9514f79c8e4fde9704ef0679ea4d5
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:268ecf5efc51af71f9b2bc2d01b410f8cfcf81a948b82d2d476fef6c0cd3e033
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:0e1ec99af5666a060176adc52612f20209e3398a9a5a2aa1dcff11ef260b2006
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:af879d8b9b5e396a5308fb675a1d6dba6d70151b9bdbe7829362e081a6adab1c
        - registry.redhat.io/mta/mta-windup-shim-rhel9@sha256:e2ea89055936d7ad229b5e890c6ba234c6cb22d8b688515ed4d6a34848efea83
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:1b5e016046d93b6a2674d63f08e0a46f8050e785bd58fd94ba7a70487efe7e2a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:d554cc36748fc05e30eb10e6fe816ff6aab2fe07b61cc60e9d93de5c3cf17748
        - registry.redhat.io/mta/mta-analyzer-lsp-rhel9@sha256:0c8f59f35a950ea8b45f16cf71362128ccc335b58a996a099384fd737dd6ff90
        - registry.redhat.io/mta/mta-operator-bundle@sha256:9eeb43af2bcab84f5261d1575f7c897903a7696dba011d256abffa1fe850eba2
        version: 7.0.3
      entries:
      - name: mta-operator.v7.0.3
        version: 7.0.3
      - name: mta-operator.v7.0.2
        version: 7.0.2
      - name: mta-operator.v7.0.1
        version: 7.0.1
      - name: mta-operator.v7.0.0
        version: 7.0.0
      name: stable-v7.0
    - currentCSV: mta-operator.v7.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
          createdAt: "2024-09-16T23:45:42Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.1.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:23c97ac122eb7cd1e953df8da968da420f6db8708baf772759517aca4929748b
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:ee2c8ce2af0be097353758d9ee7eaa7e6b3fcd5f2bf5e04280e450e2b619d2d8
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:e4e70201df23c9e1168646d3c6013a8ffa22481a85b07df491f240310c13ebe1
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:16cb28d683159ba71a5d7c233404ac1d6b7a07a374b46d24effe551a13a596c3
        - registry.redhat.io/rhel8/postgresql-15@sha256:16faf1ee82382263408f713ee4002f47ff13c82d4545e8a69d5b0ebd1d289330
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:4ad19b50cdd6be21e010155f2026abf0dc865e72aac1353e35bbb5eeb8320a4e
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:08b33b44140397785f8aa1b736796e09cf6e49ac30cf7c386a2fbd8b678d47b1
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ac38f1c3d34bc718928eb20743835c99783d846f63a3197d13839ea22f1634c8
        - registry.redhat.io/mta/mta-operator-bundle@sha256:6aef520983fe01351886e108d32b998b20e74116fc0280c754a914bfad03d0bf
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:4f8d66597feeb32bb18699326029f9a71a5aca4a57679d636b876377c2e95695
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:30a3b1d5a70dec74a9253ca34f18bd8f2cdfaa68b85b0b25dc395158c2596330
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:582f28e609c1a0466d30c5586d6b6023f1b1de3c28f76cabb498322914ab24d4
        version: 7.1.1
      entries:
      - name: mta-operator.v7.1.1
        version: 7.1.1
      - name: mta-operator.v7.1.0
        version: 7.1.0
      name: stable-v7.1
    - currentCSV: mta-operator.v7.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
          createdAt: "2025-06-17T16:04:15Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.2.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel8@sha256:c0643c05833dd1b70ed361bf56c46f56ed6536aac3106b92f0d85b83da1b9d15
        - registry.redhat.io/mta/mta-rhel8-operator@sha256:917d6cd534c158f6c4a7e9c83f0c66fa85ba81037c352d8a039e198581126a7f
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:10ef4b218d603432508b5e1f3685fb32a80cbce6c327e35952cdcccb0465468b
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:9d6c02f6463771e747a30126de95fc0728fd95fff4ddfa6c70d828c55e0b6790
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:4cde2c2df482293bd4d6b86a344336e13d2f9f81b29323698f603e656b14b190
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ed10e35f45a140e21d826bee71b3ab96e67f1b4dc8c2f7f1676380fa51b67fe
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/rhel8/postgresql-15@sha256:dd4aec2775da1cf187c0839384cb16d69268e9781d4330874bbc865f98153a7a
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:c9447659195b9eebb8f8a265fd5052ca9b2611d57080017448297a33409863c6
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:b10fdc78c6f453a34a3d306313dc29fcb3152688a6a1d54cd236d00d9fab88af
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:2dfc51d5398a90e73273b3e332f70d7ed94bac05e12b142fba22104ecc929436
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:70ff3b5e429a8fe0c68752ee05c47ead885b59bf9c8d7235337801270e7fc1de
        version: 7.2.3
      entries:
      - name: mta-operator.v7.2.3
        version: 7.2.3
      - name: mta-operator.v7.2.2
        version: 7.2.2
      - name: mta-operator.v7.2.1
        version: 7.2.1
      - name: mta-operator.v7.2.0
        version: 7.2.0
      name: stable-v7.2
    - currentCSV: mta-operator.v7.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
          createdAt: "2025-08-15T16:32:16Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.2.0 <7.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:32328fc0459099ef1de0f92e064c15d29a46d88dfe2f54c88f822d44ab82e3f4
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:9748d954ad8654ac203257286497ccf19cfdfa211fee40449cbe732ff48c1176
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:8ca00460b2fb64583a247f5b906c19743b7732b60ff8215f1baec12a766c61d1
        - registry.redhat.io/rhel8/postgresql-15@sha256:854710b428c4c1132c448cbe9ad30087a8a4b3c33ae9c0c3a4de21a15c0ae642
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:072f676323e9abcb2ff22eb5e7a70f6681e9d885d8666ddd81060fbad54c443e
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:6abd6541a955543207c1ff4b903928debcc89dd6ea3d7aeaa3e47465849f7b37
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:ef57a58e609e244333d35681a164a1d3bf382c0eee37d33c993b018439814198
        - registry.redhat.io/mta/mta-operator-bundle@sha256:68953a2a721309fd103e09f3171d0c9d50020413631b575f1a44602a5d2ac855
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:b2ff154a4b4ead22311005b7115c02346928b5ebf77d776ef909ad4af1f350bc
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:6118c1db15667851f26eb2acf6060e045ed9065459e3e649bbdd0e2625d825bd
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:80aa13813fe121da49e166d81d09d63749393f3375c3b96d762d0b09c6685ca9
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:a93cbb575ff59bc8015602165645204b9ca56e327308b11a3d75b651701db875
        version: 7.3.2
      entries:
      - name: mta-operator.v7.3.2
        version: 7.3.2
      - name: mta-operator.v7.3.1
        version: 7.3.1
      - name: mta-operator.v7.3.0
        version: 7.3.0
      name: stable-v7.3
    - currentCSV: mta-operator.v8.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "container": {
                    "image": "addon-img",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "external-provider-img",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "feature_auth_required": "true"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "openshift-mta"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Modernization & Migration
          certified: "false"
          containerImage: registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
          createdAt: "2025-09-26T03:03:31Z"
          description: MTA is an application modernization platform that helps organizations
            safely and predictably modernize applications to OpenShift at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.3.0 <8.0.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "openshift-mta"
              },
              "spec": {
                  "feature_auth_required": "true"
              }
            }
          operatorframework.io/suggested-namespace: openshift-mta
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/konveyor/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Schema
            displayName: Schema
            kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Migration Toolkit for Applications is an application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to OpenShift.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Documentation
          Documentation can be found on our [website](https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications).

          ### Getting help
          If you encounter any issues while using Migration Toolkit for Applications Operator, create a [support case](https://access.redhat.com/support/cases/) for bugs, enhancements, or other requests.

        displayName: Migration Toolkit for Applications Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        - mta
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/migration_toolkit_for_applications
        - name: Developers
          url: https://developers.redhat.com/products/mta
        - name: Konveyor
          url: https://www.konveyor.io
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/mta/mta-ui-rhel9@sha256:460d1d45ed44ae9328ed93c5fed240e154549bbec662dc1fc55569ff7256aca9
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/mta/mta-discovery-addon-rhel9@sha256:b1b38e218ea37b529f344eaa1f0f3b55b716ad7953e852c19f34f4a25ab6c5a7
        - registry.redhat.io/mta/mta-hub-rhel9@sha256:af2c1790f2b98ebcfadb5d6bdd43cb9e7e96cdabd5b653d48d0fb67901781c49
        - registry.redhat.io/mta/mta-dotnet-external-provider-rhel9@sha256:2a6f0b392b86fe16244d5bf9fcb84ed9770a22329a96e9861f03872b7209fb04
        - registry.redhat.io/mta/mta-java-external-provider-rhel9@sha256:8ae0887fa27d6f4281e5bb9b629bf6b64a9c460948a002839ba42da9b25b5521
        - registry.redhat.io/mta/mta-solution-server-rhel9@sha256:e9d48d11ae203108229d67bc96721daa1c3bb10160b9d6d957cd2e0ba68a2ef7
        - registry.redhat.io/mta/mta-cli-rhel9@sha256:aaf7bf10b1af107ad55289caab2cf9dd052e69211968d2e8d05c1d25f21ed2b3
        - registry.redhat.io/mta/mta-generic-external-provider-rhel9@sha256:a77ee9b76225b6cca667fa3402492b0845587fe5e400fa8d141b5ab636c56811
        - registry.redhat.io/mta/mta-operator-bundle@sha256:0ec977b0c9c26551db110dcf2d5069647b1a5495f869ff48a0319cb496b84cdc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:ef10b2ecbfa0c4ac8568af9ee64ee4a554e3fdc9089ee5683b6c2c3a74caafbe
        - registry.redhat.io/mta/mta-rhel9-operator@sha256:83c2b804667d28f782d64132a0c393194083d10c820b88df49c2503d3351fd8b
        - registry.redhat.io/mta/mta-platform-addon-rhel9@sha256:dcb35178810b7eddc2f2713434d377003f8f4fb7022a40f6c1926b74b13fe23a
        - registry.redhat.io/mta/mta-analyzer-addon-rhel9@sha256:616d04cfbab92341157c42eee0683f31867d81c5a289723da8da8208d1054774
        version: 8.0.0
      entries:
      - name: mta-operator.v8.0.0
        version: 8.0.0
      name: stable-v8.0
    defaultChannel: stable-v8.0
    packageName: mta-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: eG Innovations
      provider-url: https://www.eginnovations.com
    name: eginnovations-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eginnovations-operator.v0.0.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eginnovations.com/v1beta1",
                "kind": "EgUniversalAgent",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eginnovations-operator"
                  },
                  "name": "eguniversalagent",
                  "namespace": "egagent"
                },
                "spec": {
                  "agent": {
                    "env": [
                      {
                        "name": "EG_MANAGER",
                        "value": "Replace with eG Manager IP"
                      },
                      {
                        "name": "EG_MANAGER_PORT",
                        "value": "Replace with eG Manager Port"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "true"
          containerImage: docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
          createdAt: "2025-09-11T20:08:00Z"
          description: eG Innovations provides a modern monitoring and analytics platform.
            Gather metrics, logs and traces for full observability of your Kubernetes
            cluster with eG Universal Agent Operator.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eginnovations/universal-agent-operator
          support: eG Innovations
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgUniversalAgent is the Schema for the eguniversalagents
              API
            displayName: Eg Universal Agent
            kind: EgUniversalAgent
            name: eguniversalagents.eginnovations.com
            version: v1beta1
        description: |
          eG Innovations’ converged application and infrastructure performance monitoring capabilities provide end-to-end visibility and correlated analytics for every layer, every tier of your IT landscape.

          The eG Enterprise Universal Agent Operator automatically configures a host agent on every Kubernetes worker node. The host agent auto-discovers the worker nodes and application containers running as Pods on each node and tracks their performance and utilization levels. In-depth monitoring of applications running on containers is also provided using the same host agent. No additional agents are required for the containers.

          ### Capabilities
            * Rolls out the egagent pod per node to monitor its pods and the node itself
            * Automatic BTM profiler injection using webhook based injection for deep application monitoring and transection tracing.

          ### Installation
            1. Create the namespace "egagent" for installing eG Agent on OpenShift Cluster.

               <br />``` $ oc create ns egagent ```
            2. Add "egagent" serviceaccount to scc( Security Context Constraints ) of OpenShift Cluster as privileged using below command.

               <br />``` $ oc adm policy add-scc-to-user privileged -z egagent -n egagent ```
            2. Install eG Innovations Universal Agent operator from [OpenShift Container Platform](https://www.openshift.com/) OperatorHub.
            3. Create the eG Universal Agent CRD (Custom Resource Definition)
               a. Click "View YAML Example" and copy the yaml contents and save it to (filename).yaml on your host.
               b. To apply the CRD run the following command

               <br />``` $ oc apply -f (filename).yaml ```

          ### Required Parameters
            * EG\_MANAGER, EG\_MANAGER\_PORT - provide the ip and port of your eG Manager Installation.

          ### Advanced Options
            * Disable Certificate Checking - disable any certificate validation that may interact poorly with proxies within your cluster
            * Image Override - use a copy of the eG Iniversal Agent container image from a registry other that Docker's or Red Hat's
            * Environment variables - define environment variables for the eG Universal Agent container of for APM profiler.
            * Namespace Selectors- select a subset of namesapces to enable instrumentation for applications.

          For complete list of supported parameters and the detailed documentation available in [eG Innovations Docs](https://docs.eginnovations.com)
        displayName: eG Innovations Universal Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Tracing
        links:
        - name: eG Innovations Kubernetes Monitoring Overview
          url: https://www.eginnovations.com/documentation/Kubernetes/Introduction-to-Monitoring-Kubernetes.htm
        maintainers:
        - email: devops@eginnovations.com
          name: eG Innovations
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: eG Innovations
          url: https://www.eginnovations.com
        relatedImages:
        - registry.connect.redhat.com/eginnovations/universal-agent-operator@sha256:65f851074e0e1db57a62b559c0f10f9d30543011727aec3b832f9a35468139e9
        - docker.io/eginnovations/universal-agent-operator@sha256:b8e3e26dca1b94fabfd79496feb36f3f2de4434331ef9ca810125892a936dca7
        version: 0.0.11
      entries:
      - name: eginnovations-operator.v0.0.11
        version: 0.0.11
      name: beta
    defaultChannel: beta
    packageName: eginnovations-operator
    provider:
      name: eG Innovations
      url: https://www.eginnovations.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - registry.connect.redhat.com/datastax/cass-operator-bundle@sha256:8d130a5f2407931dbc73e45a53610b95ceb04cb507b8f6519f630cada1c0da45
        - registry.connect.redhat.com/datastax/cass-operator@sha256:6abc6253b50a0d7f6bdd4b394d2bf96320a8845952aeb8f0bbed99863192ec52
        - registry.connect.redhat.com/datastax/system-logger@sha256:3408620786e6265aa835be56f85aadc48b20060b2038dc21126f614685f77c6b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      name: stable
    defaultChannel: stable
    packageName: cass-operator
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubemq.io
      provider-url: ""
    name: kubemq-operator-marketplace
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubemq-operator.v1.7.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
              "apiVersion": "core.k8s.kubemq.io/v1beta1",
              "kind": "KubemqCluster",
              "metadata": {
                "name": "kubemq-cluster"
              },
              "spec": {
                "replicas": 3
              }
            },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqDashboard",
                "metadata": {
                  "name": "kubemq-dashboard"
                }
              },
              {
                "apiVersion": "core.k8s.kubemq.io/v1beta1",
                "kind": "KubemqConnector",
                "metadata": {
                  "name": "kubemq-bridges"
                },
                "spec": {
                  "type": "bridges",
                  "config": "bindings: null"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
          createdAt: 2020-03-11 12:00:00+00:00
          description: Kubernetes Messaging Platform
          support: support@kubemq.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Kubemq Cluster
            displayName: Kubemq Cluster
            kind: KubemqCluster
            name: kubemqclusters.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Cluster Dashboard
            displayName: Kubemq Dashboard
            kind: KubemqDashboard
            name: kubemqdashboards.core.k8s.kubemq.io
            version: v1beta1
          - description: Kubemq Connector
            displayName: Kubemq Connector
            kind: KubemqConnector
            name: kubemqconnectors.core.k8s.kubemq.io
            version: v1beta1
        description: |2-

          ## Kubernetes Messaging Platform
          A message-based architecture for microservices providing flexible connectivity for hybrid environments: multi-cloud, on-premises, and at the edge. Powered by enterprise-grade message broker and message queue with ready to use connectors, bridges, and control center.

          Messaging platform components:
            - **KubeMQ Cluster** - Enterprise-grade message broker and message queue, scalable, high available, and secured.
            - **KubeMQ Targets Connector** - connects a KubeMQ cluster with external systems and cloud services.
            - **KubeMQ Bridges Connector** - bridge, replicate, aggregate, and transform messages between KubeMQ clusters.
            - **KubeMQ Sources Connector** - transform external systems and cloud services data to KubeMQ cluster messages.
            - **KubeMQ Dashboard** - monitor KubeMQ cluster metrics.

          ### Kubernetes Native
          Innovative and modern message queue and message broker in a lightweight container developed to run in Kubernetes, certified in the CNCF landscape, and connect natively to the cloud-native ecosystem.

          ### Ease of use

          Simple deployment in Kubernetes in less than 1 minute. Developer friendly by simple to use SDKs and elimination of the many developers and DevOps-centered challenges to define exchanges, brokers, channels, routes, and predefined topics

          ### All messaging patterns

          KubeMQ is available with all messaging patterns such as Queue, Stream, Pub/Sub, and RPC, Saving the need to maintain multiple messaging platforms for different use cases.

          ### Hybrid Environments

          KubeMQ enables the microservices from multiple environments to communicate and build one hybrid infrastructure solution across clouds, on-prem, and at the edge.

          ### KubeMQ Cluster key features

          - All-batteries included Messaging Queue Broker for the Kubernetes environment
          - Deployed with Operator for full life cycle operation
          - Blazing fast (written in Go), small and lightweight Docker container
          - Asynchronous and Synchronous messaging with support for  `At Most Once Delivery` and `At Least Once Delivery` models
          - Supports durable FIFO based Queue, Publish-Subscribe Events, Publish-Subscribe with Persistence (Events Store), RPC Command and Query messaging patterns
          - Supports gRPC, Rest and WebSocket Transport protocols with TLS support (both RPC and Stream modes)
          - Supports Access control Authorization and Authentication
          - Supports message masticating and smart routing
          - No Message broker configuration needed (i.e., queues, exchanges)
          - .Net, Java, Python, Go, and NodeJS SDKs

          ### KubeMQ Targets Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major services types (databases, cache, messaging, serverless, HTTP, etc.)
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-targets)

          ### KubeMQ Bridges Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Build Any Topology - connects KubeMQ clusters in 1:1, 1:n , n:1, n:n
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-bridges)

          ### KubeMQ Sources Connector key features

          - Runs anywhere - Kubernetes, Cloud, on-prem, anywhere
          - Stand-alone - small docker container / binary
          - Single Interface - One interface all the services
          - Any Service - Support all major messaging services such as Kafka, ActiveMQ, SQS, and more.
          - Plug-in Architecture Easy to extend, easy to connect
          - Middleware Supports - Logs, Metrics, Retries, and Rate Limiters
          - Easy Configuration - simple yaml file builds your topology

          [Check out Github page](https://github.com/kubemq-io/kubemq-sources)

          ## Getting Started

          [Start with OpenShift](https://docs.kubemq.io/getting-started/create-cluster/openshift)

          ## Get Enterprise Trial License

          [Register](https://account.kubemq.io/login/register)

          ## Request a Demo

          [See KubeMQ in Action](https://kubemq.io/request-a-demo/)
        displayName: Kubemq Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - messaging
        - queue
        - broker
        - pubsub
        - rpc
        links:
        - name: Learn more about Kubemq
          url: https://kubemq.io/
        - name: Register for free license
          url: https://account.kubemq.io/login/register
        - name: Contact Sales
          url: https://kubemq.io/contact-sales/
        - name: Documentation
          url: https://docs.kubemq.io/
        - name: Get Help
          url: https://kubemq.io/contact-support/
        maintainers:
        - email: support@kubemq.io
          name: kubemq
        maturity: alpha
        provider:
          name: Kubemq.io
        relatedImages:
        - registry.connect.redhat.com/kubemq/kubemq-operator-bundle@sha256:9009198860c7a1e0de4f429261d442b75b5b2893a943bb77914b14d643db8829
        - registry.connect.redhat.com/kubemq/kubemq-operator:v1.7.2
        version: 1.7.8
      entries:
      - name: kubemq-operator.v1.7.8
        version: 1.7.8
      name: alpha
    defaultChannel: alpha
    packageName: kubemq-operator-marketplace
    provider:
      name: Kubemq.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Confluent, Inc.
      provider-url: https://confluent.io
    name: confluent-for-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: confluent-for-kubernetes.v2.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.8.3",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.8.3",
                    "init": "confluentinc/confluent-init-container:2.10.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
          createdAt: "2025-06-19T11:07:10Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.9.0 <2.10.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-zookeeper@sha256:c8487e18318a92a64e84a5ab29f844a0973f3b8b2aa291723014731471459176
        - docker.io/confluentinc/cp-server@sha256:072719eef75ee61294027ba77f7b9cd821437158761495a1644b5687af9305a8
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:7a271565db9c500f8c5755de397e8093251673d2fc6b8a10974f927f382b4704
        - confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:2520a514b848b53338c411f0e0dfa68491596761f8bbbb090d652f71eda010af
        - docker.io/confluentinc/cp-kafka-rest@sha256:9c19e7900669e555810d26676393737a7d573b39c31e284c0ad78807bfe05ce3
        - docker.io/confluentinc/cp-server-connect@sha256:05e77fe8e19747521e1f1d5630bbf384b0fe0ec9efe8dc798de6e4a3eb0bfa2d
        - docker.io/confluentinc/cp-schema-registry@sha256:5b6bf5b971dd8fe2c821e227391858f58f6ae2128f1ae1faa988c3ea59b86801
        - docker.io/confluentinc/confluent-init-container@sha256:550a6b92fc5aca81ad8e0d3d47c0063af9da877e45c1ecda1a9cc44a70e5d87a
        - docker.io/confluentinc/confluent-operator@sha256:1d164a46cbada9ee8fe0a6f8ace2e1eb4838e7ed1c67b1e8a2caa24480bf44ca
        - docker.io/confluentinc/cp-ksqldb-server@sha256:df1f6a99ecfe8d8acc0af222848aad13349f06071fc2bf1360d6d20f99769394
        version: 2.10.2
      entries:
      - name: confluent-for-kubernetes.v2.10.2
        version: 2.10.2
      name: "2.10"
    - currentCSV: confluent-for-kubernetes.v2.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:7.9.2",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Zookeeper",
                "metadata": {
                  "name": "zookeeper-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "image": {
                    "application": "confluentinc/cp-zookeeper:7.9.2",
                    "init": "confluentinc/confluent-init-container:2.11.2"
                  },
                  "logVolumeCapacity": "10Gi",
                  "replicas": 3
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
          createdAt: "2025-06-19T11:14:20Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.10.0 <2.11.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:b7cde5736b5275abafafbb81f4c917890abfe8f400f436c3a1f712cf200832f8
        - docker.io/confluentinc/cp-schema-registry@sha256:4f6cabeea11ae532cb47905eadaa3192d7113ea6d0a0da310cdfd5eb7142fc45
        - docker.io/confluentinc/confluent-init-container@sha256:37161a90b89dfad625bf61ba7d299707d7b2a96271e9667044c3eaba3ffc5f66
        - docker.io/confluentinc/cp-server@sha256:dfed43a1618de1fea6d202921ca4cc2168f7455c3d37fe046f3e64febf042a87
        - docker.io/confluentinc/cp-zookeeper@sha256:f1e6d3d698a6e2757e3c225b5c6f8f007251838226c46f652a4cbb4314b50277
        - confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - docker.io/confluentinc/confluent-operator@sha256:fe9f81c2aa5fc4e8bd9580eb15032e581d76cf224acda8e2181e5c333c39cdd6
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:2baa0cb5df8017d9d14c2f0c7e02ce371ba92462d8d1e93a3ca1bc548c20a74e
        - docker.io/confluentinc/cp-enterprise-control-center@sha256:507e1538a64c4232ffda86504594def45cb20bbcb6b13fc6196c1f41e71f93bf
        - docker.io/confluentinc/cp-ksqldb-server@sha256:744f5d844ac223c857116f3994c5be2105898ea3896d53026eb62e76c7240c4f
        - docker.io/confluentinc/cp-server-connect@sha256:73822ffff10d386b98c9ca2bb295b4d079deb210116c8e4283cdc26ce304526f
        version: 2.11.2
      entries:
      - name: confluent-for-kubernetes.v2.11.2
        version: 2.11.2
      name: "2.11"
    - currentCSV: confluent-for-kubernetes.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "CMFRestClass",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "cmfrestclass-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ClusterLink",
                "metadata": {
                  "name": "clusterlink-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "aclFilters": [
                    {
                      "accessFilter": {
                        "host": "example.com",
                        "operation": "read",
                        "permissionType": "allow",
                        "principal": "User:Bob"
                      },
                      "resourceFilter": {
                        "name": "coffee",
                        "patternType": "prefixed",
                        "resourceType": "topic"
                      }
                    }
                  ],
                  "configs": {
                    "connections.max.idle.ms": "620000"
                  },
                  "consumerGroupFilters": [
                    {
                      "filterType": "INCLUDE",
                      "name": "filter-example",
                      "patternType": "LITERAL"
                    }
                  ],
                  "destinationKafkaCluster": {
                    "kafkaRestClassRef": {
                      "name": "krc-example"
                    }
                  },
                  "sourceKafkaCluster": {
                    "bootstrapEndpoint": "example.com:9092",
                    "clusterID": "test_kafka_cluster_id"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ConfluentRolebinding",
                "metadata": {
                  "name": "cfrb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "clustersScopeByIds": {
                    "schemaRegistryClusterId": "id_schemaregistry_confluent"
                  },
                  "kafkaRestClassRef": {
                    "name": "default"
                  },
                  "principal": {
                    "name": "user-example",
                    "type": "user"
                  },
                  "role": "SecurityAdmin"
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connect",
                "metadata": {
                  "name": "connect-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "connect-basic-example"
                    },
                    "type": "basic"
                  },
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server-connect:8.0.0",
                    "init": "confluentinc/confluent-init-container-operator"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Connector",
                "metadata": {
                  "name": "connector-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "class": "org.apache.kafka.connect.file.FileStreamSinkConnector",
                  "configs": {
                    "file": "/tmp/test.txt",
                    "topics": "test-topic"
                  },
                  "connectClusterRef": {
                    "name": "connect",
                    "namespace": "confluent"
                  },
                  "connectRest": {
                    "authentication": {
                      "secretRef": "basic-credential-example",
                      "type": "basic"
                    }
                  },
                  "restartPolicy": {
                    "maxRetry": 20,
                    "type": "OnFailure"
                  },
                  "taskMax": 4
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "ControlCenter",
                "metadata": {
                  "name": "controlcenter-ng-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "controlcenter-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "alertManagerClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9093"
                    },
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071"
                    },
                    "prometheusClient": {
                      "url": "http://controlcenter-next-gen.confluent.svc.cluster.local:9090"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-enterprise-control-center-next-gen:2.2.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "services": {
                    "alertmanager": {
                      "image": "confluentinc/cp-enterprise-alertmanager:2.2.0"
                    },
                    "prometheus": {
                      "image": "confluentinc/cp-enterprise-prometheus:2.2.0",
                      "pvc": {
                        "dataVolumeCapacity": "10Gi"
                      }
                    }
                  },
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkapplication-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "FlinkEnvironment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "confluent-for-kubernetes"
                  },
                  "name": "flinkenvironment-sample",
                  "namespace": "confluent-operator-system"
                },
                "spec": null
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftController",
                "metadata": {
                  "name": "kraftcontroller-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10G",
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "controller": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KRaftMigrationJob",
                "metadata": {
                  "name": "kraft-migration-job",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kRaftController": {
                      "name": "kraftcontroller",
                      "namespace": "confluent"
                    },
                    "kafka": {
                      "name": "kafka",
                      "namespace": "confluent"
                    },
                    "zookeeper": {
                      "name": "zookeeper",
                      "namespace": "confluent"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Kafka",
                "metadata": {
                  "name": "kafka-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "zookeeper": {
                      "endpoint": "zookeeper.confluent.svc.cluster.local:2181"
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "listeners": {
                    "internal": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestClass",
                "metadata": {
                  "name": "krc-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "kafkaRest": {
                    "authentication": {
                      "bearer": {
                        "secretRef": "rest-credential-example"
                      },
                      "type": "bearer"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaRestProxy",
                "metadata": {
                  "name": "krp-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-kafka-rest:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 3,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KafkaTopic",
                "metadata": {
                  "name": "topic-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "configs": {
                    "cleanup.policy": "compact"
                  },
                  "kafkaClusterRef": {
                    "name": "kafka",
                    "namespace": "confluent"
                  },
                  "partitionCount": 12,
                  "replicas": 1
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "KsqlDB",
                "metadata": {
                  "name": "ksqldb-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "authentication": {
                    "basic": {
                      "secretRef": "ksqldb-basic-example"
                    },
                    "type": "basic"
                  },
                  "dataVolumeCapacity": "10Gi",
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-ksqldb-server:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 2,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "Schema",
                "metadata": {
                  "name": "schema-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "data": {
                    "configRef": "schema-config",
                    "format": "avro"
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaExporter",
                "metadata": {
                  "name": "schemaexporter-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "contextName": "mycustom",
                  "destinationCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "destination"
                    }
                  },
                  "sourceCluster": {
                    "schemaRegistryClusterRef": {
                      "name": "schemaregistry",
                      "namespace": "source"
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.confluent.io/v1beta1",
                "kind": "SchemaRegistry",
                "metadata": {
                  "name": "schemaregistry-example",
                  "namespace": "confluent-operator-system"
                },
                "spec": {
                  "dependencies": {
                    "kafka": {
                      "authentication": {
                        "jaasConfig": {
                          "secretRef": "credential-example"
                        },
                        "type": "plain"
                      },
                      "bootstrapEndpoint": "kafka.confluent.svc.cluster.local:9071",
                      "tls": {
                        "enabled": true
                      }
                    }
                  },
                  "image": {
                    "application": "confluentinc/cp-schema-registry:8.0.0",
                    "init": "confluentinc/confluent-init-container:3.0.0"
                  },
                  "replicas": 1,
                  "tls": {
                    "autoGeneratedCerts": true
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Streaming & Messaging
          containerImage: docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
          createdAt: "2025-06-13T13:02:11Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.11.0 <3.0.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Confluent Platform License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Confluent, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterLink is the schema for the ClusterLink API.
            displayName: Cluster Link
            kind: ClusterLink
            name: clusterlinks.platform.confluent.io
            version: v1beta1
          - description: CMFRestClass is the Schema for the cmfrestclasses API
            displayName: CMFRest Class
            kind: CMFRestClass
            name: cmfrestclasses.platform.confluent.io
            version: v1beta1
          - description: ConfluentRolebinding is the schema for the ConfluentRolebinding
              API.
            displayName: Confluent Rolebinding
            kind: ConfluentRolebinding
            name: confluentrolebindings.platform.confluent.io
            version: v1beta1
          - description: Connector is the schema for the Connector API.
            displayName: Connector
            kind: Connector
            name: connectors.platform.confluent.io
            version: v1beta1
          - description: Connect is the schema for the Connect API.
            displayName: Connect
            kind: Connect
            name: connects.platform.confluent.io
            version: v1beta1
          - description: ControlCenter is the schema for the Control Center API.
            displayName: Control Center
            kind: ControlCenter
            name: controlcenters.platform.confluent.io
            version: v1beta1
          - description: FlinkApplication is the Schema for the flinkapplications
              API
            displayName: Flink Application
            kind: FlinkApplication
            name: flinkapplications.platform.confluent.io
            version: v1beta1
          - description: FlinkEnvironment is the Schema for the flinkenvironments
              API
            displayName: Flink Environment
            kind: FlinkEnvironment
            name: flinkenvironments.platform.confluent.io
            version: v1beta1
          - description: KafkaRestClass is the schema for the Kafka REST API.
            displayName: Kafka Rest Class
            kind: KafkaRestClass
            name: kafkarestclasses.platform.confluent.io
            version: v1beta1
          - description: KafkaRestProxy is the schema for the Kafka REST Proxy API.
            displayName: Kafka Rest Proxy
            kind: KafkaRestProxy
            name: kafkarestproxies.platform.confluent.io
            version: v1beta1
          - description: Kafka is the schema for the Kafka API.
            displayName: Kafka
            kind: Kafka
            name: kafkas.platform.confluent.io
            version: v1beta1
          - description: KafkaTopic is the schema for the Kafka Topic API.
            displayName: Kafka Topic
            kind: KafkaTopic
            name: kafkatopics.platform.confluent.io
            version: v1beta1
          - description: KRaftController is the schema for the KRaft Controller API.
            displayName: KRaft Controller
            kind: KRaftController
            name: kraftcontrollers.platform.confluent.io
            version: v1beta1
          - description: KRaftMigrationJob is the schema for the KRaftMigrationJob
              API.
            displayName: KRaft Migration Job
            kind: KRaftMigrationJob
            name: kraftmigrationjobs.platform.confluent.io
            version: v1beta1
          - description: KsqlDB is the schema for the ksqlDB API.
            displayName: Ksql DB
            kind: KsqlDB
            name: ksqldbs.platform.confluent.io
            version: v1beta1
          - description: SchemaExporter is the schema for the SchemaExporter API.
            displayName: Schema Exporter
            kind: SchemaExporter
            name: schemaexporters.platform.confluent.io
            version: v1beta1
          - description: SchemaRegistry is the schema for the Schema Registry API.
            displayName: Schema Registry
            kind: SchemaRegistry
            name: schemaregistries.platform.confluent.io
            version: v1beta1
          - displayName: Schema
            kind: Schema
            name: schemas.platform.confluent.io
            version: v1beta1
          - description: Zookeeper is the schema for the Zookeeper API.
            displayName: Zookeeper
            kind: Zookeeper
            name: zookeepers.platform.confluent.io
            version: v1beta1
        description: Confluent for Kubernetes (CFK) is a cloud-native control plane
          for deploying and managing Confluent in your private cloud environment.
          It provides standard and simple interface to customize, deploy, and manage
          Confluent Platform through declarative API.Confluent for Kubernetes runs
          on Kubernetes, the runtime for private cloud architectures.
        displayName: Confluent for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Confluent
        - Kafka
        - Kubernetes
        - Private Cloud
        - Cloud-Native
        - Event Streaming
        - Data in Motion
        links:
        - name: Documentation
          url: https://docs.confluent.io/operator/current/overview.html
        - name: Examples and Tutorials
          url: https://github.com/confluentinc/confluent-kubernetes-examples
        - name: Product Overview (Blog)
          url: https://www.confluent.io/blog/confluent-for-kubernetes-offers-cloud-native-kafka-automation/
        maintainers:
        - email: confluent-private-cloud-team@confluent.io
          name: Confluent Private Cloud team
        maturity: stable
        provider:
          name: Confluent, Inc.
          url: https://confluent.io
        relatedImages:
        - docker.io/confluentinc/cp-kafka-rest@sha256:ba49d46654770994c302d586a609afdfc58a7cc05f204aa090b643f0680a5014
        - docker.io/confluentinc/cp-server-connect@sha256:e4cbc1836402497ea42b2c961c5d6b4420c25b3c6a5f6f71be3e99fcc0f09702
        - docker.io/confluentinc/cp-enterprise-control-center-next-gen@sha256:6378a0174433685ee52a7fe2e48424d3f598827ed2a5570ec74b916102485e0d
        - docker.io/confluentinc/cp-ksqldb-server@sha256:a283f2a72b6346febaf711e71dee1e0e3666d7265320f3f69307fc2c0b95d5ef
        - docker.io/confluentinc/cp-schema-registry@sha256:3dcb9652a4c83d4566a7334041c1ac3ec359665a9450a70c5d7b41adaea55c90
        - registry.connect.redhat.com/confluentinc/confluent-for-kubernetes-v2@sha256:35f2799f71c308d87e6a7dada00263abb9e08e895b10f10c815e7a9edf19dd49
        - docker.io/confluentinc/confluent-init-container@sha256:82277228368a783ba84367c920a9fa15fad6844723df1e16b07c9dcc829a54ad
        - docker.io/confluentinc/cp-enterprise-alertmanager@sha256:fcce6f5311c26cd58427a039530a9f8dfd009249ff5a519694aa945d34502252
        - docker.io/confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - confluentinc/confluent-operator@sha256:1cd3fdaea26536bfbe3a42b0061f11d31ffd7dc552c0630ad74e53e97bd981de
        - docker.io/confluentinc/cp-enterprise-prometheus@sha256:db0989ef9859dbc88bc59f9b8ade86f2ba13e9dc36533e654566218328e5ff57
        - docker.io/confluentinc/cp-server@sha256:ffd5961412da42c81c9b20f35e334aa8bdf351dcb452610e84156ed511116ef5
        version: 3.0.0
      entries:
      - name: confluent-for-kubernetes.v3.0.0
        version: 3.0.0
      name: "3.0"
    defaultChannel: "3.0"
    packageName: confluent-for-kubernetes
    provider:
      name: Confluent, Inc.
      url: https://confluent.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: https://linbit.com
    name: linstor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: dev
    - currentCSV: linstor-operator.v2.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorCluster",
                "metadata": {
                  "name": "linstorcluster"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorNodeConnection",
                "metadata": {
                  "name": "example-connection"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatellite",
                "metadata": {
                  "name": "example-satellite"
                },
                "spec": {
                  "clusterRef": {
                    "name": "linstorcluster"
                  }
                },
                "status": {
                  "conditions": []
                }
              },
              {
                "apiVersion": "piraeus.io/v1",
                "kind": "LinstorSatelliteConfiguration",
                "metadata": {
                  "name": "all-satellites"
                },
                "spec": {},
                "status": {
                  "conditions": []
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          com.redhat.openshift.versions: v4.7
          containerImage: drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
          createdAt: "2025-09-25T13:44:22Z"
          description: LINSTOR Operator deploys LINBIT SDS in the cluster.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: linbit-sds
          operators.openshift.io/valid-subscription: '[''LINBIT SDS Subscription'']'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/internal-objects: '["linstorsatellites.piraeus.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster-wide configuration for LINSTOR.
            displayName: Linstor Cluster
            kind: LinstorCluster
            name: linstorclusters.piraeus.io
            version: v1
          - description: Configuration for connections between specific nodes.
            displayName: Linstor Node Connection
            kind: LinstorNodeConnection
            name: linstornodeconnections.piraeus.io
            version: v1
          - description: Configuration for specific Satellites/Nodes in the LINSTOR
              Cluster.
            displayName: Linstor Satellite Configuration
            kind: LinstorSatelliteConfiguration
            name: linstorsatelliteconfigurations.piraeus.io
            version: v1
          - description: Resource to configure a single satellite, managed by the
              Operator itself.
            displayName: Linstor Satellite
            kind: LinstorSatellite
            name: linstorsatellites.piraeus.io
            version: v1
          required:
          - kind: ServiceMonitor
            name: servicemonitors.monitoring.coreos.com
            version: v1
          - kind: PrometheusRule
            name: prometheusrules.monitoring.coreos.com
            version: v1
          - kind: PodMonitor
            name: podmonitors.monitoring.coreos.com
            version: v1
        description: |
          The LINSTOR Operator manages [LINSTOR](https://github.com/LINBIT/linstor-server) clusters in Kubernetes.

          All components of the LINSTOR software stack can be managed by the operator:
          * DRBD
          * LINSTOR
          * LINSTOR CSI driver
          * LINSTOR HA Controller
        displayName: LINSTOR Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        links:
        - name: User Guide
          url: https://linbit.com/drbd-user-guide/linstor-guide-1_0-en/
        - name: LINSTOR Operator
          url: https://charts.linstor.io/
        - name: Source Code
          url: https://github.com/LINBIT/linstor-operator-builder
        maintainers:
        - email: support@linbit.com
          name: LINBIT
        maturity: stable
        minKubeVersion: 1.20.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: ""
          kind: ServiceAccount
          version: v1
        - group: apps
          kind: DaemonSet
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRole
          version: v1
        - group: rbac.authorization.k8s.io
          kind: ClusterRoleBinding
          version: v1
        - group: rbac.authorization.k8s.io
          kind: Role
          version: v1
        - group: rbac.authorization.k8s.io
          kind: RoleBinding
          version: v1
        - group: storage.k8s.io
          kind: CSIDriver
          version: v1
        provider:
          name: LINBIT
          url: https://linbit.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:624f61b1ae9d32f5a074ee8c12839d975fba9efcd3bfa7981cfe38febf5c4495
        - drbd.io/linstor-controller@sha256:841b3aa84cdf13aaaf1243bc5da8ac2205ca3fe0186fdfc7a04b69397f0020d6
        - drbd.io/linstor-csi@sha256:9c7ef452b0997d3373803201614c3b8075687e75c397359738a9e4ef5530cab3
        - drbd.io/linstor-k8s-ha-controller@sha256:ed0f0f86dad88feadf8c3c32d7d143eaa52b1130af3576315a562909b39d2afa
        - drbd.io/linstor-affinity-controller@sha256:78e5b299ef1702114528678387b5e74821d7658bc02f05f0bd7edd9a176fba39
        - drbd.io/drbd9-rhel9@sha256:4eb553336ed64e127068a4566370b040166bf11301e21ac47ac5b92237ec9d2a
        - drbd.io/ktls-utils@sha256:3351a6e0fc048fe1544499095cdb8fdd221e832cecfa8a9a01f046c271991138
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:700be620923fc99ed03f9f1c4a911e4084beb30685361270c49165b27ad1d374
        - drbd.io/drbd-reactor@sha256:77c51aeeefdcd788cb1583738f93778032fcc03c36fb289b692d3cf195c8b184
        - registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:ac3963bfd4246da386b82d810a348e9d42a9e4b746acaf4dbcbe0766643cc2fa
        - drbd.io/drbd9-rhel7@sha256:0c96b216e55a128a9944276af9080aced8bd72fdf1fd424f024ed9ace2c1feb5
        - drbd.io/drbd-shutdown-guard@sha256:7ba049ecb205a0086e425d2ff37285b00449549d89e179b4b943ace6e59a9156
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:d0ef8a6a9e878952fdbc34144afee101fd837276d9f62cab5c2dd4f8a4e5a113
        - drbd.io/drbd9-rhel8@sha256:7cddfa770de66cefcb42c2b33da57d0f1adfcc5f1cc12522977aa6993722bb2a
        - drbd.io/linstor-operator@sha256:e7f21a7195aa72075e8ed4eedae28315cc6835d3f4e752c719d3c5d67f69da3f
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:2669cde3db2268d37bc671f494eaa06480478a90b8943054af14cbffc050debf
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:cd58f1789027e32eae08f00dac7b2275496c41020fb7cdebc10e9986c625d4f4
        - drbd.io/linstor-satellite@sha256:ce0a0b05494c6e6d5f16121e5762bf58e7b78f40c00b84780e2abba2a690af1b
        - drbd.io/drbd9-rhel10@sha256:da0a7015cf15724a7c73100d505891e62262a726141347df7f2d51ec79a0a1e1
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:f8cde5b5c289f7b3d233773f4e79990ba392fe6e14e0177de80714e7d3e98862
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:5a650dc21a6a038de6267f117d01adfd904ff3f5c94f428f6a13735f18bf94d6
        version: 2.9.1
      entries:
      - name: linstor-operator.v2.9.1
        version: 2.9.1
      name: stable
    defaultChannel: dev
    packageName: linstor-operator
    provider:
      name: LINBIT
      url: https://linbit.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Internal
      provider-url: redhat.com
    name: l5-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: l5-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pets.bestie.com/v1",
                "kind": "Bestie",
                "metadata": {
                  "name": "bestie"
                },
                "spec": {
                  "image": "quay.io/opdev/bestie",
                  "maxReplicas": 10,
                  "size": 3,
                  "version": "1.3"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: quay.io/opdev/l5-operator:v0.0.3
          createdAt: "2023-06-27T16:56:10Z"
          description: L5 Operator is a demo operator that aims to demonstrate level
            5 capabilities for web applications.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          relatedImages: quay.io/opdev/bestie@sha256:0e99ada2b1de19043845dd9076789584aeddc2646402e5f0987dd00424be1ee7
          repository: https://github.com/opdev/l5-operator-demo
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Bestie is the Schema for the besties API.
            displayName: Bestie
            kind: Bestie
            name: besties.pets.bestie.com
            version: v1
        description: |
          ## Usage
          The L5 operator is meant to demonstrate a minimum viable prototype of all capability levels outilned in the operator framework. The L5 operator is compatible with both Kubernetes and OpenShift clusters. When running on vanilla kubernetes a few prerequisites need to be kept in mind which may not be available out of the box. These include an ingress controller and a prometheus operator for level 4 capability. The L5 operator also depends on the Crunchy Data PostgreSQL operator. On openshift when installing via the Operator Lifecycle Manager, this will automatically be installed when installing the operator via the embdeded operator hub. The Operator Lifecycle Manager is also available for vanilla kubernetes however it may not be installed out of the box. In this case the Crunchy Data PostgreSQL operator must be installed manually. The Crunchy Data PostgreSQL operator can be installed on vanilla kuberentes using the [instructions provided by Crunchy Data](https://access.crunchydata.com/documentation/postgres-operator/4.7.5/installation/postgres-operator/) More documentation and examples can be found on [github](https://github.com/opdev/l5-operator-demo).
        displayName: L5 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - l5-operator
        links:
        - name: L5 Operator
          url: https://opdev.github.io/l5-operator-demo/
        maintainers:
        - email: partnereng@redhat.com
          name: OpDev Team
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Internal
          url: redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/opdev/l5-operator:v0.0.3
        - quay.io/openshift-community-operators/l5-operator@sha256:e4cb13801a80753c99a47f009ca0a29416b3e04afe5314d07d97c2464a6f0d75
        version: 0.0.3
      entries:
      - name: l5-operator.v0.0.3
        version: 0.0.3
      - name: l5-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: l5-operator
    provider:
      name: Red Hat Internal
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/xcrypt-operator-rhmp/support?utm_source=openshift_console
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.marketplace.redhat.com/rhm/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:712d31214423b326ab8f107435c85431ce8d99c2e52fd8d615005a4274e0ff98
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.marketplace.redhat.com/rhm/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator-rhmp
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kubecost
      provider-url: https://www.kubecost.com
    name: kubecost-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubecost-operator.v2.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kubecost.com/v1alpha1",
                "kind": "CostAnalyzer",
                "metadata": {
                  "name": "costanalyzer-sample"
                },
                "spec": {
                  "clusterController": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256",
                      "tag": "1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1"
                    }
                  },
                  "forecasting": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd"
                  },
                  "global": {
                    "platforms": {
                      "openshift": {
                        "createMonitoringClusterRoleBinding": true,
                        "createMonitoringResourceReaderRoleBinding": true,
                        "enabled": true,
                        "scc": {
                          "clusterController": true,
                          "networkCosts": true,
                          "nodeExporter": true
                        }
                      }
                    },
                    "prometheus": {
                      "enabled": false,
                      "fqdn": "https://prometheus-k8s.openshift-monitoring.svc.cluster.local:9091",
                      "kubeRBACProxy": true
                    }
                  },
                  "grafana": {
                    "image": {
                      "repository": "registry.redhat.io/rhel9/grafana@sha256",
                      "tag": "9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1"
                    },
                    "sidecar": {
                      "image": {
                        "repository": "quay.io/kiwigrid/k8s-sidecar@sha256",
                        "tag": "d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23"
                      }
                    }
                  },
                  "kubecostAggregator": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "kubecostFrontend": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8"
                  },
                  "kubecostModel": {
                    "fullImageName": "registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef"
                  },
                  "networkCosts": {
                    "image": {
                      "repository": "registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256",
                      "tag": "c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0"
                    }
                  },
                  "prometheusRule": {
                    "enabled": true
                  },
                  "serviceMonitor": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          createdAt: "2025-08-08T19:34:34Z"
          description: Deploys Kubecost 2.8.2 via the Helm chart.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Contact Kubecost for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Kubecost.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The CostAnalyzer API describes a deployment of Kubecost and
              its dependencies.
            displayName: CostAnalyzer
            kind: CostAnalyzer
            name: costanalyzers.charts.kubecost.com
            version: v1alpha1
        description: |2-

          IMPORTANT - Before installation, you must run the following command to allow monitoring of the Kubecost namespace by the local OpenShift Promtheus. If you install in a namespace other than "Kubecost", adjust accordingly.

          `oc label namespace kubecost openshift.io/cluster-monitoring=true`

          Kubecost provides real-time cost visibility and insights for teams using Kubernetes, helping you continuously reduce your cloud costs.

          This OpenShift Operator deploys Kubecost from the official Helm chart.

          Kubecost allows you to:

          * See allocated spend across all native Kubernetes concepts and provide your teams with transparent, accurate cost data reconciled with your actual cloud bill.
          * Join in-cluster costs like CPU and memory with out-of-cluster spend from other cloud services.
          * Get complete cost visibility across OpenShift, On-Prem, AWS, GCP, Azure, Oracle, AliCloud + more.
          * Gain insights you can use to save 30-50%+ on your infrastructure spend, without exposing your private information — your data never gets shared externally, even with us.
          * Have real-time alerts and recurring reports, empowering teams to take control of their Kubernetes-enabled infrastructure and stay within budgeted limits.

          Visit the Kubecost website for more details including features, pricing, documentation, and more at www.kubecost.com.
        displayName: Kubecost
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kubernetes
        - cost
        - kubernetes costs
        - kubernetes optimization
        - k8s
        - cost monitoring
        - cloud costs
        - k8s cost
        links:
        - name: Kubecost
          url: https://www.kubecost.com
        maintainers:
        - email: support@kubecost.com
          name: Kubecost Support
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Kubecost
          url: https://www.kubecost.com
        relatedImages:
        - registry.connect.redhat.com/kubecost/kubecost-modeling@sha256:8f59f8fd8696964c1b5322ca79fae78befe97323fa56c2b509bdcc67ee4c0edd
        - quay.io/kiwigrid/k8s-sidecar@sha256:d3749777d5bedf62549b452a7e2ccb205da53588dc7f04927efc75d0da5b6d23
        - registry.connect.redhat.com/kubecost/operator@sha256:2cb5ec186af3264cba42333245616dd01f0f9868a849129b1a31afbc51ee1410
        - registry.connect.redhat.com/kubecost/kubecost-cluster-controller@sha256:1a6f11dcd7341ab3f24eba23ec06c7b75646a534a7ed960153601806425469d1
        - registry.connect.redhat.com/kubecost/kubecost-cost-model@sha256:1b9c9f309e6f41addc7fb6418724a6a6b08486290e31fe1f614efc649874bcef
        - registry.connect.redhat.com/kubecost/kubecost-network-costs@sha256:c444d475be4f1aa36680bd84358e1e07303dcfe5d13a95544ad9e59f80d4a5f0
        - registry.redhat.io/rhel9/grafana@sha256:9f71de028c18034e07a8d15dac13ef68afb91cb899b9e8e7d39703547dd0eac1
        - registry.connect.redhat.com/kubecost/kubecost-frontend@sha256:f9b2196dc6ac017b8074f8fe40f1a58e7a2630282c5b7a019ca9f3dbdb670dc8
        - registry.connect.redhat.com/kubecost/kubecost-operator-manager@sha256:582223a66a77f35d1f7ca33460032cfca1581a8ff30f75ead289d8cd4cf62bb9
        version: 2.8.2
      entries:
      - name: kubecost-operator.v2.8.2
        version: 2.8.2
      - name: kubecost-operator.v2.8.1
        version: 2.8.1
      - name: kubecost-operator.v2.8.0
        version: 2.8.0
      - name: kubecost-operator.v2.7.2
        version: 2.7.2
      - name: kubecost-operator.v2.7.1
        version: 2.7.1
      - name: kubecost-operator.v2.5.4
        version: 2.5.4
      name: alpha
    defaultChannel: alpha
    packageName: kubecost-operator
    provider:
      name: Kubecost
      url: https://www.kubecost.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Harness Inc.
      provider-url: https://harness.io
    name: harness-gitops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: harness-gitops-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gitops.harness.io/v1alpha1",
                "kind": "GitopsAgent",
                "metadata": {
                  "annotations": {
                    "containerImage": "harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde",
                    "createdAt": "2025-07-31T12:11:18Z"
                  },
                  "name": "gitops-agent-sample"
                },
                "spec": {
                  "identity": {
                    "accountIdentifier": "accountIdentifier",
                    "agentIdentifier": "agentIdentifier",
                    "agentName": "agentName"
                  },
                  "replicas": 1,
                  "secrets": {
                    "agentSecret": "agentSecret"
                  },
                  "target": {
                    "http": "https://app.harness.io/gitops/",
                    "protocol": "HTTP1"
                  },
                  "upgrade": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
          createdAt: "2025-07-31T06:41:18Z"
          description: Harness operator to manage and provision GitOps Agent
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Harness Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Schema for the Gitops Agents API
            displayName: Gitops Agent
            kind: GitopsAgent
            name: gitopsagents.gitops.harness.io
            version: v1alpha1
        description: |-
          ## Overview
          Harness Gitops Operator manages the full lifecycle for [Gitops Agent](https://developer.harness.io/docs/continuous-delivery/gitops/connect-and-manage/install-a-harness-git-ops-agent/)

          The operator aims to provide the following:

            * Easy configuration and installation of Harness GitOps Agent.
            * Provide seamless upgrades for Harness GitOps Agent.
            * Monitoring and autorecovery for GitOps Agent.

          ## Usage
          Deploy gitops-agent in namespace where operator is installed with basic setup.

          ### Prerequisites:
            * Go to your Harness account and create a new gitops agent
            * Copy account, org and project identifiers
            * Retrieve the agent secret key

          ```
          apiVersion: gitops.harness.io/v1alpha1
          kind: GitopsAgent
          metadata:
            name: gitops-agent-sample
          spec:
            identity:
              accountIdentifier: accountIdentifier
              agentIdentifier: agentIdentifier
              agentName: agentName
            secrets:
              agentSecret: agentSecret
            target:
              http: https://app.harness.io/gitops/
              protocol: HTTP1
            replicas: 1
            upgrade:
              enabled: true
          ```
          **Note:** Gitops Agent require an active ArgoCD instance running in the same namespace where agent is installed.

          See [Harness Documentation](https://developer.harness.io/) for more information.
        displayName: Harness GitOps Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - gitops
        - harness
        - agent
        - cd
        - argocd
        - kubernetes
        links:
        - name: Harness Gitops Operator
          url: https://harness-gitops-operator.domain
        maintainers:
        - email: alexandru.bosii@harness.io
          name: Alexandru Bosii
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Harness Inc.
          url: https://harness.io
        relatedImages:
        - harness/gitops-agent-controller@sha256:e4342fb1599538d2b6145e644f1e302b2783a48347c53a1da7ec8e2454df5fde
        - registry.connect.redhat.com/harness/gitops-operator@sha256:f97cdc2245b7133a04235c55540d3a87165e080b1d5b9572a663731f6ccbe597
        - docker.io/harness/gitops-agent@sha256:c9f57338be294d32a73a4cbecf90dfff50f5d36a699f3ec32d962dcb64d8d085
        version: 0.3.0
      entries:
      - name: harness-gitops-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: harness-gitops-operator
    provider:
      name: Harness Inc.
      url: https://harness.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weaveworks
      provider-url: ""
    name: tf-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tf-controller.v0.9.0-rc.8
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"infra.contrib.fluxcd.io/v1alpha1\",\n
            \   \"kind\": \"Terraform\",\n    \"metadata\": {\n        \"name\": \"helloworld-tf\",\n
            \       \"namespace\": \"flux-system\"\n    },\n    \"spec\": {\n        \"interval\":
            \"5m\",\n        \"approvePlan\": \"auto\",\n        \"path\": \"./\",\n
            \       \"sourceRef\": {\n          \"kind\": \"GitRepository\",\n          \"name\":
            \"helloworld\",\n          \"namespace\": \"flux-system\"\n        }      \n
            \   }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
          description: Weave TF-controller is a Flux controller that brings GitOps
            to Terraform resources.
          operatorframework.io/suggested-namespace: flux-system
          support: Weaveworks
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Terraform
            displayName: Terraform
            kind: Terraform
            name: terraforms.infra.contrib.fluxcd.io
            version: v1alpha1
        description: "**Weave GitOps Terraform Controller** (aka Weave TF-controller)
          is a Flux controller for reconciling Terraform resources in the GitOps way.
          \nWith the power of Flux together with Terraform, TF-controller allows you
          to GitOps-ify infrastructure, \nand application resources, in the Kubernetes
          and Terraform universe, together at your own pace.\n\nTF-controller offers
          many GitOps models:\n  1. **GitOps Automation Model:** GitOps your Terraform
          resources from the provision steps to the enforcement steps, like a whole
          EKS cluster.\n  2. **Hybrid GitOps Automation Model:** GitOps parts of your
          existing infrastructure resources. For example, you have an existing EKS
          cluster. You can choose to GitOps only its nodegroup, or its security group.\n
          \ 3. **State Enforcement Model:** You have a TFSTATE file, and you'd like
          to use GitOps enforce it, without changing anything else.\n  4. **Drift
          Detection Model:** You have a TFSTATE file, and you'd like to use GitOps
          just for drift detection, so you can decide to do things later when a drift
          occurs.\n\nFeatures\n  * GitOps Automation for Terraform\n  * Drift detection\n
          \ * Plan and Manual Approve"
        displayName: Weave GitOps Terraform Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - weave
        - flux
        - cd
        - continuous delivery
        - terraform
        - gitops
        links:
        - name: Weave GitOps Terraform Controller
          url: https://github.com/weaveworks/tf-controller
        - name: Documentation
          url: https://weaveworks.github.io/tf-controller/
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Weaveworks
        relatedImages:
        - ghcr.io/weaveworks/tf-controller:v0.9.0-rc.8
        - quay.io/openshift-community-operators/tf-controller@sha256:cdaeb470d9b331a3fa08167e5c904befb00be983e1354651670e21e13560bf9b
        version: 0.9.0-rc.8
      entries:
      - name: tf-controller.v0.9.0-rc.8
        version: 0.9.0-rc.8
      - name: tf-controller.v0.9.0-rc.3
        version: 0.9.0-rc.3
      name: stable
    defaultChannel: stable
    packageName: tf-controller
    provider:
      name: Weaveworks
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: sosivio
      provider-url: https://www.sosiv.io
    name: sosivio
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sosivio-operator.v1.5.1-97
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              },
              {
                "apiVersion": "sosivio.sosiv.io/v1alpha1",
                "kind": "Sosivio",
                "metadata": {
                  "name": "sosivio"
                },
                "spec": {
                  "platform": "ocp4",
                  "useSha": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Monitoring,Integration & Delivery,Logging & Tracing,Developer
            Tools
          com.redhat.openshift.versions: v4.1-v4.1
          containerImage: quay.io/sosivio/sosivio-operator:v1.5.1-97
          createdAt: "2021-10-28 01:01:01"
          description: Sosivio | Predictive Troubleshooting for Kubernetes
          operatorhub.io/ui-metadata-max-k8s-version: "1.25"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Sosivio/operator
          support: Sosivio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Sosivio is the Schema for the sosivios API
            displayName: Sosivio
            kind: Sosivio
            name: sosivios.sosivio.sosiv.io
            version: v1alpha1
        description: "# Sosivio Operator\nThe Sosivio Operator for RedHat’s Openshift
          Container Platform allows users to quickly and easily deploy Sosivio’s Kubernetes
          Predictive Troubleshooting Platform on your cluster.\n\n# Sosivio Overview\n\nSosivio
          is the first and only non-intrusive and predictive autonomous cloud-native
          platform built specifically for Kubernetes from the ground up. Sosivio provides
          predictive troubleshooting tools, root cause determination and instant remediation
          for any failures in Kubernetes environments.\n\nPredict and Prevent!\n\nThe
          Operator will deploy and manage Sosivio on Kubernetes and OpenShift. Included
          are the following features;\n-   **Root Cause Analysis and Remediation**
          - predictive troubleshooting to prevent failures before they occur. Instantly
          understand where the root cause is and how to fix it.\n-   **Application
          Profiling** - analyze container resource consumption and recommendations
          for requests and limits.\n-   **Real-time Metrics** - in-depth, granular,
          and accurate real-time data from the process level inside containers (and
          pods) with failure correlation.\n-   **Cluster Health Checks** - proactive
          health checks for cluster infrastructure and applications running on the
          cluster.\n-   **Multi-Cluster View** - management, monitoring, and maintenance
          capabilities for all clusters from one location. \n-   **Kubernetes Components
          Alerting** - notifications sent through Slack, email and other 3rd party
          platforms.\n\n# Installation\nSosivio is completely **non-intrusive** and
          **does not require any persistent storage**. There is **no data offloading**
          outside of your Kubernetes cluster. Applying our Data Swirling methodology
          enables the product to perform at a **remarkably low resource footprint**
          and requires **no persistent volume configurations**.\n\n\nSosivio is installed
          under one namespace (labeled: \"sosivio\") which creates all the required
          components for the product.\n\nAll users will start with 4 weeks of Sosivio
          Premium after which Sosivio will automatically convert to the Community
          Version (free forever).\n\n# Help\nMore information about Sosivio and our
          product usage instructions can be found in our documentation [here](https://docs.sosiv.io/)
          \ \n"
        displayName: Sosivio | Predictive Troubleshooting for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sosivio
        - kubernetes
        - observanility
        - monitoring
        - root cause
        links:
        - name: Sosivio
          url: https://www.sosiv.io
        - name: Documentation
          url: https://docs.sosiv.io
        maintainers:
        - email: info@sosiv.io
          name: Sosivio
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: sosivio
          url: https://www.sosiv.io
        relatedImages:
        - quay.io/sosivio/kube-rbac-proxy:v0.8.0
        - quay.io/sosivio/network-aggregator@sha256:e59d937f4d9c52acad0adadab39c6849d20b4ee72a83e239a33ea137f08323f4
        - quay.io/sosivio/sosivio-operator@sha256:11bc627773bb5e08b629e0b013e7f600fcb4aaefbd97f2de337d339475616763
        - quay.io/sosivio/analyzer@sha256:3825608fe7c43ef059108e76665acc99fe72bea77d4d9e3852f6ad45e74379b1
        - quay.io/sosivio/kube-rbac-proxy@sha256:34e8724e0f47e31eb2ec3279ac398b657db5f60f167426ee73138e2e84af6486
        - quay.io/sosivio/crud-manager@sha256:38289a28829fa737917d76dd4810ba6140e123061ead0a4b79fa79b2e246a293
        - quay.io/sosivio/sosivio-node-netpacket@sha256:84dd159b1dfdb52dee3eb1c4c93f323bae49ce517ff44663b69978e14bdd8390
        - quay.io/sosivio/classifier@sha256:dd8bf145524db23ff333efa1bcdff258dab116d3534f95f73775d8dd339e6963
        - quay.io/sosivio/sosivio-node-pinger@sha256:937de52ad7b76e6cff4ad5197de9705988d61f285ebadac5ebadb5856f59ee58
        - quay.io/sosivio/sosivio-node-collector@sha256:61ed8dd52de5edc09d16a31b18f2f65ad9344c0b6c9cfab323c3eb3d7e7ed060
        - quay.io/sosivio/discovery-engine@sha256:b87b3f7c9ed3aed80d0a77788fd8925789450d9dc05d25017e89dff611cd82da
        - quay.io/sosivio/notifier@sha256:35f5f3e8e5ccad914fcfe34fb4b4a1b044ee87f5a6141284cd5273f0b2b3bf22
        - quay.io/sosivio/correlation-entities@sha256:c00af09b2589bad4740fd44d7a81c69b2e29138c360939b2a45f65edd44b6901
        - quay.io/sosivio/actuator@sha256:a9abe9730800ca2b0a1bbd7a361b2ab3a818cbf127798314e5477fb1cb16ba4c
        - quay.io/sosivio/sosivio-dashboard@sha256:fdfca5bb2123d1008cec2b656cde73c7e9364f84c5ff4f4a6fee5751b1bfb2f8
        - quay.io/openshift-community-operators/sosivio@sha256:110d7f4fb993d65ff497eca57c42206faab07dc5f5b1f43b1b682387e7cd3a10
        - quay.io/sosivio/nsq@sha256:781aac23ef0a1ed0af2b8e93d625ae49bd11dbdd98787d3c79dde8d128f3810d
        - quay.io/sosivio/sosivio-operator:v1.5.1-97
        - quay.io/sosivio/authentication@sha256:0ea6025f333576792f5b1fc4d96d9ab1a404749f185830410e71248683004da2
        - quay.io/sosivio/sequence-recognition@sha256:7335cf8f3e40bc46e8ca0de525d99f0655833b7c52d36dc273aa7b096eb70fa2
        - quay.io/sosivio/sosiviodb@sha256:a3c6c3975bc2b0316135ee5c547f1eb215f3a1f9c8b0a9c8eceff23415f93597
        - quay.io/sosivio/communicator@sha256:fffec9a6c66594fde5f05e2dfb84361544ca9d53a345188c197a0ea6f53bc4b9
        - quay.io/sosivio/contract-testing@sha256:cad6fac6ed3fdf2597ceb09ba5dbb7d81f70cf45f483a7064e3968215dd982e3
        version: 1.5.1
      entries:
      - name: sosivio-operator.v1.5.1-97
        version: 1.5.1
      - name: sosivio-operator.v1.5.0-95
        version: 1.5.0
      - name: sosivio-operator.v1.4.4-85
        version: 1.4.4
      - name: sosivio-operator.v1.4.3-83
        version: 1.4.3
      - name: sosivio-operator.v1.4.2-82
        version: 1.4.2
      - name: sosivio-operator.v1.4.1-81
        version: 1.4.1-81
      - name: sosivio-operator.v1.4.1.1-80
        version: 1.4.1-80
      - name: sosivio-operator.v1.4.1-78
        version: 1.4.1-79
      - name: sosivio-operator.v1.3.1-76
        version: 1.3.1-76
      - name: sosivio-operator.v1.3.0-75
        version: 1.3.0-75
      - name: sosivio-operator.v1.2.1-1
        version: 1.2.1-1
      - name: sosivio-operator.v1.2.0-1
        version: 1.2.0-1
      name: stable
    defaultChannel: stable
    packageName: sosivio
    provider:
      name: sosivio
      url: https://www.sosiv.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operator-kogitocloud: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: 1.x
    - currentCSV: kogito-operator.v1.44.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "process-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogito-infinispan-infra"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "infinispan.org/v1",
                    "kind": "Infinispan"
                  }
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "process-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "app.kiegroup.org/v1beta1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/kiegroup/kogito-operator:1.44.1
          createdAt: "2019-08-22T13:12:22Z"
          description: Kogito Operator for deployment and management of Kogito services.
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.app.kiegroup.org
            version: v1beta1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.app.kiegroup.org
            version: v1beta1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.app.kiegroup.org
            version: v1beta1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.app.kiegroup.org
            version: v1beta1
        description: |-
          Kogito Operator is designed to deploy Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Creates two kinds of build configuration: Native or JVM. For more information refer to [Native build and deployment](https://docs.jboss.org/kogito/release/latest/html_single/#_native_build_and_deployment) section on our docs
          * Builds a Kogito service from a remote git repository and creates a custom image for the given runtime: Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Installs the [Kogito Data Index Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-data-index-service_kogito-deploying-on-openshift) to enable data index feature for Kogito services.
          * Installs the [Kogito Jobs Service](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-with-jobs-service_kogito-deploying-on-openshift) to enable jobs scheduling.

          For Kubernetes deployment of Kogito services, please refer to the section [Deploying Kogito services on Kubernetes](https://docs.jboss.org/kogito/release/latest/html_single/#proc-kogito-deploying-on-kubernetes_kogito-deploying-on-openshift) on our docs.

          **IMPORTANT!** Kogito Operator might need Infinispan Operator and Strimzi to implement persistence and messaging use cases if you do not provide these services beforehand.
          If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the Kogito Operator.
        displayName: Kogito
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://kogito.kie.org/
        - name: Documentation
          url: https://github.com/kiegroup/kogito-operator/blob/main/README.md
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/kogito-operator@sha256:28cd33562ab56c44d65486984cac7edfbdd1336ced1277f1af7df6c2a8073e3e
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.12.0
        - quay.io/kiegroup/kogito-operator:1.44.1
        version: 1.44.1
      entries:
      - name: kogito-operator.v1.44.1
        version: 1.44.1
      - name: kogito-operator.v1.43.0
        version: 1.43.0
      - name: kogito-operator.v1.42.0
        version: 1.42.0
      - name: kogito-operator.v1.41.0
        version: 1.41.0
      - name: kogito-operator.v1.40.0
        version: 1.40.0
      - name: kogito-operator.v1.39.0
        version: 1.39.0
      - name: kogito-operator.v1.38.0
        version: 1.38.0
      - name: kogito-operator.v1.37.0
        version: 1.37.0
      - name: kogito-operator.v1.36.0
        version: 1.36.0
      - name: kogito-operator.v1.35.0
        version: 1.35.0
      - name: kogito-operator.v1.34.0
        version: 1.34.0
      - name: kogito-operator.v1.33.0
        version: 1.33.0
      - name: kogito-operator.v1.32.0
        version: 1.32.0
      - name: kogito-operator.v1.31.0
        version: 1.31.0
      - name: kogito-operator.v1.30.0
        version: 1.30.0
      - name: kogito-operator.v1.29.0
        version: 1.29.0
      - name: kogito-operator.v1.28.0
        version: 1.28.0
      - name: kogito-operator.v1.27.0
        version: 1.27.0
      - name: kogito-operator.v1.26.0
        version: 1.26.0
      - name: kogito-operator.v1.25.0
        version: 1.25.0
      - name: kogito-operator.v1.24.0
        version: 1.24.0
      - name: kogito-operator.v1.23.0
        version: 1.23.0
      - name: kogito-operator.v1.22.1
        version: 1.22.1
      - name: kogito-operator.v1.22.0
        version: 1.22.0
      - name: kogito-operator.v1.21.0
        version: 1.21.0
      - name: kogito-operator.v1.20.0
        version: 1.20.0
      - name: kogito-operator.v1.19.0
        version: 1.19.0
      - name: kogito-operator.v1.18.0
        version: 1.18.0
      - name: kogito-operator.v1.17.0
        version: 1.17.0
      - name: kogito-operator.v1.16.1
        version: 1.16.1
      - name: kogito-operator.v1.16.0
        version: 1.16.0
      - name: kogito-operator.v1.15.0
        version: 1.15.0
      - name: kogito-operator.v1.14.0
        version: 1.14.0
      - name: kogito-operator.v1.13.0
        version: 1.13.0
      - name: kogito-operator.v1.12.0
        version: 1.12.0
      - name: kogito-operator.v1.11.1
        version: 1.11.1
      - name: kogito-operator.v1.11.0
        version: 1.11.0
      - name: kogito-operator.v1.10.0
        version: 1.10.0
      - name: kogito-operator.v1.9.1
        version: 1.9.1
      - name: kogito-operator.v1.9.0
        version: 1.9.0
      - name: kogito-operator.v1.8.0
        version: 1.8.0
      - name: kogito-operator.v1.7.0
        version: 1.7.0
      - name: kogito-operator.v1.6.0
        version: 1.6.0
      - name: kogito-operator.v1.5.0
        version: 1.5.0
      - name: kogito-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: 1.x
    packageName: kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: metallb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: metallb-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "basic",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 64512,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            },
                            "toReceive": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-neighs",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64512,
                        "neighbors": [
                          {
                            "address": "172.30.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          },
                          {
                            "address": "172.18.0.6",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "prefixes": [
                                  "192.168.2.0/24"
                                ]
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRConfiguration",
                "metadata": {
                  "name": "two-routers-vrf",
                  "namespace": "frr-k8s-system"
                },
                "spec": {
                  "bgp": {
                    "routers": [
                      {
                        "asn": 64513,
                        "neighbors": [
                          {
                            "address": "172.32.0.3",
                            "asn": 4200000000,
                            "ebgpMultiHop": true,
                            "port": 180,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ]
                      },
                      {
                        "asn": 64515,
                        "neighbors": [
                          {
                            "address": "172.31.0.3",
                            "asn": 4200000000,
                            "port": 179,
                            "toAdvertise": {
                              "allowed": {
                                "mode": "all"
                              }
                            }
                          }
                        ],
                        "prefixes": [
                          "192.168.2.0/24",
                          "192.169.2.0/24"
                        ],
                        "vrf": "red"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "frrk8s.metallb.io/v1beta1",
                "kind": "FRRNodeState",
                "metadata": {
                  "name": "frr-k8s-worker"
                },
                "status": {
                  "lastConversionResult": "success",
                  "lastReloadResult": "success",
                  "runningConfig": "Building configuration...\n\nCurrent configuration:\n!\nfrr version 9.0.2_git\nfrr defaults traditional\nhostname frr-k8s-worker\nlog file /etc/frr/frr.log informational\nlog timestamp precision 3\nno ipv6 forwarding\nservice integrated-vtysh-config\n!\nip nht resolve-via-default\n!\nipv6 nht resolve-via-default\n!\nend\n"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BFDProfile",
                "metadata": {
                  "name": "bfd-profile-sample",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "receiveInterval": 380,
                  "transmitInterval": 270
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "peers": [
                    "ebgp-single-hop0"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPoolSelectors": [
                    {
                      "matchLabels": {
                        "test": "ipv6"
                      }
                    },
                    {
                      "matchLabels": {
                        "test": "ipv4"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "65535:65282"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample4",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "communities": [
                    "NO_ADVERTISE"
                  ],
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "localPref": 50
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPAdvertisement",
                "metadata": {
                  "name": "bgp-adv-sample5",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "aggregationLength": 32,
                  "aggregationLengthV6": 128,
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ],
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    },
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-worker"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-beta1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "Community",
                "metadata": {
                  "name": "community1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "communities": [
                    {
                      "name": "NO_ADVERTISE",
                      "value": "65535:65282"
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "name": "ip-addresspool-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "192.168.10.0/24",
                    "192.168.9.1-192.168.9.5",
                    "fc00:f853:0ccd:e799::/124"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv4"
                  },
                  "name": "ip-addresspool-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "172.20.0.100/24"
                  ],
                  "autoAssign": false
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "IPAddressPool",
                "metadata": {
                  "labels": {
                    "test": "ipv6"
                  },
                  "name": "ip-addresspool-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "addresses": [
                    "2002:2:2::1-2002:2:2::100"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "ipAddressPools": [
                    "ip-addresspool-sample1"
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "L2Advertisement",
                "metadata": {
                  "name": "l2-adv-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "nodeSelectors": [
                    {
                      "matchLabels": {
                        "kubernetes.io/hostname": "kind-control-plane"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "metallb.io/v1beta1",
                "kind": "MetalLB",
                "metadata": {
                  "name": "metallb",
                  "namespace": "metallb-system"
                },
                "spec": {}
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample1",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample2",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "bfdProfile": "bfd-profile-sample",
                  "myASN": 64512,
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3"
                }
              },
              {
                "apiVersion": "metallb.io/v1beta2",
                "kind": "BGPPeer",
                "metadata": {
                  "name": "bgp-peer-sample3",
                  "namespace": "metallb-system"
                },
                "spec": {
                  "holdTime": "9m0s",
                  "keepaliveTime": "0s",
                  "myASN": 64512,
                  "passwordSecret": {
                    "name": "secretname",
                    "namespace": "metallb-system"
                  },
                  "peerASN": 64512,
                  "peerAddress": "172.30.0.3",
                  "peerPort": 180
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
          createdAt: "2023-06-06T15:25:00Z"
          description: An operator for deploying MetalLB on a kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.8.0 <4.18.0'
          operatorframework.io/suggested-namespace: metallb-system
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/metallb-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BFDProfile is the Schema for the bfdprofiles API
            kind: BFDProfile
            name: bfdprofiles.metallb.io
            version: v1beta1
          - description: BGPAdvertisement is the Schema for the bgpadvertisements
              API
            kind: BGPAdvertisement
            name: bgpadvertisements.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta1
          - description: BGPPeer is the Schema for the bgppeers API
            kind: BGPPeer
            name: bgppeers.metallb.io
            version: v1beta2
          - description: Community is the Schema for the communities API
            kind: Community
            name: communities.metallb.io
            version: v1beta1
          - description: FRRConfiguration is the Schema for the frrconfigurations
              API
            kind: FRRConfiguration
            name: frrconfigurations.frrk8s.metallb.io
            version: v1beta1
          - description: FRRNodeState is the Schema for the frrnodestates API
            kind: FRRNodeState
            name: frrnodestates.frrk8s.metallb.io
            version: v1beta1
          - description: IPAddressPool is the Schema for the ipaddresspools API
            kind: IPAddressPool
            name: ipaddresspools.metallb.io
            version: v1beta1
          - description: L2Advertisement is the Schema for the l2advertisements API
            kind: L2Advertisement
            name: l2advertisements.metallb.io
            version: v1beta1
          - description: MetalLB is the Schema for the metallbs API
            displayName: MetalLB
            kind: MetalLB
            name: metallbs.metallb.io
            version: v1beta1
          - description: L2ServiceStatus is the Schema for the l2servicestatus API
            kind: ServiceL2Status
            name: servicel2statuses.metallb.io
            version: v1beta1
        description: An operator for deploying MetalLB on a kubernetes cluster.
        displayName: MetalLB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metallb-operator
        links:
        - name: MetalLB Operator
          url: https://github.com/openshift/metallb-operator
        maintainers:
        - email: fpaoline@redhat.com
          name: Federico Paolinelli
        - email: obraunsh@redhat.com
          name: Ori Braunshtein
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-metallb-operator-bundle@sha256:c80c3467cf52fc4b2a55e444fba9fdc8cf3b1efedad5d1276a889d07aed44e68
        - registry.redhat.io/openshift4/frr-rhel9@sha256:a0db688989ed590c75c722e2891572112a1f5b1f714f894ea814b7026cd9adb4
        - registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:d563d251b32997496991820f476ed90b1e7aede56600e8d9de2c7ffd8836924c
        - registry.redhat.io/openshift4/metallb-rhel9@sha256:6f23dd225f2bccfe39d05428a51de990a4a9338c49da079384eb30fb470c04f5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509100344
      entries:
      - name: metallb-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: metallb-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: metallb-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: metallb-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: metallb-operator.v4.18.0-202506251003
        version: 4.18.0-202506251003
      - name: metallb-operator.v4.18.0-202506070804
        version: 4.18.0-202506070804
      - name: metallb-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: metallb-operator.v4.18.0-202505081435
        version: 4.18.0-202505081435
      - name: metallb-operator.v4.18.0-202504291304
        version: 4.18.0-202504291304
      - name: metallb-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: metallb-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: metallb-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: metallb-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: metallb-operator.v4.18.0-202503181802
        version: 4.18.0-202503181802
      - name: metallb-operator.v4.18.0-202503110933
        version: 4.18.0-202503110933
      - name: metallb-operator.v4.18.0-202502261233
        version: 4.18.0-202502261233
      - name: metallb-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: metallb-operator.v4.18.0-202502041632
        version: 4.18.0-202502041632
      name: stable
    defaultChannel: stable
    packageName: metallb-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: bpfman-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bpfman-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfapplication"
                  },
                  "name": "bpfapplication-sample"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "kprobe": {
                        "bpffunctionname": "kprobe_counter",
                        "func_name": "try_to_wake_up",
                        "offset": 0,
                        "retprobe": false
                      },
                      "type": "Kprobe"
                    },
                    {
                      "tracepoint": {
                        "bpffunctionname": "tracepoint_kill_recorder",
                        "names": [
                          "syscalls/sys_enter_kill"
                        ]
                      },
                      "type": "Tracepoint"
                    },
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "containernames": [
                            "bpfman",
                            "bpfman-agent"
                          ],
                          "namespace": "bpfman",
                          "pods": {
                            "matchLabels": {
                              "name": "bpfman-daemon"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "BpfNsApplication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "bpfnsapplication"
                  },
                  "name": "bpfapplication-sample",
                  "namespace": "acme"
                },
                "spec": {
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/go-app-counter:latest"
                    }
                  },
                  "nodeselector": {},
                  "programs": [
                    {
                      "tc": {
                        "bpffunctionname": "stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      },
                      "type": "TC"
                    },
                    {
                      "tcx": {
                        "bpffunctionname": "tcx_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "direction": "ingress",
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 500
                      },
                      "type": "TCX"
                    },
                    {
                      "type": "Uprobe",
                      "uprobe": {
                        "bpffunctionname": "uprobe_counter",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "func_name": "malloc",
                        "retprobe": false,
                        "target": "libc"
                      }
                    },
                    {
                      "type": "XDP",
                      "xdp": {
                        "bpffunctionname": "xdp_stats",
                        "containers": {
                          "pods": {
                            "matchLabels": {
                              "app": "nginx"
                            }
                          }
                        },
                        "interfaceselector": {
                          "primarynodeinterface": true
                        },
                        "priority": 55
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FentryProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fentryprogram"
                  },
                  "name": "fentry-example"
                },
                "spec": {
                  "bpffunctionname": "test_fentry",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fentry:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "FexitProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "fexitprogram"
                  },
                  "name": "fexit-example"
                },
                "spec": {
                  "bpffunctionname": "test_fexit",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/fexit:latest"
                    }
                  },
                  "func_name": "do_unlinkat",
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "KprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kprobeprogram"
                  },
                  "name": "kprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_kprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/kprobe:latest"
                    }
                  },
                  "func_name": "try_to_wake_up",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "offset": 0,
                  "retprobe": false
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcnsprogram"
                  },
                  "name": "tc-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcprogram"
                  },
                  "name": "tc-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tc_pass:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxnsprogram"
                  },
                  "name": "tcx-containers",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "tcx_next",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TcxProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tcxprogram"
                  },
                  "name": "tcx-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "tcx_pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tcx_test:latest"
                    }
                  },
                  "direction": "ingress",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "TracepointProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "tracepointprogram"
                  },
                  "name": "tracepoint-example"
                },
                "spec": {
                  "bpffunctionname": "enter_openat",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/tracepoint:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "names": [
                    "syscalls/sys_enter_openat"
                  ],
                  "nodeselector": {}
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobensprogram"
                  },
                  "name": "uprobe-example",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "UprobeProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "uprobeprogram"
                  },
                  "name": "uprobe-example"
                },
                "spec": {
                  "bpffunctionname": "my_uprobe",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/uprobe:latest"
                    }
                  },
                  "func_name": "syscall",
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "nodeselector": {},
                  "retprobe": false,
                  "target": "libc"
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpNsProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpnsprogram"
                  },
                  "name": "xdp-ns-pass-all-nodes",
                  "namespace": "acme"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "containers": {
                    "containernames": [
                      "nginx"
                    ],
                    "pods": {
                      "matchLabels": {
                        "app": "nginx"
                      }
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "interfaces": [
                      "eth0"
                    ]
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              },
              {
                "apiVersion": "bpfman.io/v1alpha1",
                "kind": "XdpProgram",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "xdpprogram"
                  },
                  "name": "xdp-pass-all-nodes"
                },
                "spec": {
                  "bpffunctionname": "pass",
                  "bytecode": {
                    "image": {
                      "url": "quay.io/bpfman-bytecode/xdp_pass:latest"
                    }
                  },
                  "globaldata": {
                    "GLOBAL_u32": [
                      13,
                      12,
                      11,
                      10
                    ],
                    "GLOBAL_u8": [
                      1
                    ]
                  },
                  "interfaceselector": {
                    "primarynodeinterface": true
                  },
                  "nodeselector": {},
                  "priority": 0
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/bpfman/bpfman-operator:latest
          createdAt: "2025-01-03T19:50:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: bpfman
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "bpfman",
                "labels": {
                  "pod-security.kubernetes.io/enforce": "privileged",
                  "pod-security.kubernetes.io/audit": "privileged",
                  "pod-security.kubernetes.io/warn": "privileged",
                },
                "annotations": {
                  "openshift.io/node-selector": ""
                },
              }
            }
          operators.openshift.io/infrastructure-features: '["csi", "disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/bpfman/bpfman
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BpfApplication is the Schema for the bpfapplications API
            displayName: Bpf Application
            kind: BpfApplication
            name: bpfapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsApplication is the Schema for the BpfNsApplications
              API
            displayName: Bpf Namespaced Application
            kind: BpfNsApplication
            name: bpfnsapplications.bpfman.io
            version: v1alpha1
          - description: BpfNsProgram is the Schema for the BpfNsProgram API
            displayName: Bpf Namespaced Program
            kind: BpfNsProgram
            name: bpfnsprograms.bpfman.io
            version: v1alpha1
          - description: BpfProgram is the Schema for the BpfProgram API
            displayName: Bpf Program
            kind: BpfProgram
            name: bpfprograms.bpfman.io
            version: v1alpha1
          - description: FentryProgram is the Schema for the Fentryprograms API
            displayName: Fentry Program
            kind: FentryProgram
            name: fentryprograms.bpfman.io
            version: v1alpha1
          - description: FexitProgram is the Schema for the Fexitprograms API
            displayName: Fexit Program
            kind: FexitProgram
            name: fexitprograms.bpfman.io
            version: v1alpha1
          - description: KprobeProgram is the Schema for the Kprobeprograms API
            displayName: Kprobe Program
            kind: KprobeProgram
            name: kprobeprograms.bpfman.io
            version: v1alpha1
          - description: TcNsProgram is the Schema for the Tcnsprograms API
            displayName: TC Namespaced Program
            kind: TcNsProgram
            name: tcnsprograms.bpfman.io
            version: v1alpha1
          - description: TcProgram is the Schema for the Tcprograms API
            displayName: TC Program
            kind: TcProgram
            name: tcprograms.bpfman.io
            version: v1alpha1
          - description: TcxNsProgram is the Schema for the Tcxnsprograms API
            displayName: TCX Namespaced Program
            kind: TcxNsProgram
            name: tcxnsprograms.bpfman.io
            version: v1alpha1
          - description: TcxProgram is the Schema for the Tcxprograms API
            displayName: TCX Program
            kind: TcxProgram
            name: tcxprograms.bpfman.io
            version: v1alpha1
          - description: TracepointProgram is the Schema for the Tracepointprograms
              API
            displayName: Tracepoint Program
            kind: TracepointProgram
            name: tracepointprograms.bpfman.io
            version: v1alpha1
          - description: UprobeNsProgram is the Schema for the Uprobensprograms API
            displayName: Uprobe Namespaced Program
            kind: UprobeNsProgram
            name: uprobensprograms.bpfman.io
            version: v1alpha1
          - description: UprobeProgram is the Schema for the Uprobeprograms API
            displayName: Uprobe Program
            kind: UprobeProgram
            name: uprobeprograms.bpfman.io
            version: v1alpha1
          - description: XdpNsProgram is the Schema for the Xdpnsprograms API
            displayName: Xdp Namespaced Program
            kind: XdpNsProgram
            name: xdpnsprograms.bpfman.io
            version: v1alpha1
          - description: XdpProgram is the Schema for the Xdpprograms API
            displayName: Xdp Program
            kind: XdpProgram
            name: xdpprograms.bpfman.io
            version: v1alpha1
        description: "The eBPF Manager Operator is a Kubernetes Operator for deploying
          [bpfman](https://bpfman.netlify.app/), a system daemon\nfor managing eBPF
          programs. It deploys bpfman itself along with CRDs to make deploying\neBPF
          programs in Kubernetes much easier.\n\n## Quick Start\n\nTo get bpfman up
          and running quickly simply click 'install' to deploy the bpfman-operator
          in the bpfman namespace via operator-hub.\n## Configuration\n\nThe `bpfman-config`
          configmap is automatically created in the `bpfman` namespace and used to
          configure the bpfman deployment.\n\nTo edit the config simply run\n\n```bash\nkubectl
          edit cm bpfman-config\n```\n\nThe following fields are adjustable\n\n- `bpfman.agent.image`:
          The image used for the bpfman-agent`\n- `bpfman.image`: The image used for
          bpfman`\n- `bpfman.log.level`: the log level for bpfman, currently supports
          `debug`, `info`, `warn`, `error`, and `fatal`, defaults to `info`\n- `bpfman.agent.log.level`:
          the log level for the bpfman-agent currently supports `info`, `debug`, and
          `trace` \n\nThe bpfman operator deploys eBPF programs via CRDs. The following
          CRDs are currently available, \n\n- BpfApplication\n- FentryProgram\n- FexitProgram\n
          - KprobeProgram\n- TcProgram\n- TcxProgram\n- TracepointProgram\n- UprobeProgram\n-
          XdpProgram\n - BpfNsApplication\n- TcProgram\n- TcxNsProgram\n- UprobeNsProgram\n-
          XdpNsProgram\n\n ## More information\n\nPlease checkout the [bpfman community
          website](https://bpfman.io/) for more information."
        displayName: eBPF Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ebpf
        - kubernetes
        links:
        - name: bpfman website
          url: https://bpfman.io/
        maintainers:
        - email: astoycos@redhat.com
          name: Andrew Stoycos
        - email: afredette@redhat.com
          name: Andre Fredette
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - registry.redhat.io/bpfman/bpfman-operator-bundle@sha256:b6d43f1fec6ba7336c23d1fbceab15ebb2ed9e9e2eebb08c78d7541ee47b4204
        - registry.redhat.io/bpfman/bpfman-rhel9-operator@sha256:c3d878d77091bed03d99f72943d9b78e210e2f6194ea4b5cd2aa23f8d48f3f61
        version: 0.5.6
      entries:
      - name: bpfman-operator.v0.5.6
        version: 0.5.6
      name: latest
    defaultChannel: latest
    packageName: bpfman-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-applicationautoscaling-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-applicationautoscaling-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalableTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "applicationautoscaling.services.k8s.aws/v1alpha1",
                "kind": "ScalingPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
          createdAt: "2025-09-20T21:43:01Z"
          description: AWS Application Auto Scaling controller is a service controller
            for managing Application Auto Scaling resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ScalableTarget represents the state of an AWS applicationautoscaling
              ScalableTarget resource.
            displayName: ScalableTarget
            kind: ScalableTarget
            name: scalabletargets.applicationautoscaling.services.k8s.aws
            version: v1alpha1
          - description: ScalingPolicy represents the state of an AWS applicationautoscaling
              ScalingPolicy resource.
            displayName: ScalingPolicy
            kind: ScalingPolicy
            name: scalingpolicies.applicationautoscaling.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Application Auto Scaling resources in AWS from within your Kubernetes cluster.
          **About Amazon Application Auto Scaling**
          Application Auto Scaling is a web service for developers and system administrators who need a solution for automatically scaling their scalable resources for individual AWS services beyond Amazon EC2. Application Auto Scaling allows you to configure automatic scaling for the following resources:
          - AppStream 2.0 fleets
          - Aurora replicas
          - Amazon Comprehend document classification and entity recognizer endpoints
          - DynamoDB tables and global secondary indexes
          - Amazon Elastic Container Service (ECS) services
          - Amazon EMR clusters
          - Amazon Keyspaces (for Apache Cassandra) tables
          - Lambda function provisioned concurrency
          - Amazon Managed Streaming for Apache Kafka (MSK) broker storage
          - SageMaker endpoint variants
          - Spot Fleet requests
          - Custom resources provided by your own applications or services. For more information, see the [GitHub repository](https://github.com/aws/aws-auto-scaling-custom-resource)

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Application Auto Scaling
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - applicationautoscaling
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Application Auto Scaling User Guide
          url: https://docs.aws.amazon.com/autoscaling/application/userguide/what-is-application-auto-scaling.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: application auto scaling maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/applicationautoscaling-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-applicationautoscaling-controller@sha256:4f1b92eac30d72a66d8fe8c2c44339c7a5fd6af65d48c0d8697459c7af290741
        version: 1.1.1
      entries:
      - name: ack-applicationautoscaling-controller.v1.1.1
        version: 1.1.1
      - name: ack-applicationautoscaling-controller.v1.1.0
        version: 1.1.0
      - name: ack-applicationautoscaling-controller.v1.0.30
        version: 1.0.30
      - name: ack-applicationautoscaling-controller.v1.0.29
        version: 1.0.29
      - name: ack-applicationautoscaling-controller.v1.0.27
        version: 1.0.27
      - name: ack-applicationautoscaling-controller.v1.0.26
        version: 1.0.26
      - name: ack-applicationautoscaling-controller.v1.0.25
        version: 1.0.25
      - name: ack-applicationautoscaling-controller.v1.0.24
        version: 1.0.24
      - name: ack-applicationautoscaling-controller.v1.0.23
        version: 1.0.23
      - name: ack-applicationautoscaling-controller.v1.0.22
        version: 1.0.22
      - name: ack-applicationautoscaling-controller.v1.0.21
        version: 1.0.21
      - name: ack-applicationautoscaling-controller.v1.0.19
        version: 1.0.19
      - name: ack-applicationautoscaling-controller.v1.0.18
        version: 1.0.18
      - name: ack-applicationautoscaling-controller.v1.0.17
        version: 1.0.17
      - name: ack-applicationautoscaling-controller.v1.0.16
        version: 1.0.16
      - name: ack-applicationautoscaling-controller.v1.0.15
        version: 1.0.15
      - name: ack-applicationautoscaling-controller.v1.0.11
        version: 1.0.11
      - name: ack-applicationautoscaling-controller.v1.0.9
        version: 1.0.9
      - name: ack-applicationautoscaling-controller.v1.0.8
        version: 1.0.8
      - name: ack-applicationautoscaling-controller.v1.0.7
        version: 1.0.7
      - name: ack-applicationautoscaling-controller.v1.0.6
        version: 1.0.6
      - name: ack-applicationautoscaling-controller.v1.0.5
        version: 1.0.5
      - name: ack-applicationautoscaling-controller.v1.0.4
        version: 1.0.4
      - name: ack-applicationautoscaling-controller.v1.0.3
        version: 1.0.3
      - name: ack-applicationautoscaling-controller.v1.0.2
        version: 1.0.2
      - name: ack-applicationautoscaling-controller.v1.0.0
        version: 1.0.0
      - name: ack-applicationautoscaling-controller.v0.2.16
        version: 0.2.16
      - name: ack-applicationautoscaling-controller.v0.2.15
        version: 0.2.15
      - name: ack-applicationautoscaling-controller.v0.2.14
        version: 0.2.14
      - name: ack-applicationautoscaling-controller.v0.2.13
        version: 0.2.13
      - name: ack-applicationautoscaling-controller.v0.2.12
        version: 0.2.12
      - name: ack-applicationautoscaling-controller.v0.2.10
        version: 0.2.10
      - name: ack-applicationautoscaling-controller.v0.2.9
        version: 0.2.9
      - name: ack-applicationautoscaling-controller.v0.2.8
        version: 0.2.8
      - name: ack-applicationautoscaling-controller.v0.2.7
        version: 0.2.7
      - name: ack-applicationautoscaling-controller.v0.2.6
        version: 0.2.6
      - name: ack-applicationautoscaling-controller.v0.2.5
        version: 0.2.5
      - name: ack-applicationautoscaling-controller.v0.2.4
        version: 0.2.4
      - name: ack-applicationautoscaling-controller.v0.2.3
        version: 0.2.3
      - name: ack-applicationautoscaling-controller.v0.2.2
        version: 0.2.2
      - name: ack-applicationautoscaling-controller.v0.2.1
        version: 0.2.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-applicationautoscaling-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-loadcore-agents
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-loadcore-agents.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.loadcore.keysight.com/v1alpha1",
                "kind": "Agents",
                "metadata": {
                  "name": "loadcore-agents-sample",
                  "namespace": "keysight-loadcore-agents"
                },
                "spec": {
                  "agents": [
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent1",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "replicas": 2,
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    },
                    {
                      "args": {
                        "adminInterface": "eth0",
                        "middlewareAddr": "nats-core.keysight-wap.svc.cluster.local:4222",
                        "testInterfaces": [
                          "net1"
                        ]
                      },
                      "exposeIngress": true,
                      "lifecycle": {
                        "postStart": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "echo Hello from the postStart handler"
                            ]
                          }
                        },
                        "preStop": {
                          "exec": {
                            "command": [
                              "/bin/sh",
                              "-c",
                              "killall -0 tail"
                            ]
                          }
                        }
                      },
                      "livenessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "initialDelaySeconds": 60
                      },
                      "name": "agent2",
                      "networks": [
                        "keysight-loadcore-agents/loadcore-agent-network"
                      ],
                      "privileged": true,
                      "readinessProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 30,
                        "periodSeconds": 20
                      },
                      "startupProbe": {
                        "exec": {
                          "command": [
                            "cat",
                            "/dev/termination-log"
                          ]
                        },
                        "failureThreshold": 10,
                        "periodSeconds": 5
                      },
                      "terminationMessagePolicy": "FallbackToLogsOnError"
                    }
                  ],
                  "image": {
                    "registry": "prod.harbor.keysight.digital",
                    "repository": "loadcore/loadcore-agent",
                    "tag": "UBI8-0.1-7d990d7254"
                  },
                  "imagePullSecrets": [
                    {
                      "name": "regcred"
                    }
                  ],
                  "networks": [
                    {
                      "config": "{ \"cniVersion\": \"0.3.1\", \"type\": \"ipvlan\", \"master\": \"ens192\", \"mode\": \"l2\", \"ipam\": { \"type\": \"whereabouts\", \"range\": \"192.168.1.0/24\", \"routes\": [ { \"dst\": \"0.0.0.0/0\" } ] } }",
                      "name": "loadcore-agent-network"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:53:55Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Agents
            name: agents.app.loadcore.keysight.com
            version: v1alpha1
        description: Keysight LoadCore Agents Operator
        displayName: Keysight LoadCore Agents Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Loadcore Agents
          url: https://keysight-loadcore-agents.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-loadcore-agents@sha256:f1d1e9f2aca11505407785bc33905744b667bc8aea49df73fcc50db379c0b375
        - registry.connect.redhat.com/loadcore/keysight-loadcore-agents-bundle@sha256:f16fcea424cb7b1eb647f3e9d3c08bd93c8a1b8bb7eab71a4e756da72a7a56a0
        version: 0.1.21
      entries:
      - name: keysight-loadcore-agents.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-loadcore-agents
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.90.0-beta.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0-beta.1 <42.90.0-beta.2'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:37ff2641e0d0561a5618eaaa8e078ef642ff9b55f02b4fc1e16077ab649e83ee
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:a58c9c8454ac337c5a359dfe5b7c6b06983b7050ad62a0d11054f06fc17e92c1
        version: 42.90.0-beta.2
      entries:
      - name: t8c-operator.v42.90.0-beta.2
        version: 42.90.0-beta.2
      name: beta
    - currentCSV: t8c-operator.v42.89.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"externalDbIP":
            "10.0.2.15","customImageNames": false,"repository": "icr.io/cpopen/turbonomic","tag":"8.17.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
          createdAt: "2023-03-24T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=42.53.0 <42.89.0'
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints. Turbonomic platform install and configuration
              documentation can be found online here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How do I start?

          Turbonomic platform installation and configuration documentation can be found on the IBM Turbonomic documentation site here https://www.ibm.com/docs/en/tarm/latest?topic=installation-installing-kubernetes-clusters
          Carefully read through the prerequisites before beginning the installation.  Once ready, on the `Details` tab for the Turbonomic Platform Operator, under `Provided APIs`, click `Create instance`.
          For samples of common configurations, refer to sample custom resource YAMLs provided here: https://github.com/turbonomic/t8c-install/tree/master/samples
          Looking to upgrade your current deployment? Refer to documentation here: https://www.ibm.com/docs/en/tarm/latest?topic=version-updating-using-red-hat-openshift-operatorhub

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/t8c-operator@sha256:f2cd1c661d721491c98e3fbc8b777ae90fc63eae6afc28d0b33643d32ef27f14
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:f6c37a0808c137f226b6afb6cac1e2ce05184d6a01b9234969ce63056571e991
        version: 42.89.0
      entries:
      - name: t8c-operator.v42.89.0
        version: 42.89.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: GitLab, Inc.
      provider-url: https://gitlab.com
    name: gitlab-runner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-runner-operator.v1.41.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta2",
                "kind": "Runner",
                "metadata": {
                  "name": "gitlab-runner-example",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "gitlabUrl": "https://gitlab.example.com",
                  "token": "runner-token-secret",
                  "tags": "kubernetes,example",
                  "concurrent": 10,
                  "imagePullPolicy": "IfNotPresent"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery, Developer Tools
          containerImage: registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
          createdAt: "2025-09-25T15:08:40Z"
          description: GitLab Runner operator manages lifecycle of GitLab Runner instances
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.11.0 <1.41.1'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: none
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
          support: GitLab, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Runner is the open source project used to run your jobs and
              send the results back to GitLab
            displayName: GitLab Runner
            kind: Runner
            name: runners.apps.gitlab.com
            version: v1beta2
        description: |-
          GitLab Runner is the lightweight, highly-scalable agent that runs your build jobs and sends the results back to a GitLab instance. GitLab Runner works in conjunction with GitLab CI/CD, the open-source continuous integration service included with GitLab.

              The GitLab Runner operator manages the lifecycle of GitLab Runner in Kubernetes or Openshift clusters. The operator aims to automate the tasks needed to run your CI/CD jobs in your container orchestration platform.

              ## Prerequisites

                For Kubernetes cluster, install cert-manager:

                ```shell
                kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.7.1/cert-manager.yaml
                ```

              ## GitLab Runner version

              This version of **GitLab Runner Operator** ships with **GitLab Runner v18.4.0**.

              To use a different version of **GitLab Runner** change the [`runnerImage` and `helperImage` properties](https://docs.gitlab.com/runner/configuration/configuring_runner_operator.html#operator-properties).

              ## Usage

               To link a GitLab Runner instance to a self-hosted GitLab instance or the hosted [GitLab](https://gitlab.com), you first need to:

               1. Create a secret containing the `runner-token` from your GitLab project.

                ```
                cat > gitlab-runner-secret.yml << EOF
                apiVersion: v1
                kind: Secret
                metadata:
                  name: gitlab-runner-secret
                type: Opaque
                # Only one of the following fields must be set. The Operator will fail to register Runner if both are provided.
                # # NOTE: runner-registration-token is deprecated and will be removed in GitLab 18.0. You should use runner-token instead.
                stringData:
                  runner-token: REPLACE_ME # your project runner token
                  # runner-registration-token: "" # your project runner secret
                EOF
                ```

                ```
                oc apply -f gitlab-runner-secret.yml
                ```

               2. Create the Custom Resource Definition (CRD) file and include the following information. The tags value must be openshift for the job to run.

                 ```
                 cat > gitlab-runner.yml << EOF
                 apiVersion: apps.gitlab.com/v1beta2
                 kind: Runner
                 metadata:
                   name: gitlab-runner
                 spec:
                   gitlabUrl: https://gitlab.example.com
                   buildImage: alpine
                   token: gitlab-runner-secret
                   tags: openshift
                 EOF
                 ```

                ```
                oc apply -f gitlab-runner.yml
                ```

              ## Full documentation

              Visit [Install GitLab Runner Operator](https://docs.gitlab.com/runner/install/operator.html)
        displayName: GitLab Runner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GitLab
        - CI/CD
        - DevOps
        - SAST
        - DAST
        links:
        - name: Gitlab Runner Operator
          url: https://gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator
        - name: GitLab Docs
          url: https://docs.gitlab.com
        - name: GitLab Runner
          url: https://docs.gitlab.com/runner/
        - name: GitLab CI/CD
          url: https://docs.gitlab.com/ee/ci/quick_start/
        maintainers:
        - email: support@gitlab.com
          name: Georgi N. Georgiev
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: GitLab, Inc.
          url: https://gitlab.com
        relatedImages:
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-ocp@sha256:0f2a5a3b6779d719eecbf079b99cf06b652603d74b563f57d753e32966d253dd
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:d93534eba3856b5aaa9ee3ea5bfa368642cc34b4b37ba038f3fa4efc4f022cff
        - registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - quay.io/community-operator-pipeline-prod/gitlab-runner-operator@sha256:3e0314a2695c071fe0a34f3dc7684b43b5a31b6705f42982b4dbaacabe360d2b
        - registry.gitlab.com/gitlab-org/ci-cd/gitlab-runner-ubi-images/gitlab-runner-helper-ocp@sha256:29727fae83a226bcf57607240d4046a0b8c208b5ff42ee1f8ca0f115bb188809
        version: 1.41.1
      entries:
      - name: gitlab-runner-operator.v1.41.1
        version: 1.41.1
      - name: gitlab-runner-operator.v1.41.0
        version: 1.41.0
      - name: gitlab-runner-operator.v1.40.1
        version: 1.40.1
      - name: gitlab-runner-operator.v1.40.0
        version: 1.40.0
      - name: gitlab-runner-operator.v1.39.0
        version: 1.39.0
      name: stable
    defaultChannel: stable
    packageName: gitlab-runner-operator
    provider:
      name: GitLab, Inc.
      url: https://gitlab.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerify",
                "metadata": {
                  "name": "ibmsecurityverify-sample"
                },
                "spec": {
                  "clientSecret": "--secret--",
                  "sessionLifetime": 3600,
                  "ssoPath": "/verify-sso"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
          createdAt: "2023-03-22T02:49:11Z"
          description: The IBM Security Verify operator can consistently enforce policy-driven
            security, including authentication and authorization, using the Ingress
            networking capability of OpenShift.
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.9"}]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerify is the Schema for the ibmsecurityverifies
              API.
            displayName: IBMSecurity Verify
            kind: IBMSecurityVerify
            name: ibmsecurityverifies.ibm.com
            version: v1
        description: "[IBM Security Verify](https://www.ibm.com/products/verify-saas)
          is an Identity-as-a-Service platform that allows IT, security and business
          leaders to protect their digital users, assets and data in a hybrid, multi-cloud
          world by enabling technical agility and operational efficiency. IBM Security
          Verify SaaS provides single sign-on (SSO), multi-factor authentication (MFA),
          AI-powered context for risk-based authentication for adaptive access decisions,
          user management, access recertification campaigns and identity analytics.\n\nAs
          part of using Red Hat OpenShift, you are entitled to Verify SaaS SSO for
          unlimited apps and users.  If you do not already have a Verify SaaS SSO
          tenant a new one can be [created](https://www.ibm.com/account/reg/signup?formid=urx-51255).\n\nFor
          a detailed description of IBM Security Verify refer to the [Official documentation](https://www.ibm.com/docs/en/security-verify).\n\nThe
          IBM Security Verify operator can consistently enforce policy-driven security
          by using the Ingress networking capability of OpenShift in conjunction with
          the [Nginx Ingress operator](https://www.nginx.com/blog/getting-started-nginx-ingress-operator-red-hat-openshift/).
          With this approach, you can enforce authentication and authorization policies
          for all of the applications in your cluster at the same time, without ever
          changing your application code. You can also dynamically register your application
          to start protecting them centrally from the cloud via Verify SaaS. \n\nSee
          the project [Readme](https://github.com/IBM-Security/verify-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Product Information
          url: https://www.ibm.com/products/verify-saas
        - name: Verify Documentation
          url: https://www.ibm.com/docs/en/security-verify
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Development Team
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-operator-bundle@sha256:2bab9fa63b128a3ba0a6031f70f0342f95d17dd53a3ec2930b19940b5b2e6f47
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - icr.io/isv-saas/verify-operator@sha256:c3bed2e5db48e54884d2278d24ca4864f389a56a667db481c627d01ab5ffaba7
        version: 23.3.0
      entries:
      - name: ibm-security-verify-operator.v23.3.0
        version: 23.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseWeb
      provider-url: ""
    name: cnfv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cnfv-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eweb.enterpriseweb.com/v1",
                "kind": "CloudNFV",
                "metadata": {
                  "name": "cloudnfv-sample"
                },
                "spec": {
                  "size": 1
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CloudNFV is the Schema for the cloudnfvs API
            displayName: Cloud NFV
            kind: CloudNFV
            name: cloudnfvs.eweb.enterpriseweb.com
            version: v1
        description: Manages a cluster of CloudNFV Runtime Modules
        displayName: CloudNFV
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - automation
        - NFV
        - controller
        links:
        - name: Cnfv Operator
          url: https://cnfv-operator.domain
        maintainers:
        - email: bill@enterpriseweb.com
          name: Bill Malyk
        maturity: alpha
        provider:
          name: EnterpriseWeb
        relatedImages:
        - quay.io/wmalyk28/cloudnfv-runtime@sha256:fcde8375576c9bff602d3a16ef1ba8ae88b780166649ba6ae355737b1d35a03f
        - quay.io/wmalyk28/cnfv-op@sha256:e5913ad5dd20cb1b2299d2387237ef15e645c9a8c9d4f5fe0044b86a6b3dc1df
        - registry.connect.redhat.com/eweb/cloudnfv@sha256:0bb427954fb8df5de197353908a5189922f514e121ae8cdb8242d125b1b3807d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:61a164ad711478ac603242338f7b3c55986319479878b092be128581a823598e
        version: 0.0.1
      entries:
      - name: cnfv-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cnfv-operator
    provider:
      name: EnterpriseWeb
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Odoo Community Association
      provider-url: ""
    name: odoo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odoo-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"odoo-community.org/v1alpha1","kind":"Odoo","metadata":{"name":"example"},"spec":{"postgresqlVolumeSize":"4Gi","odooVolumeSize":"4Gi"}}]'
          capabilities: Full Lifecycle
          categories: Application Runtime
          certified: "false"
          containerImage: ursa/odoo-operator:0.0.2
          createdAt: "2021-04-26T12:00:00Z"
          description: Operator to manage Odoo instances within an OpenShift cluster
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          repository: https://github.com/ursais/odoo-operator
          support: Open Source Integrators
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Odoo
            displayName: Odoo
            kind: Odoo
            name: odoos.odoo-community.org
            version: v1alpha1
        description: |
          ## About Odoo

          [Odoo](https://odoo-community.org), is a suite of business management software tools including CRM, e-commerce, billing, accounting, manufacturing, warehouse, project management, and inventory management to name a few.
          ## About the Odoo Operator

          This is an Ansible-based operator created with operator-sdk v1.5.0.

          The upgrade is handled by the Odoo container itself using [Marabunta](https://pypi.org/project/marabunta/) and [Anthem](https://pypi.org/project/anthem/).

          ## Roadmap
          * Add horizontal pod autoscaler
          * Add cronjob for backup
          ## Prerequisites for enabling this Operator

          Please refer to the README file on the [GitHub project](https://github.com/ursais/odoo-operator).
        displayName: Odoo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ERP
        - Business Software
        links:
        - name: Odoo Community Association
          url: https://odoo-community.org
        - name: GitHub Repository
          url: https://github.com/ursais/odoo-operator
        - name: Open Source Integrators
          url: https://opensourceintegrators.com
        maintainers:
        - email: mchambreuil@opensourceintegrators.com
          name: Maxime Chambreuil
        maturity: alpha
        provider:
          name: Odoo Community Association
        relatedImages:
        - quay.io/openshift-community-operators/odoo-operator@sha256:616e1067ebf8deb66636d57cae2ac1f9981b83419d8e7f023bc7df1c1b745536
        - ursa/odoo-operator:0.0.2
        version: 0.0.2
      entries:
      - name: odoo-operator.v0.0.2
        version: 0.0.2
      - name: odoo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: odoo-operator
    provider:
      name: Odoo Community Association
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Docling Maintainers
      provider-url: https://docling-project.github.io/docling/
    name: docling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: docling-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2025-05-19T20:20:14Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DoclingServe is the Schema for the doclingserves API
            displayName: Docling Serve
            kind: DoclingServe
            name: doclingserves.docling.github.io
            version: v1alpha1
        description: |-
          **Overview**

          The Docling Operator is responsible for managing the full lifecycle of Docling Serve instances in your Kubernetes or OpenShift container platforms.


          **Docling Serve**

          The Docling Serve application deploys a REST API for consuming and integrating Docling in your workflows.

          The API offers endpoints for multiple types of interaction, i.e. both synchronous and asynchronous endpoints, as well as status subscriptions via websockets.

          The options for the Docling conversion can be customized at runtime by the users (e.g. which OCR engine, which enrichment models, etc). Users can also choose which of the many DoclingDocument export formats (JSON, Markdown, HTML, DocTags, etc) to receive in the result response.

          Docling Serve allows to choose between a local compute engine, i.e. running Docling in the same container as the API server, or a distributed compute engine which leverages Kubeflow Pipelines to parallelize large-scale ingestion jobs.

          **Docling**

          Docling is a powerful library which simplifies document processing, parsing diverse formats — including advanced PDF understanding — and providing seamless integrations with the gen AI ecosystem.

          Learn more about the Docling Project by visiting the GitHub organization at [https://github.com/docling-project](https://github.com/docling-project).


          **Sample DoclingServe Custom Resources**


          To deploy with the `local` compute engine with the UI enabled and Route exposed:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-local-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              local:
                numWorkers: 2
            route:
              enabled: true
          ```

          To deploy with the `KFP` compute engine with the UI enabled and Route exposed. **Note: Kubeflow Pipeline must be installed and setup in the cluster to use this option.**:
          ```yaml
          apiVersion: docling.github.io/v1alpha1
          kind: DoclingServe
          metadata:
            labels:
              app.kubernetes.io/name: docling-operator
              app.kubernetes.io/managed-by: kustomize
            name: doclingserve-kfp-sample
          spec:
            apiServer:
              image: "quay.io/docling-project/docling-serve:latest"
              enableUI: true
              instances: 1
            engine:
              kfp:
                endpoint: "https://NAME.NAMESPACE.svc.cluster.local:8888"
            route:
              enabled: true
          ```

          **Additional Configurations**


          The CRD exposes `apiServer.configMapName` to enable additional configurations of the `docling-serve` workload. For a full list of configurations can be found on docling-serve's
          [Configuration](https://github.com/docling-project/docling-serve/blob/main/docs/configuration.md) page.
        displayName: Docling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - document
        - processing
        links:
        - name: Docling Operator
          url: https://docling-project.github.io/docling/
        maintainers:
        - email: docling-technical-discuss@lists.lfaidata.foundation
          name: Docling Maintainers
        maturity: alpha
        provider:
          name: Docling Maintainers
          url: https://docling-project.github.io/docling/
        relatedImages:
        - quay.io/community-operator-pipeline-prod/docling-operator@sha256:9a3717ee195572b29d1ee31b96bb95561f5521b032dbaa09a000413b227b5f12
        - quay.io/docling-project/docling-operator@sha256:200d8db51874c4b1e3d20d8d1423e9a4807843fbb29f514cd5d8ba872e47f4af
        version: 0.0.1
      entries:
      - name: docling-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: docling-operator
    provider:
      name: Docling Maintainers
      url: https://docling-project.github.io/docling/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: ScaleOps
      provider-url: https://scaleops.com
    name: scaleops-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: scaleops-operator.v1.21.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "analysis.scaleops.sh/v1",
                "kind": "ScaleOps",
                "metadata": {
                  "name": "scaleops",
                  "namespace": "scaleops-system"
                },
                "spec": {
                  "scaleopsToken": "${SCALEOPS_TOKEN}"
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
          createdAt: "2025-09-29T13:35:24Z"
          description: |-
            ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

            ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

            Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

            Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

            Contact [us](mailto:team@scaleops.com) to start your free trial today
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.21.8
          operators.openshift.io/valid-subscription: Contact ScaleOps for subscription
            information.
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://scaleops.com
          support: ScaleOps Support support@scaleops.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: This will install an instance of the ScaleOps Platform on
              your cluster
            displayName: ScaleOps Platform
            kind: ScaleOps
            name: scaleops.analysis.scaleops.sh
            version: v1
        description: |-
          ScaleOps is the industry-first Kubernetes Optimization Platform that automatically adjusts Compute Resources to changes in real-time, streamlining a new Kubernetes experience for engineering teams.

          ScaleOps eliminates 80% of Kubernetes cloud spending, frees engineers from repeated ongoing configurations, and proactively ensures SLAs are always achieved.

          Installation takes only 2 minutes, and value is provided immediately. Starting from read-only recommendations to continuous and automatic pod optimization during runtime with zero disruption.

          Battle-tested on some of the most critical production clusters in full automation mode. Trusted by industry leaders, such as Wiz, Outbrain, SentinelOne, Outreach, Cato Networks, and many others!

          Contact [us](mailto:team@scaleops.com) to start your free trial today
        displayName: ScaleOps
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vertical scaling
        - horizontal scaling
        - node optimization
        - application runtime
        links:
        - name: ScaleOps
          url: https://scaleops.com
        maintainers:
        - email: support@scaleops.com
          name: ScaleOps
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: ScaleOps
          url: https://scaleops.com
        relatedImages:
        - registry.connect.redhat.com/scaleops-sh/operator@sha256:528f2e09d2dad4f48e01a272b51baa1753c12069515831223e3f97c7dfe03e3c
        - registry.connect.redhat.com/scaleops-sh/scaleops-gpu@sha256:724aef7683ea4cb245ab0a464517a6d542efa4634e5469ff35b7a7b7f00b563e
        - registry.connect.redhat.com/scaleops-sh/scaleops-operator-bundle@sha256:efd483d9c826e25b0d1cd79deff4263772260c72da79bad149ef7985cc00a44b
        - registry.connect.redhat.com/scaleops-sh/scaleops-sh@sha256:57d5d06f4f4b5470562c4f24ea5d37203633e1f4b65cd497e71d9066fd2b9e41
        version: 1.21.8
      entries:
      - name: scaleops-operator.v1.21.8
        version: 1.21.8
      - name: scaleops-operator.v1.21.7
        version: 1.21.7
      - name: scaleops-operator.v1.21.6
        version: 1.21.6
      - name: scaleops-operator.v1.21.5
        version: 1.21.5
      - name: scaleops-operator.v1.21.3
        version: 1.21.3
      - name: scaleops-operator.v1.21.2
        version: 1.21.2
      - name: scaleops-operator.v1.21.1
        version: 1.21.1
      - name: scaleops-operator.v1.21.0
        version: 1.21.0
      - name: scaleops-operator.v1.20.9
        version: 1.20.9
      - name: scaleops-operator.v1.20.8
        version: 1.20.8
      - name: scaleops-operator.v1.20.7
        version: 1.20.7
      - name: scaleops-operator.v1.20.6
        version: 1.20.6
      - name: scaleops-operator.v1.20.5
        version: 1.20.5
      - name: scaleops-operator.v1.20.4
        version: 1.20.4
      - name: scaleops-operator.v1.20.2
        version: 1.20.2
      - name: scaleops-operator.v1.20.1
        version: 1.20.1
      - name: scaleops-operator.v1.19.8
        version: 1.19.8
      - name: scaleops-operator.v1.19.7
        version: 1.19.7
      - name: scaleops-operator.v1.19.6
        version: 1.19.6
      - name: scaleops-operator.v1.19.5
        version: 1.19.5
      - name: scaleops-operator.v1.19.4
        version: 1.19.4
      - name: scaleops-operator.v1.19.1
        version: 1.19.1
      - name: scaleops-operator.v1.19.0
        version: 1.19.0
      - name: scaleops-operator.v1.18.8
        version: 1.18.8
      - name: scaleops-operator.v1.18.7
        version: 1.18.7
      - name: scaleops-operator.v1.18.6
        version: 1.18.6
      - name: scaleops-operator.v1.18.5
        version: 1.18.5
      - name: scaleops-operator.v1.18.4
        version: 1.18.4
      - name: scaleops-operator.v1.18.3
        version: 1.18.3
      - name: scaleops-operator.v1.18.2
        version: 1.18.2
      name: stable
    defaultChannel: stable
    packageName: scaleops-operator
    provider:
      name: ScaleOps
      url: https://scaleops.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: EnterpriseDB Corporation
      provider-url: ""
    name: cloudnative-pg
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudnative-pg.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
          createdAt: "2025-07-31T12:14:29Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: The CloudNativePG Contributors
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:c2212e806e950ed5b667455f476481aef179c8e467e435352b624dc670000443
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:54bb383ac6c5f57ae77dcd96cdc33cea5c4c1aff9f9d356c1721591aa820e82d
        version: 1.27.0-rc1
      entries:
      - name: cloudnative-pg.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloudnative-pg.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloudnative-pg.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloudnative-pg.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloudnative-pg.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891",
                      "major": 15
                    },
                    {
                      "image": "ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.cnpg.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
          createdAt: "2025-08-12T14:08:35Z"
          description: CloudNativePG is an open source operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/cloudnative-pg/cloudnative-pg
          support: EnterpriseDB Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.cnpg.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.cnpg.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.cnpg.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.cnpg.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.cnpg.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.cnpg.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.cnpg.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.cnpg.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnpg` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: CloudNativePG
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnpg
        links:
        - name: CloudNativePG
          url: https://cloudnative-pg.io/
        - name: Documentation
          url: https://cloudnative-pg.io/docs/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EnterpriseDB Corporation
        relatedImages:
        - ghcr.io/cloudnative-pg/cloudnative-pg@sha256:53995e9d6a7760c77d77e5298f78b49029427494f4dd933d5fa7a74f7af3bb44
        - ghcr.io/cloudnative-pg/postgresql@sha256:1c97e0d6e94d8b335be7f258eeda2202966df5fde0c934eef4b13c760ec6a891
        - ghcr.io/cloudnative-pg/postgresql@sha256:ee2e3ce7f0ea935c3f4ab245c0b3959c9f1c49816ae02222ceb6bc347f48e43b
        - registry.connect.redhat.com/enterprisedb/cloudnative-pg@sha256:948ad5eb47b7809ba5dd78cb6bb30b1bc64163617f229dc33cec251c1d5bb20e
        version: 1.27.0
      entries:
      - name: cloudnative-pg.v1.27.0
        version: 1.27.0
      - name: cloudnative-pg.v1.26.1
        version: 1.26.1
      - name: cloudnative-pg.v1.26.0
        version: 1.26.0
      - name: cloudnative-pg.v1.25.1
        version: 1.25.1
      - name: cloudnative-pg.v1.25.0
        version: 1.25.0
      - name: cloudnative-pg.v1.24.1
        version: 1.24.1
      - name: cloudnative-pg.v1.24.0
        version: 1.24.0
      - name: cloudnative-pg.v1.23.3
        version: 1.23.3
      - name: cloudnative-pg.v1.23.2
        version: 1.23.2
      - name: cloudnative-pg.v1.23.1
        version: 1.23.1
      name: stable-v1
    defaultChannel: stable-v1
    packageName: cloudnative-pg
    provider:
      name: EnterpriseDB Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Carlos Eduardo de Paula
      provider-url: ""
    name: lbconfig-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lbconfig-operator.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.45"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-infra-sample-shard",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "http",
                    "path": "/healthz",
                    "port": 1936
                  },
                  "nodelabels": {
                    "kubernetes.io/region": "DC1",
                    "node-role.kubernetes.io/infra": ""
                  },
                  "ports": [
                    80,
                    443
                  ],
                  "provider": {
                    "creds": "f5-creds",
                    "host": "https://192.168.1.35",
                    "partition": "Common",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "F5_BigIP"
                  },
                  "type": "infra",
                  "vip": "192.168.1.46"
                }
              },
              {
                "apiVersion": "lb.lbconfig.carlosedp.com/v1",
                "kind": "ExternalLoadBalancer",
                "metadata": {
                  "name": "externalloadbalancer-master-sample",
                  "namespace": "lbconfig-operator-system"
                },
                "spec": {
                  "monitor": {
                    "monitortype": "https",
                    "path": "/healthz",
                    "port": 6443
                  },
                  "ports": [
                    6443
                  ],
                  "provider": {
                    "creds": "netscaler-creds",
                    "host": "https://192.168.1.36",
                    "port": 443,
                    "validatecerts": false,
                    "vendor": "Citrix_ADC"
                  },
                  "type": "master",
                  "vip": "192.168.1.40"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Networking
          certified: "false"
          containerImage: quay.io/carlosedp/lbconfig-operator:v0.5.1
          createdAt: "2025-01-09T15:45:28Z"
          description: Manage External Load Balancers allowing creation/update for
            VIPs and Servers dynamically via API.
          k8sMaxVersion: ""
          k8sMinVersion: ""
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/carlosedp/lbconfig-operator
          support: https://github.com/carlosedp/lbconfig-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalLoadBalancer is the Schema for the externalloadbalancers
              API
            displayName: ExternalLoadBalancer Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
          required:
          - description: ExternalLoadBalancer represents a configured instance of
              an external Load-Balancer for a specific group of nodes of the cluster.
              The Instance has a VIP and ports to be balanced to the cluster nodes
              based on a set of node labels.
            displayName: External Load-Balancer Configuration Instance
            kind: ExternalLoadBalancer
            name: externalloadbalancers.lb.lbconfig.carlosedp.com
            version: v1
        description: |
          ## About the Operator

          The LBConfig Operator, manages the configuration of External Load Balancer instances (on third-party equipment via it's API) and creates VIPs and IP Pools with Monitors for a set of OpenShift or Kubernetes nodes like Master-nodes (Control-Plane), Infra nodes (where the Routers or Ingress controllers are located) or based on it's roles and/or labels.

          The operator dynamically handles creating, updating or deleting the IPs of the pools in the Load Balancer based on the Node IPs for each role or label. On every change of the operator configuration (CRDs) or addition/change/removal or cluster Nodes, the operator updates the Load Balancer properly.

          ## Main Use Cases

          The main users for this operator is enterprise deployments or clusters composed of multiple nodes having an external load-balancer providing the balancing and high-availability to access the cluster in both API and Application levels.

          ### Create ExternalLoadBalancer Instances

          Create the instances for each Load Balancer instance you need (for example one for Master Nodes and another for the Infra Nodes). **If installing on OpenShift or Kubernetes with OLM (or in a different namespace), adjust the sample YAMLs to match the created namespace**.

          **The provider `vendor` field can be (case-insensitive):**

          * **`F5_BigIP`** - Tested on F5 BigIP version 15
          * **`Citrix_ADC`** - Tested on Citrix ADC (Netscaler) version 13
          * **`HAProxy`** - HAProxy with Dataplane API. ([Docs](./docs/haproxy/))
          * **`Dummy`** - Dummy backend used for testing to only print log messages on operations

          Create the secret holding the Load Balancer API user and password:

          ```sh
          oc create secret generic f5-creds --from-literal=username=admin --from-literal=password=admin123 --namespace lbconfig-operator-system
          ```

          #### Sample CRDs and Available Fields

          Master Nodes using a Citrix ADC LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-master-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.40"
            type: "master"
            ports:
              - 6443
            monitor:
              path: "/healthz"
              port: 6443
              monitortype: "https"
            provider:
              vendor: Citrix_ADC
              host: "https://192.168.1.36"
              port: 443
              creds: netscaler-creds
              validatecerts: false
          ```

          Infra Nodes using a F5 BigIP LB:

          ```yaml
          apiVersion: lb.lbconfig.carlosedp.com/v1
          kind: ExternalLoadBalancer
          metadata:
            name: externalloadbalancer-infra-sample
            namespace: lbconfig-operator-system
          spec:
            vip: "192.168.1.45"
            type: "infra"
            ports:
              - 80
              - 443
            monitor:
              path: "/healthz"
              port: 1936
              monitortype: http
            provider:
              vendor: F5_BigIP
              host: "https://192.168.1.35"
              port: 443
              creds: f5-creds
              partition: "Common"
              validatecerts: false
          ```

          To choose the nodes which will be part of the server pool, you can set either `type` or `nodelabels` fields. The yaml field `type: "master"` or `type: "infra"` selects nodes with the role label `"node-role.kubernetes.io/master"` and `"node-role.kubernetes.io/infra"` respectively. If the field `nodelabels` array is used instead, the operator will use nodes which match all labels.

          Clusters with sharded routers or using arbitrary labels to determine where the Ingress Controllers run can be configured like:

          ```yaml
          spec:
            vip: "10.0.0.6"
            ports:
              - 80
            nodelabels:
              "node.kubernetes.io/ingress-controller": "production"
              "kubernetes.io/region": "DC1"
            ...
          ```

          ## Getting Help

          This project fully open-source and is hosted at [GitHub](https://github.com/carlosedp/lbconfig-operator). Always check the project page and issues for the latest updates and/or solutions to problems encountered. If a problem is detected, please fill an issue. Feature request and improvement issues are welcome.

          Some fields inside `providers` are optional and depend on the used backend. Check the [API docs](https://pkg.go.dev/github.com/carlosedp/lbconfig-operator/api/lb.lbconfig.carlosedp.com/v1?utm_source=gopls#Provider) which fields are backend-specific.

          ## Disclaimers

          * The operator does not check if the requested configuration (names, IPs) already exists and/or conflicts with existing configuration in the Load Balancer. The user is responsible for these checks before deployment;
          * I am not responsible if the operator changes/deletes existing configuration on the Load Balancer if existing names are already configured.
          * The operator creates the entries(Pools, VIPs, Monitors) in the provided Load Balancer with the `name` of the instance configured in the CustomResource prefixed with the type. Eg. For a CR with name `externalloadbalancer-master-sample`, the operator creates a server pool named `Pool-externalloadbalancer-master-sample-6443` (suffixed with the port), a monitor named `Monitor-externalloadbalancer-master-sample` and a VIP named `VIP-externalloadbalancer-master-sample-6443` (suffixed with the port).
        displayName: External Load-Balancer Configuration Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 'load-balance '
        - infrastructure
        links:
        - name: GitHub
          url: https://github.com/carlosedp/lbconfig-operator
        - name: Documentation
          url: https://github.com/carlosedp/lbconfig-operator/docs
        - name: Issues
          url: https://github.com/carlosedp/lbconfig-operator/issues
        maintainers:
        - email: carlosedp@gmail.com
          name: Carlos Eduardo de Paula
        maturity: beta
        minKubeVersion: 1.18.0
        provider:
          name: Carlos Eduardo de Paula
        relatedImages:
        - quay.io/carlosedp/kube-rbac-proxy:v0.15.0
        - quay.io/carlosedp/lbconfig-operator:v0.5.1
        - quay.io/community-operator-pipeline-prod/lbconfig-operator@sha256:b0bb724a5d9905633962a3e8267e75d371f754bce866a29cba3ac0d0c671ebff
        version: 0.5.1
      entries:
      - name: lbconfig-operator.v0.5.1
        version: 0.5.1
      - name: lbconfig-operator.v0.5.0
        version: 0.5.0
      - name: lbconfig-operator.v0.4.1
        version: 0.4.1
      - name: lbconfig-operator.v0.4.0
        version: 0.4.0
      - name: lbconfig-operator.v0.2.0
        version: 0.2.0
      name: beta
    defaultChannel: beta
    packageName: lbconfig-operator
    provider:
      name: Carlos Eduardo de Paula
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-3-operator:3.0.15
          createdAt: "2025-09-22T13:04:40Z"
          description: Deploy and manage Apicurio Registry 3 on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio Registry
              version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL, SQL Server, in-memory
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          ## Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Apicurio Registry 3 on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio Registry 3
          - Perform a rolling upgrade of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-3-operator:3.0.15
        - quay.io/community-operator-pipeline-prod/apicurio-registry-3@sha256:e207a4f7f53486b78f95ad958f7e0053a8a8561c3dcb1463a7176f27eb9ea082
        version: 3.0.15
      entries:
      - name: apicurio-registry-3.v3.0.15
        version: 3.0.15
      - name: apicurio-registry-3.v3.0.14
        version: 3.0.14
      - name: apicurio-registry-3.v3.0.12
        version: 3.0.12
      - name: apicurio-registry-3.v3.0.9
        version: 3.0.9
      - name: apicurio-registry-3.v3.0.8
        version: 3.0.8
      - name: apicurio-registry-3.v3.0.7
        version: 3.0.7
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudwatch-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudwatch-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
          createdAt: "2025-09-19T21:59:41Z"
          description: AWS CloudWatch controller is a service controller for managing
            CloudWatch resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MetricAlarm represents the state of an AWS cloudwatch MetricAlarm
              resource.
            displayName: MetricAlarm
            kind: MetricAlarm
            name: metricalarms.cloudwatch.services.k8s.aws
            version: v1alpha1
          - description: MetricStream represents the state of an AWS cloudwatch MetricStream
              resource.
            displayName: MetricStream
            kind: MetricStream
            name: metricstreams.cloudwatch.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudWatch resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudWatch**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon CloudWatch
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudwatch
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudWatch Developer Resources
          url: https://aws.amazon.com/CloudWatch/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudwatch maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudwatch-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudwatch-controller@sha256:254df5f0e6821a65277bc3d029e7af87b10800b192f9724997996224525a9943
        version: 1.2.1
      entries:
      - name: ack-cloudwatch-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudwatch-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudwatch-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudwatch-controller.v1.0.13
        version: 1.0.13
      - name: ack-cloudwatch-controller.v1.0.12
        version: 1.0.12
      - name: ack-cloudwatch-controller.v1.0.11
        version: 1.0.11
      - name: ack-cloudwatch-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudwatch-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudwatch-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudwatch-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudwatch-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudwatch-controller.v1.0.5
        version: 1.0.5
      - name: ack-cloudwatch-controller.v1.0.4
        version: 1.0.4
      - name: ack-cloudwatch-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudwatch-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudwatch-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudwatch-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudwatch-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudwatch-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudwatch-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudwatch-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudwatch-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudwatch-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudwatch-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:610a5e66cf1b6c7f5451c9a0f5e0a6cc9435f2548335fbd50f44b8656adadd99
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle@sha256:bc0d8b62d5814407e08817ece44626400cc3d9d29b14aa35dfe711b907fddf21
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: cpd-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cpd-platform-operator.v6.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cpd.ibm.com/v1",
                "kind": "Ibmcpd",
                "metadata": {
                  "name": "ibmcpd-sample"
                },
                "spec": {
                  "license": {
                    "accept": false,
                    "license": "EE"
                  },
                  "storageClass": "managed-nfs-storage"
                }
              }
            ]
          capabilities: Seamless Upgrades
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.0 <6.1.3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Ibmcpd
            name: ibmcpds.cpd.ibm.com
            version: v1
          required:
          - kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: "IBM Cloud Pak for Data is a cloud-native solution that enables
          you to put your data to work quickly and efficiently. \nCloud Pak for Data
          includes a suite of services that support you on your journey to AI.\n#
          Introduction\nWith Cloud Pak for Data, your users can collaborate from a
          single, unified interface that supports a large number of \nservices that
          are designed to work together. You can use the Cloud Pak for Data Operator
          to quickly provision services for \nimmediate use. These are just a few
          of the services that are available in Cloud Pak for Data:     \n- Watson
          Studio\n- Watson Machine Learning\n- Watson Knowledge Catalog\n- Watson
          OpenScale\n- Watson Machine Learning Accelerator\n- Data Virtualization\n-
          Db2 Data Management Console\n- Db2 Warehouse\n- Cognos® Dashboards\n- Analytics
          Engine Powered by Apache Spark\n- IBM Match 360 with Watson\n- Data Refinery\n-
          Db2® Big SQL\n- Decision Optimization\n- SPSS® Modeler\n- Db2 Data Gate\n\n#
          Details\n\n## Prerequisites\n- Red Hat OpenShift Container Platform 4.6
          or newer installed on x86_64\n- A user with cluster administrator role\n-
          No storage is required to install the operator, but you need storage and
          resources to install Cloud Pak for Data services. For more information,
          see [Storage considerations](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-storage-considerations).\n-
          See [Pre-installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)\n-
          To install the Cloud Pak for Data control plane, review the [system\nrequirements
          for IBM Cloud Pak for Data](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=planning-system-requirements).\n-
          Before you install Cloud Pak for Data services, review the [system requirements\nfor
          services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n-
          For air-gapped installations, see [Installing the Cloud Pak for Data Operator
          on air-gapped clusters](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-mirroring-images-your-container-registry).\n##
          Resources Required\n- The operator requires 100m CPU and 256Mi memory.\n-
          For the minimum requirements for services, see the [system requirements
          for services](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest?topic=requirements-system-services).\n##
          Installing\n- [Installing Cloud Pak for Data Control Plane] (https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-cloud-pak-data)\n-
          [Installing Cloud Pak for Data Service Subscriptions](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=tasks-creating-operator-subscriptions)\n-
          [Installing Cloud Pak for Data Services](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=integrations-services)\n##
          Configuration\n See [Cloud Pak for Data Post Installation](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-post-installation-tasks)\n##
          Limitations\n- `x86-64`, `PPCLE`, and `x390` platforms are supported by
          this operator.\n- See [limitations](https://www.ibm.com/docs/en/search/limitations?scope=SSQNUZ_4.0)
          in Cloud Pak for Data services\n## SecurityContextConstraints Requirements\nThe
          `restricted` security context constraint (SCC) is used for the operator.\n##
          Documentation\nSee the [Cloud Pak for Data documentation](https://www.ibm.com/docs/en/cloud-paks/cp-data/latest)
          and [license information](https://ibm.biz/BdfVHT).\n"
        displayName: Cloud Pak for Data Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cloud Pak
        - Data
        - Analytics
        - IBM
        - AI
        - Machine Learning
        - Cloud Pak for Data
        - Auto AI
        - Db2
        - Watson
        - Data Governance
        links:
        - name: Cloud Pak for Data Operator
          url: https://www.ibm.com/products/cloud-pak-for-data/deployment-model-options
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:96d382b1b26a1e1e4074b7a1aa48b973cd4865bb5c3cf02795306e92c752a8b4
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:dadfa9ef372b8d165276fc5f1ee9c63c3f01367d156700810d63e1ba19f3da79
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:777e090e0394ffdf39abdde285b5b2749d6835e61458f117dbdac108186c34c8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:78fffd8a74eb6495f104ef346bdd03a5636134e2d5e8da9ef5c2647a9ebc7624
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7418ddd0cb77b45dbaa77f98d1bc578f6e55967652c23c7bc0ce14228866a6fa
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:6bec687b2bd5c4d04403a25f56d2245f822e914907771f26c25166462976258e
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:787e01e29abcbb3d1fe58e0b6051067b309085459983f38d6bd18d3811d608d9
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:7de2c2f7cd478baba2011461c8d28c1a4c70aa485404d4523d1166dcb94c0a79
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:74aadd4fe99e66b89471222f853e442fa8338a23d9c24b4e908d38536c66456a
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:2a50cd6ee3cb352ba186a57ba73fb9df321cf98f0b79b3d81024cbe57e0c7f35
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4184f25155643f01850f286f15cc92294a92d8fdc8e5020b1b522977ba429c2c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b7468b824d519e46e3ea23f5a803a8ace030b74d71d4cd282819027665ff0e12
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/runtime-manager-api@sha256:a37773df07fe757dd698e3f09f8687902e7b3592d97855e703dd92ecc832aab0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:23ba1b277cf76bc28f952ce6f5bb02d6177f76eca8a8b4ff7925ef119a994a85
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:fd35c33c7361bf04a66e69112b05a078ad65fdbe8f8ca8b7f4ecc2143b1643e2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:0d427ac013c895782653f5ce5d9644425edab7fcd0c6faeed78a2d0aa631dc50
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:2f2590734343c99183dba7aa9fa4debc73377ee707f9bb090680750e62acdd68
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:2f45aabc99eb9646db4c74686569a7f2af3f27cc44362a30c92d171540bb274c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:ca4da5a8a3c1e7add6ac630c3fe4713a08739f6b891b2290bb17026b94c31317
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:d7ca81372914a28c8298de5f9d3c35cdbd6d8d8582ac9b8fda5a85d22e3c0def
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:7505eb74c6bf165d40f6a8f3511f8e891cc34c0201fee38240c693a08e002ab1
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:934bc42fb29cbe616edcd93d2a060dbffab5e5a8a2ef4caa7b2e538f26167ee3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:67bfdf9115fbee782e35ac61dd93801bcb9e622d441806f1144208b4447f1430
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:07142ce0d259412d0a9908b63dd78d845ef6b7c5f8a24442b93964b4c38372cd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:2472053d912f693a5fcdd00b7d9f0a7ce98697c47eae955e31c0b602fd8a4a37
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:46bbfdc1f5e418db95ba806228fa5696be22c4cf9eefc185ce27bfa68ec1d247
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:a3dc5fcff6a2aeee7c7e562fe9f6e0cf0652fd7ad4ec61496021d366206e2544
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:2e70036d7299168331c94af70a22e3b2063c56160f616362b1a2187c12d003a4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:631a18f4b437e5092b9671fb4c86af05e99effe5540c4d6f964dc626466764a3
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:f662be48eeb6dc93a8a4b0c3374f59bc342b5b8f5360437978cbac8f74c3a084
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f6e8de9f11696ca22302b8576ffd5bc2c024df0c8dc3734c404cc799565f351a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20f05c662659be9b2b54407c02f1805b93f09a3c204011981b7f17a76caad8ee
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:0ae22914dbaf09d6679bd06bc9b6a5728e744c9ce2309cbd5f01e28308c403ea
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:c078bd65fb18b31ecfe03c832643bd6fd85f7330b94a1ca26f557d32d89903c1
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:40064d2893ce54b358c02023f4cc3603d3d28202ba56a177b20636abd74cb559
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:63d4a9fa16231cc0c01f1b8bb9363221ac1e93f1571ea132866f170b156731e8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:20ceaa70a3ab5ed8a2e451a1e2960d04c201f0087fba1679fb80c57c6be6117a
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:245f81b08980bfd4790cc3661e609e2cf53f3dd4d4c1f988e5db4dcf316d90fb
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:25c98c67d09cfaa3a4d332c461fb6766700a65876a48a1aabbd73b58bc2653f3
        - registry.marketplace.redhat.com/cp/cpd/environment-specifications-sync@sha256:894b4624d8ead7286cf9068c6477300992c32d96339644fe764bcb29b1c8ef5c
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:209ea4b77b3b0863c591bcdec6f53c4c39dafcb10ff13e27ae4e1cb986a59727
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:2754384c6936b36ffd64f89df751cbe47993f73dd700a184af35f412b3408d72
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:498ad7832e989e9356aeffbee96f310e3edaa6a7466320fda9b4f75034e73eec
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:45194b56f9fcb3af83158eb0297248364db40e8164fb4fe4f5656c58aeac3fa5
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:2dd02d8f15dca1dce7cecad098cd8791cf1cbeddd2abd8e8c6d8f066d85f9e2e
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:e4eb6a75e0d207232e479693d9c99ba863f8e62830f3191ca4d1e7ceaf3586d1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:10a935e41cc2dc09f8f3939cd512d9f6bc71d68457f77793a6957c5420cf4bf5
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:be5679938413dfb7766f765fba7923d3079d2d9b6f464485a8309e8e306f19e6
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e12599ff99db058fdc8f9c1eebfba31160c636ff092138deb1f6f6705f44fd62
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:0157fe80f2c67786cacaaf1302c334a6e0ec0b941afd870e0c4ba163ce394216
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:2668a562eb8df1117c40acbc3600d6735469d210e5aae4aa03bf9f54d0ebfe46
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:8438c2229b500fc9c40629280701856345e228fc645a5b468bee2135c61b2387
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bcf193d6f30ce1c7f9337ba8efa8d9b853eb7961e0008aa82635aa383fbe9d76
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:cb3620eb9aec24f7ee54c3ec6709fc223b685177500c503af8ffeb23642dcdb5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:42ffa27ae47b577cdda214e0fbc46c10bd108211e7dc58777d8236c59cb78766
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:aea8fcb72664a41c3881839086a47fef1b54dd672300d4aaf583fd0b575a64b4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:13a9ea6ee24d9b0ac1d5fb09c3e42eff731779867e7e71ce40429b43ba2779f9
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:21dffe4500172f74345fdafb9beb476e0ccffa55f7bca59f372f312f8d28d247
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:be2ddb821274938e5ab926627bd998279b6d19b7789d9683513e85a4b05bc2d8
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0f01a67524fa46829e455f654d3b9b00ceac44f3d870b21c022051dfcc8a00eb
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:c0459e5d3acb3b2a6d9863efe6ca1a24e529b38a6ddd3cc1e13d143924e872e7
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:78682d9b9f19eae363cec0b6d31e696b2a5fd1bd1f8429df6f65a197310b6778
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:48828f128c98e73f8e345d87bc228f0920a3ff2cdc325d4a4ffe73ad95c5c5e0
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:96ebe954a21f697ebaecfe5aa2c497c60b2ddec17e9ec4b376f435e6c53b3c29
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:603a6bf0e616f16b609bf157b9e21415802478aefc9c606e124008d0224a1188
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:da38999d66199df0eeabd7c9e934205fb858cf2c52be8bbaf6b34adb894dd38e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:9a970756cd76ec748b555d5bed97889b8c9f410db6765eedee84fad4976cd523
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:ecac5e0327d939ca84d304b2fe95739ae43a06745ef775ee59e4e5378ced65c2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9186c350fa54e8b792d72e47be6b54a67b610f726175c9641795010d45ad51f9
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:fcb79d76f965e74f7e3459abb92e3561ba2f016b34e2ebb6f88298bb19f7bc11
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:4f0cd4920e7cef64e9f49ae1abc5d87303102fc96819e0c3b699d86b8f8a3980
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:582ced3f60212f736f720558524225b0d4a1c79ad4890656c47dd58009ffe0de
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:0341c499ae144ee81dc2b92e8b99ee313ff6759374b7c1a6469156c20f522921
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:ff5a5be492347e75b309194661d7742d1c909c3765c9947c8b41f4c3af62b08c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:ecbc65705de76804a3c7642339ecfb4cc0f5a61891058e07c20bf2b63f1b82b8
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:c0b6ad3b88f216379fc8f14aa5f075f5adc06ad06cc46495d3a40cc765f3deba
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:aed24dcc7ec24707c978a17b86501b65c6f0a77bd3ea0d173c28684b9df9cd91
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fa3d8a54b1ba33465171240d66f62271286cb120c51c54267d2b4f8b731c6177
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:5f63e062489b8e1bf49b5943e7a836174fb47af8dbbc8092ace5c8b72baabb99
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:3f6e9eeb659857799e30011f222924cef35301c59eaa5327ab5e8d212af75ce4
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:f88a7294aa200d3c633c0a25c4e4ccda52c75de174ba508d847ffa3092d5b8d1
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:772f98763527c5452c662b2ae516de59c1f6edcf339da46231f4be8aab6c580d
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:d34d561bbf8afbfba9362d2a9a58f2919faed1502abf72e11882dd9cd4fc9711
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:427e431d5b8b56011a91bc4f1e1586cd5490944615011fa6ac20abc8f45f92a7
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:3d5c6573faf4ae00cbb21cbfe9b2ad5cd8ffb658df0c5ba09bd98b3bf54f9979
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:bbb457d4f6b7bab065e909756c2dae31af7162dce10bfb92f274aedbc98d0762
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:6cd1825ba40071bca7f46c1af3e813ae7ea44a393dd3d10e3c83e62834563ffd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:8ed09b7d65178aab93b77b9d39371bc86b1022f32c5e63a9192baef02c71016d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:8160c688f6b66206f8428f78dcf5e21b65c3a1cadc38b765f22ba8403a87f4d6
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:ae5e6858ed698d67b4b3917c0d41d0b4f7e2564585694f834099584c21a3c012
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:e21a346d15b286956f62bdd1fbb6d7b77a554ee10a475cb698a8796c42f3a267
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ee161954766211651d7cb13f3f5a3c025ac04094ba742be453ce789c6f0e24a
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:bcc579eac4be59609102836252ded82a17ee0088e5ce28c684fff41c80232a0e
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:83ca1235dc054bf9ead64932c85333ea0dbade076b4a0cd1c3fe2b6bc2e05896
        - registry.marketplace.redhat.com/cpopen/ds-operator-bundle@sha256:e30b09797e341a5fdcc3ec377657e0e65be5855fb850dfc31f35b200f43130d8
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:ad027c91c719fd1d931cbac8c76735474f676b228f9ab23d1f243201246b3238
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:860c17d2b2c6b0e013d0f4c035a388e45d8ef1483f53aa7ad022913b025f73b0
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:b0c8bb70a01bb19c685d019e89f83d8ed327bfa670a13b78d3e803fed77499f3
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:83038bb7e3f4b4637f0d5095b42dc906e7fc27bbffcbd66f6ad93dc9b1cd6bbb
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:3129d9c7b096f9fce9cc12002eafb5c3d731a162ddef61578455dcf8acd31fac
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:a34169228887ff93161584753c990bb82a5520ad7945dfa1ff4078e70aede592
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:cf2b88be76a4aad436f45a79694f32d2e2175587c3630543a6d24d7725b60c5a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:107911ed6a60a024d151ec3724ddd6d2d1402aa70a519b9dbec1a519d88231fe
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:2346f67ffdc1a802720b325e010018cdb908aa458f372fe50ac6d8a92d8fb733
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:31470f873a0931d130484d5255692d445e3127398aae0e777a3c3cb702917743
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:0b640bda3afa672011d5da390af798bbd3e6dd9126029a9eef03846d41582743
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0ed81ba4ab2caa7ffd1cb90e11aa8915df1c5bc1b76aeb4c87532c84097bb102
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:5f086cdc7a272384c8a9b680f8f90fd471799e9e96175b40c7f16126b73169e7
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:62e2b059f565f87beded59b36962c457c9662d1e401a78766fda328ea1aeb78f
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:37853d2979d0ca5f88fed76fed5194b628507ef0480e668dfd6eb034fa1d2023
        - registry.marketplace.redhat.com/cp/cpd/environments-ui@sha256:2923664475c7a8ea5e63055c90748db37825fe76c7975b845f1185d4b32ee4ea
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:009d98ff8e3cf353cf8ba71a973228e1e409e4de21373498534a6cce01c07ba5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:719b0871f0d32da894ec80dc2a23483623b7802a5c6f935cf3672468f99faa40
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:bf4052b099393b563332acd5dcba6ba37ac424aa431d2d4f21532838b5ed6d6f
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:aa0068aec14e37986620c982295c083df6fc6bbe95facba582ae3c6f769a8e3a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:be5fd8a8b43ae3792c58de1342cca91836287fb584bab003f7f9aeeeaca3f65f
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:234a8dcd8a9c5a68533bce6e0f793324e3ad7a660fdedf95895c5ed73a4cee3d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3af0967b0fb0a28a83a3e007b2c78fe27ad56a5e9c153806c6a7b17d7d9b1df5
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-usage@sha256:fe5bc4a40472926ec0df3023777b2aa8105507128bd9ac02eac2c68ad08723c1
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:e3138406cfe81021370924992dca891cba22b3be2eef84546faa0e8f8aa57ad6
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:4f0165926685f49cde22a93387fea401bc1799ee846041fe54bca148dfc2f62f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:2f40d8188dc8b41fb072256ce8d109167095469af416b322ac023071814cda8f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4c311560fefda9dee6744bfdaed0237b88ef427b890f2f7e14c1e74ad0b49d29
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:aa8d647f007d921db565a505d52583511e25de4a6a1187fc6b4df15427944364
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:f00158405adec00c7e992efa7d7a19a520df485cbbcf39972e579ed3c6e4d289
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:47cd9776a46ae5fdac3a101d79a779baefc802ddf8e1f8fb02216f47feb55fa9
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:4a87dbcb7d2ca2d121512fbf57b645eaa9ab745439b79722f474b84efa9eb391
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:c6862158a2a5ae0a4e4206bfdc7f73e1c2359544966c72cf4fc3a617acb13697
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:9d4ae40d34e38a48ad2ad6556e997a42f6308006d5f6b8638cb7bc4ce5b1e40a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:7ae4dac24ea054ce527a507dea6c77a3233aca0047672ba5a367fd5742448517
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99589ee11b50af8ed6355a5f9c2272ce1fc7f9a13e80fda8eda5883a332782a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:6df671937bc55c410a9df13aba15687b2832c7c19c0eff4686c3f83648528068
        - registry.marketplace.redhat.com/cpopen/cpfs/cpdbr-hook-utils@sha256:63a5606df81013b8a7b5cb07bd1f29bf032c8998db843885184fe86a76fb93bb
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:aadebf0598c492639e7c1109ce681256731215b686fec808578bb09558a40315
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:4b3ffae1ddb7a2c28f98199be6343ed826f8fd21b64fc9c584566341f44e44d9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:e3d8d4d5285d058130345d6108ed4056ea9c871ed8c0575601a94629e86ec691
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:6e0a23c3434046195428792c45c95869d81fb4ac4ba75444e4343963fda6014b
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:113fdfb9395c96f7eb65f9a20e18ede8a2eae4aa6c91660be8fd3e517656e156
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:ab31869bdab5180b816a89c45008328cb7013c26b09c531bdbe5b00382f0dd7d
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:3112fa03dc9a530642d329666a0cc64c06a04308beb1ddca0f7472915719ef5d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:483bc8f74d02efcbfb40ceb6abb0ab0b6581523cdea576b34512709b98879233
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bd237f42078822b038c13271f35e6710273c2a4f3fe2f3eb74429c93cd2a0e41
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:e8eee5083cbe317b368c002f47ce09f98e0c5e3407e5a4712e1c5473c4090ba3
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:ef03bb6ed647d92b9c84d34810376da7049aad88589eb2e35798d82a22f5b994
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.connect.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:15e8cbce994b783fa6dac737ac28f1b713d1c75891dbce4987d2c09ff295376b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:25420a21686a32c62e83b99fcb215df015dd13fec10f73f397f927c61226f964
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:bcd879a7cc3442ef62bb9677cc5534a2bfc9e7ea248e566aa4f1f6a0242431a8
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:99d3b5a88d20380bcb4161e3b182f305af2e4cb405460d0f2a91965dd6f99f80
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:aba8f1d08d143a8901b4d2cd8790e067fc7b9eca9987696a1ec6f9f505a686ba
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:6caf5203489492755e5cf6bffcba53be0bd32662e8cd331dada68a85b2c083ca
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:55d4c6a13452f76bbab485363288efcace23975623b8a4b6b83f3c549f844c58
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-bundle@sha256:94c7a1dd290e2abd55c5b8998cf08ad174d3fbfc659ae6fa6fb6cbab61d5fe14
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:838f7649df5a7c13d6155394e211ac565aa86233bd30a3a2aab9706ec4a74eec
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8da8b240176c00f8f6eadf32e5614c663cc1d56d546c73996cca614b715acf4a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:3ddc6bbed1ddae1be71043d177cbcb58830d0dad542b06567fcfa254dd11e626
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:96e6db93b2d16ba590252d6568b68f254548aaed73ec17e4bdd8d530ea668be6
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:c8d2825596b7e675c247158bdd06911e1cc0ad9af3364c9c175a5b4cf16d5d8e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:00c30fe372524fa82b5ed4a637509dd9d4373dc727a9abc706221f832858f9bd
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-velero-plugin@sha256:fe562192e86558a17d1c88a8fe21148bc000fa93c8928f7a3a897acce62e6798
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:c411456aa41fe705b47dec60eb49c6d65d4b7bd1a64d1e0545817efc95afc67f
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:6f2e7cc917adb54b3112dc4909f645ac08c33e8c2c33266fbc8f5133cbcd1973
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:cab91a231068a6bbef03ee8db32fadf11654317c4573c6714721a15d0e2a98dd
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:8d744af7dfeb9892378b7cc67d0e85e29d12b6f79c8b6b2a53f7a53aeffe8231
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:6ccbb20c9fc7ce00ddfbac19c8808417957481473eeb1ee62ecda165e0e275cc
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:044b663e4b7508c62dbb28b7fb2aa267c9771fc62ab509b08d46ea1499d95431
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:bb45d0217b4bc4352b209e3771756cb82923b8d260846c75b33ec270b4a8b8dd
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:6e055b6e07cb1adaa652b8e7515c4ce91438593a56ab0ae051435f55373d96ab
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:78e33fff4bfc2e33ea179eaed9902df9bbd11463b9152d0281d1e0924fe09563
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:7dc67256c002e86ee416c60ebb2942b35d8dfe0a5f96b4db495eff7361d21466
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:cdfb097f2438b5739bb88905b8318d15d15c4c64d37cfa6fdb06f44dfe976e6d
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:9fa6eea66a3daff39b7c0c7839ef5349b483b29dc219e1db384aef6c4c31ecb4
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:d6cdb71edd7b7451a206d1620b35c321f88b3ba8028468329595cd538a45abf6
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:bb6492c452fa6b481551628da26bb85bad151d12afd34a89c1a272a9587ecd1c
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:b2ca6bc1119571cce732f1ce2ccb9ce406b298550b55144f950bb46cf985df3a
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:508412a22fbd65074964150d6b270f38d6df898e17eb7ec5e28f279c4c2e1062
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:48d237cd3d363da8a206ca5e092d0f91e8a5bc931ea81763498dcf6f28c0434b
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:1019e13e6f072c07da6d31277c8ac639e8ff2200599ece1666a3e34745e763d4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:9c635194a9c63fed440e7c0ab0be1fe6cf59b1c3f954216dfb9150fe9e0fc207
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:7661b04bc59d64a4fafb2f660cffb99e3d326772749759ed0d2a1940ba147078
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:9c24d02b1cbb3857d017453744bbb2e68fbe133210c35e7da5ce0341921d5b08
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:51e1275f6e8297418997b0ca9615642153371c29b3466f155e15d7fb9e80cda9
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:82f60c881048618473c07970d6463b3186518bc18e4b263e16ca61d15448ab30
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:722944bfd0eba280255f4c198f65f562423a26e2c4ad3580591ee71cb6e9c2ef
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:786b21db2a9ce6cb0eabd5cb7c93dc41a8b1db6c1f2d07725a5bdd75b3a1b7d5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:0a773245bec514cb893f730086f39a3789d0204b82cd11f5d11a5c270b92f70f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:57e3433f860ec2b65f5c5f86574614fa9273c09b6b37db96a9e1888f4f6932f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:749dd365ce85a4c9ac8e6ea7f90a34b791a7bc5300319daa85d427b04f40af65
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:b5a8fb14c3bb7c89ff970ca85f06dfcdca832e2b45ad86e8bb1b6c0a0e7e0764
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:66c11636fdf68631e37bda7467536093b9402e883999e7989dbdcc4b4d2ba7d7
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:c41e4d3314fac0d2cf5f08e7db85801cb2098d37d0aa8104a347e9e82526e914
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:07c5e4b3dc18b921339a7a5c78b8fe9179205678eb662377e8124e46479d7f25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:351375de35d7fc9d940b905b867ce8a00b7ff1416e70d3a100918797fe3d8c56
        - registry.marketplace.redhat.com/cp/cpd/spaces-ui-job@sha256:95dade51444e30ec034c12207ecbef675b70889b81970f49f10aa75160b2f947
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:67515618d631b0ec00ad95ecb9bcb48d7f199855af9a64e86c4b46e7ae3be69e
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:267b7da5936493b0daf8ce824e3e6b50ef3b2bcb6e3126ed18fa9e09cdb0da43
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:f5248b19c4a9b605a9160ecdb5dbbe0906841b667ede6fd6cf8e1d6c4369fc8e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:04f9ed67696fc193877f9da3d73ee3f6f4cadc7b8d09fd07fa12d0ee15f1152d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:b9a255ea269a11c85f52fe556291b0e869efbff9918b488b65e4297b958c6059
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:1b29e04308bb8e3c990a8c41a5d3d1e854e197bbdf0cffb2833fea4866772ba9
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:9460c98c7202ab368e732951c13bb6f5064cf18f69d36b3d26eedcfed233ec81
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:b646ea3330f3fd9bb3174224de69f20ef58589a464d50d1c9399211141f67bd5
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:2c39ff1850b7a9d32a11e99393039ef74ab363f3fffec1e4fd542f2ae148a6b4
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:0ab7d6651cdeda78fd0416137eff72eb58a621c8cd55bbbcb16f269808dec7cd
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:9f7e6a1ca3faa5af1db2afe8693f6bf9182296be2a23e305ba6c990aee3fc301
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:12eece1fc0a2e14244a2f57dcf7c198a792ced9ad0d02c9606645bce1f295877
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:b9c8a65c1742c72c6351b1c8f4a858940c2bfa00ac5ec271c57b2ee7e5e7a66b
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:20863de342e07dddc9aa8eeb214fba5cb17478bb88613e4e1f167d77cd0af214
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:61ba2b05a6826f618853ba048496e2c9558e509a01648c74b50cce6d28d0d764
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:60dbe69693314c937a69bd213de438c3af0a8a290aa9b1cfaad8a57e9d04c448
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:8305f26c2ebc493b54abfea0757d65b679f6c96bc3a0840c89f08a034327ed9c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:d6575de3156c7a9b0553887abaf8713a758d9c9916f9cdcef611923c50632867
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:bc39ff644108a458d1b7e29fcd88bf53982ab4070ce2ea0e75b0908678b9205d
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:1f82599d91ab20d2024abe134d2f56caca1f9bca1f02d64352c81b5f1e0f1730
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:9cd2ba7ab3ecab7094c892086a5276bbde22ff96d26ce0d087df2c1791235b02
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:a65ed5ac55cd5be99cc43c5a8d33bf127e8694917fce5f83fb8401c04e31b1f1
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:6bbcb64cf9adb4d3b6a4c259eb910e7da5c3c58732643e60d416399b2083cc0d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:4e3ec61a34fc7c71d213999b68b1f5c3f866f97e19c12fb001d506d4166a50f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:a99b1f5a4eb0306b83562de0184fc826c34835ffbf46d929ceb9e70e0d325361
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:268c763fec7defe810624bbd31c96407fa37fef47b904758bce4c58d95aac043
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a4daf28771178b0af391965684337805f8de0019a677753112961710f6011991
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:5e5176376a664755fa776efe623a1e11bc5d3c31a3d0af980b165e54cf06620d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:474187fe2837da4b86ed9482a96fbca0a7f83db149859f8ef23db84de9a6d7fb
        - registry.marketplace.redhat.com/cp/cpd/environments-init-job@sha256:6af8f900af7c11eee6894e2c3e52f740b3c700dac0ee753fadef330230bd42dd
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:7693312b2bcee7b8502de0da8786be985607fbfe6616bc9b0da4af8a6fdedea8
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:81d8b25221d827160dd5e88efd383453f532057ad02d7a2ee3dcd9c6eb16ba65
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:c330a321438943fdd264b292b68fdfb8be89c3bde93a7bd987bd92c9dd2512b0
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:f561f089c4e5037f7c04971d96fa024ba4df5e6d17834f0c44545d0e53c772d7
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:d06ef794046dac2c58163c201f55489b69081dd15dd91cab16387510a50e24b2
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:b19bbc36033c3e60f387a9924711aab1ed81f150ca9d8f435e5433fed7f6a5c8
        - registry.marketplace.redhat.com/cp/opencontent-os-base-plugins-2.19.1@sha256:68a98f52b7eba7e41c78aa91602c4e88d21d4a2a2a11b9b7592f181ce30e41ea
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:b82a10ce8f75d4441e869bf62fbfb5da349a2f228b52178bb29723af62f13938
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:6298c1561c066eae3d420f20212a3c7a8f9069a6996408685437588bf8335dec
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:4a53892a469c6b9b751a4cc2449378bfb0b15bfe1f3c0dd5056eeaf1587c82a4
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:fb15cbade10fb37b14afe063a93decefd9a40583895b7b6b436f36f08bd73aee
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:bcfb9b7d3b379ee2bb0b449cdf5af3fbc88f99c77641441ca392ebed04403e00
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:3be006596f8e1d36865a9000c77df3223c264e807bf0aae6049a23307e8079e8
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:d5feb4aecf8f36531db4c52e672558c04f7e314a1a791b27b5638202f9477bf5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:fb850a3d660729fe8281967c2cb466fbad3f01fdc302db17988c27682217ad29
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:65d327900e3b7a874dfe4bfbd0f2c7a1ee402346aba0e8cbc628b406e5b5329b
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:fada47a5f4824d3cd5bdc262b92018ca0e697e18dd373f92fd93186fb8719a1e
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:93f73afa23f9fe8b73fd389081281fe1b0fd74bba6aa9bf7a6df663053afd3b8
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:ee15594f82d374da81dab59c559567af74153524a6d9bfeb28ee085bd4109b37
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:0cc0283b768c2968849f24cbf7178213926e084da5d0c90c18d38e2260e19138
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:132670412a0ce6dd010869e569504d2e89aee0983ff221da371c9e30d5f8d8a9
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:ad19b65042fb26cc1244a92f81a96e5ceb4313f8c823181efe28b66e9fecf9c8
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:9c2093089ca0651bcabc1d01ee2b69d2fb0cf2a277a8b966cec9533950fdcdae
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:c70d3c86d4c61bd87a427f7a7571f37fc4d0966ff4f420b73ac4a2af73ed1379
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:b87fa8374976ccd3dad90ddbf6b65cd6ae5a028ba896562f284f089380ebe44a
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:2011f32eb1e68ed5b75a4e8a99c992eac546995e219701d17a511596a5a09361
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:2698150edf65a99a1d24bf3e8336f02b57fd964fcad283e182794634b033af89
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:351358f551dfcf817c5936c0b43c2e28db61adcde65d7d2c81cae666630426b8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:3f2f97e6b2b18526255e88ff02d1039b24312eca51ee22b7cbf76d5680f7aee1
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:54d8ceaaa06196e1ff524b7be595cbe90811f3be1a57c767b1ba74f53edb2f0e
        - registry.marketplace.redhat.com/cp/cpd/environments-api@sha256:6dd60b530f81d7a2d55a534211f7f128814abf87b8cd3575d528dba53442c44e
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:a1a5dc5fb8e18460f9f6e0069eba39e42f082119159dadfd69b0711b1bc18c1d
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:07c13b18197ed0bdd8b97d65312a342f95d7a7fc43d50af746d788032a67f83a
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:010bd5bb9943ffdffa8fd00cc8226e59d9583bb3baaeacfe70e5b00e805878f1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:a10a15f07b856a7803ce0dcd30a542741dcb591804dc74c2a5fd50c55de0a983
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:ec137115210dae9b69cb2899ed8a5437d94c8a658ae1ae5eb295232b34b0925e
        - registry.marketplace.redhat.com/cp/cpd/job-trigger@sha256:f2ea487072393a24e4c6864b2107c1fce072ebba07f9dbc2c139bd37fbd062da
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:f40a9375fe5722972f995a82b50c00f254e3fe37f202b0aa1bebf66b0504099f
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:8cba5f256747002b62fcedb9066f7ea6dfdb1071eb359e9137fd413d20b29ff8
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:556a7c9cfd0b6a9eb39be798760f361684d9f9276c6bee23dc5e8d14a559eaf4
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:6d903fd4bd0ef3ef361d22b209ba7b686601b49d6f8d4e29b5d030da8d949dd5
        - registry.marketplace.redhat.com/cp/cpd/portal-job-manager@sha256:6fe90f051d9261ad224dccf4d9f6bc9f19ce0e8d22fa010a46b2752374bce2b3
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:4dd54195a3283a820d260eaeaf83463722e49bee05d875333c1b99201a7e2b85
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:9f835bb2984329db9db7eb2ac5521371ce0ced216a7875e45ef3ec31d255b648
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:26500e941ccdddc08dea07d3a4b9aa9eae7840fd5b683191320060f6dfc88c97
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:8d9e35c3f29f8f56d55d81a2e6cc4d1b0cc352ddc8d33d83dced83c4a111a0cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:d76476d2ecd88ea7b4fd857e591cc9b71be9d31fe0fb0057baa1b21e6113f730
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:2d8f924a7e4ef373df9760b4f1f3ad8050facec7de0c0efccd8c27d46b756e91
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:3a872814d01ded278afc8dfa6b1569d9e933e8fbda28c55a991da010b4304c69
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:4794a0fdae9d060def1e0d75a98560e713ca18de2625e4d9e5b5fa4c4877fac2
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:c8786cffb4a27decb5dff3d2a9682d07d11b7096a689bda0d7cf105aa3156299
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b341d9af40ad6aceb4ecc792ec75295313128a1eea3a0974aac879aed24a89ba
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-catalog@sha256:23337786ca20ab624eb428b0d53490975e2b08a2db642ccc76d974dc0e455f55
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:130872119947fe91cbfb8dd82b3baaf78f7adcaca5933ca394eca3f9ee84c8cc
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:f6f3758287dd31889933ed0461fcf02dfa85d872c3d193669608573ca5269524
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:4a5ed208b68613d5d88936687f6912b3aad0e93c75f0d311120a8e55200f5795
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:525924934f74b28783dacf630a6401ccd4c88d253aafa2d312b7993d8b9a14c1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:24dff64a80b910f0b4c5442d4f611f9df7602a16c5231eff2c9d08f07b2a9f9a
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:b2034fae4f2287c8997104154064f2712b98d3c4382d1faea3df777338711f23
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:d30397b74780cdc14930a3a02dfcf6b99264847702afe0513ce536983c32e2b1
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:f5527ddbbd0b9878555fce852996b13020fd91b1604d8c03d34a61ca2fdb2d1d
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator-bundle@sha256:e4340bd9481da998a6b89e94c884c46e17cfb50d9d0c1099f39a3c54e500d2b2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:8ccecc00f3a9656a3dd1b07af76ef065e301e4101bda4ae8dc5bb91097cad7ad
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connector@sha256:b10ca594f31751faad4050b4f7717e0e98c7977fb7e26d4866e1c14125422225
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:46893c166cabdd8b3422a190e220ee2e34861e797aa068100ccf53da349e3fe7
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:bb2f19e1797a900d60fb9f05d9f115faf20d514000aa3acac55fdba701581d81
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:56315ce0a0a70f5899191a7eab856105489143f0e889c8cdd1b6b8c225f02eb7
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:075b6bc4866b812c4b5394c8902b012978449a710ce71a30f55231fead941105
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator@sha256:78206c6634f2be2bd36806eb717bb8e4db2c755263cefa6d026a425a19329324
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-network-performance@sha256:30e0b58531f876a47174c14bc6f94a3c38a40f757403775a06745d4ae992060c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:276d5d06661ed401f0b0672dbcfa3d9f07fa2cde662fbb532fcdffbe3e332227
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:47a3745c78280f98f2ee352bb9192cdcd2fcd6d4c1a3627e7aa42178a0f107fd
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:1279b116fbcebb1c052eb27a72c67011b44a12fe673a86c1ece28202afd911d4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:849ecfeaa36fe365c7dd8f1a76fa0d263a1196fd16af805dc66035643e8e9d9e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:53d1fba68aa0227c2f4c8de26837f6c114c56ca4a397826e1f2e93bb857e5b7a
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:894d67753dd0c5b8625369e53635a1ff53066d3910b73d91e5625dd8907bff43
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:b87b6c32a2050f1d5ff311d163a7b45efa7d5de1d352984b2f70fa578301a952
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:b6500764d4f6d281f1b42458eb223d21f6a413b5db55f2a2ba1a7973aad8465c
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:9aec2bab8c7a6941e97c18352cb9e838a99aec08b06269e8053ac764638a90b9
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:39700bf47b93c7f0db1894148c16c60eaf406196034351f54573f0a94c4c8aac
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:7655f531c354cfcce438c61cc5cde2c7690bfee4e2eadf9f99e4b157f163cd73
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:095c969026e191bc4fceb020b5760e7595e439cca91b88f0d8c3e422591d56aa
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:b3387bae415fe3a66db2264ee766cf43aa8f31e2b615b200baa9228216596a0a
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:de7094acd65a80daa67669d8eab809220dc715b9cd6aaa96c58e5fcbe34c460a
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:670a36f3400164cc9eec5d23422d184340e0c1edfb13a634e58f0345c325c21b
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:66c1d2493e6721094e85a0c65011523004f9589f58b7f68cc6f33c5e5adb6984
        - registry.marketplace.redhat.com/cp/cpd/ds-cas-lite@sha256:39e6aee0bb64fe6dc18ce81496936b113de6d6746221655cea102e04c11dfa33
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:b368c22956707b8e594441c0e9bac20981339a5713d4fdfbe3c64bbfdf677b35
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:a76d3114418a3df1d4d3cfe03da587c37028ca1697795e2535cac9f6f190457d
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:523dcc671f1c05e690d5e578a90300d28d76570f605d4de7f80645b590f4a462
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:86cdb5b6543e7184361cbf0b0fa556277232460737ad254e920756c6d365e47c
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:5e161ff695c0552d1333a1f993048bd055168a9be9e0a75adb7c4464c9c2b603
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:5ec262d1ada13a7d5d519ad98cf82dfbbd915664bb7560408558001ea6bb8451
        - registry.marketplace.redhat.com/cp/cpd/wml-asset-registration@sha256:70d23c45cc636af2148fadc1c7a08802692fed14e1a31893214b8d2d82f06e1e
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:5aecbb4de6fca34f026de72b7480f35dbb2ee8d8ef1f22e524a85a71bea99a1d
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:2565a8ba9aa103956d6902f2bcb53ac13328b647a2d7a150bf80ea2749d5b972
        - registry.marketplace.redhat.com/cp/cpd/jobs-ui@sha256:49daf607918b378593485da16ac417056ee295bca94f1a5cac37b334381ae096
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator@sha256:ad892ce5e6a3f69ef484eed68e92d9c972597a86420e22ff80c247689ba02905
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:a0273c013c5943c9b9a223cf1aafbc96b3cdd2dbb147c73d2a9ed49befdbdfc2
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:5d889be3dff96d1e888eff94bba94228aa46f4e94e5f30ba33077ef00692509b
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:811d1cbda1d0287dab48cff90c7091dd1bc8aea385f562c896f2b09f075cd1f6
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-connection@sha256:6f96b8b6afb00cb26fcf903b8e4c30b0d6344bb86d8bce2f2026fc8cb0b575ec
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-api@sha256:dd9e4909317512a6fada5ecfc7cdb47c6807f266fa456a01b60d7b3486faeeb7
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4eab846d6e80d68ce488e2cda36329cec1a6bf2f81feadf666b1005fed294e48
        - registry.marketplace.redhat.com/cp/cpd/portal-common-api@sha256:da0190fee910cb8245cf7a61d6e7aa2bf599d41c679eeedc183950918ea2611f
        - registry.marketplace.redhat.com/cpopen/cpfs/common-web-ui@sha256:c2aa3566f8fce992f00f43867173e5f47593e8618357369b23044fdf3c67ba53
        - registry.marketplace.redhat.com/cp/cpd/ds-runtime@sha256:2712755961a1c1e23f30d49341c037858c0965dbfc065aebda88c84526291f21
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:8542b9e449b054e455abf6e7884dcfda708a6213015e932844568d1853b65eb2
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:0a32e127dd305fc445808d0469147db6c9704cd15d924d57d791e035475ecb60
        - registry.marketplace.redhat.com/cp/cpd/portal-catalog@sha256:690e9b5218a200de93a7c8477fd2980d8d341f22be75002967201db09b5a5524
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:21ce62d80df2575c3aa7c8dfde706cf8af01b2fc86f72ea68f767bc1770fe733
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:be55d9ad6ae76bd061046af448d032e65926abaeef95c831c978e01e2b0355c6
        - registry.marketplace.redhat.com/cpopen/cpfs/spoke-nginx@sha256:183ee860bf48076ea155db08032eed7004c1204a363625f54218734b00175441
        - registry.marketplace.redhat.com/cpopen/cpd/ibm-service-functionality@sha256:ea4c32a19b1549698c8a68fa030d7036d3ea3cf2d000f3523b6e9fc7c6c708f1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:847751338f3b78cc06e29a4a26a18dc0f5d0bc24cbda23b765198a6a5c4503c1
        - registry.marketplace.redhat.com/cp/cpd/dap-base-job@sha256:f208686a3f56d2980e7b8640b6ebdea9057109e7520c6817a731e68d08506a36
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:8a6ce0dada706056efda73ad0cc1547566003e0a7aab8ed379b50e4776861b94
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:697baab21258166a1897ce4e056e39bc4667ff837bb97cadf92dad5863242ceb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:481391c273cffdeca36f91543732ba1f082dc409ee851ae95d3c1d205a7a6a08
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:36e3b8e88cbb895ca89288ed385fb800b888743844f50fb2233ebe5e6c7498c3
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:62102f79bec5a8f66c600d184478f2b42c8acfde059fc71bc262b16dd46975cb
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:4b6bc5640d9e68519a7f543ddbe3f16ca200f32a74ce9c07e996cf122da3b59f
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:186e849f3cf782f0d78660eb4433853e1e61296234d230a9dc5921fa4bd41060
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:7d8fb7d8e2551bf949edd33589393fd12f634871b486e26464cd089e3e4b429d
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:369f66883e040accb73c3deae984ed479e90eed242051b7035723f0ba21b51ac
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:0dc719e58694fb501b8b36f3a375b3c0462624d27fe49afe6b835821e31393f3
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:993ef97d93a7885626df345260140a527b490a7508e926b9a881405c260043a2
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:6c226a0ccde697b56c81427a970062872929eab73d99791c9dc670f4aa176ab8
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:411a4c7164b59ab3eed86e67aa457e45bd028067c937d35d0a227de63870cf87
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:740a1d87a43e61e32cfdea032d852af33100d959c8b57ff2f165f4baf77fc011
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:673320deb0ae96c6132545032f6b3e569d40204d3a5565a568dd97c5c1a4a552
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:173ef1b032594c4e070e2e013db469115ec58d970fa8b62f3b748163b46d6d3e
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:08dcf99453b4dc6dfb5620410b8f7e4eeb42fa824aedb2e480e2fb40ec8ce347
        - registry.marketplace.redhat.com/cp/cpd/couchdb3@sha256:c91fc0992cff4b674f5e3690398e7dfb70662c807b930b9b0e65ba507ff925d2
        - registry.marketplace.redhat.com/cp/cpd/ds-migration@sha256:b0735e396c09fbd08f328ce0686cd55fe9eafd5b759f36cf5677326b24cbb859
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:8cb37bb0f0461efc603cab89b5420c56990aa104fe39c9b77f0393bc6b641172
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:e296fa15dc14ee827d5046e8b1434d7f6838059c41d355471e1275e09f087411
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:c59837981b9a2a645c9ffd84c40d4f9ddc3797911f7a785ce4e3a2bfdd0bbe5c
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e924abf025141f3b3437ed89830957c5915152d2610213090e33fde988bd879e
        - registry.marketplace.redhat.com/cp/cpd/ds-px-compute@sha256:e0feb4d305c5ec9501b30ccdf8f5826cd8eb1ba5ba1d62d4d685101c785a708f
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:4e65e96603d5043e78a517a4c6e5debc5df7e3747fad4d8025ef1fdf1cd6a9da
        - registry.marketplace.redhat.com/cp/cpd/ds-assets@sha256:68ae72930a346d0fa0871cff573c64285266311208f2bf64101697a540f60272
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:edfb732843a53254ebf7d4a2f109af109623be8fb9bbae94366d30febd2d0494
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-watchdog@sha256:96451d4a9d1c4b759d82dff2eaa9aba2a6fd58518f6ac6306cc082ebb7766a64
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:978cf4bf299cd3e8cde9d7ee7aa75b73fd9d9e6755567651d6c1c49036b8978b
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:17f016c11039c34f90f67279b45ee488740cc3d033112a3bd589bb1a60ce6812
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-catalog@sha256:ec5e4a002cfc63c1ff1bec340111e8a1755c8624066445db36c23fdd1c3061f0
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:dcd74c128502dc301482eb302b34c21dc02173ae1cf522356d5b500ba42b4e25
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-bundle@sha256:00cc7d072ea6b6f5e9f3e5b95b9cb4ab8ad7bd942d8c259e70f8ba0e9c09a464
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:75595bf2753fb43391e2a496fb6d59faa8785babebb4aeb1da34cd5e72c93699
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:21c81ff7c4b8809accaf62b242fd9b3d52f235489ae427556b7f71cc21bf8926
        - registry.marketplace.redhat.com/cp/cpd/qs-ruleset@sha256:650e68c41252300836740ad495ad6776a7da94331e8070520e60a1c5e2c5f3ca
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:0cdba17b95a959b6d04889179d105515ed3176ea3449bc104f18b86374876048
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:85669f2a35ecc4ebf6b46fd793ba6a6bdf996164c00662e65cd7ceece8d60cda
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:1a1bc54977b160b70cc00cc4eb1380250dcd639fa3d741363212b8081b35f72b
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:bb85f2567aaf7b612343fb9819fe05e2e771ccca8383316d34282ec37fcd1383
        - registry.marketplace.redhat.com/cp/cpd/haproxy@sha256:54c86367bfe0ab658611856cc0a6c5de24c14a336bd129ac9f16293945a55238
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:75ca37701dea1bd079df8f8cf302cf95e8dd4ed817fe126b192ca368e92fb78c
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:c35eaf4921bfd046985ea7a03fc325f79766bf73e3ddba69408f4467fbd7994f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:887348b5035f5a70f1c31aede992a5c76397f5eddd2aebeb27074c3ea6248c00
        - registry.marketplace.redhat.com/cp/cpd/portal-notifications@sha256:bcdf9b7eea8753b3e463973a2690140ae7aa8c02c3c7726511f116bd751ec2d8
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:b372540b630507f4cb74f5eab38d0335d7b5bd64197e3f83852b2c7c5ba9067a
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c730fd4a7baa3f46c03f6c73e2c6c7c0ff24a8900b79dfc0f2fd319909edce74
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:4ccd786884e39641a74096d783742c219fb277caafe80b667a80396c3a9cf6c3
        - registry.marketplace.redhat.com/cp/cpd/runtime-assemblies-operator@sha256:ef24cc79265473a8b6fea231b1930c5763293dca79eedd46c89ec63dd95812c5
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:6a996123e69c7f2c3254f334e278ee0daf75fc53980183e1a72358323e41349c
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:d28e6a3c65b580051ffbe287492dfad28dbb1c669cb76aa3df23bd9428f4bbe9
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:095aae63ced410e033d8f7f3d27bd14dc5b93a10068ef5a389f181115c5d07c2
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:1f68dadaf2ab779dece300037f3abda1b0bb33603436b84336f8fb7d53d33644
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-catalog@sha256:9669ed5ba26ef32ad36a6784b2658553c00ac2a89cf4e9caddadf7bf46c1592e
        - registry.marketplace.redhat.com/cp/cpd/ds-nginx-proxy@sha256:ef38f12712c406f8206742dffae9378e344d66c774e5b37b4fcb0b657db456ca
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:a6e98064effd12ffda0613db98e795e056b83ef77b5deee932a762b178ceec3f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:94662b9dba5d4bafeec53b8ad334e7a1235385b21ac0952e12975f4f0a7623a3
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:31edab3e5662489dc2fe9528d078966663c8d49836425e648b7e112f871808dc
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:fa01c7d11c23a47bb84375389e69619455c2879b5ed6bb5b9b753b945dcb02c3
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:1cf983f9b9cda18257066fc82e199bd3e22e2fcba13daf0f60c92b50a4a0e078
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d51033417b786dd2dfbc94e648db00ba68f316c23b06a0bc9e36a5c22f579124
        - registry.marketplace.redhat.com/cpopen/cpfs/icp4data-nginx-repo@sha256:9c160303552a9b8a97b951311018de2ada410883468680dd61fb41519b9a9e7a
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:deb66a64288a370c5da5c7cc0ba4e2ba4ab816d7c5a080a0c505b05f55fa6072
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:c957787ae587a31e6a497b46649da9cd46102a2a3dff29243fd5cbac96e7e2d5
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator-bundle@sha256:cd92fd1c85a72560241fffcb3bb62cf965f96bf2a4e1f5cca84c0b52624ae962
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-catalog@sha256:7dbff355db7739152961cb6a97887d2e43bc960ac58837c126e03ed1a4480a3a
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:8b7941a3f518e952766a2e6f88acfa7c9914b6d2bcd4ad9ddd54df313fa30347
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:0fb78efb9e92dd0c2e52ca2488e6f1e7431e197f87e6994e15f7d621b90319f2
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:932825fc7af8b96c381d504cd0976c751661ce078257430e5c79d565f416bacb
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:7ca8d085dd9aeb568eec2c4da974c03aa72bbfe4648f5d66a62d0444a2d1c839
        - registry.marketplace.redhat.com/cp/cpd/task-credentials@sha256:e162828d5ad07ceae4db0589f930979842714e3b766552102bee373a3a570af3
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:f7e68137e1ca58a63924967a8e4c872912557fa68af8a7392e3e0d4aaaef1e78
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0711e1c6001586c950e74fbdeef8f04d20ce7b93de2ad033e82d2d5421338936
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:65108559a5edc9a7c4f82962ab2293017e0873872730d60d8487963970761941
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:716b5823661ac8f1eacda11ffbe5f0ff08a0e9d2296d04acffb646b7526128a9
        - registry.marketplace.redhat.com/cpopen/cpd/cpd-toolbox@sha256:1b3469f6fcd037aacabde3e05605b908e22f012d331d7365d0450f6d933d3efd
        - registry.marketplace.redhat.com/cp/cpd/ds-canvas@sha256:3d83ebdfddcbe6eca38705fc3d575f2e70580cf9f7e8735941afee0a33d3b5c5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:dd76a8484628970b62238334e122768aeea820ab18e24d62d23a6328c83b868c
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-bundle@sha256:d61b7b8e6a9785b59630d9c5776f72a29ffbf2e81eb753d7b11dcf2d9d77f1f0
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:bcadf3d367f76d3454e119a9d39f6fdeeb6cb806c282f21a1bc3c38bed41b427
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator@sha256:24ca69c2f80a4369e29aa382a94d64fd3546b110787df1142c45a20b8246e844
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:f2f256a2fb787419b25e99c765003d25d1edbbe60042151d91b46a5d029fc0e7
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:1d6dc7a3039e6fa044053d4e89ef54ba2e83ceadd57bcc4613a6d447d027ae2f
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cpopen/ibm-events-operator@sha256:3743af183b7b801f72b279a658b0d32da0bdbfa526a00d191a9a44de6df2284f
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:5436a25b6c5b419b77643a76eb073bbbf8f09517299dbf4a992ae14c1965972c
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-perf@sha256:ab99b62d40ba05cef8e699ebc7cc65271ea6fcb5900a76fe17526e7d55b50b08
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator-bundle@sha256:a7c050ae29cb93754f59d8f83676275ee9a91ae9a9956395b97abf8c5e20e96e
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:b8c83fe5215b7412765181aaba84162b6cb188b5765e45972a6f29b2a9f19826
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:562281dbb2bec5191a7c6b595ab853092e6fe3ff1432e77b2471f71ccdffcbc8
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-catalog@sha256:dc257df716b270858ca5d07909bda9ab635a388c9887d30b015391f60cc7b5c1
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:ca175109562cd9140d98bea3be88d6b34490ccb8d79f0cbbde08c821db9c5fe9
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:17bac3307ad3843afea77ed213279603b5f775f24dcf4857095baf5029409c18
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:0ceaaab514f65f9a1c40717e97ca652506f8b3595186222d470a99d653afcead
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:b797e00eb32ff9a6a56e71d355bea653ecdecfecc121ddb197a6e347b89dbf8c
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core-api@sha256:e6ce3c998642b235080daa6c0b954733bd3fe973fc8d85a10ef0cdae2a73aeb0
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:b2ce782efc5349a49d9c7326e2f72b2c1d99d30eccfb6975f7cba10008f9b2d3
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:70ba7f283bde348bce68487ccfb833a418612f063c73bc8ef785836129739928
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1402@sha256:7ac32270044a9ec08ddaa8dee62b61618d9363c8773fe44ba1dd86e1b53acbd5
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cp/cpd/dataview-api-service@sha256:d0e0144eaf27e03aec1499e53cae5366e11a3da284536cd3b27ba5691b4f70d0
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:ceee171c5ce0a8e580dfeccfe0dbfb73474c109becd45a562411678f3ab08595
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:0e563d51d601adbd710833cbde26ab8f857092f85686213adfd6fb3fe5e2828f
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:14da175e95c91864142ae00e9087ac2fe0e2083fc1a3bcb20a6b29dbab99d8a3
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:a6d5bd253959be1f91b50f5f2c56ed77e1c82b06483e62094aa86ff57cd46eec
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator-catalog@sha256:3c1e08c56430e9bfc0c831be2394fa28e395a3341ce5cb5201b08828298deb60
        - registry.marketplace.redhat.com/cpopen/ibm-zen-operator-bundle@sha256:552fc53664722903eccaa7d3782a3e84413509fd357725fe0b0f4a077c48ab2c
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:a28b71ef64c1967cdd36a16e01b611fff6a387b6a741dbcaf7556ed6be9b1f69
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:4deb399f28d3385c50a256c6e927cf27115c311be5943021733e4b281c12a419
        - registry.marketplace.redhat.com/cpopen/cpfs/privatecloud-usermgmt@sha256:abd96661823553d623eb0ea1fac8271d64391243dd0f13eb4a4fcc858dfe10a5
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:27210840cbff8a7c61acac2ccc328a415a405b4738422b4f47eb9a0110127d22
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr@sha256:43cf735b51e2ef45a911ac7032f03c74cf0125e2d2801258d4aa04831e1c6ce0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-ccs-operator@sha256:ce57bc323b95db548dcd2f90ec938e708988cf6103579e793ee39d36359d60cf
        - registry.marketplace.redhat.com/cpopen/cpd/cpdtool@sha256:714caed1042ee03461efc9256a3da1046bd8f43819caf3a91ec38b118d459869
        - registry.marketplace.redhat.com/cp/cpd/dsx-requisite@sha256:9ad526bd56829ef24834b8a790b29c50809866c1dfb60cf8e0600fb9a903d388
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:88d3c1ee02acbffedbf6ddcc27b80be6727f85425c361d45cc5e5c4056cf6733
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:30a5b2c4e3c231373fa18d8aaab16d3505cc11a8a5d4be36eaf99d7c0dee35f2
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:e76b61f9f0efe16facbddd80808d80bc32d0d41b4c16aed566116f1366152987
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:443b51f8b10acc85bdefde7193e0f45b1c423fa7fbdcaa28342805815c43db3d
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:b63e2c10478735b51b6828bb11709569a2fce62a19b70157a8a5b1964820164d
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.6.1@sha256:545b41cc6c044d6807bfeb7453a045e074a5670cfff9f5c311f4918946901fcb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:9ced60f18740063ab281b836d4caf21035242ac3f659c6bb022ddb546770bff9
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-rsi-adm-controller@sha256:e857323acd10f7b99c5b7fb4b0a5b89a808f68036a1d9a224eff55a7ae7744b6
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-openresty-base@sha256:e8e0ea5cb1a070443011794988f2f6c9d67758605dc2847bf0183e19bfde52eb
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-events-kafka-3.8.0@sha256:381170c9c497b1bda33eef3ddeb1bafefa07be1e6ca75e75c9eaef9dc87b8d4a
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:489d7f80d83c9890600e3c9c4cb9442a6f42b9c6c4f13dafdf0f614b72e67143
        - registry.marketplace.redhat.com/cpopen/cpfs/icpd-requisite@sha256:8fe85b93b9dcab9c33306ae358e6910685e109f32662d02e2b6bcd8dae243968
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:fb37fe521693bdb8b96400fd491a39c573545f35b4131fdfab501b7ba29713f6
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-hook-utils@sha256:ec44be69b4c585e9a3d507e29bd287f5b464360d82c8e0969c5fb4ed3e95abc0
        - registry.marketplace.redhat.com/cp/cpd/wdp-dataview@sha256:f42395ea365edc2c1bbef914a55b091e6442d17bd87bf3730a8f9f6d401bf4f5
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-minio@sha256:ea57c7cb1ccd5bce17f6cb35e326ab974718be8cd19b0cd38ea8112e91c51939
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator@sha256:7d9b2d61a4eaac645490d1c9e6db97ba193d65015b6efa8d5a1f3ec6cccb1ad1
        - registry.marketplace.redhat.com/cp/cpd/jobs-api@sha256:173b88f90078e1817686ac949f48387af70a706e23e6fdab4a7c6217db5ad86e
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:c268d2852580286e5c6770857fcfc3b06a698521f5acd17ed83b8e2cb9ab55cf
        - registry.marketplace.redhat.com/cp/cpd/portal-main@sha256:4fb2127ba7c525a959bbbcf024de364ef266ebb3ca7e696f74d990d2552cda59
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:bdbbd76c93f0fb8f838edd9f124b02cc646a6f735a68a430a00f025a06c5f407
        - registry.marketplace.redhat.com/cp/cpd/ds-metrics@sha256:d904761e70d3969a4ccacd71a60ad4d77dda939c556615df8d65782ccc3273aa
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:166a3918b942a4ea7b6c371a1e30fd4c928586d7ffd5b6919c2fc728a086ccbb
        - registry.marketplace.redhat.com/cp/opencontent-search-haproxy-fips1403@sha256:7a90c8e0180da12172b23f3b54536bd1afe35affe37d908ddd78996a81435751
        - registry.marketplace.redhat.com/cpopen/ds-operator@sha256:119be0bf4c60230c43bd64c0ea252108a0173adfffc9f05e5257ec260b335599
        - registry.marketplace.redhat.com/cpopen/opencontent-elasticsearch-operator-bundle@sha256:f9c9202e106d5a159c8f6b8fd883da4d6ac0d81e75b558fa52e0f7776269404c
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:a16b822b41f5a745b475ceb353a459e998f9f43bb9bb867ca1640866ba51780e
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:9021a04805bb7f21d68fc6211f844a21c1d06f700c332ec3238073065b7c224b
        - registry.marketplace.redhat.com/cp/cpd/event-logger-api@sha256:d72cef9b19ea48f03f7658dd55cabe6e118be0209852025845d426b1fbf0e0ca
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:3c18707afb90a7dc1bd2ed1fff05c53cce89b7c85062e72261f538a12834ab37
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:a95e42a4f9138f1f97937a74982c98d466c171d0e3e79c522ededba6c25b39a3
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:973eda4612c905661ff5ebc28c7cac6d9fe16fe5edaa544548d2311496ed938e
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-cloud-native-postgresql-operator@sha256:a72c2b70a9f0a93047d47247cddc1616138cbc87337289ddb208565af6e4369d
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:f3f8c6c395ae860fed5761f67ce092a122e7073edbc205c9167c27a4ba85094f
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:36f4c57c6594f997383c4731e1e62c4765b8642c0b5c7373ba22118220928d09
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:e1b10435f94bbb7165703720f58aa8c68acbdaa18006a2c36d1e24ffdd72cfc3
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:8bd9fe1d35785fa0125b8fb3a60333219a50ee8be7a7ec07386e8e9d65316061
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-acmesolver@sha256:7b2e9c373f9bfd91e16eff68ef2313171c69b94fe5a21d3d3158470c74f54a05
        - registry.marketplace.redhat.com/cp/cpd/edb-postgres-license-provider@sha256:cff7466aa6063a2bca677815df01208822930f8c12784c1c406fa352d6bd6cb4
        - registry.marketplace.redhat.com/cpopen/ibm-licensing-operator-bundle@sha256:d2680c5f05c886b44df644aef6cb2fdc9534bb6a238066f6ebbb97bbeaac431e
        - registry.marketplace.redhat.com/cp/cpd/ngp-projects-api@sha256:df420e9f616ad8a123d9eedd117fb6cb007b30a772138db534f47ad8801fa96d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:ac128758328325ae56273800ea010f9559902281cc60226c8f99de51b636b180
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:9a0c76fd9ef9b2dd9af5b0f29e1813567ab7f7d35c1efdf71da1a315f2942b45
        - registry.marketplace.redhat.com/cp/cpd/dc-main@sha256:06403c4ad99655cfc10fde46b21e84f90c05ae143bb6d08464dbc8954486897a
        - registry.marketplace.redhat.com/cpopen/ds-operator-catalog@sha256:d4738f64e36a8f0c88d40043eebaae1d561193895d438fb6d8bf1a042d0a1007
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:eb157bfe50005f9ffd3560ce1ef7ee0fc283ddfe265c42b35a71f18409e3f561
        - registry.marketplace.redhat.com/cpopen/cpd/k8s-storage-test@sha256:8f9d3438f58604cf29d643691b125a4138275b2c09cca4afc8e4b29289febd77
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:19ee4cb8e3e60d9c1e8fa63b0bf3e218bd683b6c05b9e9553d9477c68c493a9d
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-utils@sha256:587af68ec8b4974b29da46ede525c70054703fce8d5934f0cbe68739c6058fe4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-webhook@sha256:67a02a9264d9c6e82edc15ae36d4f99b3fa9099697c550088d27c76511189f37
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-jobs@sha256:993f16d744a81c95e73339d3c8d5880b5cdbcf29beb914fcb2c6403af6d628a1
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:5f0974565de8b045d3faf6acac7c8ebce6cd35327dd76290b39fe48b27ad5b62
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:8444703b2a7822f82141afcf4f36bf255650d755321b9e9fe2cbd1bdea41d5c7
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cp/cpd/spaces@sha256:a1754e003811afb578dace78f9eaef91e7917e5468e590fcb41920fa71bcc22d
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:ab686e9e4c38b7909994f9fb496943b745e1a07492fe61a61b9e8a78f756eb1c
        - registry.marketplace.redhat.com/cp/cpd/portal-projects@sha256:257bf301338f7b123099e2ffe31c3f3b793bf28d6a0570a25b889271239579ba
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:ded16bf3cc993bfa5af8907a1517cc34ecb511948e4d9ac73bab68888af5543b
        - registry.marketplace.redhat.com/cp/cpd/wkc-init-container-ccs@sha256:cb9a33db3707c0c6045e6c823ecb21b0f229c17ff868da7d36602d0026eff351
        - registry.marketplace.redhat.com/cp/cpd/ds-dataservice@sha256:c2b9d370b6c3e3c8e30079465502ce376055261cbf0f95c48d8a2ae10ca36c76
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:e9e408a13bd103fb46536f896c51e7a27cba38d9b224231b20328a28efe1e616
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:62dda55cc870279d491679f6b43be177986991b6eb85d9dfeba3ed2e72b3f10e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:e673f1d0c5a0e32e9abc7ac40eff905ce411f9f3d8ca0b03160e23ba5661e041
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-controller@sha256:4ca30e2017a0ff2d64100a03a84bee8d601c765f0e624e1f485cbb97748bb8a2
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        - registry.marketplace.redhat.com/cpopen/edb/postgresql@sha256:99c863002ae1e93e97792fa606128d05063abbedad9e9404acf28c745ae1a5c4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:d733c5a10a21cac2df2fff6378035151e5d0df57e3ed1f96bb3ac17c0c95b8eb
        - registry.marketplace.redhat.com/cp/cpd/wkc-search@sha256:d023f35b65816f244b6d699b41362e6031553f7d4c416efcee4a7f2395a61355
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-core@sha256:686700e561181ebe1431f71eb9aa11b550bc4361dab8a3668ce0c7ead217f79a
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90d45b7a3c5ffcd30a1ca1447693ba4aedfb400983cf49779a8421a42dc0eecd
        - registry.marketplace.redhat.com/cp/cpd/ds-flows@sha256:a31e0fb7a05e969e31c10a74cc858c483d5d4da14ce3abc36f5318100e60e792
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/rabbitmq@sha256:0621eb9b83c9b2b8bfe5cd09797fe08fa37a854a5506914935ff6f0df599b631
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/asset-files-api@sha256:6cffef7631ee54612fba71efc0eff4e17bc08defc3e2137aee9881d685444298
        - registry.marketplace.redhat.com/cp/cpd/wml-main@sha256:a3a0d58f895fca4d832e602d0dee80018df4ad7bad8800ad4058eea07fda4817
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/cpd/olm-utils-v3@sha256:866ebafb730e660062cf7401023d82c5242a2b80cca0126d47a68193c84670e0
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/catalog_master@sha256:9782a9de863ad10ed36341d07ea5087e40f6f213cd4453c3ac2ec100d8733ce2
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:4d24fb274b5429ba61fc4bc6833e93b14b92b8d525556db64e2172b45433de31
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:abec4867d6b9294d674cc02b9f7f254e65e7de8eea0231fa94f50513f2b7c45d
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-platform-operator-catalog@sha256:077107dd9cede2c49beec429decb66760372a229c476ca1089e42e4ed5ef40d9
        - registry.marketplace.redhat.com/cpopen/ibm-commonui-operator@sha256:3be3ee728cbae6bb8cf29d47c37ed1a4cd5791b54e1262135089424e5bda7ea1
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-audit@sha256:4aa81bd0043d0bac60c9f1b0ab56e45a3035c88314614e60eaea48b2a29ffa16
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:a64a93104dee64e0e7e3e381144340511f903155f72b7f619c82c1d440712227
        - registry.marketplace.redhat.com/cp/cpd/catalog-api-aux_master@sha256:7a0e784b0722ab0a4c62754a4e817a1f25e5a2d5e960de76387971fa1ff7da41
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator-catalog@sha256:7861185d55a5fd18d896dbc57501ae5cbc1de0bc02f1ae3fe558a36f0f80b4fc
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:80c0913e2c84530d9b7c6a2152de76ded5754c134f6768754d00944f46d0ed4f
        - registry.marketplace.redhat.com/cp/cpd/wdp-connect-flight-svc@sha256:fbc6eab07fcfc2d0c1895f18e17f751e9481ad111875959e9ff08e14f81b9a35
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing@sha256:17f7039e49f28abb7239e0c4f3689de58c75f34c9bde9e7b33a6ffb93ba92677
        - registry.marketplace.redhat.com/cpopen/cpfs/zen-iam-config@sha256:f1aa6add76f59afeccdff84c9ea89a46b49c6696f62acd75ef8ff2f31a3080be
        - registry.marketplace.redhat.com/cp/cpd/redis@sha256:0e0ae96aff8540323da8a2379af014d3af1770e4b0043b84b9b415dda44501f7
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:73c40c6c72b7f2399fb0ed7e462baa72c3e8b5aa246fd2fe1082904b4afc0796
        - registry.marketplace.redhat.com/cpopen/ibm-cert-manager-operator@sha256:758f9e49dccaec5ebd125fc0c383fb5a656ba9f2f301984f993d50b103208d12
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:c43770715b76e0aa3fd9b3c2b21998ee369f2142e3c5d04320c677c9e9b29add
        - registry.marketplace.redhat.com/cp/opencontent-opensearch-2.19.1@sha256:9b91aaeacb196806646a1f196d4adc83b5159de767fad1a2eab699f735938911
        - registry.marketplace.redhat.com/cp/cpd/ds-px-runtime@sha256:2d320dee224cc4818a84493723f2262d259891bef89dedef09e1f1e6ae5686cb
        - registry.marketplace.redhat.com/cpopen/cpd/cpdbr-oadp@sha256:35594a16332ba3ac00226414be9b85243bb25dd0648e256d7ef9d42137dcf914
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-cert-manager-cainjector@sha256:e6fc9d7a70877100b42e290cd091011f2e1c674a27985fce180ee245373391dc
        version: 6.1.3
      entries:
      - name: cpd-platform-operator.v6.1.3
        version: 6.1.3
      name: v6.1
    defaultChannel: v6.1
    packageName: cpd-platform-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-sds-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-sds-operator.v8.0.230
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.230"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleStoragePool",
                "metadata": {
                  "name": "infoscalestoragepool-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
          createdAt: "2023-10-26T20:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          olm.skipRange: <8.0.230
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.230"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:12d6ae3f139daf33dbf104499c195d9915470b957f401b5f622cfde92350a6c1
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:f4947eab2b31fccc86b5d5d0d3fd4eb1f7513fd8325e379cc259f5ccd1dfc3e4
        version: 8.0.230
      entries:
      - name: infoscale-sds-operator.v8.0.230
        version: 8.0.230
      - name: infoscale-sds-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-sds-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.330"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
          createdAt: "2024-06-24T09:03:32Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.330"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:965ec7875eccc6a5fc56c3b8f0da668d826d34580117ba1438a607a9d0d13b73
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:dab501b99373c840c7e8e3ab479fa273b39f9bf8a896f3f1e4e945f4ab81f64e
        version: 8.0.330
      entries:
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-sds-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.410"
                }
              },
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleClusterSet",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infoscale-sds-operator",
                    "app.kubernetes.io/instance": "infoscaleclusterset-dev",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infoscaleclusterset",
                    "app.kubernetes.io/part-of": "infoscale-sds-operator"
                  },
                  "name": "infoscaleclusterset-dev"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
          createdAt: "2024-07-30T09:00:57Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/internal-objects: '["infoscaleclustersets.infoscale.veritas.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleClusterSet is the Schema for the infoscaleclusterset
              API
            displayName: InfoScale Cluster Set
            kind: InfoScaleClusterSet
            name: infoscaleclustersets.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Arctera InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/168721626-168743109-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.410"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Arctera recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Arctera
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.arctera.io/infoscale-kubernetes
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-storage-operator-bundle@sha256:a51a626126460cbf8e614fd3287bce3419fc5b728788c5cb9b1ab1fa1cf98259
        - registry.connect.redhat.com/arctera/infoscale-storage-operator@sha256:43397ed6430691b35c71bc4a4ff4eddcbd685e84cf0a3e3604e19feb4f363d9f
        version: 8.0.410
      entries:
      - name: infoscale-sds-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-sds-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-sds-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-sds-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-sds-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-sds-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-sds-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-sds-operator.v8.0.210
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infoscale.veritas.com/v1",
                "kind": "InfoScaleCluster",
                "metadata": {
                  "name": "infoscalecluster-dev"
                },
                "spec": {
                  "clusterInfo": [
                    {
                      "nodeName": null
                    },
                    {
                      "nodeName": null
                    }
                  ],
                  "version": "8.0.210"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
          createdAt: "2022-11-26T16:36:59Z"
          description: InfoScale™ SDS Operator manages the lifecycle of the InfoScale™
            cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.22.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InfoScaleCluster is the Schema for the infoscaleclusters
              API
            displayName: InfoScale Cluster
            kind: InfoScaleCluster
            name: infoscaleclusters.infoscale.veritas.com
            version: v1
          - description: InfoScaleStoragePool is the Schema for the infoscalestoragepools
              API
            displayName: InfoScale Storage Pool
            kind: InfoScaleStoragePool
            name: infoscalestoragepools.infoscale.veritas.com
            version: v1
        description: |
          ## InfoScale™ SDS Operator

          InfoScale™ SDS Operator manages the lifecycle of the InfoScale™ cluster

          ## Overview

          - Veritas InfoScale™ delivers Infrastructure resiliency and persistent storage for your critical containerized applications for OpenShift® and Kubernetes Native deployments
          - Engineered to support stateful workloads generated for mission-critical containerized applications.

          ---

          ## Data Services & Benefits

          **1. Software-Defined Persistent Storage Volumes:** Enables customers to support multiple container application requirements leveraging existing SAN or DAS storage

          **2. CSI API Driver:** Facilitates static and dynamic provisioning for applications with RWX, RWO and ROX access modes

          **3. Life Cycle Management:** Enables automated deployment, configuration and upgrades of InfoScale Software-defined container images. Certified and Integrated with Red Hat OpenShift for a single-click deployment

          **4. Availability:** Provides scaling, mounting and/or movement of InfoScale persistent storage volumes on cluster nodes with minimal disruption

          **5. Data Integrity:** Prevents data corruption by allowing only the active cluster nodes to write to the volume. The I/O fencing feature recovers from cluster disruptions quickly by ensuring that application pods are moved to another node to continue normal operations

          **6. Point-in-Time Data Copies:** Create snapshots of Persistent Volumes for backup products, data analytics or forensic discovery and analysis

          **7. Disaster Recovery (DR) Tech Preview:** This DR feature provides the ability to test and validate disaster recovery capabilities by migrating Kubernetes cluster metadata and application components to peer cluster in case of a local or remote disaster

          ---

          ## Pre-requisites

          - [Please refer to pre-requisite section from official documentation](https://www.veritas.com/support/en_US/doc/151215298-151215302-0)


          ## InfoScale Cluster custom resource

          ```
          kind: InfoScaleCluster
          metadata:
            name:  < Infoscale Cluster Name >

          spec:
            version: "8.0.210"

            clusterInfo:
            - nodeName: <Name of the first node>
            ip:
            - <Optional - First IP address of the first node >
            - <Optional - Second IP address of the first node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            - nodeName: <Name of the second node>
            ip:
            - <Optional - First IP address of the second node >
            - <Optional - Second IP address of the second node>
            excludeDevice:
            - <Optional - Device path of the disk on the node that you want to exclude from Infoscale disk group.>

            # You can add up to 16 nodes.

            customImageRegistry: < Optional – Registry for Infoscale Container images>
          ```

          #### Note

          You can specify up to 16 worker nodes in CR. Although cluster configuration is allowed even with one Network Interface Card,
          Veritas recommends a minimum of two physical links for performance and High Availability (HA). Number of links for each network link must be same on all
          nodes. Optionally, you can enter node level IP addresses. If IP addresses are not provided, IP addresses of OpenShift cluster nodes are used.
        displayName: InfoScale™ SDS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Infoscale
        - Storage foundation
        - Veritas
        - Data storage
        - Storage
        - Data
        - SDS
        - Software Defined Storage
        - Clustered storage
        - CFS
        links:
        - name: Infoscale Operator
          url: https://www.veritas.com/solution/virtualization/containers
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator-bundle@sha256:f8a8cfcebba4b17fa6cfa99b3ef430af272c5237476dba4665803a90adfd5d2a
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:1667227b5111fe52ed304ab19ff9fadfc677239e56b2c93a3b89b1bfbc31354d
        - registry.connect.redhat.com/veritas-technologies/infoscale-sds-operator@sha256:26573b98743245c0b607abbe0d981abce6ce9fb1c5c0cd9fe735351b7136f4e8
        version: 8.0.210
      entries:
      - name: infoscale-sds-operator.v8.0.210
        version: 8.0.210
      - name: infoscale-sds-operator.v8.0.201
        version: 8.0.201
      - name: infoscale-sds-operator.v8.0.200
        version: 8.0.200
      - name: infoscale-sds-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-sds-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-sds-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Randoli
      provider-url: https://www.randoli.io
    name: app-insights-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: app-insights-operator.v1.17.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "appinsights.randoli/v1beta1",
                "kind": "RandoliAgent",
                "metadata": {
                  "name": "randoli",
                  "namespace": "randoli-agents"
                },
                "spec": {
                  "costManagement": {
                    "opencost": {
                      "exporter": {
                        "image": {
                          "tag": "1.112.1"
                        }
                      }
                    },
                    "vpaOperator": {
                      "enabled": false
                    }
                  },
                  "logs": {
                    "vector": {
                      "agent": {
                        "containerSecurityContext": {
                          "privileged": true
                        }
                      }
                    }
                  },
                  "network": {
                    "standaloneConsole": {
                      "enable": true
                    }
                  },
                  "observability": {
                    "netobserv": {
                      "enabled": false
                    },
                    "otel": {
                      "operator": {
                        "enabled": false
                      }
                    }
                  },
                  "resources": {
                    "limits": {
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "512Mi"
                    }
                  },
                  "securityContext": {
                    "fsGroup": 1001
                  },
                  "tags": {
                    "costManagement": false,
                    "observability": false
                  }
                }
              }
            ]
          categories: Developer Tools
          containerImage: docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
          createdAt: "2025-09-23T19:45:43Z"
          description: Randoli Agent connects OpenShift clusters to Randoli Observability
            & Cost Management.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Randoli
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: RandoliAgent
            name: randoliagents.appinsights.randoli
            version: v1beta1
        description: |-
          Randoli Agent connects OpenShift clusters to Randoli Observability & Cost Management platforms.

          This operator provides a container-based agent to be installed in your OpenShift clusters for connecting to Randoli Observability & Cost Management platforms

          **Prerequisites**

          If you are using the Randoli Observability platform, you will need to install the following Operators:
          - Red Hat build of OpenTelemetry Operator
          - Network Observability

          If you are using the Randoli Cost Management platform, you will need to install the following Operator:
          - VerticalPodAutoscaler

          For additional information please use our Openshift installation guide here: [https://docs.insights.randoli.io/agent/openshift](https://docs.insights.randoli.io/agent/openshift)

           **Randoli Observability for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-observability)

          **Randoli Cost Management for Kubernetes**

          [Product Page](https://www.randoli.io/product/kubernetes-cost-management)
        displayName: Randoli Agent
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - developer tools
        - insights
        - monitoring
        links:
        - name: Randoli Observability for Kubernetes
          url: https://www.randoli.io/product/kubernetes-observability
        - name: Randoli Cost Management for Kubernetes
          url: https://www.randoli.io/product/kubernetes-cost-management
        maintainers:
        - email: contact@randoli.ca
          name: Randoli Inc.
        maturity: beta
        minKubeVersion: 1.24.0
        provider:
          name: Randoli
          url: https://www.randoli.io
        relatedImages:
        - docker.io/randoli/app-insights-operator:1.17.5-4@sha256:22e7150c29557f2a71186389e5d7d60bfe82a4ea022e1403033c6ad61bc95f7a
        - registry.connect.redhat.com/randoli/app-insights-operator@sha256:21ed3d9b8b8466665c1cb784d0bc244279d0730f6319625caf46af111c25d333
        version: 1.17.5-4
      entries:
      - name: app-insights-operator.v1.17.5-4
        version: 1.17.5-4
      - name: app-insights-operator.v1.17.5-3
        version: 1.17.5-3
      - name: app-insights-operator.v1.17.5-2
        version: 1.17.5-2
      - name: app-insights-operator.v1.17.5-1
        version: 1.17.5-1
      - name: app-insights-operator.v1.17.3-1
        version: 1.17.3-1
      - name: app-insights-operator.v1.17.2-1
        version: 1.17.2-1
      - name: app-insights-operator.v1.17.1-1
        version: 1.17.1-1
      - name: app-insights-operator.v1.17.0-2
        version: 1.17.0-2
      - name: app-insights-operator.v1.17.0-1
        version: 1.17.0-1
      name: stable
    defaultChannel: stable
    packageName: app-insights-operator
    provider:
      name: Randoli
      url: https://www.randoli.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: alpha
    - currentCSV: silicom-sts-operator.v0.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 3,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-BC-8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "nodeSelector": {
                    "sts.silicom.com/mode": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "hp-ocp",
                  "namespase": "openshift-operators"
                },
                "spec": {
                  "dummy": 0
                },
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicom-sts/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicom-sts/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/rmr/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - registry.connect.redhat.com/silicomltd/silicom-sts-operator@sha256:8e96eee44c38ad13d232cffb229a3f33a58acd48a730e3d2fff30befa192cb4e
        - quay.io/silicom/sts-operator@sha256:a79a5cf99f5238ee08c509efe0b6142197ef0b79a6b209c125d9f7b7235758e5
        - quay.io/silicom/ice-driver-src@sha256:569c8333e21c7c022ea71fa5b418404899040d5b3a3a4128eeac21f62aacaf3e
        - quay.io/silicom/phc2sys@sha256:b067167a1a5d93e9fe715c7774f6cdbbcd93cac27780f022e44dcb8a4d4d7bb9
        - quay.io/silicom/sts-plugin@sha256:233bc95f8617a06bd452548ec116d91551dd4933970860d8d5ccb81f75ad2220
        - quay.io/silicom/tsyncd@sha256:3f1cbd4a8b5d62d7d3e525af5f26d9f27b8c08e886b0639a1e368163e166b537
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - quay.io/silicom/sts-operator@sha256:7d474d0aa91ac57d8f5817b033b67b511c8601e0dbd83d808a69004bd24b4e60
        - quay.io/silicom/gpsd@sha256:e479c542e061b5422865f2617ed0563794f265c6e6660e5a9d204c9e1cb2f346
        - quay.io/silicom/grpc-tsyncd@sha256:631bb754742e1a24376913baed60e480a6e6ae988ad9a3892254ff96996c15d1
        version: 0.1.3
      entries:
      - name: silicom-sts-operator.v0.1.3
        version: 0.1.3
      - name: silicom-sts-operator.v0.0.9
        version: 0.0.9
      - name: silicom-sts-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/multiarch-tuning-operator
    name: multiarch-tuning-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multiarch-tuning-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multiarch.openshift.io/v1beta1",
                "kind": "ClusterPodPlacementConfig",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "logVerbosity": "Normal",
                  "namespaceSelector": {
                    "matchExpressions": [
                      {
                        "key": "multiarch.openshift.io/exclude-pod-placement",
                        "operator": "DoesNotExist"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Other
          console.openshift.io/disable-operand-delete: "false"
          containerImage: registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
          createdAt: "2025-05-09T05:49:47Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operator.openshift.io/uninstall-message: You must remove all Operands before
            uninstalling the operator.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-multiarch-tuning-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/multiarch-tuning-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1alpha1
          - description: ClusterPodPlacementConfig defines the configuration for the
              architecture aware pod placement operand. Users can only deploy a single
              object named "cluster". Creating the object enables the operand.
            displayName: Cluster Pod Placement Config
            kind: ClusterPodPlacementConfig
            name: clusterpodplacementconfigs.multiarch.openshift.io
            version: v1beta1
        description: |
          The Multiarch Tuning Operator optimizes workload management within multi-architecture clusters and in
          single-architecture clusters transitioning to multi-architecture environments.

          By default, the scheduler does not consider the architecture of a pod's container images
          when determining the placement of new pods onto nodes.

          The Multiarch Tuning Operator introduces the ClusterPodPlacementConfig custom resource to
          enable architecture-aware scheduling, ensuring that pods are assigned to nodes with
          a corresponding architecture for their container images.

          To enable this functionality, you must create a ClusterPodPlacementConfig resource.

          When the ClusterPodPlacementConfig object is created, the Multiarch Tuning Operator deploys the necessary operands
          to support architecture-aware workload scheduling.
        displayName: Multiarch Tuning Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multiarch-tuning-operator
        - pod-placement
        links:
        - name: Multiarch Tuning Operator Documentation
          url: https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/postinstallation_configuration/configuring-multi-architecture-compute-machines-on-an-openshift-cluster#multiarch-tuning-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.27.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/multiarch-tuning-operator
        relatedImages:
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-operator-bundle@sha256:f49b7a2c60e79edaba1a5faabc3983f87f9123b6d84121a438d179ccbb8102e3
        - registry.redhat.io/multiarch-tuning/multiarch-tuning-rhel9-operator@sha256:d532619b2090be5cc69a70225ca3dcbbc701af35a533be277783beb8ffdcdc91
        version: 1.1.1
      entries:
      - name: multiarch-tuning-operator.v1.1.1
        version: 1.1.1
      - name: multiarch-tuning-operator.v1.1.0
        version: 1.1.0
      - name: multiarch-tuning-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: multiarch-tuning-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/multiarch-tuning-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: apicurio-registry-3
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-registry-3.v3.0.9-r3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "simple"
                },
                "spec": {
                  "app": {
                    "ingress": {
                      "host": "simple-app.apps.cluster.example"
                    }
                  },
                  "ui": {
                    "ingress": {
                      "host": "simple-ui.apps.cluster.example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
          createdAt: "2025-08-05"
          description: Deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Application Foundations",
            "Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Red Hat build
              of Apicurio Registry version 3.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Red Hat build of Apicurio Registry 3

          Apicurio Registry stores and retrieves API designs and event schemas,
          and gives you control of their evolution.

          **Features**
          - Supports: Apache Avro, AsyncAPI, GraphQL, JSON Schema, Kafka Connect Schema, OpenAPI, Protobuf
          - Allows you manage event schemas and API designs using the Apicurio Registry web console, REST API, Maven plug-in, or Java client
          - Includes Serializers and Deserializers for Kafka client integration
          - Configurable rules to control schema validity and evolution (compatibility)
          - Storage options: Kafka, PostgreSQL, MySQL\*, SQL Server\*, in-memory\*
          - Compatible with Confluent APIs
          - Runs on a lightweight Quarkus platform

          _Note: Storage options marked with * have not yet been tested by Red Hat for production deployments._

          ## Red Hat build of Apicurio Registry 3 Operator

          Provides a quick and easy way to deploy and manage Red Hat build of Apicurio Registry 3 on OpenShift.

          **Features**
          - Supports installation and configuration of Red Hat build of Apicurio Registry 3
          - Perform a rolling upgrade of Red Hat build of Apicurio Registry 3

          ## Prerequisites

          This operator does not deploy storage for Red Hat build of Apicurio Registry 3. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Red Hat build of Apicurio Registry 3 and Apicurio Registry 3 Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Red Hat build of Apicurio Registry 3
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - integration
        - streaming
        - messaging
        - api
        - schemas
        - registry
        - apicurio
        - apicurio-registry
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        - email: jsenko+apicurio@redhat.com
          name: Jakub Senko
        - email: ewittman+apicurio@redhat.com
          name: Eric Wittmann
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/apicurio/apicurio-registry-rhel9@sha256:327a9899adfe89af9aebdcfac651a9eddec9ecf8a345caee1cef1b4a1fb279d5
        - registry.redhat.io/apicurio/apicurio-registry-ui-rhel9@sha256:df0493f7edbdedf754c70e4efc256fa4d246a24d6c8882c562e56ef3844278c1
        - registry.redhat.io/apicurio/apicurio-registry-3-operator-bundle@sha256:728354aee3729a1fba233c2aa92cacb21d27efd4869706ddd214a2d3688e33b2
        - registry.redhat.io/apicurio/apicurio-registry-rhel9-operator@sha256:2f26805a2b77578d2238b4c5454bc20b7a717a333c64c75b57c55a47bc779c61
        version: 3.0.9-r3
      entries:
      - name: apicurio-registry-3.v3.0.9-r3
        version: 3.0.9-r3
      - name: apicurio-registry-3.v3.0.9-r2
        version: 3.0.9-r2
      - name: apicurio-registry-3.v3.0.9-r1
        version: 3.0.9-r1
      - name: apicurio-registry-3.v3.0.7-r2
        version: 3.0.7-r2
      - name: apicurio-registry-3.v3.0.7-r1
        version: 3.0.7-r1
      - name: apicurio-registry-3.v3.0.6-beta.r1
        version: 3.0.6-beta.r1
      - name: apicurio-registry-3.v3.0.6-beta.r1-0.1740495889.p
        version: 3.0.6-beta.r1+0.1740495889.p
      - name: apicurio-registry-3.v3.0.6-beta
        version: 3.0.6-beta
      name: 3.x
    defaultChannel: 3.x
    packageName: apicurio-registry-3
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: rh-service-binding-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: service-binding-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ServiceBinding expresses intent to bind an operator-backed
              service with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding for Kubernetes. Describes the connection
              between a Provisioned Service and an Application Projection. Implements
              Service Binding Specifcation for Kubernetes (https://github.com/k8s-service-bindings/spec).
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:d9930900b0d2f92ae813246fc322290234ad8bbd165b10e36671b53852228718
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:fe6629401bc8447da95ec55949b5a27e620364bb6e02f86230262699ac8a5d55
        version: 0.9.1
      entries:
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: beta
    - currentCSV: service-binding-operator.v0.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindablekinds-sample"
                }
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:e799bd06043028d214baeeffc683710ad2eed7b83272fc653e3b89f69711319b
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:87448055a162e215a4d4982a8494fec6e7cc59fe6704494ddfdcfea076a2100c
        version: 0.11.0
      entries:
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: candidate
    - currentCSV: service-binding-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "service.binding/v1alpha2",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "Kind": "Ddeployment",
                    "apiVersion": "apps/v1",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.service.binding
            version: v1alpha2
        description: " The Service Binding Operator enables application developers
          to more easily bind applications together with operator managed backing
          services such as databases, without having to perform manual configuration
          of secrets, configmaps, etc. The Service Binding Operator accomplishes this
          through automatically collecting binding information and sharing with an
          application to bind it with operator managed backing services. The binding
          is performed through a new custom resource called a ServiceBindingRequest.\n###
          Example\nA set of examples, each of which illustrates a usage scenario for
          the Service Binding Operator, is being developed in parallel with the Operator.
          Each example includes documentation and can be run either through the OpenShift
          web console or command line client. The examples are available [here](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md#example-scenarios)\n###
          Documentation\nRefer to the [documentation](https://github.com/redhat-developer/service-binding-operator/blob/master/README.md)\n###
          \ Help\nRaise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)\n###
          Licence\nService Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE) "
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Blog post
          url: https://developers.redhat.com/blog/2019/12/19/introducing-the-service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: candidate
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator@sha256:3d7a3648e86848402577dd19c1f8fd130a5556ebdfbc0f663a72f9e141024812
        - registry.redhat.io/ocp-tools-4-tech-preview/service-binding-rhel8-operator-metadata@sha256:a7c6f1c9260c79ebbdfb25ea996516ba1e2bcbedf25401e9e6848b955a2d3d78
        version: 0.10.0
      entries:
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: preview
    - currentCSV: service-binding-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "BindableKinds",
                "metadata": {
                  "name": "bindable-kinds"
                },
                "spec": {}
              },
              {
                "apiVersion": "binding.operators.coreos.com/v1alpha1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "example-servicebinding"
                },
                "spec": {
                  "application": {
                    "group": "apps",
                    "name": "nodejs-rest-http-crud",
                    "resource": "deployments",
                    "version": "v1"
                  },
                  "services": [
                    {
                      "group": "postgresql.example.dev",
                      "kind": "Database",
                      "name": "pg-instance",
                      "version": "v1alpha1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1alpha3",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "application": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  },
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  }
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ClusterWorkloadResourceMapping",
                "metadata": {
                  "name": "cronjobs.batch"
                },
                "spec": {
                  "versions": [
                    {
                      "containers": [
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.containers[*]"
                        },
                        {
                          "path": ".spec.jobTemplate.spec.template.spec.initContainers[*]"
                        }
                      ],
                      "version": "*",
                      "volumes": ".spec.jobTemplate.spec.template.spec.volumes"
                    }
                  ]
                }
              },
              {
                "apiVersion": "servicebinding.io/v1beta1",
                "kind": "ServiceBinding",
                "metadata": {
                  "name": "servicebinding-sample"
                },
                "spec": {
                  "service": {
                    "apiVersion": "postgresql.example.dev/v1alpha1",
                    "kind": "Database",
                    "name": "pg-instance"
                  },
                  "type": "postgres",
                  "workload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "name": "nodejs-rest-http-crud"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          containerImage: registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
          description: An operator to support binding capabilities between imported
            apps and operator backed services
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.17.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-developer/service-binding-operator
          support: Service Binding Operator Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BindableKinds contains a list of bindable type (group, version,
              kind) found in the cluster. It is cluster-scoped and there is only single
              instance named 'bindable-kinds'.
            displayName: Bindable Kinds
            kind: BindableKinds
            name: bindablekinds.binding.operators.coreos.com
            version: v1alpha1
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1alpha3
          - description: Cluster Workload Resource Mapping defines the mapping for
              a specific version of an workload resource to a logical PodTemplateSpec-like
              structure. It provides a way to define exactly where binding data needs
              to be projected. Use this method when you are not able to configure
              custom path locations correctly by any other methods.
            displayName: Cluster Workload Resource Mapping
            kind: ClusterWorkloadResourceMapping
            name: clusterworkloadresourcemappings.servicebinding.io
            version: v1beta1
          - description: Service Binding expresses intent to bind a backing service
              with an application workload.
            displayName: Service Binding
            kind: ServiceBinding
            name: servicebindings.binding.operators.coreos.com
            version: v1alpha1
          - description: (Tech Preview) Service Binding implementing community specification
              (https://github.com/k8s-service-bindings/spec) describing the connection
              between a backing service and an application workload. Available as
              tech preview but subjected to change in near future.
            displayName: Service Binding (Spec API Tech Preview)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1alpha3
          - description: Service Binding implementing community specification (https://github.com/k8s-service-bindings/spec)
              describing the connection between a backing service and an application
              workload.
            displayName: Service Binding (Spec API)
            kind: ServiceBinding
            name: servicebindings.servicebinding.io
            version: v1beta1
        description: |
          The Service Binding Operator manages the data plane for applications and backing services. It enables developers to connect their
          applications to backing services (REST API, databases, event buses and others) with a consistent and predictable
          experience. The operator reads data made available by the control plane of backing services and projects the bindings
          data to applications either as accessible files mounted into the application’s container or environment variables.

          In Kubernetes, each service suggests a different way to access their secrets and developers will consume them in their
          application in a custom way. While this provides a good deal of flexibility, it also creates discrepancies and
          complexity dealing with each specific services, which leads large development teams to lose overall velocity.

          Service Binding Operator removes the error-prone manual configuration of binding information, provides service operators
          a low-touch administrative experience to provision and manage access to services and enriches development lifecycle.

          For additional details refer to the project [documentation](https://redhat-developer.github.io/service-binding-operator).

          Get Started with Service Binding Operator
          * Find out about SBO’s [prerequisites](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/installing-service-binding.html).
          * Refer to the [Quick Start Guide](https://redhat-developer.github.io/service-binding-operator/userguide/getting-started/quick-start.html) to see SBO in action in a simple scenario.

          Need Help
          * Raise a ticket for bugs, features and enhancement [here](https://github.com/redhat-developer/service-binding-operator/)

          Licence
          * Service Binding Operator is licensed under [Apache License 2.0](https://github.com/redhat-developer/service-binding-operator/blob/master/LICENSE)

          Service Binding Operator is providing and supporting two different APIs.
          * `binding.operators.coreos.com/v1alpha1`: This API is compliant with the Service Binding specification. It is fully supported and considered as the mature API. It’s the API we recommend in production environments.
          * `servicebinding.io/v1beta1`: This API is the one defined in the Service Binding specification. It is available in Service Binding Operator 1.x.

          We’ll consider supporting only the API from the specification, once the Service Binding specification will be officially GA. Of course, that will be a breaking change and handle in a safe way for you.
        displayName: Service Binding Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ""
        links:
        - name: Documentation
          url: https://redhat-developer.github.io/service-binding-operator
        maintainers:
        - email: service-binding-support@redhat.com
          name: Openshift Application Services
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - registry.redhat.io/ocp-tools-4/service-binding-operator-bundle@sha256:9ab3d5eb70c654270fef9e8e9d4e5b68e1ac3c86bd0e0dcd36cc6f8741be1603
        - registry.redhat.io/ocp-tools-4/service-binding-rhel8-operator@sha256:9c8fb0c8f9d2f54685561af21e595aa88bf309eaebebdf23fb6997b75d026b18
        version: 1.3.3
      entries:
      - name: service-binding-operator.v1.3.3
        version: 1.3.3
      - name: service-binding-operator.v1.3.1
        version: 1.3.1
      - name: service-binding-operator.v1.3.0
        version: 1.3.0
      - name: service-binding-operator.v1.2.0
        version: 1.2.0
      - name: service-binding-operator.v1.1.1
        version: 1.1.1
      - name: service-binding-operator.v1.1.0
        version: 1.1.0
      - name: service-binding-operator.v1.0.1
        version: 1.0.1
      - name: service-binding-operator.v1.0.0
        version: 1.0.0
      - name: service-binding-operator.v0.11.0
        version: 0.11.0
      - name: service-binding-operator.v0.10.0
        version: 0.10.0
      - name: service-binding-operator.v0.9.1
        version: 0.9.1
      - name: service-binding-operator.v0.9.0
        version: 0.9.0
      - name: service-binding-operator.v0.8.0
        version: 0.8.0
      - name: service-binding-operator.v0.7.1
        version: 0.7.1
      - name: service-binding-operator.v0.7.0
        version: 0.7.0
      - name: service-binding-operator.v0.6.0
        version: 0.6.0
      - name: service-binding-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: rh-service-binding-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: self-node-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: self-node-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
          createdAt: "2023-10-20 16:38:06"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.4.0 <0.5.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the heath detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-self-node-remediation-operator.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:36561e5888cd34f8fb83ceff2910141fceaa6ea290bd246125930fe183e2ebc3
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:cb03286b9d9cda4a47e5b1c0058ca315bf082bcc8697506ea88d97ba77452bea
        version: 0.5.1
      entries:
      - name: self-node-remediation.v0.5.1
        version: 0.5.1
      name: 4.12-eus
    - currentCSV: self-node-remediation.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
          createdAt: "2023-10-20 15:32:20"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          olm.skipRange: '>=0.5.0 <0.7.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:9e8bcbff46b97f3bc1071146b8db233abd1630016e6fc0efb679df8243c6fd03
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:e2048a8a2ec37e10b57dafeba2e131311e61982242cda4beb989e5e30c1a5537
        version: 0.7.1
      entries:
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.14-eus
    - currentCSV: self-node-remediation.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
          createdAt: "2024-07-10 00:59:51"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.9.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:518691d157b3e3b5031a2b1e4cc3f415839536079bb8570ae8690ac92d302ba0
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:a64f67f29a4eb8ce426ed842f1a5a5fcd83fd9d91acebb064cf65c89c9d8bfb6
        - registry.redhat.io/workload-availability/self-node-remediation-rhel8-operator@sha256:771c5ba544599518e60c1c50e01cc7c951299255ce7f30a8b6e1bb9b127c6697
        version: 0.9.0
      entries:
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.16-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: 4.18-eus
    - currentCSV: self-node-remediation.v0.10.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediation",
                "metadata": {
                  "name": "selfnoderemediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationConfig",
                "metadata": {
                  "name": "self-node-remediation-config",
                  "namespace": "self-node-remediation"
                },
                "spec": {}
              },
              {
                "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                "kind": "SelfNodeRemediationTemplate",
                "metadata": {
                  "name": "selfnoderemediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
          createdAt: "2025-07-02 18:34:25"
          description: Self Node Remediation Operator for remediate itself in case
            of a failure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/self-node-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SelfNodeRemediationConfig is the Schema for the selfnoderemediationconfigs
              API in which a user can configure the self node remediation agents
            displayName: Self Node Remediation Config
            kind: SelfNodeRemediationConfig
            name: selfnoderemediationconfigs.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediation is the Schema for the selfnoderemediations
              API
            displayName: Self Node Remediation
            kind: SelfNodeRemediation
            name: selfnoderemediations.self-node-remediation.medik8s.io
            version: v1alpha1
          - description: SelfNodeRemediationTemplate is the Schema for the selfnoderemediationtemplates
              API
            displayName: Self Node Remediation Template
            kind: SelfNodeRemediationTemplate
            name: selfnoderemediationtemplates.self-node-remediation.medik8s.io
            version: v1alpha1
        description: This self node remediation operator is using an alternate mechanism
          for a node in a cluster to detect its health status and take actions to
          remediate itself in case of a failure. While not all remediation events
          can result in the node returning to a healthy state, the operator does allow
          surviving parts of the cluster to assume the node has reached a safe state
          so that it’s workloads can be automatically recovered. A prerequisite for
          the Self Node Remediation operator is a functioning health detection system
          that implements the external remediation API, such as Node Health Check
          Operator or Machine Health Check Controller (in OCP or Cluster-API based
          clusters). Once a node/machine is unhealthy, the detection system will create
          the SelfNodeRemediation CR, which triggers the Self Node Remediation Operator.
          For this the configuration of the health detection system needs to refer
          to a SelfNodeRemediationTemplate CR. Such a template is created by default
          in the operator's namespace, for usage in other namespaces it needs to be
          created manually. The Node Health Check Operator is configured to use Self
          Node Remediation by default.
        displayName: Self Node Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fencing
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - SNR
        links:
        - name: Self Node Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.3/html/remediation_fencing_and_maintenance.html
        - name: Source Code
          url: https://github.com/medik8s/self-node-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/self-node-remediation-rhel9-operator@sha256:a774c6cef3941555f22631096670c6c861be34df319052e3dcdb6ee4aca6fb72
        - registry.redhat.io/workload-availability/self-node-remediation-operator-bundle@sha256:1267b3f113652d675c9fb24bb191e8d23f3b88a36dff89469fd88da80006e463
        version: 0.10.2
      entries:
      - name: self-node-remediation.v0.10.2
        version: 0.10.2
      - name: self-node-remediation.v0.10.1
        version: 0.10.1
      - name: self-node-remediation.v0.10.0
        version: 0.10.0
      - name: self-node-remediation.v0.9.0
        version: 0.9.0
      - name: self-node-remediation.v0.8.0
        version: 0.8.0
      - name: self-node-remediation.v0.7.1
        version: 0.7.1
      name: stable
    defaultChannel: stable
    packageName: self-node-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhacs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        version: 3.74.9
      deprecation:
        message: |
          The `latest` channel is no longer supported.  Please switch to the `stable` channel.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: latest
    - currentCSV: rhacs-operator.v3.62.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
          createdAt: "2021-08-13T17:10:01.410173+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.61.0 < 3.62.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.5.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main page before deploying, or consult the RHACS
              documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create an cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead. If you want to deploy Secured Cluster Service in another namespace, additionally pass the `-n <namespace>` flag.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:558c0dabdc52efaee7bc889ef397ac2ac30f7c54562c2cc0070d0ef991a8b31b
        - registry.redhat.io/rh-acs/collector@sha256:98cfab15684506f1d0978eb701e8de478d91fc5bda3013ffd3b436aa95f07cc0
        - registry.redhat.io/rh-acs/main@sha256:9932304bb62b084ee1ae7e75bb21d6dfb737260c851f2d8daf05082f79230c66
        - registry.redhat.io/rh-acs/scanner-db@sha256:ff877f503e08ecca0ce7a7d8db760b2e4004e8e8ae2f27fb9b9c00a05b58c93a
        - registry.redhat.io/rh-acs/scanner@sha256:266a575c2547a3223890e8b73db0f14d69dc5a22217c7e4e3dc79644e6ab0258
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:7fd7595e6a61352088f9a3a345be03a6c0b9caa0bbc5ddd8c61ba1d38b2c3b8e
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:56905c2e7df0e29ca2b6301cd81b4f9bbe02b483374708f7f7da219a4b29428d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e30f7835cf52dbce9afdbaf55e5a93d8baf9c4b3a3ae488d1d42c1adba7f46af
        version: 3.62.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.62.1
        version: 3.62.1
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.62
    - currentCSV: rhacs-operator.v3.64.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
          createdAt: "2021-09-22T05:10:01.286657+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.63.0 < 3.64.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/main@sha256:6aeb2223124540ac217c38b016dea99d770ab897dde4906a821d8776d91ea03d
        - registry.redhat.io/rh-acs/scanner-db@sha256:67e81bf896267b7acab8cb6755cde552173de4f0881c38eaca8e491bd926cc53
        - registry.redhat.io/rh-acs/scanner@sha256:8b7121bce8f9c7133d29245ebb265a4b8f9831a66a57f11b2d68b9b8b5cf79cc
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:5187b04f8010a274d6bc017fc82f1b00d886124e520c1b7de8b4da8cd2f37bc0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fa478167f2f738f244bed6a06450679be73f8464c40f86acde4ffd390505692b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        version: 3.64.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.64.2
        version: 3.64.2
      - name: rhacs-operator.v3.64.1
        version: 3.64.1
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.64
    - currentCSV: rhacs-operator.v3.65.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
          createdAt: "2021-09-21T13:26:28.711610+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.64.0 < 3.65.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:229c7e5f967c70dfbc3cf638d521267ea6acf875b055499dfd30cb19c0b1f69c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0cacd00c0a147cc8295138565ef2dc2ddbd9ad1fb25f6be817719a75ea4c6448
        - registry.redhat.io/rh-acs/collector@sha256:2b3a8bd3af62516a6714fe612e5fdc0433b91d873182e290d3ddc2b32ce291f0
        - registry.redhat.io/rh-acs/collector@sha256:3eb099da081cf208d46be5ccfa7b5e9ac0e2cad49f127fa4db91e49e3f26db42
        - registry.redhat.io/rh-acs/main@sha256:e672ef9a1239065f8cf0f331757861fea082e971a9e9760d1e65795f399cbde1
        - registry.redhat.io/rh-acs/scanner-db@sha256:20f258500ac2ba4e626df92dbcd193cf7ab8687ac251591be2442b8a46871835
        - registry.redhat.io/rh-acs/scanner@sha256:bfe9b2205bfca6aebed14827b4bda098bfad574e67ffa967ad339543062200fe
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:d3bf33899ca8b199ca8782978618ca343164e43fd0ae2ac6a0cfbf697efa6428
        version: 3.65.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.65.1
        version: 3.65.1
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.65
    - currentCSV: rhacs-operator.v3.66.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
          createdAt: "2021-10-20T16:38:38.091764+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.65.0 < 3.66.1'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:d93e13ee1d1cc4b7f82e644a7d43e560aac42bd1e06d25b476cea9a4338bf9ce
        - registry.redhat.io/rh-acs/main@sha256:97230c01e76d90c7fe35ffeaa591be3d9175f350a8fd88599aded09a1edf930c
        - registry.redhat.io/rh-acs/scanner-db@sha256:490b0bf49161e4bc6de2504726aaf360e0ca9a214aea6b886fff13153bdb263d
        - registry.redhat.io/rh-acs/scanner@sha256:323ee86d6a9cbeb32270c659c1154f3a62c36535c6ea894e4e29e93eb1f4a84f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6536ee795e8aaeab9321ee4820e55933195983a768ba24fab40508d03e50a6e4
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:10cdb0ff3e67b65737cde420536494bd84c2b7ccf65192bbed2eb43eaed66a41
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bd74f0b95ff2881faf6eb4520feecd067b69a759f09bbfecd467447ce2a2a25e
        - registry.redhat.io/rh-acs/collector@sha256:5673c27554a71502e06e08945a4424c6ce5da27f489854f3aa452e6d31240aee
        version: 3.66.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.66.1
        version: 3.66.1
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.66
    - currentCSV: rhacs-operator.v3.67.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
          createdAt: "2021-12-16T12:54:10.117625+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.66.0 < 3.67.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-acs/collector@sha256:57c2dfa3dbc8368076291fbc627baa01dcb5ebc50f838e02ce7ac38a128d027b
        - registry.redhat.io/rh-acs/collector@sha256:e4648723ccbd43fa5acf005d9734bfc06d41e19e1be0bde839880e28fa361e16
        - registry.redhat.io/rh-acs/main@sha256:8db7630d094b661411ebb3f35248848d3147c735677d0922f7a04de9a41eacc9
        - registry.redhat.io/rh-acs/scanner-db@sha256:6f8d47277732eb56317d576c2d06fb1f539d46d16520e6ca34abec290392cef9
        - registry.redhat.io/rh-acs/scanner@sha256:2860ed271da6064149bb53568c5ad60e129918d55a2f244f9ade4a0bea4c4933
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:ad88773ab141f45fe24c61e83ec736aab27f652fb82ca2a85fee6f885568c5b9
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:893ebcb2d0c96647e56bce3b16fe32851df59cd71f2c0fc8f6e66ca3a669bb70
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:86e5fa1fa294987114be200890c2e516501e424aee0fb98ece25c95e7716295b
        version: 3.67.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.67.2
        version: 3.67.2
      - name: rhacs-operator.v3.67.1
        version: 3.67.1
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.67
    - currentCSV: rhacs-operator.v3.68.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
          createdAt: "2022-06-15T22:13:22.427644+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.67.0 < 3.68.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:f0d2f5f11c2a96d34d6b4baf249cbf821fe064eade73c3f95c22f3c6981ac4d1
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:195a694dc96096166fe82e81ecb3ceb493f96cfb5029112e576e44ab39b56abd
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9b9b9146d9683d6da0ff2abbc8ea82b96d945b7a8c3740d72d8c018fb2b04fba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c9d0a5a2871c24c0b1058e419f2b090429363a319856397855412fd2c0037395
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:081f2198ad70bfb2ec66153b50b48b950b9a5550ba996de413c5a2aff4052783
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6a10a4859527946f61f497ff84b80f973cb7be5982d428423836b5aca81a1f5d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0bb76363a3a25916dd4493e177e9b2d36eac9633473b07fe0fffb853fd072da3
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:105d10bd50a6062c998ed107dea388dd78f0caddb0b57713c8692b1d07042774
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:711293438bab35b17d2e1f162e7f74f667a6e16dd32e45122f917b6817fefb4e
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b5fc3dcaa72bbc8dea005189c940e1cf2d00c94a6cf7676cf0e1790e1191d45c
        version: 3.68.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.68.2
        version: 3.68.2
      - name: rhacs-operator.v3.68.1
        version: 3.68.1
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.68
    - currentCSV: rhacs-operator.v3.69.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
          createdAt: "2022-06-21T09:07:40.705944+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.68.0 < 3.69.2'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:9ae03669cfdc0f4fe56b021d911edb17159c07757720ae30a1f22d7f7a897c07
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:da6dd486a20fdba61ea12d2c77ea299d9795a9f22d9057d65ff992b1f5896e07
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b254746ed3faa1f1cbcafdf1343a1957bf0214311b36f4e892be3000cc2aecf3
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:297e7f863106c6b41f882bbd5b691a12c797c24d832473a04e604d4639e4a68e
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:8db5206532e95a7ad7cb23f389155d41757aaf5f445ddce92b008a06627d3fd5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:cf5629823f2dd583495bcbbcf3b83b4f29f855ace3c388f5cb7b24a6d34ed267
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:867613178d491f416e57975843817251490c5fbca2fc69d8b55b169bea675279
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9b2beceb255124855ba1b352d16009b2f2cd422901a19c311f0783dab1fcc3a0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:f167e928740d1c614e26948ea58f49e569f6a4fddc438cd0abe52f60538432d0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:77f5db63e17b140a8d6b84e3822fd77463035be9c0c832b55b4977696da13ccc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:2afb2823edc6f3434e18db60ee37381b5b79f51234c1a261b6d0f8ed74cc8b63
        version: 3.69.2
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.69.2
        version: 3.69.2
      - name: rhacs-operator.v3.69.1
        version: 3.69.1
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.69
    - currentCSV: rhacs-operator.v3.70.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
          createdAt: "2022-06-20T12:30:26.999648+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.69.0 < 3.70.1'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:e71f2f2347dd46dec790405b1bb98d54a348a4583949f142f5703be9a790f96e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3d7c8b77671740e324ee6c309a84bbaa4d925e6750979fe4e46c561bc79f1b4d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:dfeb239931f7a33ade5c829a102cabc01f5c348e1aa35a7ddc24757cfd549629
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:df8cfdaa40b50ff2f2670d97c2115b47ae4bea443ae2831caaa468dd62f3c75f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6c32b86ed8bc18e384706eb6188406e66910bf34c8869bd155a60c3525d0fea7
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ee55a233d2a8fd77d89ec2c69f357721c7debead6470a9f494b3cd6a001829c4
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:9b35ae6df2c19935783735d213483f3fcecead2f49f873a4f7844ef9fb5c7f1a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:d2e827a209840b7197960c8115720972237532eb9cf9d6fee7a36107053408ce
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:b47c96af1c9ffe4d8728da510a93efe21adc221225fe2b822cfcace6c6672ba2
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fb903bf9e110ba03e9c63977ebaf707e6884dd8fd413bcd6b793a0ed795cdd0e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:6b14e850cda9a54618847858292ea669abff599a40692e27fed0e337c4a60526
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:08a7d8c892fad69c589b81ad8fbd21d36c8fa4078530dba361cfa9a8f9f05950
        version: 3.70.1
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.70.1
        version: 3.70.1
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.70
    - currentCSV: rhacs-operator.v3.71.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
          createdAt: "2022-12-07T08:31:36.777654+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.70.0 < 3.71.3'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c3ec88cd99908894d0e67153779062f90ac1b007980d66040a79b2be13798a10
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:df7b41999b542a477848823a1d0e7e5c29ae1b00fbac035007a5a1f33fac139f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:4a0dc8c337299d7f6440ff75c2d5219d3ab68183f4f117f008ef065757dd5d4b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:cb2d0f470968c2b725b0d176d423cc1866fd56893344d077d28f2c16581d3614
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e001b9f844a6c298dd74f1ab683f3dbbf0ad66044cec419c373ef4f97d47be73
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:469b6c1592226eb9a521131ef758a33e6a6ac7c96bc67fcc903cea30469c125e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:f16f92e074f7125b88e9e2033c7c12b44421145b93d040034fa4688626869af1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:5fdfd3690af60bd5ecf04741cd1f73259491da501700b066120976960567f0db
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:4d8bac9b3857af6ba351c7768f70ef30ef7cb99ae4df60cb226b5a793353a052
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:7f279bb6e0f34a5208d8711ba4d7b62c78f08b683815ce33f6aa28c23c4246da
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:9bc25e81c21cda6d8b535d8ac67c1286fe57a1596001850e730f8f73bf371111
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:e282a8e5e606003ea344d44ac5ee7699563848f7160eeae183aa58ad740e0e94
        version: 3.71.3
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.71.3
        version: 3.71.3
      - name: rhacs-operator.v3.71.2
        version: 3.71.2
      - name: rhacs-operator.v3.71.1
        version: 3.71.1
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.71
    - currentCSV: rhacs-operator.v3.72.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
          createdAt: "2023-02-28T17:02:08.443613+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.71.0 < 3.72.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:622f30b7f591194d86a5bf95deede0225e66f7b624db197f6289a5fbbbe887a9
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:c07fae3616cbc681e4d31839abea6623b4bf4dade93273201c8db7ba0d635aad
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:0b69595405e4e598183bbf4197ac5b7593724fa6162413f225aa01a542af18c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:3ff61e1cc77b66a22c8a8c0ab306f0e122cd74e762005285aede23ebb6c2f6a2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:42e5f6c4ba789f124ac3ebe7b6d1a607ede8e763a35888c97ccbd0259bec4a1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:575b6c3e2914751df1668ffff83cbda187ce6e56f32214414cdf036b7a62bbd8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:cd475a6c2356d63cfe736a11114de7fedfbd8fdcfecc56224a00b6c819103be6
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:cd0c205ee44eddcdf99b84e80fcf4403851d956efc50532b9d12013bafe44988
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:cb5df858e0c5cae8f77b285e23daeb83062bd72570c90ed5ea3ce4a02d2f5f08
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:878ed77d07d3f40f049a594af44376797305786a2a264dae3d0c50401643bfa5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d56a00b3e20e60b7169f5c35d03fa6a1bd3b9980dd3ebc9d89d759e1d6d20ad2
        version: 3.72.4
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.72.4
        version: 3.72.4
      - name: rhacs-operator.v3.72.3
        version: 3.72.3
      - name: rhacs-operator.v3.72.2
        version: 3.72.2
      - name: rhacs-operator.v3.72.1
        version: 3.72.1
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.72
    - currentCSV: rhacs-operator.v3.73.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
          createdAt: "2023-05-26T21:27:04.033276+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.72.0 < 3.73.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:5df359156b01505ce43e5f14c3b08722856bdb86f3d0d9b4fba785a2e580f346
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:082bd6f4b40209e0591be001dea358f328cd86a08691cdceb7f71ea59d861d9d
        - registry.redhat.io/advanced-cluster-security/rhacs-docs-rhel8@sha256:ee2a9e4ab74f5a554d1e4ff148833a53bdaab70ac1e2668cc9ee2fceca290ba9
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:459a8d91beff91638872d377a8b82c279ad2905ce3f4ad4bfc00f6822418ac57
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:c020a80ff82bf4d9b5cec3eeac5de9990de76041e22c444e664a20e15c46ce3b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:648b14e4158da2a85eedb46b3456fb502d680d52d375ad34bf45b4b94f8b2ef6
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e4a27d3ffb173c87728fb637ca7fde92a9e78ea9ee3edcecccc2a5e3d3748b8c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:e2c89f735969d438f75a89ff63c404a8ec4361a03fde48e880fee73006b9c669
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a8486b66f977a639d77ef110492e5ecdde47be8acaefbaa1cbb5417085c56839
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:7bf9415aa4b0746d11ca5a59d4179bc382a0fac4ded969f0c4294171fde27d58
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dce5db4f162dd40c09cad114cb802cd6d7c72d0c6986581f2e74c977493f75b0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:4d0ead8969f120a629104a92a454842bb7103aec137deb6c8209b7db9c48a1b2
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:e6687c12d070069f655c17832d94c619c18274a259ccc59715036f9908e162f7
        version: 3.73.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.73.5
        version: 3.73.5
      - name: rhacs-operator.v3.73.4
        version: 3.73.4
      - name: rhacs-operator.v3.73.3
        version: 3.73.3
      - name: rhacs-operator.v3.73.2
        version: 3.73.2
      - name: rhacs-operator.v3.73.1
        version: 3.73.1
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.73
    - currentCSV: rhacs-operator.v3.74.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
          createdAt: "2024-02-06T23:06:13.271846+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.73.0 < 3.74.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:e714fea6deb82225c54de32cf7ff3e91e1966470c58d04a221c7d90a501fdb8a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:a108e9a7967001e0bdfed7124b01e9a3dbf1dec308444762973030aec57e1d5a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:2c0840f4573373a19eb8c9a26501138ef0e76efbd508fd69a6260e78194d86c8
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:120b9891ee9db839ab0b2029a62e67ec4747c47c6ffe15af38e9035160f57b78
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:d546f2d3f668ce2f20e594394ade0a86d808f6c8c88cadbba227083d4abac940
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:22bf00e1457b4a9e63901892aa4375a1d4a9d1b2c067c1f054d918a24b222b10
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:71b9fdcaf4cc761b1499be8e6ccff6c463cc6b9856e9deb1f93baad45ea59eb0
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4980165af1d43837bf5c0ae5dd271c8d482845670d608db81652ab886898b593
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:264102d4bca7d9bd384dc04e5bbff1a7b43343ea8e59cc79047cdda50a3f299f
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:87eb62fe8aa7952d0d69dc13c4867ac2f51e821f56d149a9b588c826caec4c16
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:74624389ac3261ed699b1f64e5717353da68895709d30a011bc8813b61a4a249
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:0482d4464d6d39c01abab8df882604dd227c52b5420d36fe7414aaff8067cbec
        version: 3.74.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v3.74.9
        version: 3.74.9
      - name: rhacs-operator.v3.74.8
        version: 3.74.8
      - name: rhacs-operator.v3.74.7
        version: 3.74.7
      - name: rhacs-operator.v3.74.6
        version: 3.74.6
      - name: rhacs-operator.v3.74.5
        version: 3.74.5
      - name: rhacs-operator.v3.74.4
        version: 3.74.4
      - name: rhacs-operator.v3.74.3
        version: 3.74.3
      - name: rhacs-operator.v3.74.2
        version: 3.74.2
      - name: rhacs-operator.v3.74.1
        version: 3.74.1
      - name: rhacs-operator.v3.74.0
        version: 3.74.0
      - name: rhacs-operator.v3.73.0
        version: 3.73.0
      - name: rhacs-operator.v3.72.0
        version: 3.72.0
      - name: rhacs-operator.v3.71.0
        version: 3.71.0
      - name: rhacs-operator.v3.70.0
        version: 3.70.0
      - name: rhacs-operator.v3.69.0
        version: 3.69.0
      - name: rhacs-operator.v3.68.0
        version: 3.68.0
      - name: rhacs-operator.v3.67.0
        version: 3.67.0
      - name: rhacs-operator.v3.66.0
        version: 3.66.0
      - name: rhacs-operator.v3.65.0
        version: 3.65.0
      - name: rhacs-operator.v3.64.0
        version: 3.64.0
      - name: rhacs-operator.v3.63.0
        version: 3.63.0
      - name: rhacs-operator.v3.62.0
        version: 3.62.0
      name: rhacs-3.74
    - currentCSV: rhacs-operator.v4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
          createdAt: "2023-10-20T07:04:50.531740+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 3.74.0 < 4.0.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:9bf3a3e0fd573f91f03794eaf5291364a90392e5e7d4b5d4ed550ef4600ab826
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:5a492b856e0b6d9fdb020b556d13186ef5e7095b0c736a2031500c9c00dad0c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:139d80d398e7e5cfc85839e3039a3630b48c7ae6650d755754430fccd94d6a27
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:cfa10bb6b461f302674f185b26c8d6efaaeba7f3f65d2cf05ae1f069aa2710d3
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:08f537cba262605a0eb0a510b6faf78e3ba26465f636d03cc1f7a6ed6b40d355
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0aedd8e865f72ffa8b79fd2bab48974caec3eea3e3a6f77c89fe6dacb8947057
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:a0d9cd31d953aaf7e84f5a39f405c5aa2ea62e68d6c728a936ca1ffc4dc1b93e
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:09e79b5db8977520f9951cc43c339c8e40b4659d7da513347d5a88b2500f809d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:3250d89e78de852a2627593f29469f0b3ff639ea3644b4b7d6d057d0c1ad3659
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:eee1aa054d4d15b23be293f31d3ea659e0ffc1d05174ef89a6c977dd3d4377cc
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:c39621837cbc8419d187a43a9575ec8cb4021c58ea1b56f215003d7264eb5d03
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:4c848eb7465e928251e05d32ffbc93cd14e7bb77972d22b8721bd6eaa4a93d2f
        version: 4.0.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.0
    - currentCSV: rhacs-operator.v4.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
          createdAt: "2024-01-13T18:55:22.011709+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          olm.skipRange: '>= 4.0.0 < 4.1.6'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:b26133843687cc50e2c52d6067d8b3369b5cb28dfbfd15dd4b6062096673195a
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:a8ad0e76ce4cae9f9b36aecf125c4ed56fc2d776fe3df113d41bfc1c2f8b3602
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:8ffb6f0ff1bf016c3919c1eefc6c3db3d4a8fbe0e79919f72b4c281afe6d4b1c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:8760d6dc68c86f0c5b53d8f0bfada0df012f80bc8e0023ba3522fe6829a05a41
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:01b9f9798d3ecbefe3da1ac2ac782e8e8b930e7eeceb82fe3473b5bfc1ccd40a
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2ec62f3da93a1346ad215744257820377ccf4ead84e7471c52e348a31a431d3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:bdbe08ef2d01b114702780fab76905e6eb1851684692ffa0a4c84604bda3ddb3
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:364c107c34694ca33ac724de78db3e41ee55f9098d9ade0775a51b11f1533065
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8547b76cf0533dfe6da2aafbbd063021bad4ff78b32b2dfed4a7607496e0a904
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:762ad216e79be0837e98873202b0072c7fd399925ee94bde57cc61c68cbd5f4e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:34694b9f4fe7fffc0c547be4e311579b5bb0c1fa92d8257bf6d1581ffd655c46
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7ac504b7824d17bd7ac55b1e59c26be308506f315a0ab0d8c45d3b4934c0a16c
        version: 4.1.6
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.1
    - currentCSV: rhacs-operator.v4.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
          createdAt: "2024-03-13T14:51:55.084122+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.1.0 < 4.2.5'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:4253eca84cbb0c5fc033507ddc2deef235b054c0ba944fa271ebbba21545b872
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:4b35f5f39865238a8e95d3cba7ebb5f15ccfbf6a2563c34f57d91a0b5c927066
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:59f1862c8ce188906499f08201ae05deeb47bd4fc5eab4be0386ca44ec065da8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:12db245af3b8475c9b5680bb0035c566109819e05cfe62bd1dfdfb5d6ed25bea
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:e2b07cb7d690477d78a2b8e1875545998d4bdb946f80dd0533a14187da99dbbf
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:3858ab936d4626a11bbf895d41bad181364ee5df3312f6dd29ebeb9b7d31ec77
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:08aebe6e8275d3314b6756d522ec568ccdfb9c52df98c9822f187f6c5a3af19c
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:2f326cabf03f780a3e6120995d5442641764641f62430d82e05edbe21b30c7b2
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b5a15f2985b114bf9bc24a0d9942db5d0cb818e60fd15a2f9094aa00904cce77
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:ff27f03064633d6d08f8383649069e39c868def4aeca8f2622750397e049d0ec
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:b82f7f654235688fc07fb037422a73e1106e27024d39a68ef7fac1422b69496d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:e79a5e5681ebfe2d6349e2763ee0c1fa521b20319846958e402d23229b1f257b
        version: 4.2.5
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.2
    - currentCSV: rhacs-operator.v4.3.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
          createdAt: "2024-06-06T22:31:19.365878+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.2.0 < 4.3.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:30f9384244819da013358325cec364bb210b3b33b4cf9680698531fe50562aac
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:266cbbc270efafa26d5d15f3e333c96eaedee3b47854ff337ae1c572d34901ce
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5092c3fbb4a38c06a2f42d125dc718ef0e1005f304ad8046afdd2a5e7ccbc580
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:9fba889021a876ab627ff8c496e4b6c46c221c61142cd0c76daf33c89762e2c1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b0c9b3858fc0547dcb4eb3065dedd5bbcf1e1f6707a8ab812ed06882d4f6d6ad
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:10d0cc82b6eddd089f3a7af37b872c33248513d0b4d3f094b6f62c01f4460825
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:bdf9785bff51194636c6e8a518448c6d12a297bd31c4bb2e79d07084409c18b4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:b7eb17fbcf6c4a6e4205ad66f1ffd9b10a4634f6ef5045d407c266566b1f255d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c36a557c9dc2ae4479c432933a9a4dbef38f2be53360a68a5b6befb614d01dac
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:6b55a25eb3a611c648d9be883828834592229c16f4166fd28ef40a25fc667061
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:2bb643651c8d51525223df3b2b675947174bbe9f1eaea8688e1fa83044cfa41f
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:2c35cb6f518393d002edd2bf5f714c1d053cd0fab55f6152a2f35676e4979202
        version: 4.3.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.3
    - currentCSV: rhacs-operator.v4.4.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
          createdAt: "2025-02-10T19:50:13.597164+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.3.0 < 4.4.8'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:467d069ee957aea07882b83b60ae9fcc62ccca159f5d643aac13e531ccf7df88
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:2b87569d7f4ef5e06c989b273c378e007c6d87b26eec41774ac1fc67dcd9e4e4
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:65707fdad55767ff5598a8d3679fd0779a37f59501e58971b81bebcb23f1f9fc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:76f9dc2e2ee5595c36530d97dec216e2dbefada2fe0747e3f6a3a725b75722ab
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:0df4e95c7406aee7812d145cdd4d8c61cec32857fd2079dbc4244954a9d8288a
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:664a5dc40dc1fa06c47a11e13b0ab5e0612a981d7b69bcfb29422abf3d1a13b2
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:39cfd01236e905034f6645a73063427436c14e351e0e0bc735e73e62aa5a1c3c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:d6b4f8caf342b105c8df14c5c9180cb36123e08d34756ee8fe4f2f76c34f8064
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:ba6e25a413c79131b966a097f72c36478933ea52deb55a62acda7011a96b3983
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:69c12230a785619c765e3749a7d56f76dce9ce4f530c20e8a9387b99eb65627b
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:183c75920ab8e7cb947cb966f1543e39b6fdb5e1bb0caa1b0cf5f89671f0328f
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:123c0108bd7a190dd2accfd21786600f787147f040e57fb5f840861e2caa441d
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:763e7568704971cb4656727dc223374de13e82b4304dd3f5387ddc0f4e42b8b6
        version: 4.4.8
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.4
    - currentCSV: rhacs-operator.v4.5.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
          createdAt: "2025-04-09T22:51:15.157672+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.4.0 < 4.5.9'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Central is the configuration template for the central services.
              This includes the API server, persistent storage, and the web UI, as
              well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: SecuredCluster is the configuration template for the secured
              cluster services. These include Sensor, which is responsible for the
              connection to Central, and Collector, which performs host-level collection
              of process and network events.<p> **Important:** Please see the _Installation
              Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d928600f75461465598a10de8df2495a275fd5232a92b078f609d1f5774eb778
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:f07cba0738897e0e7ae74971be0470ce2a855657f133a6e1cdd8a7092f6151b0
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:cc4b329cc795752ecbb178d2e4efc152bde54b04181334eb5bd70304c3354cad
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:ed33e7396357f1db03a065573c020846a5872b12a21284439770686689222434
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:a7dfbe3590668f91fe6694335b5a35b2fca07b4735348cc5202921755d2f654d
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:0ff948b5517df3773cae2df02b708c348d8707409899b1de4108588e31601c0c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:5109b304eb9458f97af85c2ba63e08f51672fa3f9406226b7f357b86b5828dc0
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:dc7f23341c8b209f1846f29f889d00ef232fb58da9d5b4bd561f3933093b2786
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:5914548f6ce2160e92c5d96580ffba23cdf1ae940c1f73f3f6424d15c6e97c37
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:fde78d34ecd2944c7c16b4e823c76b3667bf6c5b468e9991128ecc4272ec0e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:41351e88688d23a443ad4f141caf0433b2c2789e6a1a85b4f5b36abeb5656bb1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:c2656b794ec4236e36f48f71bf26bdedd9de27fc4acef39deba2bdcfcddab03b
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:1817cf8d2c08b04391792321702ef53d16a4b1ca06459c00c9c053128f534358
        version: 4.5.9
      deprecation:
        message: |
          This version is no longer supported. Please switch to the `stable` channel or a channel for a version that is still supported.
      entries:
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.5
    - currentCSV: rhacs-operator.v4.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
          createdAt: "2025-09-26T09:24:28.046251+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.5.0 < 4.6.10'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:19db8c5d2425dcf18ae63b025f946fc46adaec0578e9505415cb585797069e95
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-slim-rhel8@sha256:16b61ca1b9ee4e0821e7fc82fe4def75ed22950e502730f5704ec4fe63bb2e59
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:899da9c23000bb821bc9d544ebb56bcda4f3245ceee5b1b9c12d09f580b4b9a1
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:713e61d3c8b55de1edc9d7834d3e8576718e1327665f94a6b58178fa7496c0cc
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:b5cd08349931669055f03d21331499ef2ea04597f8a90fb201615d43000271bc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:8fad8ab2efdcbc99d9efa736a90cca6f236f0a9d45fdb355ae0b6b4dc42d77f0
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:ffe41363dc00b05b0daa0c6d16efb30e534086aa4e3d7ccc599c86d33813138c
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:b630d53ad0b862151be7e14f0483ca5adac570216deae66f60e75af87a4dfd46
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:c4d434e1ff052508e5594d2b95dafa8726fd8adf5317754774c013a46de224d8
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:6d6b2696a12d49cf3403e338a0e028187473b6f43ef2b0bd72c7701200ac8e2e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:90c47c6f95571cfcca04aa4e05a18befad75c1b44a6f3a055fcdfdab749d29c5
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:6da59c5b82625894f6efa3fa0bc5ea0378edd4f95cf42dee772e9c8a565dba81
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:65793c05aec8085a53fab8bcf33a85192c17b46c91a7d86e0a457d1e3a113878
        version: 4.6.10
      entries:
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.6
    - currentCSV: rhacs-operator.v4.7.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
          createdAt: "2025-09-26T11:05:05.900682+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.6.0 < 4.7.7'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:98553e6f18d3c03a29212b4f7ea1d30a42bcf8b52ca12aecbcbc47769fc35b53
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:5b60d8271d6fa70df3c8f7e0ecd0d64feecfdf674dc84036500fe5f145c9415a
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:5ed9902b52830be93436d7977c10e73d346b6290532f6d7245cfe7cca9c93b31
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8882c00c9be9b579654a9ee8cd0c38537cf92d2f4bd8ebded1678aa0b6ae827
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:34ea97c6e9fc2d23427d0c6f2628075c0eaa0e151ca1c381539b039c249217f7
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:c7f894ab773d1d389656d13ed88fbf575faddd95c5b8396c6593ab3f993f331e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:def71e1ddea5b8c1e54c8582f39dc9ecca855b38aa9423690150aa6ef0ff189f
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:0c32318ee4cbd547e2c1fd64433cd23a629d1b06523d7c0f07ddb98c2baaa708
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:6bb454d64a9e339b1aeb81ca17b26f33db4a84f19ced4ad81858e40fd01878dc
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:bac79889d4717233e89bb6c8fae6732ed7bec5b9a55fd3b3de1f1311d870d2b1
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:7596529cbdce7118226b79d69e0667ab35ee692569e4014b74abdf0fa7415222
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:d8fcfe1267495801d23a0b94101c7b4d06b48346a949c97946ebcbffe148f4b6
        version: 4.7.7
      entries:
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.7
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: rhacs-4.8
    - currentCSV: rhacs-operator.v4.8.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "Central",
                "metadata": {
                  "name": "stackrox-central-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "central": {
                    "exposure": {
                      "route": {
                        "enabled": true
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "platform.stackrox.io/v1alpha1",
                "kind": "SecuredCluster",
                "metadata": {
                  "name": "stackrox-secured-cluster-services",
                  "namespace": "stackrox"
                },
                "spec": {
                  "clusterName": "my-cluster"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          containerImage: registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
          createdAt: "2025-09-12T11:06:11.361461+00:00"
          description: Red Hat Advanced Cluster Security (RHACS) operator provisions
            the services necessary to secure each of your OpenShift and Kubernetes
            clusters.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 4.7.0 < 4.8.4'
          operatorframework.io/suggested-namespace: rhacs-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Security"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              Central is the configuration template for the central services. This includes the API server, persistent storage,
              and the web UI, as well as the image scanner.
            displayName: Central
            kind: Central
            name: centrals.platform.stackrox.io
            version: v1alpha1
          - description: |-
              SecuredCluster is the configuration template for the secured cluster services. These include Sensor, which is
              responsible for the connection to Central, and Collector, which performs host-level collection of process and
              network events.<p>
              **Important:** Please see the _Installation Prerequisites_ on the main RHACS operator page before deploying, or
              consult the RHACS documentation on creating cluster init bundles.
            displayName: Secured Cluster
            kind: SecuredCluster
            name: securedclusters.platform.stackrox.io
            version: v1alpha1
          - description: SecurityPolicy is the schema for the policies API.
            displayName: Security Policy
            kind: SecurityPolicy
            name: securitypolicies.config.stackrox.io
            version: v1alpha1
        description: |
          ## Why use Red Hat Advanced Cluster Security for Kubernetes?

          Protecting cloud-native applications requires significant changes in how we approach security—we must apply controls earlier in the application development life cycle, use the infrastructure itself to apply controls, and keep up with increasingly rapid release schedules.


          Red Hat® Advanced Cluster Security for Kubernetes, powered by StackRox technology, protects your vital applications across build, deploy, and runtime. Our software deploys in your infrastructure and integrates with your DevOps tooling and workflows to deliver better security and compliance. The policy engine includes hundreds of built-in controls to enforce DevOps and security best practices, industry standards such as CIS Benchmarks and National Institute of Standards Technology (NIST) guidelines, configuration management of both containers and Kubernetes, and runtime security.

          Red Hat Advanced Cluster Security for Kubernetes provides a Kubernetes-native architecture for container security, enabling DevOps and InfoSec teams to operationalize security.

          ## Features and Benefits

          **Kubernetes-native security:**
          1. Increases protection.
          1. Eliminates blind spots, providing staff with insights into critical vulnerabilities and threat vectors.
          1. Reduces time and costs.
          1. Reduces the time and effort needed to implement security and streamlines security analysis, investigation, and remediation using the rich context Kubernetes provides.
          1. Increases scalability and portability.
          1. Provides scalability and resiliency native to Kubernetes, avoiding operational conflict and complexity that can result from out-of-band security controls.

          ## Using the RHACS Operator

          **RHACS Operator channels:**

          Red Hat provides the RHACS Operator by using the following update channels in the Red Hat Operator catalog:

          * `stable`: Provides the most recent version and patches to the most recent version.
            Using the `stable` channel and configuring automatic operator upgrades ensures that the most recent RHACS version is deployed.
          * `rhacs-x.yy` (for example, `rhacs-3.74`): Channels follow a specific RHACS version and include all patches to that version (e.g. `3.74.0`, `3.74.1`, ...).

          Note that the `latest` channel is deprecated and is not updated after RHACS version 3.74. Newer versions are published to the `stable` channel.

          **RHACS comes with two custom resources:**

          1. **Central Services** - Central is a deployment required on only one cluster in your environment. Users interact with RHACS via the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. Users may select exposures for Central that best meet their environment.

          2. **Secured Cluster Services** - Secured cluster services are placed on each cluster you manage and report back to Central. These services allow users to enforce policies and monitor your OpenShift and Kubernetes clusters. Secured Cluster Services come as two Deployments (Sensor and Admission Controller) and one DaemonSet (Collector).

          ### Central Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Central                          | Deployment      | Users interact with Red Hat Advanced Cluster Security through the user interface or APIs on Central. Central also sends notifications for violations and interacts with integrations. |
          | Central DB                       | Deployment      | Central DB is a PostgreSQL-based persistent storage for the data collected and managed by Central. |
          | Scanner                          | Deployment      | Scanner is a Red Hat developed and certified image scanner. Scanner analyzes and reports vulnerabilities for images. Scanner uses HPA to scale the number of replicas based on workload. |
          | Scanner DB                       | Deployment      | Scanner DB is a cache for vulnerability definitions to serve vulnerability scanning use cases throughout the software development life cycle. |

          ### Secured Cluster Services Explained

          | Service                          | Deployment Type | Description     |
          | :------------------------------- | :-------------- | :-------------- |
          | Sensor                           | Deployment      | Sensor analyzes and monitors Kubernetes in secured clusters. |
          | Collector                        | DaemonSet       | Analyzes and monitors container activity on Kubernetes nodes.|
          | Admission Controller             | Deployment      | ValidatingWebhookConfiguration for enforcing policies in the deploy lifecycle. |

          ### Central Custom Resource

          Central Services is the configuration template for RHACS Central deployment. For all customization options, please visit the RHACS documentation.

          ### SecuredCluster Custom Resource

          SecuredCluster is the configuration template for the RHACS Secured Cluster services.

          #### Installation Prerequisites

          Before deploying a SecuredCluster resource, you need to create a cluster init bundle secret.

          - **Through the RHACS UI:** To create a cluster init bundle secret through the RHACS UI, navigate to `Platform Configuration > Clusters`, and then click `Manage Tokens` in the top-right corner. Select `Cluster Init Bundle`, and click `Generate Bundle`. Select `Download Kubernetes secrets file`, and store the file under a name of your choice (for example, `cluster-init-secrets.yaml`).
          - **Through the `roxctl` CLI:** To create a cluster init bundle secret through the `roxctl` command-line interface, run `roxctl central init-bundles generate <name> --output-secrets <file name>`. Choose any `name` and `file name` that you like.

          Run `oc project` and check that it reports the correct namespace where you intend to deploy SecuredCluster. In case you want to install SecuredCluster to a different namespace, select it by running `oc project <namespace>`.
          Then, run `oc create -f init-bundle.yaml`. If you have chosen a name other than `init-bundle.yaml`, specify that file name instead.

          #### Required Fields

          The following attributes are required to be specified. For all customization options, please visit the RHACS documentation.

          | Parameter          | Description     |
          | :----------------- | :-------------- |
          | `clusterName`      | The name given to this secured cluster. The cluster will appear with this name in RHACS user interface. |
          | `centralEndpoint`  | This field should specify the address of the Central endpoint, including the port number. `centralEndpoint` may be omitted if this SecuredCluster Custom Resource is in the same cluster and namespace as Central. |
        displayName: Advanced Cluster Security for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - advanced-cluster-security
        - stackrox
        - security
        - vulnerabilities
        - compliance
        - devsecops
        - monitoring
        - scanning
        - runtime-security
        - network policy
        - configuration
        - risk
        links:
        - name: Red Hat Advanced Cluster Security Documentation
          url: https://docs.openshift.com/acs/welcome/
        - name: DataSheet
          url: https://www.redhat.com/en/resources/advanced-cluster-security-for-kubernetes-datasheet
        - name: Support Policy
          url: https://access.redhat.com/node/5822721
        - name: Community Site
          url: https://www.stackrox.io/
        maintainers:
        - email: rhacs-pm@redhat.com
          name: Advanced Cluster Security product team
        maturity: alpha
        minKubeVersion: 1.15.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-rhel8@sha256:d1758b76c32f03b5a1a44951d18fb2bb0c4fd0a99d24f8c42fe9e6b059574889
        - registry.redhat.io/advanced-cluster-security/rhacs-central-db-rhel8@sha256:e5c200f28187c1f6971562e41cae969183ee15f3d0feb076bb202b22ff67cbcb
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-rhel8@sha256:97df59280c12727a99fdec0c9bf56eb874e6a1cacd12d849709797d4be635ce9
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-db-slim-rhel8@sha256:92ff7691963e620dc5bf6da0a7694e6bcd68c2545afc1642a59a443ba2ab5611
        - registry.redhat.io/advanced-cluster-security/rhacs-operator-bundle@sha256:0c0a839b097ccbf758b084b85f8f9285a7e621fd44e8a62752be7b6bcae86669
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-v4-db-rhel8@sha256:b8964b1441251607b23f92918e9158724174013dc4af423988f56d6098dfb2d5
        - registry.redhat.io/advanced-cluster-security/rhacs-main-rhel8@sha256:adf47cb6615d344533fcca0aee41774f5e8df9b952b9b89a57c2be2b7e0fc03c
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-rhel8@sha256:e949830f3034c41aa5434a3654b2b1ccba7dac0373bf43d7c50a9ea43bb686fb
        - registry.redhat.io/advanced-cluster-security/rhacs-rhel8-operator@sha256:fe56f7733778f2ba9072ee35df562e82283eaeaa7bed26333d2fb916ecb76929
        - registry.redhat.io/advanced-cluster-security/rhacs-roxctl-rhel8@sha256:16912df2c26e5fc2e6051c0501170722856db274415063608de716cf53851038
        - registry.redhat.io/advanced-cluster-security/rhacs-scanner-slim-rhel8@sha256:db691065e46f604b980b0b202f6c2e14c8b204094d784c4a15db1e4fecb2965e
        - registry.redhat.io/advanced-cluster-security/rhacs-collector-rhel8@sha256:0c92d5b5508032d77559606927e45686179eaafc59e16d5496e088f2f212d258
        version: 4.8.4
      entries:
      - name: rhacs-operator.v4.8.4
        version: 4.8.4
      - name: rhacs-operator.v4.8.3
        version: 4.8.3
      - name: rhacs-operator.v4.8.2
        version: 4.8.2
      - name: rhacs-operator.v4.8.1
        version: 4.8.1
      - name: rhacs-operator.v4.8.0
        version: 4.8.0
      - name: rhacs-operator.v4.7.7
        version: 4.7.7
      - name: rhacs-operator.v4.7.6
        version: 4.7.6
      - name: rhacs-operator.v4.7.5
        version: 4.7.5
      - name: rhacs-operator.v4.7.4
        version: 4.7.4
      - name: rhacs-operator.v4.7.3
        version: 4.7.3
      - name: rhacs-operator.v4.7.2
        version: 4.7.2
      - name: rhacs-operator.v4.7.1
        version: 4.7.1
      - name: rhacs-operator.v4.7.0
        version: 4.7.0
      - name: rhacs-operator.v4.6.10
        version: 4.6.10
      - name: rhacs-operator.v4.6.9
        version: 4.6.9
      - name: rhacs-operator.v4.6.8
        version: 4.6.8
      - name: rhacs-operator.v4.6.7
        version: 4.6.7
      - name: rhacs-operator.v4.6.6
        version: 4.6.6
      - name: rhacs-operator.v4.6.5
        version: 4.6.5
      - name: rhacs-operator.v4.6.4
        version: 4.6.4
      - name: rhacs-operator.v4.6.3
        version: 4.6.3
      - name: rhacs-operator.v4.6.2
        version: 4.6.2
      - name: rhacs-operator.v4.6.1
        version: 4.6.1
      - name: rhacs-operator.v4.6.0
        version: 4.6.0
      - name: rhacs-operator.v4.5.9
        version: 4.5.9
      - name: rhacs-operator.v4.5.8
        version: 4.5.8
      - name: rhacs-operator.v4.5.7
        version: 4.5.7
      - name: rhacs-operator.v4.5.6
        version: 4.5.6
      - name: rhacs-operator.v4.5.5
        version: 4.5.5
      - name: rhacs-operator.v4.5.4
        version: 4.5.4
      - name: rhacs-operator.v4.5.3
        version: 4.5.3
      - name: rhacs-operator.v4.5.2
        version: 4.5.2
      - name: rhacs-operator.v4.5.1
        version: 4.5.1
      - name: rhacs-operator.v4.5.0
        version: 4.5.0
      - name: rhacs-operator.v4.4.8
        version: 4.4.8
      - name: rhacs-operator.v4.4.7
        version: 4.4.7
      - name: rhacs-operator.v4.4.6
        version: 4.4.6
      - name: rhacs-operator.v4.4.5
        version: 4.4.5
      - name: rhacs-operator.v4.4.4
        version: 4.4.4
      - name: rhacs-operator.v4.4.3
        version: 4.4.3
      - name: rhacs-operator.v4.4.2
        version: 4.4.2
      - name: rhacs-operator.v4.4.1
        version: 4.4.1
      - name: rhacs-operator.v4.4.0
        version: 4.4.0
      - name: rhacs-operator.v4.3.8
        version: 4.3.8
      - name: rhacs-operator.v4.3.7
        version: 4.3.7
      - name: rhacs-operator.v4.3.6
        version: 4.3.6
      - name: rhacs-operator.v4.3.5
        version: 4.3.5
      - name: rhacs-operator.v4.3.4
        version: 4.3.4
      - name: rhacs-operator.v4.3.3
        version: 4.3.3
      - name: rhacs-operator.v4.3.2
        version: 4.3.2
      - name: rhacs-operator.v4.3.1
        version: 4.3.1
      - name: rhacs-operator.v4.3.0
        version: 4.3.0
      - name: rhacs-operator.v4.2.5
        version: 4.2.5
      - name: rhacs-operator.v4.2.4
        version: 4.2.4
      - name: rhacs-operator.v4.2.3
        version: 4.2.3
      - name: rhacs-operator.v4.2.2
        version: 4.2.2
      - name: rhacs-operator.v4.2.1
        version: 4.2.1
      - name: rhacs-operator.v4.2.0
        version: 4.2.0
      - name: rhacs-operator.v4.1.6
        version: 4.1.6
      - name: rhacs-operator.v4.1.5
        version: 4.1.5
      - name: rhacs-operator.v4.1.4
        version: 4.1.4
      - name: rhacs-operator.v4.1.3
        version: 4.1.3
      - name: rhacs-operator.v4.1.2
        version: 4.1.2
      - name: rhacs-operator.v4.1.1
        version: 4.1.1
      - name: rhacs-operator.v4.1.0
        version: 4.1.0
      - name: rhacs-operator.v4.0.5
        version: 4.0.5
      - name: rhacs-operator.v4.0.4
        version: 4.0.4
      - name: rhacs-operator.v4.0.3
        version: 4.0.3
      - name: rhacs-operator.v4.0.2
        version: 4.0.2
      - name: rhacs-operator.v4.0.1
        version: 4.0.1
      - name: rhacs-operator.v4.0.0
        version: 4.0.0
      name: stable
    defaultChannel: stable
    packageName: rhacs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision-rhmp.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/model-builder-for-vision-certified-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.connect.redhat.com/ibm/modelbuilder-bundle-rhm@sha256:5d14e0c44ffb543b7cf59817fe587c8e819542d8edfa261a44e58aea0ecdb79b
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3658954f199040b0f244945c94955f794ee68008657421002e1b32962e7c30fc
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision-rhmp.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified-rhmp
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-keyspaces-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-keyspaces-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Keyspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "keyspaces.services.k8s.aws/v1alpha1",
                "kind": "Table",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
          createdAt: "2025-09-20T01:23:26Z"
          description: AWS Keyspaces controller is a service controller for managing
            Keyspaces resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Keyspace represents the state of an AWS keyspaces Keyspace
              resource.
            displayName: Keyspace
            kind: Keyspace
            name: keyspaces.keyspaces.services.k8s.aws
            version: v1alpha1
          - description: Table represents the state of an AWS keyspaces Table resource.
            displayName: Table
            kind: Table
            name: tables.keyspaces.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Keyspaces resources in AWS from within your Kubernetes cluster.

          **About Amazon Keyspaces**

          Amazon Keyspaces (for Apache Cassandra) is a scalable, highly available, and managed Apache Cassandra–compatible database service. With Amazon Keyspaces, you don’t have to provision, patch, or manage servers, and you don’t have to install, maintain, or operate software.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Keyspaces
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - keyspaces
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Keyspaces Developer Resources
          url: https://aws.amazon.com/keyspaces/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: keyspaces maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/keyspaces-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-keyspaces-controller@sha256:6e56400162867a4f55a70c340aeeb074726fd3dca13b6da60c395d6f35a48b48
        version: 1.1.1
      entries:
      - name: ack-keyspaces-controller.v1.1.1
        version: 1.1.1
      - name: ack-keyspaces-controller.v1.1.0
        version: 1.1.0
      - name: ack-keyspaces-controller.v1.0.14
        version: 1.0.14
      - name: ack-keyspaces-controller.v1.0.12
        version: 1.0.12
      - name: ack-keyspaces-controller.v1.0.11
        version: 1.0.11
      - name: ack-keyspaces-controller.v1.0.10
        version: 1.0.10
      - name: ack-keyspaces-controller.v1.0.9
        version: 1.0.9
      - name: ack-keyspaces-controller.v1.0.7
        version: 1.0.7
      - name: ack-keyspaces-controller.v1.0.6
        version: 1.0.6
      - name: ack-keyspaces-controller.v1.0.5
        version: 1.0.5
      - name: ack-keyspaces-controller.v1.0.3
        version: 1.0.3
      - name: ack-keyspaces-controller.v1.0.2
        version: 1.0.2
      - name: ack-keyspaces-controller.v1.0.1
        version: 1.0.1
      - name: ack-keyspaces-controller.v1.0.0
        version: 1.0.0
      - name: ack-keyspaces-controller.v0.0.16
        version: 0.0.16
      - name: ack-keyspaces-controller.v0.0.15
        version: 0.0.15
      - name: ack-keyspaces-controller.v0.0.14
        version: 0.0.14
      - name: ack-keyspaces-controller.v0.0.13
        version: 0.0.13
      - name: ack-keyspaces-controller.v0.0.12
        version: 0.0.12
      - name: ack-keyspaces-controller.v0.0.11
        version: 0.0.11
      - name: ack-keyspaces-controller.v0.0.10
        version: 0.0.10
      - name: ack-keyspaces-controller.v0.0.9
        version: 0.0.9
      - name: ack-keyspaces-controller.v0.0.8
        version: 0.0.8
      - name: ack-keyspaces-controller.v0.0.7
        version: 0.0.7
      - name: ack-keyspaces-controller.v0.0.6
        version: 0.0.6
      - name: ack-keyspaces-controller.v0.0.5
        version: 0.0.5
      - name: ack-keyspaces-controller.v0.0.4
        version: 0.0.4
      - name: ack-keyspaces-controller.v0.0.3
        version: 0.0.3
      - name: ack-keyspaces-controller.v0.0.2
        version: 0.0.2
      - name: ack-keyspaces-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-keyspaces-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Automotive SIG CentOS
      provider-url: ""
    name: automotive-infra
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: automotive-infra.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ContainerBuild",
                "metadata": {
                  "name": "cb-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "ImageBuild",
                "metadata": {
                  "name": "imagebuild-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "automotive.sig.centos.org/v1",
                "kind": "OperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "automotive-infra"
                  },
                  "name": "operatorconfig-sample",
                  "namespace": "automotive-infra-operator"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
          createdAt: "2025-03-31T14:06:33Z"
          description: Streamline CI/CD for RHIVOS and its ecosystem using OpenShift
            Pipelines
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: automotive-infra-operator
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
          support: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Client
            name: clients.jumpstarter.dev
            version: v1alpha1
          - kind: ContainerBuild
            name: containerbuilds.automotive.sig.centos.org
            version: v1
          - kind: ExporterAccessPolicy
            name: exporteraccesspolicies.jumpstarter.dev
            version: v1alpha1
          - kind: Exporter
            name: exporters.jumpstarter.dev
            version: v1alpha1
          - kind: ImageBuild
            name: imagebuilds.automotive.sig.centos.org
            version: v1
          - kind: Lease
            name: leases.jumpstarter.dev
            version: v1alpha1
          - kind: OperatorConfig
            name: operatorconfigs.automotive.sig.centos.org
            version: v1
        description: This operator is designed to streamline CI/CD for RHIVOS and
          its ecosystem using OpenShift Pipelines
        displayName: Automotive-CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Automotive
        - RHIVOS
        - Pipelines
        - CI
        links:
        - name: Automotive infra Operator
          url: https://gitlab.com/fedora/sigs/robotics/infra/infra-ci-operator
        maintainers:
        - email: automotive-sig@lists.centos.org
          name: Automotive SIG CentOS Community
        maturity: alpha
        provider:
          name: Automotive SIG CentOS
        relatedImages:
        - quay.io/fedora-sig-robotics/infra-ci-operator:v0.0.10
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/automotive-infra@sha256:6d4180ff6899d4de8ba2cd4a06ece234bbd8f0ec1bd0b4b0ae4b7a0c047121da
        version: 0.0.10
      entries:
      - name: automotive-infra.v0.0.10
        version: 0.0.10
      - name: automotive-infra.v0.0.9
        version: 0.0.9
      - name: automotive-infra.v0.0.8
        version: 0.0.8
      - name: automotive-infra.v0.0.7
        version: 0.0.7
      - name: automotive-infra.v0.0.6
        version: 0.0.6
      - name: automotive-infra.v0.0.5
        version: 0.0.5
      - name: automotive-infra.v0.0.4
        version: 0.0.4
      - name: automotive-infra.v0.0.3
        version: 0.0.3
      - name: automotive-infra.v0.0.2
        version: 0.0.2
      - name: automotive-infra.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: automotive-infra
    provider:
      name: Automotive SIG CentOS
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"aws-efs.managed.openshift.io/v1alpha1","kind":"SharedVolume","metadata":{"name":"sv1"},"spec":{"accessPointID":
            "fsap-0123456789abcdef", "fileSystemID": "fs-0123cdef"}}]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/app-sre/aws-efs-operator:9c87267
          createdAt: "2021-12-20T19:51:50Z"
          description: Manage read-write-many access to AWS EFS volumes in an OpenShift
            cluster.
          repository: https://github.com/openshift/aws-efs-operator
          support: https://github.com/openshift/aws-efs-operator/issues/new
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Shared Volume
            displayName: SharedVolume
            kind: SharedVolume
            name: sharedvolumes.aws-efs.managed.openshift.io
            version: v1alpha1
        description: |-
          ## About the managed application
          AWS provides Elastic File System, an NFS-based shared storage solution. A couple of mechanisms exist to enable mounting such a file system in containers. Curating these options and making them work in a cluster is nontrivial.
          ## About this Operator
          The aws-efs-operator minimizes the barrier to entry for container use of EFS by:
          - Installing the [AWS EFS CSI Driver](https://github.com/kubernetes-sigs/aws-efs-csi-driver) and supporting infrastructure.
          - Reducing the identification of a shared storage resource to a simple Custom Resource Definition with only two fields.
          - Creating the necessary PersistentVolume and PersistentVolumeClaim resources that can simply be referenced from a Pod specification.

          More information can be found in the operator's [source repository](https://github.com/openshift/aws-efs-operator).
          ## Prerequisites for enabling this Operator
          An OpenShift cluster backed by AWS.
        displayName: AWS EFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AWS
        - EFS
        - NFS
        - Shared storage
        links:
        - name: README
          url: https://github.com/openshift/aws-efs-operator/blob/master/README.md
        - name: Design
          url: https://github.com/openshift/aws-efs-operator/blob/master/DESIGN.md
        maintainers:
        - email: support@redhat.com
          name: Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/app-sre/aws-efs-operator:9c87267
        - quay.io/openshift-community-operators/aws-efs-operator@sha256:6372847df81faddc23c938928157c3ac280e537a4eec9f7edd599480260bc566
        version: 0.0.8
      entries:
      - name: aws-efs-operator.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: aws-efs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Performance
      provider-url: ""
    name: cluster-impairment-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-impairment-operator.v1.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "all-impairments"
                },
                "spec": {
                  "duration": 480,
                  "egress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "ingress": {
                    "bandwidth": 10000,
                    "latency": 50,
                    "loss": 0.02
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 30,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "duplication"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "duplication": 25
                  },
                  "ingress": {
                    "duplication": 25
                  },
                  "interfaces": [
                    "ens3"
                  ],
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "two-min-flap"
                },
                "spec": {
                  "duration": 480,
                  "interfaces": [
                    "ens2f0"
                  ],
                  "linkFlapping": {
                    "downTime": 120,
                    "enable": true,
                    "upTime": 120
                  },
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "typical-scenario"
                },
                "spec": {
                  "duration": 30,
                  "egress": {
                    "latency": 50
                  },
                  "ingress": {
                    "latency": 50
                  },
                  "interfaces": [
                    "ens2f0",
                    "eno1"
                  ],
                  "loss": 0.02,
                  "startDelay": 5
                }
              },
              {
                "apiVersion": "apps.redhat.com/v1alpha1",
                "kind": "ClusterImpairment",
                "metadata": {
                  "name": "uneven-latency"
                },
                "spec": {
                  "duration": 60,
                  "egress": {
                    "latency": 100
                  },
                  "ingress": {
                    "latency": 10
                  },
                  "interfaces": [
                    "ens2f0"
                  ],
                  "nodeSelector": {
                    "key": "node-role.kubernetes.io/worker",
                    "value": ""
                  },
                  "startDelay": 5
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Developer Tools
          containerImage: quay.io/redhat-performance/cluster-impairment-operator:1.0.4
          createdAt: "2021-08-11 12:15:00"
          description: Run Impairments on Kubernetes/Openshift Nodes
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/redhat-performance/cluster-impairment-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterImpairment
            name: clusterimpairments.apps.redhat.com
            version: v1alpha1
        description: "cluster-impairment-operator is an operator designed to automate
          the application of impairments to the nodes in a cluster.\n\n## Features\n\nTraffic
          Direction:\n\n* Egress impairments affect the traffic going out.    \n*
          Ingress impairments affect the traffic coming in. (Ingress impairments require
          the kernel module IFB.)\n\n### Impairments\n\n| Impairment    | Description
          \                               | Unit    | Uses Netem | Correlation Supported
          |\n|---------------|--------------------------------------------|---------|------------|-----------------------|\n|
          Bandwidth     | The bandwidth limit                        | kbit/s  | Yes
          \       | No                    |\n| Latency       | The delay of the packets
          \                  | ms      | Yes        | Yes                   |\n| Packet
          Loss   | The percent of packets that are dropped    | percent | Yes        |
          Yes                   |\n| Corruption    | The percent of packets that are
          corrupted  | percent | Yes        | Yes                   |\n| Duplication
          \  | The percent of packets that are duplicated | percent | Yes        |
          Yes                   |\n| Link Flapping | Turns the interface on and off
          \            | bool    | No         | No                    |\n\n\nOn the
          tested environment (RHEL CoreOS 48.84), the impairments can be used alongside
          link flapping.\n\n\n#### Latency Options\n\nIn addition to simply delaying
          the packets, there are advanced latency options.\n\n##### Jitter\n\nSemi-randomly
          adds or subtracts from the latency according to the distribution up to the
          amount specified.\nIf latency is 100ms, and jitter is 10ms, the actual latency
          will vary 100±10ms.\n\n\n##### Distribution\n\nThe distribution of the jitter.
          The options are:\n* Normal\n* Uniform\n* Pareto\n* Paretonormal\n\n#####
          Jitter correlation\n\nThe percent chance that the next latency's jitter
          value will correlate with the preceeding latency.\n\n##### Reorder\n\nThe
          percentage of packets that are not delayed, causing a reorder due to them
          being sent before the delayed ones.\nApplying jitter itself has the potential
          to also cause reordering.\n\n##### Reorder correlation\n\nThe percet chance
          that the value for the next reorder will correlate with the preceeding value.\n\n####
          Loss options\n\n##### Correlation\n\nThe percent chance that the previous
          loss value for a packet correlates with the loss value for the next packet.\n\n####
          Corruption options\n\n###### Corruption\n\nThe percent chance that each
          packet will be corrupted.\n\n##### Corruption correlation\n\nThe percet
          chance that the value for the next corruption will correlate with the preceeding
          value.\n\n#### Duplication options\n\n##### Duplication\n\nThe percent chance
          that each packet will be duplicated.\n\n##### Duplication correlation\n\nThe
          percet chance that the value for the next duplication will correlate with
          the preceeding value.\n\n\n\n## Configuration\n\nHere is an example of the
          ClusterImpairment custom resource.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: test-impairment-cr\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds. It typically takes about 2-3 seconds
          for the Daemonset to run\n  interfaces:\n  - \"ens2f0\"\n  ingress: # uses
          ifb\n    bandwidth: 0 # kbit\n    latency: 10 # ms\n    latencyOptions:\n
          \     jitter: 5 # ms\n      jitterCorrelation: 25 # percent\n      distribution:
          normal\n      reorder: 25 # percent of packets that will skip the delay\n
          \     reorderCorrelation: 25 # percent\n    loss: 0 # percent\n    lossOptions:\n
          \     correlation: 25 # percent\n    corruption: 0.1 # percent\n    corruptionOptions:\n
          \     correlation: 25 # percent\n    duplication: 1 # percent\n    duplicationOptions:\n
          \     correlation: 25 # percent\n  egress:\n    bandwidth: 0 # kbit\n    latency:
          100 # ms\n    loss: 0 # percent\n  linkFlapping:\n    enable: false\n    downTime:
          3 # Seconds\n    upTime: 3 # Seconds\n  nodeSelector:\n    key: \"node-role.kubernetes.io/worker\"\n
          \   value: \"\"\n    \n```\n\n#### Interfaces\n\nThe interfaces configuration
          option is a list of all interfaces that the impairments should be applied
          to. The current implementation will error out once it hits an invalid interface.\n\nIf
          an invalid interface is found, it will print out the list of interfaces.\n\n####
          Node Selector\n\nThe default node selector is all worker nodes, but you
          can change it to whatever node selector you want by setting the key and
          value.\nThere is a toleration to allow it to run on master nodes if the
          node selector includes master nodes.\n\n#### Duration\n\nThe number of seconds
          the impairments should run. It will try to sync the start and end time between
          all pods.\nIf set to -1, it will run indefinitely (a year), until you remove
          the ClusterImpairment custom resource. This is good for testing that requires
          steady impairments.\n\nIf the script is link flapping, set this value to
          be short enough so that if the link flap interrupts communication between
          the nodes, the daemonset will remove itself.\n\n#### Start Delay\n\nThe
          delay before starting the script. If you want the pods to be in sync, a
          minimum of a few seconds should be used because the pods take time to start
          up.\n\nYou can also utilize this feature to run an separate impairment after
          another. Just apply two resources at the same time, but add the duration
          and start delay of the first to the start delay of the second.\n\n#### Ingress
          and Egress\n\nThe configuration sections \"ingress\" and \"egress\" apply
          to each direction.\n\n##### Examples:\n\n##### Example 1\n\n\nIn this example,
          egress latency is set to 100ms, but the ingress latency is set to 10ms.
          So the latency to the interface will end up being 10ms, but 100ms going
          out. When pinging, this will result in 110ms of latency.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: uneven-latency\nspec:\n
          \ duration: 60\n  startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  ingress:\n
          \   latency: 10 # ms\n  egress:\n    latency: 100 # ms\n\n```\n\n##### Example
          2\n\n\nIn this example, link flapping will be enabled, and it will turn
          the interface on and off every 2 minutes. That is long enough for kubernetes
          to determine that a node is unavailable.\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: two-min-flap\nspec:\n  duration: 480\n
          \ startDelay: 5\n  interfaces:\n  - \"ens2f0\"\n  linkFlapping:\n    enable:
          true\n    downTime: 120 # Seconds\n    upTime: 120 # Seconds\n\n```\n\n#####
          Example 3\n\n\nIn this example, a realistic set of impairments is applied
          to `ens2f0` and `eno1` for 30 seconds:\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind:
          ClusterImpairment\nmetadata:\n  name: typical-scenario\nspec:\n  duration:
          30 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  - \"ens2f0\"\n
          \ - \"eno1\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \ loss: 0.02 # percent\n\n```\n\n##### Example 4\n\n\nIn this example, a
          combination of latency, packet loss, bandwidth, and link flapping is applied.\n\n```\n\napiVersion:
          apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n  name: all-impairments\nspec:\n
          \ duration: 480 # seconds\n  startDelay: 5 # seconds\n  interfaces:\n  -
          \"ens2f0\"\n  egress:\n    latency: 50 # ms. Bidirectional, so total of
          100ms\n    loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about
          1 mbit/s\n  ingress:\n    latency: 50 # ms. Bidirectional, so total of 100ms\n
          \   loss: 0.02 # percent\n    bandwidth: 1000 # 1000 kbit/s, about 1 mbit/s\n
          \ linkFlapping:\n    enable: true\n    downTime: 30 # Seconds\n    upTime:
          120 # Seconds\n\n```\n\n##### Example 5\n\n\nIn this example, packet duplicaiton
          is applied.\n\n\n```\n\napiVersion: apps.redhat.com/v1alpha1\nkind: ClusterImpairment\nmetadata:\n
          \ name: duplication\nspec:\n  duration: 60\n  startDelay: 5\n  interfaces:\n
          \ - \"ens3\"\n  ingress:\n    duplication: 25 # percent\n  egress:\n    duplication:
          25 # percent\n\n```\n\n## Setup\n\n### Requirements\n\n1. You need `make`
          installed.\n2. You need access to the kubernetes cluster with a kubeconfig.\n\n###
          Installation\n\nTo run using the current latest image:\n1. Clone the repository.\n2.
          Run `make deploy` with KUBECONFIG environment variable set to the location
          of the kubeconfig.\n\nTo run with your own code, there are more steps.\n1.
          Fork the repository.\n2. Clone to a machine that has access to the Kubernetes
          cluster and the kubeconfig.\n3. Modify the makefile to change the `IMG`
          variable to your image repository. If you do not have podman installed,
          also change podman to docker.\n4. Run `make docker-build` then `make docker-push`.\n5.
          Then run `make deploy`\n\n### Deploying from operator-hub\n\n*Important*
          You must manually apply the serviceaccount!\n\nFirst, Follow the instructions
          on this site to deploy from here.\n\nNext, apply the manifest \"serviceaccount.yaml\"
          from the manifests folder.\n\nThen you can apply your ClusterImpairment
          resources.\n\n## Running impairments\n\nFirst, configure a ClusterImpairment
          type's spec with the information for the impairment you want to run. Most
          importantly, set the interface(s). You can list the interfaces with `ifconfig`.
          If an invalid interface is set, the pod's logs will show `ifconfig` output.\n\n\nOnce
          the clusterimpairment type is set, apply it and it will work.\n\n\nTo re-run
          an impairment, delete the impairment with `oc delete -f path/to/manifest.yaml`,
          then apply it again.\n\n\n## Limitations\n\n\n### Multiple Impairments\n\n\nYou
          should avoid any configurations that apply impairments to the same interface
          on the same node. There are potential conflicts. The worker pod will attempt
          to remove all impairments on that interface before applying new ones.\n\n\nInstead,
          take advantage of the full control of both ingress and egress impairments
          from within the same ClusterImpairment resource.\n\n\nThere is currently
          a configured limit of 20 ClusterImpairment resources working at the same
          time. If that is a limitation for your use case, you can change it in `config/manager/manager.yaml`
          or you can try to find a way to make each ClusterImpairment do more.\n\n###
          Traffic Control (TC)\n\nTraffic control is how cluster-impairment-operator
          applies the latency, bandwidth, and packet loss impairments. The limitation
          is due to Linux not being realtime operating system, so the impairment will
          not be perfectly consistent.\n\n### Link Flapping\n\nWhen link flapping,
          if you flap the link that Kubernetes uses to communicate with the pods,
          you may be unable to remove the pod until the link is on long enough for
          Kubernetes to communicate with the impaired node.\n\n\nIn this case, it
          is helpful to set the duration properly instead of running for an indefinite
          or large amount of time, because the node will properly go back to the unimpaired
          state at that time.\n\n\nIf the cluster becomes offline due to the link
          flapping when you do not want it to be offline, soft restarting the nodes
          after removing the custom resource should remove all impairments.\n"
        displayName: cluster-impairment-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - impairment
        - kubernetes
        - openshift
        - tc
        - trafficcontrol
        links:
        - name: Cluster Impairment Operator
          url: https://github.com/redhat-performance/cluster-impairment-operator
        maintainers:
        - email: joconnel@redhat.com
          name: Jared O'Connell
        - email: akrzos@redhat.com
          name: Alex Krzos
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat Performance
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/cluster-impairment-operator@sha256:6fd114eb8f525f670e505373c2999a4e66b53beda988e99b267204d2fe0e1b4f
        - quay.io/redhat-performance/cluster-impairment-operator:1.0.4
        version: 1.0.4
      entries:
      - name: cluster-impairment-operator.v1.0.4
        version: 1.0.4
      - name: cluster-impairment-operator.v1.0.3
        version: 1.0.3
      - name: cluster-impairment-operator.v1.0.2
        version: 1.0.2
      - name: cluster-impairment-operator.v1.0.1
        version: 1.0.1
      name: beta
    defaultChannel: beta
    packageName: cluster-impairment-operator
    provider:
      name: Red Hat Performance
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: keepalived-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keepalived-operator.v1.5.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KeepalivedGroup",
                "metadata": {
                  "name": "keepalivedgroup-workers"
                },
                "spec": {
                  "blacklistRouterIDs": [
                    1,
                    2
                  ],
                  "image": "registry.redhat.io/openshift4/ose-keepalived-ipfailover",
                  "interface": "ens3",
                  "nodeSelector": {
                    "node-role.kubernetes.io/loadbalancer": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
          createdAt: "2023-03-11T18:48:22Z"
          description: This operator provides a simple way to create self-hosted load
            balancers using keepalived.
          operatorframework.io/suggested-namespace: keepalived-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/keepalived-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KeepalivedGroup
            name: keepalivedgroups.redhatcop.redhat.io
            version: v1alpha1
        description: "The objective of the keepalived operator is to allow for a way
          to create self-hosted load balancers in an automated way. From a user experience
          point of view the behavior is the same as of when creating [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services with a cloud provider able to manage them.\n\nThe keepalived operator
          can be used in all environments that allows nodes to advertise additional
          IPs on their NICs (and at least for now, in networks that allow multicast),
          however it's mainly aimed at supporting LoadBalancer services and ExternalIPs
          on bare metal installations (or other installation environments where a
          cloud provider is not available).\n\nOne possible use of the keepalived
          operator is also to support [OpenShift Ingresses](https://docs.openshift.com/container-platform/4.5/networking/configuring_ingress_cluster_traffic/overview-traffic.html)
          in environments where an external load balancer cannot be provisioned.\n\n##
          How it works\n\nThe keepalived operator will create one or more [VIPs](https://en.wikipedia.org/wiki/Virtual_IP_address)
          (an HA IP that floats between multiple nodes), based on the [`LoadBalancer`](https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer)
          services and/or services requesting [`ExternalIPs`](https://kubernetes.io/docs/concepts/services-networking/service/#external-ips).\n\nFor
          `LoadBalancer` services the IPs found at `.Status.LoadBalancer.Ingress[].IP`
          will become VIPs.\n\nFor services requesting a `ExternalIPs`, the IPs found
          at `.Spec.ExternalIPs[]` will become VIPs.\n\nNote that a service can be
          of `LoadBalancer` type and also request `ExternalIPs`, it this case both
          sets of IPs will become VIPs.\n\nDue to a [keepalived](https://www.keepalived.org/manpage.html)
          limitation a single keepalived cluster can manage up to 256 VIP configurations.
          Multiple keepalived clusters can coexists in the same network as long as
          they use different multicast ports [TODO verify this statement].\n\nTo address
          this limitation the `KeepalivedGroup` [CRD](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/)
          has been introduced. This CRD is supposed to be configured by an administrator
          and allows you to specify a node selector to pick on which nodes the keepalived
          pods should be deployed. Here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  image:
          registry.redhat.io/openshift4/ose-keepalived-ipfailover\n  interface: ens3\n
          \ nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n  blacklistRouterIDs:\n
          \ - 1\n  - 2  \n```\n\nThis KeepalivedGroup will be deployed on all the
          nodes with role `loadbalancer`. One must also specify the network device
          on which the VIPs will be exposed, it is assumed that all the nodes have
          the same network device configuration.\n\nServices must be annotated to
          opt-in to being observed by the keepalived operator and to specify which
          KeepalivedGroup they refer to. The annotation looks like this:\n\n`keepalived-operator.redhat-cop.io/keepalivedgroup:
          <keepalivedgroup namespace>/<keepalivedgroup-name>`\n\nThe image used for
          the keepalived containers can be specified with `.Spec.Image` it will default
          to `registry.redhat.io/openshift4/ose-keepalived-ipfailover` if undefined.
          \n\n## Requirements\n\n### Security Context Constraints\n\nEach KeepalivedGroup
          deploys a [daemonset](https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/)
          that requires the [privileged scc](https://docs.openshift.com/container-platform/4.5/authentication/managing-security-context-constraints.html),
          this permission must be given to the `default` service account in the namespace
          where the keepalived group is created by and administrator.\n\n```shell\noc
          adm policy add-scc-to-user privileged -z default -n keepalived-operator\n```\n\n###
          Cluster Network Operator\n\nIn Openshift, use of an external IP address
          is governed by the following fields in the `Network.config.openshift.io`
          CR named `cluster`\n\n* `spec.externalIP.autoAssignCIDRs` defines an IP
          address block used by the load balancer when choosing an external IP address
          for the service. OpenShift supports only a single IP address block for automatic
          assignment.\n\n* `spec.externalIP.policy` defines the permissible IP address
          blocks when manually specifying an IP address. OpenShift does not apply
          policy rules to IP address blocks defined by `spec.externalIP.autoAssignCIDRs`\n\nThe
          following patch can be used to configure the Cluster Network Operator:\n\n```yaml\nspec:\n
          \ externalIP:\n    policy:\n      allowedCIDRs:\n      - ${ALLOWED_CIDR}\n
          \   autoAssignCIDRs:\n      - \"${AUTOASSIGNED_CIDR}\"\n```\n\nHere is an
          example of how to apply the patch:\n\n```shell\nexport ALLOWED_CIDR=\"192.168.131.128/26\"\nexport
          AUTOASSIGNED_CIDR=\"192.168.131.192/26\"\noc patch network cluster -p \"$(envsubst
          < ./network-patch.yaml | yq r -j -)\" --type=merge\n```\nAdditionally, the
          fields can be edited manually via `oc edit Network.config.openshift.io cluster`\n\n##
          Blacklisting router IDs\n\nIf the Keepalived pods are deployed on nodes
          which are in the same network (same broadcast domain to be precise) with
          other keepalived the process, it's necessary to ensure that there is no
          collision between the used routers it.\nFor this purpose it is possible
          to provide a `blacklistRouterIDs` field with a list of black-listed IDs
          that will not be used.\n\n## OpenShift RHV, vSphere, OSP and bare metal
          IPI instructions\n\nWhen IPI is used for RHV, vSphere, OSP or bare metal
          platforms, three keepalived VIPs are deployed. To make sure that keepalived-operator
          can work in these environment we need to discover and blacklist the corresponding
          VRRP router IDs.\n\nTo discover the VRRP router IDs being used, run the
          following command, you can run this command from you laptop:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids <cluster_name>\n```\n\nIf
          you don't know your cluster name, run this command:\n\n```shell\npodman
          run quay.io/openshift/origin-baremetal-runtimecfg:4.5 vr-ids $(oc get cm
          cluster-config-v1 -n kube-system -o jsonpath='{.data.install-config}'| yq
          -r .metadata.name)\n```\n\nThen use these [instructions](#Blacklisting-router-IDs)
          to blacklist those VRRP router IDs.\n\n## Verbatim Configurations\n\nKeepalived
          has dozens of [configurations](https://www.keepalived.org/manpage.html).
          At the early stage of this project it's difficult to tell which one should
          be modeled in the API. Yet, users of this project may still need to use
          them. To account for that there is a way to pass verbatim options both at
          the keepalived group level (which maps to the keepalived config `global_defs`
          section) and at the service level (which maps to the keepalived config `vrrp_instance`
          section).\n\nKeepalivedGroup-level verbatim configurations can be passed
          as in the following example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          KeepalivedGroup\nmetadata:\n  name: keepalivedgroup-router\nspec:\n  interface:
          ens3\n  nodeSelector:\n    node-role.kubernetes.io/loadbalancer: \"\"\n
          \ verbatimConfig:  \n    vrrp_iptables: my-keepalived\n```\n\nthis will
          map to the following `global_defs`:\n\n```\n    global_defs {\n        router_id
          keepalivedgroup-router\n        vrrp_iptables my-keepalived\n    }\n```\n\nService-level
          verbatim configurations can be passed as in the following example:\n\n```yaml\napiVersion:
          v1\nkind: Service\nmetadata:\n  annotations:\n    keepalived-operator.redhat-cop.io/keepalivedgroup:
          keepalived-operator/keepalivedgroup-router\n    keepalived-operator.redhat-cop.io/verbatimconfig:
          '{ \"track_src_ip\": \"\" }'\n```\n\nthis will map to the following `vrrp_instance`
          section\n\n```\n    vrrp_instance openshift-ingress/router-default {\n        interface
          ens3\n        virtual_router_id 1  \n        virtual_ipaddress {\n          192.168.131.129\n
          \       }\n        track_src_ip\n    }\n```\n\n## Metrics collection\n\nEach
          keepalived pod exposes a [Prometheus](https://prometheus.io/) metrics port
          at `9650`. Metrics are collected with [keepalived_exporter](github.com/gen2brain/keepalived_exporter),
          the available metrics are described in the project documentation.\n\nWhen
          a keepalived group is created a [`PodMonitor`](https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#podmonitor)
          rule to collect those metrics. All PodMonitor resources created that way
          have the label: `metrics: keepalived`. It is up to you to make sure your
          Prometheus instance watches for those `PodMonitor` rules. Here is an example
          of a fragment of a `Prometheus` CR configured to collect the keepalived
          pod metrics:\n\n```yaml\n  podMonitorSelector:\n    matchLabels:\n      metrics:
          keepalived\n```\n"
        displayName: Keepalived Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadbalancer
        - keepalived
        - ipfailover
        - externalIP
        - self-hosted
        links:
        - name: Keepalived Operator
          url: https://github.com/redhat-cop/keepalived-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/keepalived-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/keepalived-operator@sha256:5a46fcb72f02a1f0d9cbc72f44948ef4b09149a4b3fe2e45fd31df90d20b3c56
        - quay.io/redhat-cop/keepalived-operator@sha256:cd0c2f7df82df1178795e4a18a0942da33adb9670b192a7b7a2b1483a5b4ccf8
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.5.4
      entries:
      - name: keepalived-operator.v1.5.4
        version: 1.5.4
      - name: keepalived-operator.v1.5.3
        version: 1.5.3
      - name: keepalived-operator.v1.5.2
        version: 1.5.2
      - name: keepalived-operator.v1.5.1
        version: 1.5.1
      - name: keepalived-operator.v1.5.0
        version: 1.5.0
      - name: keepalived-operator.v1.4.2
        version: 1.4.2
      - name: keepalived-operator.v1.4.1
        version: 1.4.1
      - name: keepalived-operator.v1.4.0
        version: 1.4.0
      - name: keepalived-operator.v1.3.4
        version: 1.3.4
      - name: keepalived-operator.v1.3.3
        version: 1.3.3
      - name: keepalived-operator.v1.3.2
        version: 1.3.2
      - name: keepalived-operator.v1.3.1
        version: 1.3.1
      - name: keepalived-operator.v1.3.0
        version: 1.3.0
      - name: keepalived-operator.v1.2.2
        version: 1.2.2
      - name: keepalived-operator.v1.2.1
        version: 1.2.1
      - name: keepalived-operator.v1.2.0
        version: 1.2.0
      - name: keepalived-operator.v1.1.0
        version: 1.1.0
      - name: keepalived-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: keepalived-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: marin3r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        - quay.io/3scale/marin3r:v0.13.1
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: alpha
    - currentCSV: marin3r.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyConfig",
                "metadata": {
                  "name": "envoyconfig-example",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "nodeID": "example",
                  "resources": [
                    {
                      "type": "cluster",
                      "value": {
                        "connect_timeout": "0.01s",
                        "dns_lookup_family": "V4_ONLY",
                        "lb_policy": "ROUND_ROBIN",
                        "load_assignment": {
                          "cluster_name": "example",
                          "endpoints": [
                            {
                              "lb_endpoints": [
                                {
                                  "endpoint": {
                                    "address": {
                                      "socket_address": {
                                        "address": "example",
                                        "port_value": 8080
                                      }
                                    }
                                  }
                                }
                              ]
                            }
                          ]
                        },
                        "name": "example",
                        "type": "STRICT_DNS"
                      }
                    },
                    {
                      "type": "route",
                      "value": {
                        "name": "local",
                        "virtual_hosts": [
                          {
                            "domains": [
                              "*"
                            ],
                            "name": "all",
                            "routes": [
                              {
                                "match": {
                                  "prefix": "/"
                                },
                                "route": {
                                  "cluster": "example"
                                }
                              }
                            ]
                          }
                        ]
                      }
                    },
                    {
                      "type": "listener",
                      "value": {
                        "address": {
                          "socket_address": {
                            "address": "0.0.0.0",
                            "port_value": 8443
                          }
                        },
                        "filter_chains": [
                          {
                            "filters": [
                              {
                                "name": "envoy.filters.network.http_connection_manager",
                                "typed_config": {
                                  "@type": "type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager",
                                  "access_log": [
                                    {
                                      "name": "envoy.access_loggers.file",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.access_loggers.file.v3.FileAccessLog",
                                        "path": "/dev/stdout"
                                      }
                                    }
                                  ],
                                  "http_filters": [
                                    {
                                      "name": "envoy.filters.http.router",
                                      "typed_config": {
                                        "@type": "type.googleapis.com/envoy.extensions.filters.http.router.v3.Router"
                                      }
                                    }
                                  ],
                                  "rds": {
                                    "config_source": {
                                      "ads": {},
                                      "resource_api_version": "V3"
                                    },
                                    "route_config_name": "local"
                                  },
                                  "stat_prefix": "https"
                                }
                              }
                            ],
                            "transport_socket": {
                              "name": "envoy.transport_sockets.tls",
                              "typed_config": {
                                "@type": "type.googleapis.com/envoy.extensions.transport_sockets.tls.v3.DownstreamTlsContext",
                                "common_tls_context": {
                                  "tls_certificate_sds_secret_configs": [
                                    {
                                      "name": "example.default.svc",
                                      "sds_config": {
                                        "ads": {},
                                        "resource_api_version": "V3"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          }
                        ],
                        "name": "https"
                      }
                    },
                    {
                      "blueprint": "tlsCertificate",
                      "generateFromTlsSecret": "example.default.svc",
                      "type": "secret"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "DiscoveryService",
                "metadata": {
                  "name": "discoveryservice-example",
                  "namespace": "my-namespace"
                },
                "spec": {}
              },
              {
                "apiVersion": "operator.marin3r.3scale.net/v1alpha1",
                "kind": "EnvoyDeployment",
                "metadata": {
                  "name": "envoydeployment-sample",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "discoveryServiceRef": "discoveryservice-example",
                  "envoyConfigRef": "envoyconfig-example",
                  "ports": [
                    {
                      "name": "https",
                      "port": 8443
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking
          certified: "false"
          containerImage: quay.io/3scale/marin3r
          createdAt: "2024-08-05T16:08:25Z"
          description: Lighweight, CRD based Envoy control plane for Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["envoyconfigrevisions.marin3r.3scale.net","discoveryservicecertificates.operator.marin3r.3scale.net"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/3scale-ops/marin3r
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DiscoveryServiceCertificate is an internal resource used
              to create certificates. This resource is used by the DiscoveryService
              controller to create the required certificates for the different components.
              Direct use of DiscoveryServiceCertificate objects is discouraged.
            displayName: DiscoveryServiceCertificate
            kind: DiscoveryServiceCertificate
            name: discoveryservicecertificates.operator.marin3r.3scale.net
            version: v1alpha1
          - description: DiscoveryService represents an envoy discovery service server.
              Only one instance per namespace is currently supported.
            displayName: DiscoveryService
            kind: DiscoveryService
            name: discoveryservices.operator.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfigRevision is an internal resource that stores a
              specific version of an EnvoyConfig resource. EnvoyConfigRevisions are
              automatically created and deleted by the EnvoyConfig controller and
              are not intended to be directly used. Use EnvoyConfig objects instead.
            displayName: EnvoyConfigRevision
            kind: EnvoyConfigRevision
            name: envoyconfigrevisions.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyConfig holds the configuration for a given envoy nodeID.
              The spec of an EnvoyConfig object holds the Envoy resources that conform
              the desired configuration for the given nodeID and that the discovery
              service will send to any envoy client that identifies itself with that
              nodeID.
            displayName: EnvoyConfig
            kind: EnvoyConfig
            name: envoyconfigs.marin3r.3scale.net
            version: v1alpha1
          - description: EnvoyDeployment is a resource to deploy and manage a Kubernetes
              Deployment of Envoy Pods.
            displayName: EnvoyDeployment
            kind: EnvoyDeployment
            name: envoydeployments.operator.marin3r.3scale.net
            version: v1alpha1
        description: |
          MARIN3R implements a control plane to deploy, configure and operate a fleet of envoy instances within a Kubernetes cluster.
          It operates a group of gateways and Pod sidecar containers to provide general purpose functionality required to operate production systems.

          Features:

          * Deploy and manage an Envoy xDS server using the DiscoveryService custom resource.
          * Inject Envoy sidecar containers based on Pod annotations.
          * Deploy Envoy as a Kubernetes Deployment using the EnvoyDeployment custom resource.
          * Dynamic Envoy configuration and re-configuration using the EnvoyConfig custom resource.
          * Use Kubernetes Secrets as certificate sources.
          * Syntactic validation of Envoy configurations.
          * Lifecycle management of Envoy containers (graceful shutdown and connection draining).
          * Self-healing.

          Check the [docs](https://github.com/3scale-ops/marin3r#readme) for more information.

          ## License
          MARIN3R is licensed under the [Apache 2.0 license](https://github.com/3scale/prometheus-exporter-operator/blob/master/LICENSE)
        displayName: MARIN3R
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - envoy
        - discovery
        - controlplane
        - proxy
        - sidecar
        - xds
        - network
        - gateway
        links:
        - name: GitHub
          url: https://github.com/3scale-ops/marin3r
        maintainers:
        - email: 3scale-operations+marin3r@redhat.com
          name: 3scale Operations Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/marin3r:v0.13.1
        - quay.io/community-operator-pipeline-prod/marin3r@sha256:68875b0da5165c5fc0d39b5c079a3706b57879f444dba77f4ed04f4db4f1c664
        version: 0.13.1
      entries:
      - name: marin3r.v0.13.1
        version: 0.13.1
      - name: marin3r.v0.12.3
        version: 0.12.3
      - name: marin3r.v0.11.1
        version: 0.11.1
      - name: marin3r.v0.10.0
        version: 0.10.0
      name: stable
    defaultChannel: stable
    packageName: marin3r
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OnGres
      provider-url: https://ongres.com
    name: stackgres-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: candidate
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      - name: stackgres-community.v1.17.0-rc1
        version: 1.17.0-rc1
      name: fast
    - currentCSV: stackgres-community.v1.17.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGCluster",
                "metadata": {
                  "name": "demo-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "instances": 3,
                  "pods": {
                    "persistentVolume": {
                      "size": "20Gi"
                    }
                  },
                  "postgres": {
                    "version": "latest"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGConfig",
                "metadata": {
                  "name": "stackgres-operator"
                },
                "spec": {
                  "authentication": {
                    "type": "jwt",
                    "user": "admin"
                  },
                  "containerRegistry": "quay.io",
                  "extensions": {
                    "repositoryUrls": [
                      "https://extensions.stackgres.io/postgres/repository"
                    ]
                  },
                  "grafana": {
                    "autoEmbed": false,
                    "datasourceName": "Prometheus",
                    "password": "prom-operator",
                    "schema": "http",
                    "user": "admin"
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "prometheus": {
                    "allowAutobind": true
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgCluster": "demo-db"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGDistributedLogs",
                "metadata": {
                  "name": "distributedlogs",
                  "namespace": "demo-db"
                },
                "spec": {
                  "persistentVolume": {
                    "size": "20Gi"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGInstanceProfile",
                "metadata": {
                  "name": "size-m",
                  "namespace": "demo-db"
                },
                "spec": {
                  "cpu": "4",
                  "memory": "8Gi"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPoolingConfig",
                "metadata": {
                  "name": "poolconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "pgBouncer": {
                    "pgbouncer.ini": {
                      "databases": {
                        "demo": {
                          "dbname": "demo",
                          "pool_size": 400,
                          "reserve_pool": 5
                        },
                        "postgres": {
                          "dbname": "postgres",
                          "pool_size": 10,
                          "reserve_pool": 5
                        }
                      },
                      "pgbouncer": {
                        "default_pool_size": "100",
                        "max_client_conn": "2000",
                        "pool_mode": "session"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGPostgresConfig",
                "metadata": {
                  "name": "pgconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "postgresVersion": "15",
                  "postgresql.conf": {
                    "effective_cache_size": "5GB",
                    "hot_standby_feedback": "on",
                    "log_min_duration_statement": "1000",
                    "maintenance_work_mem": "2GB",
                    "max_connections": "600",
                    "shared_buffers": "3GB",
                    "work_mem": "16MB"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "create-db-script",
                  "namespace": "demo-db"
                },
                "spec": {
                  "continueOnError": false,
                  "managedVersions": true,
                  "scripts": [
                    {
                      "name": "create-demo-database",
                      "script": "CREATE DATABASE demo WITH OWNER postgres;\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGScript",
                "metadata": {
                  "name": "stream-db",
                  "namespace": "demo-db"
                },
                "spec": {
                  "source": {
                    "sgCluster": {
                      "name": "demo-db"
                    },
                    "type": "SGCluster"
                  },
                  "target": {
                    "cloudEvent": {
                      "http": {
                        "url": "http://cloudevents-nodejs"
                      }
                    },
                    "type": "CloudEvent"
                  }
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedBackup",
                "metadata": {
                  "name": "demo-backup",
                  "namespace": "demo-db"
                },
                "spec": {
                  "managedLifecycle": false,
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1",
                "kind": "SGShardedDbOps",
                "metadata": {
                  "name": "demo-restart-op",
                  "namespace": "demo-db"
                },
                "spec": {
                  "op": "restart",
                  "sgShardedCluster": "demo-shardeddb"
                }
              },
              {
                "apiVersion": "stackgres.io/v1alpha1",
                "kind": "SGShardedCluster",
                "metadata": {
                  "name": "demo-shardeddb",
                  "namespace": "demo-db"
                },
                "spec": {
                  "coordinator": {
                    "instances": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "database": "sharded",
                  "postgres": {
                    "version": "15.3"
                  },
                  "shards": {
                    "clusters": 3,
                    "instancesPerCluster": 2,
                    "pods": {
                      "persistentVolume": {
                        "size": "10Gi"
                      }
                    }
                  },
                  "type": "citus"
                }
              },
              {
                "apiVersion": "stackgres.io/v1beta1",
                "kind": "SGObjectStorage",
                "metadata": {
                  "name": "backupconfig",
                  "namespace": "demo-db"
                },
                "spec": {
                  "gcs": {
                    "bucket": "stackgres-backups",
                    "gcpCredentials": {
                      "secretKeySelectors": {
                        "serviceAccountJSON": {
                          "key": "gcloudkey",
                          "name": "backups-gcp"
                        }
                      }
                    }
                  },
                  "type": "gcs"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
          createdAt: "2025-08-31T22:20:22Z"
          description: |
            The most advanced Postgres Enterprise Platform.
            Fully Open Source.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: 1.33.999
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: quarkus.javaoperatorsdk.io/v1-alpha
          repository: https://gitlab.com/ongresinc/stackgres
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Handle to a performed (or to be performed, if run manually)
              backup
            displayName: StackGres Backup
            kind: SGBackup
            name: sgbackups.stackgres.io
            version: v1
          - description: Main CRD, manages Postgres clusters (one or more Postgres
              pods)
            displayName: StackGres Cluster
            kind: SGCluster
            name: sgclusters.stackgres.io
            version: v1
          - description: Operator configuration for OLM-based installations (equivalent
              to values.yaml with Helm)
            displayName: StackGres Operator Configuration
            kind: SGConfig
            name: sgconfigs.stackgres.io
            version: v1
          - description: Day 2 Operations, including upgrades, restarts, vacuum, repack,
              etc
            displayName: StackGres Database Operation
            kind: SGDbOps
            name: sgdbops.stackgres.io
            version: v1
          - description: Multi-tenant logs server, to aggregate Postgres logs. Fully
              managed
            displayName: StackGres Distributed Logs
            kind: SGDistributedLogs
            name: sgdistributedlogs.stackgres.io
            version: v1
          - description: Instance Profiles are like "t-shirt" sizes, used for pods
              sizing
            displayName: StackGres Instance Profile
            kind: SGInstanceProfile
            name: sginstanceprofiles.stackgres.io
            version: v1
          - description: Handle to an existing Object Storage (e.g. S3), used to store
              backups
            displayName: StackGres Object Storage
            kind: SGObjectStorage
            name: sgobjectstorages.stackgres.io
            version: v1beta1
          - description: Strongly validated customized Postgres configuration (postgresql.conf)
            displayName: StackGres Postgres Configuration
            kind: SGPostgresConfig
            name: sgpgconfigs.stackgres.io
            version: v1
          - description: Customized PgBouncer (connection pooler) configuration
            displayName: StackGres Connection Pooling Configuration
            kind: SGPoolingConfig
            name: sgpoolconfigs.stackgres.io
            version: v1
          - description: Managed SQL Scripts, used for initial SQL commands or migrations
            displayName: StackGres Script
            kind: SGScript
            name: sgscripts.stackgres.io
            version: v1
          - description: Handle to a performed (or to be performed, if run manually)
              backup for sharded clusters
            displayName: StackGres Sharded Backup
            kind: SGShardedBackup
            name: sgshardedbackups.stackgres.io
            version: v1
          - description: Manages Postgres sharded clusters (two or more SGClusters)
            displayName: StackGres Sharded Cluster
            kind: SGShardedCluster
            name: sgshardedclusters.stackgres.io
            version: v1alpha1
          - description: Day 2 Operations for sharded clusters, including restarts,
              resharding, etc
            displayName: StackGres Sharded DbOps
            kind: SGShardedDbOps
            name: sgshardeddbops.stackgres.io
            version: v1
          - description: Stream Change Data Captured (CDC) events
            displayName: StackGres Stream
            kind: SGStream
            name: sgstreams.stackgres.io
            version: v1alpha1
        description: |2

          [StackGres](https://stackgres.io) is the **Stack** required for enterprise production Post**Gres**. A fully-featured platform to run Postgres on Kubernetes. Fully Open Source, StackGres supports both a declarative approach suitable for GitOps workflows and a complete Web Console for the best user experience.

          Built by [OnGres](https://ongres.com) ("**On** Post**Gres**"), StackGres requires little to no prior Postgres experience. StackGres can perform fully automated deployments; fully automated database operations ("Day 2 operations") and comes with advanced database tuning by default. Yet remaining highly customizable for Postgres expert DBAs.

          [StackGres features](https://stackgres.io/features/) include, among others:

          * **High Availability with automated failover**. StackGres relies on [Patroni](https://github.com/zalando/patroni), and its built-in and fully automatic.
          * **Integrated connection pooling**. Built-in, by default, like it should be for production workloads.
          * **Automatic backups with lifecycle policies**. Backup your clusters automatically to any object store. Apply retention policies. Restoration supports PITR.
          * **Advanced replication modes**, including async, sync and group replication. It also supports cascading replication and standby clusters on separate Kubernetes clusters for disaster recovery.
          * **More than 150 Postgres extensions**. The Postgres platform with [the largest number of extensions in the world](https://stackgres.io/extensions/). With new extensions added continuously.
          * **Observability**. Fully integrated with the Prometheus stack. Includes pre-defined, Postgres-specific dashboards and alerts.
          * **Fully-featured Web Console**. Perform any operation from the Web Console. Supports SSO, fine-grained RBAC and a REST API.
          * **Distributed Logs**. StackGres developed a mechanism to ship logs from all pods to a central log server, managed by StackGres, that store logs in Postgres. Query your logs with SQL or from the Web Console!
          * **Automated Day 2 Operations**. Minor and major version upgrades, container upgrades, controlled restart, vacuum, repack, even benchmarks!
          * **Expertly tuned by default**. From the creators of [CONF](https://postgresqlco.nf), StackGres pre-tunes your Postgres servers with more than 40 parameters tuned by default.
          * **100% Open Source**. No "premium version with advanced features", no production usage restrictions. Just Open Source.
          * **[24/7 Support](https://stackgres.io/pricing/) Available from OnGres**

          ## Installation and documentation

          Installation:
          * For a quick test, you can follow our [quickstart](https://stackgres.io/doc/latest/demo/quickstart/).
          * [Production installations](https://stackgres.io/doc/latest/install/).

          All the documentation is available at [stackgres.io/doc](https://stackgres.io/doc/latest/install/).

          Join the [Slack](https://slack.stackgres.io) and/or [Discord](https://discord.stackgres.io) Public Communities for Community support.
        displayName: StackGres
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - rdbms
        - open source
        - ongres
        links:
        - name: StackGres Web
          url: https://stackgres.io
        - name: StackGres Docs
          url: https://stackgres.io/doc
        maintainers:
        - email: stackgres@ongres.com
          name: OnGres
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: OnGres
          url: https://ongres.com
        relatedImages:
        - quay.io/stackgres/restapi@sha256:3f44d544bc812214d5fafe12b94a2cb7c2f32cbb0da53b8cd543ae92046caef5
        - quay.io/community-operator-pipeline-prod/stackgres-community@sha256:71c65b0d16bf87e51645e5c4041623dbf7a4090318a7ae70553c54b1d27f44c4
        - quay.io/ongres/kubectl@sha256:3bb27ccffb13242f3af4ca995f87ac1950d3daf7e3f6a7eb302a0adb35741b3a
        - quay.io/stackgres/admin-ui@sha256:5f90cf8c186a1aa0e3a74fc55cd8f353fd4db926c36e5dfef5dc61052ef7cfbc
        - quay.io/stackgres/jobs@sha256:f18b1c1b8c3351ae8e5240e71d8e551ea5b5e7e0ab6b7f2586a265478604a86c
        - quay.io/stackgres/operator@sha256:b43c9aa82fda32ccfdef11284600fbfbb01925ba6186768dba0ca8f9c24b087b
        version: 1.17.2
      entries:
      - name: stackgres-community.v1.17.2
        version: 1.17.2
      - name: stackgres-community.v1.17.1
        version: 1.17.1
      - name: stackgres-community.v1.17.0
        version: 1.17.0
      name: stable
    defaultChannel: stable
    packageName: stackgres-community
    provider:
      name: OnGres
      url: https://ongres.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
          createdAt: "2025-03-12T11:19:10Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b94833f55644c289fe2c1297b68a4f9931baf977587c6a2c4c7a3e484c704f64
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:731f92c146f299bc3662b309a3068bde18fad51f65a7d29063e582b61e07042f
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:081d7f4f20332599fdbbd07f22cc18eca6c5625b619efeae6e861239cde4b12a
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:0bde3c1a0b667a693f443688ed3926eef1705c45ecf4cce3a3dedbddca7a2784
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:c34676932ddd14db213dbf42a1d341f51692f3f47f1b751d06187e292c99a170
        version: 6.1.8
      entries:
      - name: loki-operator.v6.1.8
        version: 6.1.8
      - name: loki-operator.v6.1.7
        version: 6.1.7
      - name: loki-operator.v6.1.6
        version: 6.1.6
      - name: loki-operator.v6.1.5
        version: 6.1.5
      - name: loki-operator.v6.1.4
        version: 6.1.4
      - name: loki-operator.v6.1.3
        version: 6.1.3
      - name: loki-operator.v6.1.2
        version: 6.1.2
      - name: loki-operator.v6.1.1
        version: 6.1.1
      - name: loki-operator.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: loki-operator.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
          createdAt: "2025-02-18T18:01:39Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b8c5b46a94a6c85c9d382eeeb65a1d2b4014a7dc765e9abfba706c92857462dc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:99d1e7c5c1f0609b33de576884782550f969b28dc622810e6661d4c60a664215
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:cf7ccc1a60d6787ebcd1d76415cca4cc55c2c3ee18d52c0511e66ca6bf9310f8
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f401be018c5ba35e8ea1e68831cc20067f0f4d1be031d05b222548c1bdfb5f7e
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:8955d8186ae98095623f2083b6d5389d3ac9f2612fcfcf695835219c3e1216c4
        version: 6.2.5
      entries:
      - name: loki-operator.v6.2.5
        version: 6.2.5
      - name: loki-operator.v6.2.4
        version: 6.2.4
      - name: loki-operator.v6.2.3
        version: 6.2.3
      - name: loki-operator.v6.2.2
        version: 6.2.2
      - name: loki-operator.v6.2.1
        version: 6.2.1
      - name: loki-operator.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: loki-operator.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
          createdAt: "2025-06-11T13:21:40Z"
          description: |
            The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
            ## Prerequisites and Requirements
            ### Loki Operator Namespace
            The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
            ### Memory Considerations
            Loki is a memory intensive application.  The initial
            set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
            to the OCP cluster if you desire to run with the recommended (or better) memory.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-operators-redhat
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: loki-operator
          operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
          operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io.test.config.v1: tests/scorecard/
          operators.operatorframework.io.test.mediatype.v1: scorecard+v1
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Cluster Logging
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |
          The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
          ## Prerequisites and Requirements
          ### Loki Operator Namespace
          The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
          ### Memory Considerations
          Loki is a memory intensive application.  The initial
          set of OCP nodes may not be large enough to support the Loki cluster.  Additional OCP nodes must be added
          to the OCP cluster if you desire to run with the recommended (or better) memory.
        displayName: Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Loki Operator
          url: https://github.com/grafana/loki
        maintainers:
        - email: team-logging@redhat.com
          name: Red Hat, AOS Logging
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:b311677919183c6c91c6da40b0ddb2529f92ed52b2c46a07e399a4bad5dea280
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
        - registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:fce287241be64b7b2c20a971c4f2d468c079ae81556aaab756572ebad27a17f5
        - registry.redhat.io/openshift-logging/loki-operator-bundle@sha256:ff98532bd8f21ecf375062d465f2aee4cdcd70b647d258c1e09f7e2286f998c9
        - registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:28657e7bc56be5c79c315c5556de0dbb6363e0c07aff9091cd5173be42479c9a
        - registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:ea2d995744cd4a4bd6e7467c2d6927bfc89a47d551ef0d6b048c0c57652f81be
        version: 6.3.1
      entries:
      - name: loki-operator.v6.3.1
        version: 6.3.1
      - name: loki-operator.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: loki-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: quay-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: quay-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
          createdAt: 2021-11-02 11:11 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-10@sha256:b2695e9af379624621bef9ec785f0cb618bfdabae98c5f17eea82ff5f8d2d5cb
        - registry.redhat.io/rhel8/redis-5@sha256:8d7e66927ce4b3050ca15ab4df4dbaeb593cb855e16fb0a6f5dde950e9db7db9
        - registry.redhat.io/quay/clair-rhel8@sha256:af5e2fd1543365b365b1a6cce00b86e44fe5e66cde6491e9b113324c21f629c8
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:0e6bef51caf479074925115fa11621c41bcada375cc1925d0511a73084b93543
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:2a471300f70d3995ab7e9fcc3df35c5513e73fcc1d140fb4d651a1c3eae949eb
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b793aef88881aa75d47a22377d8932efc0b497bb01b2f3f376d64891d1dace54
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:d73321dfa6dcced943405fae45c2ccf841663e0ab57e8be8cc1ddba251b7ce08
        - registry.redhat.io/quay/quay-rhel8@sha256:ed8c4249d376b4226e35bf088ffe5bf688bf68b059718c5e218aa1184fe048d0
        version: 3.4.7
      entries:
      - name: quay-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: quay-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
          createdAt: 2021-09-09 20:09 UTC
          description: Opinionated deployment of Red Hat Quay on Kubernetes.
          olm.skipRange: '>=3.4.x <3.5.7'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.5.7
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:1f5e4cc9ec2d9b8b89d82f1a698acfa1ac2764ca4ebc5c85b3417da7730b5867
        - registry.redhat.io/quay/quay-rhel8@sha256:c599892160ac20c744d833603d9375923af7aec1edfd86982b7513e02fb6d463
        - registry.redhat.io/rhel8/postgresql-10@sha256:a723bb3a7a5046527b9a5b628ddfb211c526e3814cc4242262c4f8a016f38edd
        - registry.redhat.io/rhel8/redis-5@sha256:7916760ff03af5ca230df5e96b2bf353fbf1829386de1c063c05f623837be7f7
        - registry.redhat.io/quay/clair-rhel8@sha256:d242bd9d37f95b2a5a8267b3f1eba02c671d46121712baae43466c86d1da89f9
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5d72577ccfd63c6d9ad712b376b604c40d2d3ecb53e391e2ee0d18caa6e0225a
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7fe408a1e428b2d0660cb578c1dcd4b961f838e37166fe7b21b2bed70d288594
        - registry.redhat.io/quay/quay-operator-bundle@sha256:722cf3c05eb5663a737cb1385b971b6d1f0c8b2595d5ac425f7b100c25e2fd26
        version: 3.5.7
      entries:
      - name: quay-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: quay-operator.v3.10.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.10.15
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:3c0201e3cf88e711c921870324002a02cbd42c0d2eecafb4f242aa0601e24602
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/clair-rhel8@sha256:efa6e69500ea643042b544e6205f088214d79ab09a36864a612e73d07979d1c5
        - registry.redhat.io/quay/quay-rhel8@sha256:fc20b8bdac88c1bfce9aec993bdce6dd2cc132f8a9d4062d174b098819c74013
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:774cd1f181527787ba8631c7c14437eb2250c15ddb4a9c75fdfa35c5c1a19a2e
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e720c6a332ec4032e8576c31421af404f61d9baaf8e88f69f8333c52e5d1683
        - registry.redhat.io/quay/quay-operator-bundle@sha256:ed97b11fa0f0f7aadf193810a5a599bf26015a0af8ee6fd90061110dcb0d50f4
        version: 3.10.15
      entries:
      - name: quay-operator.v3.10.15
        version: 3.10.15
      - name: quay-operator.v3.10.14
        version: 3.10.14
      - name: quay-operator.v3.10.13
        version: 3.10.13
      - name: quay-operator.v3.10.12
        version: 3.10.12
      - name: quay-operator.v3.10.11
        version: 3.10.11
      - name: quay-operator.v3.10.10
        version: 3.10.10
      - name: quay-operator.v3.10.9
        version: 3.10.9
      - name: quay-operator.v3.10.8
        version: 3.10.8
      - name: quay-operator.v3.10.7
        version: 3.10.7
      - name: quay-operator.v3.10.6
        version: 3.10.6
      - name: quay-operator.v3.10.5
        version: 3.10.5
      - name: quay-operator.v3.10.4
        version: 3.10.4
      - name: quay-operator.v3.10.3
        version: 3.10.3
      - name: quay-operator.v3.10.2
        version: 3.10.2
      - name: quay-operator.v3.10.1
        version: 3.10.1
      - name: quay-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: quay-operator.v3.11.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.11.13
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:35d86936c539007a78a89869a14b87b76788fe117df442a670f63936b978ed5b
        - registry.redhat.io/quay/clair-rhel8@sha256:3746677ef1cb0ddd4f237ab598b14cf0bc76dcce080ffde127b5bba87ffe6693
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f19ae051041c46fe7587f903a9c6221bcbe8be2c4cc0235739f7592eb93ae164
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:cb3a41caa3a572b7981d424be2a32b8e881127cc03b557ee0a36ab13f430b897
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:1a083cd72502609427428ed22fb60846613ddb41a4613496c5cddc797b382d3a
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:11b8c9d64c42f0f0c7b2df491bba859d0410666a8ef1d2fd41eb77e4ad9d2dc8
        - registry.redhat.io/quay/quay-rhel8@sha256:7230fa1d4970e877cb6de8716e99a9bd43302c0ff23b4501f2aa806a8d496d9b
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        version: 3.11.13
      entries:
      - name: quay-operator.v3.11.13
        version: 3.11.13
      - name: quay-operator.v3.11.12
        version: 3.11.12
      - name: quay-operator.v3.11.11
        version: 3.11.11
      - name: quay-operator.v3.11.10
        version: 3.11.10
      - name: quay-operator.v3.11.9
        version: 3.11.9
      - name: quay-operator.v3.11.8
        version: 3.11.8
      - name: quay-operator.v3.11.7
        version: 3.11.7
      - name: quay-operator.v3.11.6
        version: 3.11.6
      - name: quay-operator.v3.11.5
        version: 3.11.5
      - name: quay-operator.v3.11.4
        version: 3.11.4
      - name: quay-operator.v3.11.3
        version: 3.11.3
      - name: quay-operator.v3.11.2
        version: 3.11.2
      - name: quay-operator.v3.11.1
        version: 3.11.1
      - name: quay-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: quay-operator.v3.12.12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.12.12
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:28e37ec35de1307405264cd4f79aaa6ce4d4ecbd6cbc27f4fef5db415c678d64
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:8cdd677449f3c723dad79bcd9c4191c8917d6eb0beeeeeda68fe19a0988bb2a8
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:accb705b6327fb2dcc1ce657a99f61e343eca752c9873d63d89fea89c10c3187
        - registry.redhat.io/quay/quay-rhel8@sha256:ccbd83e05aea0bb96bf099f63420175b36ed413bfc8a618f4c90b6a2abc53895
        - registry.redhat.io/quay/clair-rhel8@sha256:dc3d5fb35f2bd5d00480351dca705fe260a225d9a97958cf0276801af08d8307
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:8e5dd171e6e94de0137bb6a972533b1f93b56efc1731e19b80869d5e9b8a26a6
        version: 3.12.12
      entries:
      - name: quay-operator.v3.12.12
        version: 3.12.12
      - name: quay-operator.v3.12.11
        version: 3.12.11
      - name: quay-operator.v3.12.10
        version: 3.12.10
      - name: quay-operator.v3.12.9
        version: 3.12.9
      - name: quay-operator.v3.12.8
        version: 3.12.8
      - name: quay-operator.v3.12.7
        version: 3.12.7
      - name: quay-operator.v3.12.6
        version: 3.12.6
      - name: quay-operator.v3.12.5
        version: 3.12.5
      - name: quay-operator.v3.12.4
        version: 3.12.4
      - name: quay-operator.v3.12.3
        version: 3.12.3
      - name: quay-operator.v3.12.2
        version: 3.12.2
      - name: quay-operator.v3.12.1
        version: 3.12.1
      - name: quay-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: quay-operator.v3.13.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.13.8
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:069ccf36346cb14f4e4ea54de1d59afb72548912b140cfef67a6bcb7534d25e9
        - registry.redhat.io/rhel8/redis-6@sha256:966a3dafa2dd31ac1a39e826e08c1b988fe694189fe79449bbe14a48b2631213
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:31a9095d8f291ce68e8e688877692e74a17143bdda8d998fe3a07bcee32575dd
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:f62cbd6632454c227b7cbb01dadc8e501e8288b2d8e0abf8e7842b806fa3195c
        - registry.redhat.io/quay/quay-operator-bundle@sha256:4a004584fa6c7187481e8c8480f17e0e4edd18c32f52a2e271c68d02f1a9655f
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-15@sha256:a5f36e7893a519c12752c6118233c92d4e83618670bc9ea7bd3a8a34a63bd50b
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:5481a4b914ce4aed6ba3ceee8432eddab6e264cad2caef528404795f904100eb
        - registry.redhat.io/rhel8/postgresql-13@sha256:f17fd8dd4d56d8e916cb2ad3f403d05ad259f79fd6918d67a38a6e9e4a8896cf
        - registry.redhat.io/quay/quay-rhel8@sha256:8887c17547c7a4be61471523eade9820e5095cf966b1885a3758081da559871c
        version: 3.13.8
      entries:
      - name: quay-operator.v3.13.8
        version: 3.13.8
      - name: quay-operator.v3.13.7
        version: 3.13.7
      - name: quay-operator.v3.13.6
        version: 3.13.6
      - name: quay-operator.v3.13.5
        version: 3.13.5
      - name: quay-operator.v3.13.4
        version: 3.13.4
      - name: quay-operator.v3.13.3
        version: 3.13.3
      - name: quay-operator.v3.13.2
        version: 3.13.2
      - name: quay-operator.v3.13.1
        version: 3.13.1
      - name: quay-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: quay-operator.v3.14.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.14.5
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:43ef8adddbf9b007b6e2396f53bd6a5475150d28c2f4254f031a0fe6c7de5826
        - registry.redhat.io/quay/quay-operator-bundle@sha256:f16c0a135dd7380e2a489f963b38bd76a63a973f482d06ece8d3656aff223703
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:a12a0918f6dea114b0aae231d7119a4acfd0cd297074d895f4ecf190d2aec43a
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/quay/clair-rhel8@sha256:e11e8a819d17bd14c67f093b5303b4958855e731f89ce67dcca70fcb7b77f45a
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8ebc1f0511f7d394ceaa39a4c993e955c560dc04588072cb3f8c8fab8c85e32b
        - registry.redhat.io/quay/quay-rhel8@sha256:33e984892650d315065313fcb5f8808e145aadec5a287219c7031d85b713235e
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        version: 3.14.5
      entries:
      - name: quay-operator.v3.14.5
        version: 3.14.5
      - name: quay-operator.v3.14.4
        version: 3.14.4
      - name: quay-operator.v3.14.3
        version: 3.14.3
      - name: quay-operator.v3.14.2
        version: 3.14.2
      - name: quay-operator.v3.14.1
        version: 3.14.1
      - name: quay-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: quay-operator.v3.15.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.15.2
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures ODF for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay Operator expects ODF to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For production deployment of
          Red Hat OpenShift Data Foundation, please refer to the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_openshift_container_storage/).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploying_the_red_hat_quay_operator_on_openshift_container_platform/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-rhel8@sha256:aa8c9f7b6ce57441ca1fe9dbaaefaf2bb83fd44aa367647b663bb822bf890362
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-operator-bundle@sha256:b499eebed32c0261db414e307f96f8aef9e743035544c80c048526e7ba4fbdc2
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:956f3110f944d880bc27bb6df9472d0804cd75f5dcf4ed11bcf721e9d1d4542b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:873a3daf1dadf2438eb96e151072326c31ba34ed48cac041614b604413d5f8e7
        - registry.redhat.io/quay/clair-rhel8@sha256:861be31d30142bae125cd059bb716d79b28694e9c4336333c68adf4a3823b61d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:05fb981d9848281b97b3683093f752b1c1dc5ced439a2b4ba18b26ff8e11231e
        version: 3.15.2
      entries:
      - name: quay-operator.v3.15.2
        version: 3.15.2
      - name: quay-operator.v3.15.1
        version: 3.15.1
      - name: quay-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: quay-operator.v3.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
          createdAt: 2021-10-12 22:10 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:7a6e592d7f5f760b4451f9c974dbb96f56b1d3de92f039adcbf284f78ba13a02
        - registry.redhat.io/quay/quay-rhel8@sha256:e75e4d084eba9bb65b9cc72e38294af314d49316b12a0beb75db808dc057830b
        - registry.redhat.io/rhel8/postgresql-10@sha256:4b912c80085b88a03309aeb7907efcc29dd3342fa3952b6ea067afb1914bfe53
        - registry.redhat.io/rhel8/redis-6@sha256:d7c7852338717308cbb59e9303e1ea35cc8e5c01ceb2818569be20c15f3f943d
        - registry.redhat.io/quay/clair-rhel8@sha256:cbafddd1eb50c1c0f9159ffafd86f44a8c0b976194903c96c6f144bfe62e749d
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:75df6b8e6bef14844a08f7d994470fa6ec4ffa66d4b15faaa0a3ba373f612fbb
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:ec36241f0a9a2983b1894cd8501898e6b70379af51a81b1e09b0ba0e9b54b33f
        - registry.redhat.io/quay/quay-operator-bundle@sha256:55a5e6297109aea029bd40fac09a9fbc10c28a8c5558f9afc9b9b86c5dc7a4d4
        version: 3.6.10
      entries:
      - name: quay-operator.v3.6.10
        version: 3.6.10
      - name: quay-operator.v3.6.9
        version: 3.6.9
      - name: quay-operator.v3.6.8
        version: 3.6.8
      - name: quay-operator.v3.6.7
        version: 3.6.7
      - name: quay-operator.v3.6.6
        version: 3.6.6
      - name: quay-operator.v3.6.5
        version: 3.6.5
      - name: quay-operator.v3.6.4
        version: 3.6.4
      - name: quay-operator.v3.6.2
        version: 3.6.2
      - name: quay-operator.v3.6.1
        version: 3.6.1
      - name: quay-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: quay-operator.v3.7.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.7.14
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/clair-rhel8@sha256:d676af83e96740220e6f18721c98817ba071fd023a62ebdc3815424e94838083
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:04ac405c152bffb43b33fb2e7f45df8b3940eb21e055ebf6aa17bd81ebdeeab4
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:c54c77e09c95dc7d80d7b31e796857c8d06f2f733426629b53d3298206d23b20
        - registry.redhat.io/quay/quay-operator-bundle@sha256:e3f63ea56de7ca5388a20c17ac63e16f07935ab09bcdcd38a86639c9d59c8a7d
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:a8283c2be661e07c5f2f2e5b683c520846bcb91d580d9a3a0aa27ce994ddfff6
        - registry.redhat.io/quay/quay-rhel8@sha256:7bc9e115e9bace07505ffd225242cab83d5eb6d50358188d18fd9161bdf1a5a6
        - registry.redhat.io/rhel8/postgresql-10@sha256:a52ad402458ec8ef3f275972c6ebed05ad64398f884404b9bb8e3010c5c95291
        - registry.redhat.io/rhel8/redis-6@sha256:b15335d88b240bd6bbedeb7fe26f0a1de2f6c330af9758b0fb2edab7ca69f6df
        version: 3.7.14
      entries:
      - name: quay-operator.v3.7.14
        version: 3.7.14
      - name: quay-operator.v3.7.13
        version: 3.7.13
      - name: quay-operator.v3.7.12
        version: 3.7.12
      - name: quay-operator.v3.7.11
        version: 3.7.11
      - name: quay-operator.v3.7.10
        version: 3.7.10
      name: stable-3.7
    - currentCSV: quay-operator.v3.8.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
          createdAt: 2021-04-23 10:04 UTC
          description: Opinionated deployment of Red Hat on Kubernetes.
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          operators.operatorframework.io/internal-objects: '["quayecosystems.redhatcop.redhat.io"]'
          quay-version: v3.8.15
          repository: https://github.com/quay/quay-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
          - description: '[DEPRECATED] Old representation of a full Quay installation.'
            displayName: Quay Ecosystem
            kind: QuayEcosystem
            name: quayecosystems.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhel8/redis-6@sha256:444200241cb9186cd97d564d7acf077d83fc997a8ee5ee7fdae41b604895085c
        - registry.redhat.io/quay/clair-rhel8@sha256:b327fcfd7472cdb9d83a564904976b26b285fa576dc0a76e5590990ec77fd871
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:3c047c7b22fb42557bd3a9b9c9ac333f0df7d5e4a4d333a2b02a4313f57085be
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:7e4e57071439d671971b0de46b597cf6cff8ef3879ab711125a6ae7ecb9c6d29
        - registry.redhat.io/quay/quay-operator-bundle@sha256:53fcddadaa31a5f7d74c38f0bbd0d80b4c09796a580ef8494f60ff82fb6fe6c1
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:abec0fa7ee641c8e424e995b5fe07b98e522ad1eb92c3b90426f69941f33eb2e
        - registry.redhat.io/quay/quay-rhel8@sha256:de2b65506f99eebb5eaafca6bb42b4ebec08e9479dd0d50aa1286f197e42d6de
        - registry.redhat.io/rhel8/postgresql-10@sha256:b7fa644277d5e406b642685ff2ace1b1de7aea2e2a1421b3cbcda7d60405f44d
        version: 3.8.15
      entries:
      - name: quay-operator.v3.8.15
        version: 3.8.15
      - name: quay-operator.v3.8.14
        version: 3.8.14
      - name: quay-operator.v3.8.13
        version: 3.8.13
      - name: quay-operator.v3.8.12
        version: 3.8.12
      - name: quay-operator.v3.8.11
        version: 3.8.11
      - name: quay-operator.v3.8.10
        version: 3.8.10
      - name: quay-operator.v3.8.9
        version: 3.8.9
      - name: quay-operator.v3.8.8
        version: 3.8.8
      - name: quay-operator.v3.8.7
        version: 3.8.7
      - name: quay-operator.v3.8.6
        version: 3.8.6
      - name: quay-operator.v3.8.5
        version: 3.8.5
      - name: quay-operator.v3.8.4
        version: 3.8.4
      - name: quay-operator.v3.8.3
        version: 3.8.3
      - name: quay-operator.v3.8.2
        version: 3.8.2
      - name: quay-operator.v3.8.1
        version: 3.8.1
      - name: quay-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: quay-operator.v3.9.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "quay.redhat.com/v1",
                "kind": "QuayRegistry",
                "metadata": {
                  "name": "example-registry"
                },
                "spec": {
                  "components": [
                    {"kind": "clair", "managed": true},
                    {"kind": "postgres", "managed": true},
                    {"kind": "objectstorage", "managed": true},
                    {"kind": "redis", "managed": true},
                    {"kind": "horizontalpodautoscaler", "managed": true},
                    {"kind": "route", "managed": true},
                    {"kind": "mirror", "managed": true},
                    {"kind": "monitoring", "managed": true},
                    {"kind": "tls", "managed": true},
                    {"kind": "quay", "managed": true},
                    {"kind": "clairpostgres", "managed": true}
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          containerImage: registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
          createdAt: "2021-04-23T10:04:00Z"
          description: Opinionated deployment of Red Hat on Kubernetes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: v3.9.16
          repository: https://github.com/quay/quay-operator
          support: Project Quay
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a full Quay registry installation.
            displayName: Quay Registry
            kind: QuayRegistry
            name: quayregistries.quay.redhat.com
            version: v1
        description: |-
          The Red Hat Quay Operator deploys and manages a production-ready
          [Red Hat Quay](https://www.openshift.com/products/quay) private container registry.
          This operator provides an opinionated installation and configuration of Red Hat Quay.
          All components required, including Clair, database, and storage, are provided in an
          operator-managed fashion. Each component may optionally be self-managed.

          ## Operator Features

          * Automated installation of Red Hat Quay
          * Provisions instance of Redis
          * Provisions PostgreSQL to support both Quay and Clair
          * Installation of Clair for container scanning and integration with Quay
          * Provisions and configures RHOCS for supported registry object storage
          * Enables and configures Quay's registry mirroring feature

          ## Prerequisites

          By default, the Red Hat Quay operator expects RHOCS to be installed on the cluster to
          provide the _ObjectBucketClaim_ API for object storage. For instructions installing and
          configuring the RHOCS Operator, see the "Enabling OpenShift Container Storage" in the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html-single/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index#enabling_openshift_container_storage).

          ## Simplified Deployment

          The following example provisions a fully operator-managed deployment of Red Hat Quay,
          including all services necessary for production:

          ```
          apiVersion: quay.redhat.com/v1
          kind: QuayRegistry
          metadata:
            name: my-registry
          ```

          ## Documentation

          See the
          [official documentation](https://access.redhat.com/documentation/en-us/red_hat_quay/3/html/deploy_red_hat_quay_on_openshift_with_the_quay_operator/index)
          for more complex deployment scenarios and information.
        displayName: Red Hat Quay
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - registry
        links:
        - name: Source Code
          url: https://github.com/quay/quay-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-operator-bundle@sha256:9d0ed6d279800a189c1466e0b32f80af2f8bad96a65b7dc0322c64337dddf39c
        - registry.redhat.io/quay/quay-operator-rhel8@sha256:8a16803946272156b7aeb00eb2202ab6cb87d875fe4a7c327554f770f0f2ecb5
        - registry.redhat.io/quay/quay-rhel8@sha256:f87b7e7ee6c31f6ae80b08aef0e10729b21fe1b8656fe18f6dee2310bae28ec1
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/quay/quay-builder-qemu-rhcos-rhel8@sha256:5a42689198701d845ff3f05502cbbebceb7e6ad7b12762ccc038c0131cf6479e
        - registry.redhat.io/rhel8/postgresql-10@sha256:eec360a32519d45c885efbfc6539dc5ff74447e1aceeb5115d1aa30b40c52aeb
        - registry.redhat.io/rhel8/postgresql-13@sha256:1d1364c675bd1f7cbcaf34d695ddbd385ddc59e16fb00334ddf7a9b5ba47ec70
        - registry.redhat.io/quay/clair-rhel8@sha256:6c67163c783d7d9fa4d187963379885a8b76685347ecf1894c4f8ebd6656544b
        - registry.redhat.io/quay/quay-builder-rhel8@sha256:af8ada0c88b6f19991e61ee86d04b49a148c19dacb57c282e4ab98630e7499b4
        version: 3.9.16
      entries:
      - name: quay-operator.v3.9.16
        version: 3.9.16
      - name: quay-operator.v3.9.15
        version: 3.9.15
      - name: quay-operator.v3.9.14
        version: 3.9.14
      - name: quay-operator.v3.9.13
        version: 3.9.13
      - name: quay-operator.v3.9.12
        version: 3.9.12
      - name: quay-operator.v3.9.11
        version: 3.9.11
      - name: quay-operator.v3.9.10
        version: 3.9.10
      - name: quay-operator.v3.9.9
        version: 3.9.9
      - name: quay-operator.v3.9.8
        version: 3.9.8
      - name: quay-operator.v3.9.7
        version: 3.9.7
      - name: quay-operator.v3.9.6
        version: 3.9.6
      - name: quay-operator.v3.9.5
        version: 3.9.5
      - name: quay-operator.v3.9.4
        version: 3.9.4
      - name: quay-operator.v3.9.3
        version: 3.9.3
      - name: quay-operator.v3.9.2
        version: 3.9.2
      - name: quay-operator.v3.9.1
        version: 3.9.1
      - name: quay-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: quay-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.12
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:9d21b738b0d357ef064690eae66bd4b3810664bcde57237822f1e90e9fd6ad03
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:9aed98a061b2c76f3c512c1d6453df5920f0c779b33d7168c1e7ab9d8bc928a4
        version: 0.0.12
      entries:
      - name: joget-dx8-openshift-operator.v0.0.12
        version: 0.0.12
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat CCIT
      provider-url: ""
    name: reportportal-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: reportportal-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortal",
                "metadata": {
                  "name": "reportportal-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "analyzer_image": "quay.io/waynesun09/service-auto-analyzer:5.3.5",
                  "api_image": "quay.io/waynesun09/service-api:5.3.5-rootless",
                  "api_replicas": 1,
                  "app_domain": "apps.test-example.com",
                  "bucket_prefix": null,
                  "default_bucket_name": null,
                  "enable_pg_restic_backup": "yes",
                  "es_backup_schedule": "@daily",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "gateway_replicas": 1,
                  "index_image": "quay.io/waynesun09/service-index:5.0.10",
                  "index_replicas": 1,
                  "integration_salt_path": null,
                  "migration_image": "quay.io/waynesun09/migrations:5.3.5",
                  "pg_restic_password": "rp_user",
                  "pg_restic_s3_bucket_init": "yes",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_enabled": false,
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY",
                  "uat_image": "quay.io/waynesun09/service-authorization:5.3.5",
                  "ui_image": "quay.io/waynesun09/service-ui:5.3.5",
                  "ui_replicas": 1
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ReportPortalRestore",
                "metadata": {
                  "name": "reportportalrestore-sample"
                },
                "spec": {
                  "accesskey": "AKIAIOSFODNN7EXAMPLE",
                  "bucket_prefix": null,
                  "es_restore": "yes",
                  "es_restore_date": "2020-08-07",
                  "es_s3_backup_dir": "s3_backup",
                  "es_snapshot_bucket": "es-snapshot-123123",
                  "pg_recovery": "yes",
                  "pg_recovery_target_time": "2020-08-07 00:00:00",
                  "pg_restic_password": "rp_user",
                  "pg_s3_bucket_name": "pgbackup-123123",
                  "s3_endpoint": null,
                  "secretkey": "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
                }
              },
              {
                "apiVersion": "rp5.reportportal.io/v1",
                "kind": "ServiceApiServiceMonitor",
                "metadata": {
                  "name": "serviceapiservicemonitor-sample"
                },
                "spec": {
                  "bearer_token_secret_key": "token",
                  "bearer_token_secret_name": "rp-user-auth"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/waynesun09/rp5-operator:v0.0.8
          createdAt: "2021-04-22 14:39:00"
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/waynesun09/reportportal-operator
          support: waynesun09
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ReportPortal Restore from s3 backup
            kind: ReportPortalRestore
            name: reportportalrestores.rp5.reportportal.io
            version: v1
          - description: ReportPortal v5 Deployment with ReportPortal Operator
            kind: ReportPortal
            name: reportportals.rp5.reportportal.io
            version: v1
          - description: ReportPortal service-api ServiceMonitor for metrics
            kind: ServiceApiServiceMonitor
            name: serviceapiservicemonitors.rp5.reportportal.io
            version: v1
        description: |-
          ReportPortal is a test automation analytics platform. ReportPortal Operator support instantiating the ReportPortal services and required dependencies.

          What services are included:

              - Report Portal v5.3
              - PostgreSQL
              - Elasticsearch
              - RabbitMQ
              - Minio

          The Operator also support data backup & restore and metrics monitoring.

          ## Install

          Install the Operator to a namespace, then create ReportPortal v5 instance under Provided APIs on the UI, all Report Portal v5 services and dependencies will be installed.

          ## Data backup and restore

          The Report Portal v5 data backup and restore in the operator include two parts, PostgreSQL database data and Elasticsearch index data.

          The Operator support automatically data backup and restore for both.

          ### PostgreSQL database WAL archiving and recovery

          The Operator supports PostgreSQL WAL archiving and recovery point-in-time via Minio S3 using restic.

          #### WAL archiving

          When deploy Report Portal v5 instance, make sure `enable_pg_restic_backup` option is set as `yes`, then the WAL archiving will be enabled when initiate the database, the WAL archive data will be automatically backup to S3 storage via Minio using restic command.

          Parameter `pg_restic_s3_bucket_init` is for init the bucket if not exist, `pg_s3_bucket_name` is the bucket name and `pg_restic_password` is the restic password for encryption.

          #### WAL recovery

          WAL archive recovery will download all WAL archive files to the Postgres container and restore point-in-time.

          To restore PostgreSQL database point-in-time, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          After the ReportPortalRestore operand created, the PG WAL restore will get started, as restore time will base on the database size, the process will last in minutes, hours or days.

          **Note:** Current Postgres container liveless probe timeout is set as one hour in the postgresql statefulset, it might cause restore fail if restore time is longer than one hour, so it might need be adjusted to be longer.

          ### Elasticsearch snapshot and restore

          The operator supports elasticsearch snapshot and restore with data saved in S3 bucket via Minio, the elasticsearch container have installed the s3 repository plugin.

          #### Snapshot

          When deploy Report Portal v5 instance, make sure following elasticsearch parameters are set:

              es_s3_backup_dir: s3_backup
              es_snapshot_bucket: es-snapshot-123123
              es_backup_schedule: '@daily'

          #### Restore

          Elasticsearch restore support restore to certain date as default snapshot schedule is daily.

          To restore elasticsearch from snapshot, on Operator UI select `ReportPortalRestore` under Provided APIs to create ReportPortalRestore operand.

          The backup dir and bucket name should be same as when create the RP instance, `es_restore` need set to `yes` to enable the restore and `es_restore_date` is the restore target date.

          The restore time will also be depend on the snapshot size.

          ## Metrics

          The operator by default have enabled metrics, the default ansible operator port 8080 is used for the operator container.

          All Report Portal v5 depended services, rabbitmq, minio, postgresql and elasticsearch have metrics enabled in this operator.

          The Report Portal gatway service have metrics enabled.

          By default several ServiceMonitors have been defined when create ReportPortal v5 operand:

              $ oc get servicemonitor
              NAME                    AGE
              elasticsearch-metrics   1d21h
              gateway-metrics         1d21h
              minio-metrics           1d21h
              postgres-metrics        1d21h
              rabbitmq-metrics        1d21h

          To export Report Portal v5 service-api metrics with ServiceMonitor, create a `secret` with RP user api token. The api access token could be found with login to Report Portal -> Click user icon -> Profile.

          Then select `ServiceApiServiceMonitor` under provided APIs to Create Instance on operator UI.

          Input the secret name and key name that have been created, and click create, a api-metrics servicemonitor will be created.

          To check the metrics, if [user workload monitoring](https://docs.openshift.com/container-platform/4.5/monitoring/monitoring-your-own-services.html) is enabled on the cluster, click to Monitoring on OpenShift UI and navigate to Metrics, where you could query the metrics.

          You could also deploy Promethues Operator in the current project and query via it.

          ## Doc

          More detail with installation, data backup & restore and metrics could be found at https://github.com/waynesun09/reportportal-operator#reportportal-operator
        displayName: reportportal-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - reportportal
        - failure analysis
        - test automation
        links:
        - name: Documentation
          url: https://github.com/waynesun09/reportportal-operator/blob/master/README.md
        - name: ReportPortal.io
          url: https://reportportal.io
        maintainers:
        - email: gsun@redhat.com
          name: Wayne Sun
        maturity: alpha
        provider:
          name: Red Hat CCIT
        relatedImages:
        - quay.io/openshift-community-operators/reportportal-operator@sha256:67a7ffaf6a303b92541170f266d7b2871f5c2fe73f6b6c9a2b7b33dbc4555fa5
        - quay.io/waynesun09/rp5-operator:v0.0.8
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.7.0
        version: 0.0.8
      entries:
      - name: reportportal-operator.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: reportportal-operator
    provider:
      name: Red Hat CCIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-community-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.8.2
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.8.2/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.8.2
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:8197e4a96ac8949d43cda62ac2257a9b1e71f0b54f42283ee4925f88c369f873
        version: 0.8.2
      entries:
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.11
    - currentCSV: 3scale-community-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.9.0
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.9.0/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.9.0
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:728074151b245f5e0d65bd8956149f05ba22a1b6438120ebca2d85b0d7d0c837
        version: 0.9.0
      entries:
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.12
    - currentCSV: 3scale-community-operator.v0.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/3scale/3scale-operator:v0.10.1
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |
          The 3scale community Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          3scale API Management makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          Documentation can be found on our [website](https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md).

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Improving [documentation](https://github.com/3scale/3scale-operator/blob/master/doc/operator-user-guide.md)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: 3scale API Management
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://github.com/3scale/3scale-operator/blob/v0.10.1/doc/operator-user-guide.md
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/3scale/3scale-operator:v0.10.1
        - quay.io/openshift-community-operators/3scale-community-operator@sha256:4109698147632b858fe2787964772a1339cdebee72b75b42d9b1c5dbbbecb45c
        version: 0.10.1
      entries:
      - name: 3scale-community-operator.v0.10.1
        version: 0.10.1
      - name: 3scale-community-operator.v0.9.0
        version: 0.9.0
      - name: 3scale-community-operator.v0.8.2
        version: 0.8.2
      name: threescale-2.13
    defaultChannel: threescale-2.13
    packageName: 3scale-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.10.0
    - currentCSV: t8c-operator.v42.10.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.10
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:6e5cbb3287ad3ec69f3064fd5dd88f0e4794fef985bab390c5135c7162c2872c
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.10
        version: 42.10.0
      entries:
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.12.0
    - currentCSV: t8c-operator.v42.13.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.13
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:088d0eebf60c8f6961ca72c268f61d12648d575aa41552ae6398fefc6d9700cb
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.13
        version: 42.13.0
      entries:
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.13.0
    - currentCSV: t8c-operator.v42.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.7"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.14
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d6d444e642bfd567eb40c7e084f5df743c23b47935e52e35ef541c67fcc00433
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.14
        version: 42.14.0
      entries:
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.14.0
    - currentCSV: t8c-operator.v42.15.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.15
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:b3469cd953ef2015fad2f7f96d960cee5ecaa6c19df631a9c882154ab5b3de98
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.15
        version: 42.15.0
      entries:
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.15.0
    - currentCSV: t8c-operator.v42.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.16
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:f49d3eddbfcec1018fd3716cd0ec2addef9961d2d308099636541868566979b0
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.16
        version: 42.16.0
      entries:
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.16.0
    - currentCSV: t8c-operator.v42.17.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.6.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.17
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.17
        - quay.io/openshift-community-operators/t8c@sha256:403b29a6262ba0e06de117f788aea7224268b798868c3e5b738b05e96df20456
        version: 42.17.0
      entries:
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.17.0
    - currentCSV: t8c-operator.v42.18.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.18
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4bfaf97ab09a1aff1316536b99957b1f82763b009dde6d08fc9eea1fdd290fbf
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.18
        version: 42.18.0
      entries:
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.18.0
    - currentCSV: t8c-operator.v42.19.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.2"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.19
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.19
        - quay.io/openshift-community-operators/t8c@sha256:61f1ea9b3851d7272703f5bc21203ff03bb35fd52ed9780d0cf996379c78a00b
        version: 42.19.0
      entries:
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.19.0
    - currentCSV: t8c-operator.v42.20.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.20
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:48a5a9fbfa92b4b87b3653613467a3fcceacadbbabd274e367b40f6871039cc8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.20
        version: 42.20.0
      entries:
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.20.0
    - currentCSV: t8c-operator.v42.21.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.21
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d65834d8f9946c2764d45666c90c50f428b74afe7b066d666c18e9545994c9a8
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.21
        version: 42.21.0
      entries:
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.21.0
    - currentCSV: t8c-operator.v42.22.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.7.5"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.22
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.22
        - quay.io/openshift-community-operators/t8c@sha256:3c48e0bb9fd04d86d92bddb5bc2367f7ee4b86af74421c8dcbd809a7ba657770
        version: 42.22.0
      entries:
      - name: t8c-operator.v42.22.0
        version: 42.22.0
      - name: t8c-operator.v42.21.0
        version: 42.21.0
      - name: t8c-operator.v42.20.0
        version: 42.20.0
      - name: t8c-operator.v42.19.0
        version: 42.19.0
      - name: t8c-operator.v42.18.0
        version: 42.18.0
      - name: t8c-operator.v42.17.0
        version: 42.17.0
      - name: t8c-operator.v42.16.0
        version: 42.16.0
      - name: t8c-operator.v42.15.0
        version: 42.15.0
      - name: t8c-operator.v42.14.0
        version: 42.14.0
      - name: t8c-operator.v42.13.0
        version: 42.13.0
      - name: t8c-operator.v42.12.0
        version: 42.12.0
      - name: t8c-operator.v42.10.0
        version: 42.10.0
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.22.0
    - currentCSV: t8c-operator.v42.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.3"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.6
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:4964196aba839fde569ce58ab7e4eef438c9638ae4057194f2d7a58eaa5a98d9
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.6
        version: 42.6.0
      entries:
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.6.0
    - currentCSV: t8c-operator.v42.7.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.4.6"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.7
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.7
        - quay.io/openshift-community-operators/t8c@sha256:46c815fa635d58820757f41b1d8f04333638c39471f4d6c901a6500379680b57
        version: 42.7.0
      entries:
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.7.0
    - currentCSV: t8c-operator.v42.8.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.0"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.8
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:d8a575a84dd86d241acaaec53af1e164bc1ff1b95b5eb710851ae698bbed6bf5
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.8
        version: 42.8.0
      entries:
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.8.0
    - currentCSV: t8c-operator.v42.9.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"tag":"8.5.1"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/turbonomic/t8c-operator:42.9
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/t8c@sha256:dfdc8fbdfc0ee76b2d5c62f6fd5deac2c50756661871d8b4357e25043e5ce66a
        - registry.connect.redhat.com/turbonomic/t8c-operator:42.9
        version: 42.9.0
      entries:
      - name: t8c-operator.v42.9.0
        version: 42.9.0
      - name: t8c-operator.v42.8.0
        version: 42.8.0
      - name: t8c-operator.v42.7.0
        version: 42.7.0
      - name: t8c-operator.v42.6.0
        version: 42.6.0
      - name: t8c-operator.v42.5.0
        version: 42.5.0
      - name: t8c-operator.v42.4.0
        version: 42.4.0
      - name: t8c-operator.v42.3.0
        version: 42.3.0
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: v42.9.0
    defaultChannel: stable
    packageName: t8c
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: JWS team
      provider-url: ""
    name: jws-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: jws-operator.v2.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "web.servers.org/v1alpha1",
                "kind": "WebServer",
                "metadata": {
                  "name": "webserver-example"
                },
                "spec": {
                  "applicationName": "webapp",
                  "replicas": 1,
                  "useInsightsClient": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
          createdAt: "2025-09-12T08:31:52.558719Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat JBoss Web Server",
            "Red Hat Runtimes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/web-servers/jws-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Web Server is the schema for the webservers API
            displayName: Web Server
            kind: WebServer
            name: webservers.web.servers.org
            version: v1alpha1
        description: WebServer operator can manage JBoss Web Server images.
        displayName: WebServer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - WebServer
        - JBoss
        links:
        - name: Jws Operator
          url: https://jws-operator.domain
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: JWS team
        relatedImages:
        - registry.redhat.io/jboss-webserver-5/jws5-rhel8-operator@sha256:ea042277bdcdb70b2dc32083ea868efd4bf474d5a8ec0f9b47f5bea35aaeae1e
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:021ff671b0fb72eed635b05489ec86342fcb7ada3ba0353591fdb9328871d067
        - registry.redhat.io/insights-runtimes-tech-preview/runtimes-inventory-rhel8-operator@sha256:58e01d11c3d30abc5b0ee3795511bd04e39aea21e2853b4f4cc6d1c8a2bd7bb0
        - registry.redhat.io/jboss-webserver-5/jws5-operator-bundle@sha256:ce4060138117dbc8fc21abb60f0b8ef1bfa5ebc0dd8eb803c0e79d7447341964
        version: 2.2.0
      entries:
      - name: jws-operator.v2.2.0
        version: 2.2.0
      - name: jws-operator.v2.1.3
        version: 2.1.3
      - name: jws-operator.v2.1.2
        version: 2.1.2
      - name: jws-operator.v2.1.1
        version: 2.1.1
      - name: jws-operator.v2.1.0
        version: 2.1.0
      - name: jws-operator.v2.0.16
        version: 2.0.16
      - name: jws-operator.v2.0.15
        version: 2.0.15
      - name: jws-operator.v2.0.14
        version: 2.0.14
      - name: jws-operator.v2.0.13
        version: 2.0.13
      - name: jws-operator.v2.0.12
        version: 2.0.12
      - name: jws-operator.v2.0.11
        version: 2.0.11
      - name: jws-operator.v2.0.10
        version: 2.0.10
      - name: jws-operator.v2.0.9
        version: 2.0.9
      - name: jws-operator.v2.0.8
        version: 2.0.8
      - name: jws-operator.v2.0.7
        version: 2.0.7
      - name: jws-operator.v2.0.6
        version: 2.0.6
      - name: jws-operator.v2.0.5
        version: 2.0.5
      - name: jws-operator.v2.0.4
        version: 2.0.4
      - name: jws-operator.v2.0.3
        version: 2.0.3
      - name: jws-operator.v2.0.2
        version: 2.0.2
      - name: jws-operator.v2.0.1
        version: 2.0.1
      name: alpha
    defaultChannel: alpha
    packageName: jws-operator
    provider:
      name: JWS team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: rook-ceph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rook-ceph-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9y
            dCgzMzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQ
            LmFkZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAg
            ICAgICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24i
            LAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJh
            c2gpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3Zp
            ZGVkIGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAog
            ICAgICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmls
            ZXN5c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IG1ldGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBk
            ZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2Vw
            aGZzIGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1l
            bnQoCiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92
            aWRlcyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0
            cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9v
            bC1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRh
            IHBvb2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBp
            biB0aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAg
            KQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1l
            bmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZh
            bHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+
            OjxQT1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCki
            LAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAg
            ICAiLS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAg
            ICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBv
            aW50IFRMUyBjZXJ0aWZpY2F0ZSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRf
            YXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1za2lwLXRscyIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJJ
            Z25vcmUgVExTIGNlcnRpZmljYXRpb24gdmFsaWRhdGlvbiB3aGVuIGEgc2VsZi1zaWduZWQgY2Vy
            dGlmaWNhdGUgaXMgcHJvdmlkZWQgKE5PVCBSRUNPTU1FTkRFRCIsCiAgICAgICAgKQogICAgICAg
            IG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLW1vbml0b3JpbmctZW5k
            cG9pbnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgZW5k
            cG9pbnRzIChjb21tYSBzZXBhcmF0ZWQgbGlzdCBvZiAoZm9ybWF0IGA8SVB2ND5gIG9yIGA8W0lQ
            djZdPmAgb3IgYDxGUUROPmApIGVudHJpZXMgb2YgYWN0aXZlIGFuZCBzdGFuZGJ5IG1ncnMpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tbW9uaXRvcmluZy1lbmRwb2ludC1wb3J0IiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAg
            ICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9IkNlcGggTWFuYWdlciBw
            cm9tZXRoZXVzIGV4cG9ydGVyIHBvcnQiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAu
            YWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1za2lwLW1vbml0b3JpbmctZW5kcG9pbnQiLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBhY3Rpb249InN0b3JlX3RydWUi
            LAogICAgICAgICAgICBoZWxwPSJEbyBub3QgY2hlY2sgZm9yIGEgbW9uaXRvcmluZyBlbmRwb2lu
            dCBmb3IgdGhlIENlcGggY2x1c3RlciIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgZXJhc3VyZSBjb2RlZCBSQkQgbWV0YWRh
            dGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAg
            ICAgICAgICAgICItLWRyeS1ydW4iLAogICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAg
            ICAgICBhY3Rpb249InN0b3JlX3RydWUiLAogICAgICAgICAgICBoZWxwPSJEcnkgcnVuIHByaW50
            cyB0aGUgZXhlY3V0ZWQgY29tbWFuZHMgd2l0aG91dCBydW5uaW5nIHRoZW0iLAogICAgICAgICkK
            ICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yYWRvcy1u
            YW1lc3BhY2UiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1G
            YWxzZSwKICAgICAgICAgICAgaGVscD0iRGl2aWRlcyBhIHBvb2wgaW50byBzZXBhcmF0ZSBsb2dp
            Y2FsIG5hbWVzcGFjZXMsIHVzZWQgZm9yIGNyZWF0aW5nIFJCRCBQVkMgaW4gYSBDZXBoQmxvY2tQ
            b29sUmFkb3NOYW1lc3BhY2UgKHNob3VsZCBiZSBsb3dlciBjYXNlKSIsCiAgICAgICAgKQogICAg
            ICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXN1YnZvbHVtZS1n
            cm91cCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgc3Vidm9sdW1lIGdy
            b3VwIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAg
            ICAgICAgIi0tcmd3LXJlYWxtLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAg
            ICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhlIG5hbWUg
            b2YgdGhlIHJndy1yZWFsbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJn
            dW1lbnQoCiAgICAgICAgICAgICItLXJndy16b25lLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlk
            ZXMgdGhlIG5hbWUgb2YgdGhlIHJndy16b25lIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cC1uYW1lIiwKICAg
            ICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAg
            ICAgIGhlbHA9InByb3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSByZ3ctem9uZWdyb3VwIiwKICAgICAg
            ICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9w
            b2xvZ3ktcG9vbHMiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJl
            ZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdG9wb2xv
            Z3ktY29uc3RyYWluZWQgcmJkIHBvb2xzIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3Vw
            LmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdG9wb2xvZ3ktZmFpbHVyZS1kb21haW4tbGFi
            ZWwiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwK
            ICAgICAgICAgICAgaGVscD0iazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gbGFiZWwgKGV4YW1w
            bGU6IHpvbmUsIHJhY2ssIG9yIGhvc3QpIGZvciB0aGUgdG9wb2xvZ3ktcG9vbHMgdGhhdCBtYXRj
            aCB0aGUgY2VwaCBkb21haW4iLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2Fy
            Z3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi12YWx1ZXMiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAg
            ICAgICAgaGVscD0iY29tbWEtc2VwYXJhdGVkIGxpc3Qgb2YgdGhlIGs4cyBjbHVzdGVyIGZhaWx1
            cmUgZG9tYWluIHZhbHVlcyBjb3JyZXNwb25kaW5nIHRvIGVhY2ggb2YgdGhlIHBvb2xzIGluIHRo
            ZSBgdG9wb2xvZ3ktcG9vbHNgIGxpc3QiLAogICAgICAgICkKCiAgICAgICAgdXBncmFkZV9ncm91
            cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJ1cGdyYWRlIikKICAgICAgICB1cGdyYWRlX2dy
            b3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tdXBncmFkZSIsCiAgICAgICAgICAgIGFj
            dGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAg
            IGhlbHA9IlVwZ3JhZGVzIHRoZSBjZXBoQ1NJS2V5cmluZ3MoRm9yIGV4YW1wbGU6IGNsaWVudC5j
            c2ktY2VwaGZzLXByb3Zpc2lvbmVyKSBhbmQgY2xpZW50LmhlYWx0aGNoZWNrZXIgY2VwaCB1c2Vy
            cyB3aXRoIG5ldyBwZXJtaXNzaW9ucyBuZWVkZWQgZm9yIHRoZSBuZXcgY2x1c3RlciB2ZXJzaW9u
            IGFuZCBvbGRlciBwZXJtaXNzaW9uIHdpbGwgc3RpbGwgYmUgYXBwbGllZC4iCiAgICAgICAgICAg
            ICsgIlNhbXBsZSBydW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rl
            ci1yZXNvdXJjZXMucHkgLS11cGdyYWRlYCwgdGhpcyB3aWxsIHVwZ3JhZGUgYWxsIHRoZSBkZWZh
            dWx0IGNzaSB1c2Vycyhub24tcmVzdHJpY3RlZCkiCiAgICAgICAgICAgICsgIkZvciByZXN0cmlj
            dGVkIHVzZXJzKEZvciBleGFtcGxlOiBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lci1vcGVu
            c2hpZnQtc3RvcmFnZS1teWZzKSwgdXNlcnMgY3JlYXRlZCB1c2luZyAtLXJlc3RyaWN0ZWQtYXV0
            aC1wZXJtaXNzaW9uIGZsYWcgbmVlZCB0byBwYXNzIG1hbmRhdG9yeSBmbGFncyIKICAgICAgICAg
            ICAgKyAibWFuZGF0b3J5IGZsYWdzOiAnLS1yYmQtZGF0YS1wb29sLW5hbWUsIC0tazhzLWNsdXN0
            ZXItbmFtZSBhbmQgLS1ydW4tYXMtdXNlcicgZmxhZ3Mgd2hpbGUgdXBncmFkaW5nIgogICAgICAg
            ICAgICArICJpbiBjYXNlIG9mIGNlcGhmcyB1c2VycyBpZiB5b3UgaGF2ZSBwYXNzZWQgLS1jZXBo
            ZnMtZmlsZXN5c3RlbS1uYW1lIGZsYWcgd2hpbGUgY3JlYXRpbmcgdXNlciB0aGVuIHdoaWxlIHVw
            Z3JhZGluZyBpdCB3aWxsIGJlIG1hbmRhdG9yeSB0b28iCiAgICAgICAgICAgICsgIlNhbXBsZSBy
            dW46IGBweXRob24zIC9ldGMvY2VwaC9jcmVhdGUtZXh0ZXJuYWwtY2x1c3Rlci1yZXNvdXJjZXMu
            cHkgLS11cGdyYWRlIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxpY2Fwb29sIC0tazhzLWNsdXN0
            ZXItbmFtZSByb29rc3RvcmFnZSAgLS1ydW4tYXMtdXNlciBjbGllbnQuY3NpLXJiZC1ub2RlLXJv
            b2tzdG9yYWdlLXJlcGxpY2Fwb29sYCIKICAgICAgICAgICAgKyAiUFM6IEFuIGV4aXN0aW5nIG5v
            bi1yZXN0cmljdGVkIHVzZXIgY2Fubm90IGJlIGNvbnZlcnRlZCB0byBhIHJlc3RyaWN0ZWQgdXNl
            ciBieSB1cGdyYWRpbmcuIgogICAgICAgICAgICArICJVcGdyYWRlIGZsYWcgc2hvdWxkIG9ubHkg
            YmUgdXNlZCB0byBhcHBlbmQgbmV3IHBlcm1pc3Npb25zIHRvIHVzZXJzLCBpdCBzaG91bGRuJ3Qg
            YmUgdXNlZCBmb3IgY2hhbmdpbmcgdXNlciBhbHJlYWR5IGFwcGxpZWQgcGVybWlzc2lvbiwgZm9y
            IGV4YW1wbGUgeW91IHNob3VsZG4ndCBjaGFuZ2UgaW4gd2hpY2ggcG9vbCB1c2VyIGhhcyBhY2Nl
            c3MiLAogICAgICAgICkKCiAgICAgICAgIyBBZGQgY29tbWFuZC1saW5lIGFyZ3VtZW50cwogICAg
            ICAgIGNvbmZpZ19ncm91cCA9IGFyZ1AuYWRkX2FyZ3VtZW50X2dyb3VwKCJjb25maWciKQogICAg
            ICAgIGNvbmZpZ19ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLWNvbmZpZy1maWxl
            IiwKICAgICAgICAgICAgdHlwZT1zdHIsCiAgICAgICAgICAgIGhlbHA9IlBhdGggdG8gdGhlIGNv
            bmZpZ3VyYXRpb24gZmlsZSwgUHJpb3JpdHk6IGNvbW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmlu
            aSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFs
            c2UsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgKQoKICAgICAgICBpZiBhcmdzX3Rv
            X3BhcnNlOgogICAgICAgICAgICBhc3NlcnQgKAogICAgICAgICAgICAgICAgdHlwZShhcmdzX3Rv
            X3BhcnNlKSA9PSBsaXN0CiAgICAgICAgICAgICksICJBcmd1bWVudCB0byAnZ2VuX2FyZ19wYXJz
            ZXInIHNob3VsZCBiZSBhIGxpc3QiCiAgICAgICAgZWxzZToKICAgICAgICAgICAgYXJnc190b19w
            YXJzZSA9IHN5cy5hcmd2WzE6XQogICAgICAgIHJldHVybiBhcmdQLnBhcnNlX2FyZ3MoYXJnc190
            b19wYXJzZSkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZShzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWU6
            CiAgICAgICAgICAgIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2Vy
            LnJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCgogICAgICAgICAgICBpZiByYmRfcG9v
            bF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtZGF0YS1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBpZiByYmRf
            bWV0YWRhdGFfZWNfcG9vbF9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiRmxhZyAnLS1yYmQtbWV0
            YWRhdGEtZWMtcG9vbC1uYW1lJyBzaG91bGQgbm90IGJlIGVtcHR5IgogICAgICAgICAgICAgICAg
            KQoKICAgICAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJvc2QgZHVtcCIsICJmb3JtYXQi
            OiAianNvbiJ9CiAgICAgICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5f
            Y29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            IG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJ7Y21kX2pzb25bJ3ByZWZpeCdd
            fSBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgICAgICBmIkVycm9yOiB7ZXJyX21z
            ZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZX09VVFBVVF9MSVNUfSIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCwgcG9vbF9leGlzdCA9IEZh
            bHNlLCBGYWxzZQoKICAgICAgICAgICAgZm9yIGtleSBpbiBqc29uX291dFsicG9vbHMiXToKICAg
            ICAgICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgZW1wdHkgYW5kIHBvb2wg
            bmFtZSBleGlzdHMgdGhlbiBpdCByZXBsaWNhIHBvb2wKICAgICAgICAgICAgICAgIGlmICgKICAg
            ICAgICAgICAgICAgICAgICBrZXlbImVyYXN1cmVfY29kZV9wcm9maWxlIl0gPT0gIiIKICAgICAg
            ICAgICAgICAgICAgICBhbmQga2V5WyJwb29sX25hbWUiXSA9PSByYmRfbWV0YWRhdGFfZWNfcG9v
            bF9uYW1lCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIG1ldGFkYXRhX3Bv
            b2xfZXhpc3QgPSBUcnVlCiAgICAgICAgICAgICAgICAjIGlmIGVyYXN1cmVfY29kZV9wcm9maWxl
            IGlzIG5vdCBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IGlzIGVjIHBvb2wKICAg
            ICAgICAgICAgICAgIGlmIGtleVsiZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSBhbmQga2V5WyJwb29s
            X25hbWUiXSA9PSByYmRfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgICAgIHBvb2xfZXhpc3Qg
            PSBUcnVlCgogICAgICAgICAgICBpZiBub3QgbWV0YWRhdGFfcG9vbF9leGlzdDoKICAgICAgICAg
            ICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAg
            ICAgIlByb3ZpZGVkIHJiZF9lY19tZXRhZGF0YV9wb29sIG5hbWUsIgogICAgICAgICAgICAgICAg
            ICAgIGYiIHtyYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lfSwgZG9lcyBub3QgZXhpc3QiCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBwb29sX2V4aXN0OgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            IlByb3ZpZGVkIHJiZF9kYXRhX3Bvb2wgbmFtZSwge3JiZF9wb29sX25hbWV9LCBkb2VzIG5vdCBl
            eGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgcmV0dXJuIHJiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWUKCiAgICBkZWYgZHJ5X3J1bihzZWxmLCBtc2cpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcHJpbnQoIkV4ZWN1dGU6ICIgKyAiJyIg
            KyBtc2cgKyAiJyIpCgogICAgZGVmIHZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydChzZWxm
            KToKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoOgogICAgICAg
            ICAgICB3aXRoIG9wZW4oc2VsZi5fYXJnX3BhcnNlci5yZ3dfdGxzX2NlcnRfcGF0aCwgZW5jb2Rp
            bmc9InV0ZjgiKSBhcyBmOgogICAgICAgICAgICAgICAgY29udGVudHMgPSBmLnJlYWQoKQogICAg
            ICAgICAgICAgICAgcmV0dXJuIGNvbnRlbnRzLnJzdHJpcCgpCgogICAgZGVmIF9jaGVja19jb25m
            bGljdGluZ19vcHRpb25zKHNlbGYpOgogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGUgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZToKICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICJF
            aXRoZXIgJy0tdXBncmFkZScgb3IgJy0tcmJkLWRhdGEtcG9vbC1uYW1lIDxwb29sX25hbWU+JyBz
            aG91bGQgYmUgc3BlY2lmaWVkIgogICAgICAgICAgICApCgogICAgZGVmIF9pbnZhbGlkX2VuZHBv
            aW50KHNlbGYsIGVuZHBvaW50X3N0cik6CiAgICAgICAgIyBleHRyYWN0IHRoZSBwb3J0IGJ5IGdl
            dHRpbmcgdGhlIGxhc3Qgc3BsaXQgb24gYDpgIGRlbGltaXRlcgogICAgICAgIHRyeToKICAgICAg
            ICAgICAgZW5kcG9pbnRfc3RyX2lwLCBwb3J0ID0gZW5kcG9pbnRfc3RyLnJzcGxpdCgiOiIsIDEp
            CiAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oZiJOb3QgYSBwcm9wZXIgZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9IikK
            CiAgICAgICAgdHJ5OgogICAgICAgICAgICBpZiBlbmRwb2ludF9zdHJfaXBbMF0gPT0gIlsiOgog
            ICAgICAgICAgICAgICAgZW5kcG9pbnRfc3RyX2lwID0gZW5kcG9pbnRfc3RyX2lwWzEgOiBsZW4o
            ZW5kcG9pbnRfc3RyX2lwKSAtIDFdCiAgICAgICAgICAgIGlwX3R5cGUgPSAoCiAgICAgICAgICAg
            ICAgICAiSVB2NCIgaWYgdHlwZShpcF9hZGRyZXNzKGVuZHBvaW50X3N0cl9pcCkpIGlzIElQdjRB
            ZGRyZXNzIGVsc2UgIklQdjYiCiAgICAgICAgICAgICkKICAgICAgICBleGNlcHQgVmFsdWVFcnJv
            cjoKICAgICAgICAgICAgaXBfdHlwZSA9ICJGUUROIgogICAgICAgIGlmIG5vdCBwb3J0LmlzZGln
            aXQoKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIlBvcnQg
            bm90IHZhbGlkOiB7cG9ydH0iKQogICAgICAgIGludFBvcnQgPSBpbnQocG9ydCkKICAgICAgICBp
            ZiBpbnRQb3J0IDwgMSBvciBpbnRQb3J0ID4gMioqMTYgLSAxOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiT3V0IG9mIHJhbmdlIHBvcnQgbnVtYmVyOiB7cG9y
            dH0iKQoKICAgICAgICByZXR1cm4gaXBfdHlwZQoKICAgIGRlZiBlbmRwb2ludF9kaWFsKHNlbGYs
            IGVuZHBvaW50X3N0ciwgaXBfdHlwZSwgdGltZW91dD0zLCBjZXJ0PU5vbmUpOgogICAgICAgICMg
            aWYgdGhlICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15IG9uZSwKICAgICAgICAjIGRvbid0
            IHRyeSB0byByZWFjaCBvdXQgdG8gdGhlIGVuZHBvaW50CiAgICAgICAgaWYgaXNpbnN0YW5jZShz
            ZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gIiIsICIiLCAiIgog
            ICAgICAgIGlmIGlwX3R5cGUgPT0gIklQdjYiOgogICAgICAgICAgICB0cnk6CiAgICAgICAgICAg
            ICAgICBlbmRwb2ludF9zdHJfaXAsIGVuZHBvaW50X3N0cl9wb3J0ID0gZW5kcG9pbnRfc3RyLnJz
            cGxpdCgiOiIsIDEpCiAgICAgICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBm
            Ik5vdCBhIHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSAhPSAiWyI6CiAgICAgICAgICAgICAg
            ICBlbmRwb2ludF9zdHJfaXAgPSAiWyIgKyBlbmRwb2ludF9zdHJfaXAgKyAiXSIKICAgICAgICAg
            ICAgZW5kcG9pbnRfc3RyID0gIjoiLmpvaW4oW2VuZHBvaW50X3N0cl9pcCwgZW5kcG9pbnRfc3Ry
            X3BvcnRdKQoKICAgICAgICBwcm90b2NvbHMgPSBbImh0dHAiLCAiaHR0cHMiXQogICAgICAgIHJl
            c3BvbnNlX2Vycm9yID0gTm9uZQogICAgICAgIGZvciBwcmVmaXggaW4gcHJvdG9jb2xzOgogICAg
            ICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBlcCA9IGYie3ByZWZpeH06Ly97ZW5kcG9pbnRf
            c3RyfSIKICAgICAgICAgICAgICAgIHZlcmlmeSA9IE5vbmUKICAgICAgICAgICAgICAgICMgSWYg
            dmVyaWZ5IGlzIHNldCB0byBhIHBhdGggdG8gYSBkaXJlY3RvcnksCiAgICAgICAgICAgICAgICAj
            IHRoZSBkaXJlY3RvcnkgbXVzdCBoYXZlIGJlZW4gcHJvY2Vzc2VkIHVzaW5nIHRoZSBjX3JlaGFz
            aCB1dGlsaXR5IHN1cHBsaWVkIHdpdGggT3BlblNTTC4KICAgICAgICAgICAgICAgIGlmIHByZWZp
            eCA9PSAiaHR0cHMiIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd19za2lwX3RsczoKICAgICAgICAg
            ICAgICAgICAgICB2ZXJpZnkgPSBGYWxzZQogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQsIHZlcmlmeT1GYWxzZSkKICAgICAgICAgICAgICAg
            IGVsaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIGNlcnQ6CiAgICAgICAgICAgICAgICAgICAgdmVy
            aWZ5ID0gY2VydAogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0cy5oZWFkKGVwLCB0aW1l
            b3V0PXRpbWVvdXQsIHZlcmlmeT1jZXJ0KQogICAgICAgICAgICAgICAgZWxzZToKICAgICAgICAg
            ICAgICAgICAgICByID0gcmVxdWVzdHMuaGVhZChlcCwgdGltZW91dD10aW1lb3V0KQogICAgICAg
            ICAgICAgICAgaWYgci5zdGF0dXNfY29kZSA9PSAyMDA6CiAgICAgICAgICAgICAgICAgICAgcmV0
            dXJuIHByZWZpeCwgdmVyaWZ5LCAiIgogICAgICAgICAgICBleGNlcHQgRXhjZXB0aW9uIGFzIGVy
            cjoKICAgICAgICAgICAgICAgIHJlc3BvbnNlX2Vycm9yID0gZXJyCiAgICAgICAgICAgICAgICBj
            b250aW51ZQogICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgIGYidW5hYmxlIHRv
            IGNvbm5lY3QgdG8gZW5kcG9pbnQ6IHtlbmRwb2ludF9zdHJ9LCBmYWlsZWQgZXJyb3I6IHtyZXNw
            b25zZV9lcnJvcn0iCiAgICAgICAgKQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICIiLAog
            ICAgICAgICAgICAiIiwKICAgICAgICAgICAgKCItMSIpLAogICAgICAgICkKCiAgICBkZWYgcGFy
            c2VfY29uZmlnX2ZpbGUoc2VsZiwgY29uZmlnX2ZpbGUpOgogICAgICAgIGNvbmZpZyA9IGNvbmZp
            Z3BhcnNlci5Db25maWdQYXJzZXIoKQogICAgICAgIGNvbmZpZy5yZWFkKGNvbmZpZ19maWxlKQog
            ICAgICAgIGZvciBhcmcgaW4gbGlzdCh2YXJzKHNlbGYuX2FyZ19wYXJzZXIpKToKICAgICAgICAg
            ICAgIyBweXRob24gdHJlYXRzIGZsYWctbmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28g
            Y29udmVydGluZyBiYWNrIHRvIGZsYWctbmFtZSwKICAgICAgICAgICAgIyBzbyB3ZSBjYW4gZ2V0
            IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAgIGFyZ3VtZW50ID0gYXJn
            LnJlcGxhY2UoIl8iLCAiLSIpCiAgICAgICAgICAgIGFyZ3VtZW50VmFsdWUgPSBzdHIoZ2V0YXR0
            cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcpKQogICAgICAgICAgICBjb25maWdfdmFsdWUgPSBjb25m
            aWcuZ2V0KCJDb25maWd1cmF0aW9ucyIsIGFyZ3VtZW50LCBmYWxsYmFjaz1Ob25lKQogICAgICAg
            ICAgICAjIGdpdmUgcHJpb3JpdHkgdG8gY29tbWFuZCBsaW5lIGFyZ3VtZW50LCBpZiBjb21tYW5k
            IGxpbmUgYXJndW1lbnQgaXMgbm90IHByZXNlbnQgdXNlIGNvbmZpZy5pbmkgdmFsdWUsCiAgICAg
            ICAgICAgIGlmIChzdHIoc3lzLmFyZ3YpLmZpbmQoYXJndW1lbnQpID09IC0xKSBhbmQgKAogICAg
            ICAgICAgICAgICAgKGNvbmZpZ192YWx1ZSAhPSBOb25lKSBhbmQgKGNvbmZpZ192YWx1ZSAhPSAi
            IikKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuX19zZXRh
            dHRyX18oYXJnLCBjb25maWdfdmFsdWUpCgogICAgICAgIHJldHVybiBjb25maWcKCiAgICBkZWYg
            X19pbml0X18oc2VsZiwgYXJnX2xpc3Q9Tm9uZSk6CiAgICAgICAgc2VsZi5vdXRfbWFwID0ge30K
            ICAgICAgICBzZWxmLl9leGNsdWRlZF9rZXlzID0gc2V0KCkKICAgICAgICBzZWxmLl9hcmdfcGFy
            c2VyID0gc2VsZi5nZW5fYXJnX3BhcnNlcihhcmdzX3RvX3BhcnNlPWFyZ19saXN0KQogICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGU6CiAgICAgICAgICAgIHNlbGYuY29uZmln
            ID0gc2VsZi5wYXJzZV9jb25maWdfZmlsZShzZWxmLl9hcmdfcGFyc2VyLmNvbmZpZ19maWxlKQog
            ICAgICAgIHNlbGYuX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoKQogICAgICAgIHNlbGYucnVu
            X2FzX3VzZXIgPSBzZWxmLl9hcmdfcGFyc2VyLnJ1bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRw
            dXRfZmlsZSA9IHNlbGYuX2FyZ19wYXJzZXIub3V0cHV0CiAgICAgICAgc2VsZi5jZXBoX2NvbmYg
            PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhfY29uZgogICAgICAgIHNlbGYuY2VwaF9rZXlyaW5nID0g
            c2VsZi5fYXJnX3BhcnNlci5rZXlyaW5nCiAgICAgICAgIyBpZiB1c2VyIG5vdCBwcm92aWRlZCwg
            Z2l2ZSBhIGRlZmF1bHQgdXNlcgogICAgICAgIGlmIG5vdCBzZWxmLnJ1bl9hc191c2VyIGFuZCBu
            b3Qgc2VsZi5fYXJnX3BhcnNlci51cGdyYWRlOgogICAgICAgICAgICBzZWxmLnJ1bl9hc191c2Vy
            ID0gc2VsZi5FWFRFUk5BTF9VU0VSX05BTUUKICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVwZ3JhZGU6CiAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4ID0gc2VsZi5ERUZBVUxUX1JH
            V19QT09MX1BSRUZJWAogICAgICAgIGlmIHNlbGYuY2VwaF9jb25mOgogICAgICAgICAgICBrd2Fy
            Z3MgPSB7fQogICAgICAgICAgICBpZiBzZWxmLmNlcGhfa2V5cmluZzoKICAgICAgICAgICAgICAg
            IGt3YXJnc1siY29uZiJdID0geyJrZXlyaW5nIjogc2VsZi5jZXBoX2tleXJpbmd9CiAgICAgICAg
            ICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKGNvbmZmaWxlPXNlbGYuY2VwaF9jb25mLCAq
            Kmt3YXJncykKICAgICAgICBlbHNlOgogICAgICAgICAgICBzZWxmLmNsdXN0ZXIgPSByYWRvcy5S
            YWRvcygpCiAgICAgICAgICAgIHNlbGYuY2x1c3Rlci5jb25mX3JlYWRfZmlsZSgpCiAgICAgICAg
            c2VsZi5jbHVzdGVyLmNvbm5lY3QoKQoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToKICAgICAgICBp
            ZiBzZWxmLmNsdXN0ZXIuc3RhdGUgPT0gImNvbm5lY3RlZCI6CiAgICAgICAgICAgIHNlbGYuY2x1
            c3Rlci5zaHV0ZG93bigpCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIGlmIHNlbGYu
            X2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2Vw
            aCBmc2lkIikKICAgICAgICByZXR1cm4gc3RyKHNlbGYuY2x1c3Rlci5nZXRfZnNpZCgpKQoKICAg
            IGRlZiBfY29tbW9uX2NtZF9qc29uX2dlbihzZWxmLCBjbWRfanNvbik6CiAgICAgICAgY21kID0g
            anNvbi5kdW1wcyhjbWRfanNvbiwgc29ydF9rZXlzPVRydWUpCiAgICAgICAgcmV0X3ZhbCwgY21k
            X291dCwgZXJyX21zZyA9IHNlbGYuY2x1c3Rlci5tb25fY29tbWFuZChjbWQsIGIiIikKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLnZlcmJvc2U6CiAgICAgICAgICAgIHByaW50KGYiQ29tbWFu
            ZCBJbnB1dDoge2NtZH0iKQogICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgIGYiUmV0
            dXJuIFZhbDoge3JldF92YWx9XG5Db21tYW5kIE91dHB1dDoge2NtZF9vdXR9XG4iCiAgICAgICAg
            ICAgICAgICBmIkVycm9yIE1lc3NhZ2U6IHtlcnJfbXNnfVxuLS0tLS0tLS0tLVxuIgogICAgICAg
            ICAgICApCiAgICAgICAganNvbl9vdXQgPSB7fQogICAgICAgICMgaWYgdGhlcmUgaXMgbm8gZXJy
            b3IgKGkuZTsgcmV0X3ZhbCBpcyBaRVJPKSBhbmQgJ2NtZF9vdXQnIGlzIG5vdCBlbXB0eQogICAg
            ICAgICMgdGhlbiBjb252ZXJ0ICdjbWRfb3V0JyB0byBhIGpzb24gb3V0cHV0CiAgICAgICAgaWYg
            cmV0X3ZhbCA9PSAwIGFuZCBjbWRfb3V0OgogICAgICAgICAgICBqc29uX291dCA9IGpzb24ubG9h
            ZHMoY21kX291dCkKICAgICAgICByZXR1cm4gcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cKCiAg
            ICBkZWYgZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoc2VsZik6CiAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJxdW9ydW1fc3RhdHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9y
            dW4oImNlcGggIiArIGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291
            dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAg
            IyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIidxdW9ydW1fc3RhdHVzJyBjb21tYW5k
            IGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwg
            IT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICApCiAgICAgICAg
            cV9sZWFkZXJfbmFtZSA9IGpzb25fb3V0WyJxdW9ydW1fbGVhZGVyX25hbWUiXQogICAgICAgIHFf
            bGVhZGVyX2RldGFpbHMgPSB7fQogICAgICAgIHFfbGVhZGVyX21hdGNoaW5nX2xpc3QgPSBbCiAg
            ICAgICAgICAgIGwgZm9yIGwgaW4ganNvbl9vdXRbIm1vbm1hcCJdWyJtb25zIl0gaWYgbFsibmFt
            ZSJdID09IHFfbGVhZGVyX25hbWUKICAgICAgICBdCiAgICAgICAgaWYgbGVuKHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3QpID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oIk5vIG1hdGNoaW5nICdtb24nIGRldGFpbHMgZm91bmQiKQogICAgICAgIHFfbGVhZGVy
            X2RldGFpbHMgPSBxX2xlYWRlcl9tYXRjaGluZ19saXN0WzBdCiAgICAgICAgIyBnZXQgdGhlIGFk
            ZHJlc3MgdmVjdG9yIG9mIHRoZSBxdW9ydW0tbGVhZGVyCiAgICAgICAgcV9sZWFkZXJfYWRkcnZl
            YyA9IHFfbGVhZGVyX2RldGFpbHMuZ2V0KCJwdWJsaWNfYWRkcnMiLCB7fSkuZ2V0KCJhZGRydmVj
            IiwgW10pCiAgICAgICAgaXBfYWRkciA9IHN0cihxX2xlYWRlcl9kZXRhaWxzWyJwdWJsaWNfYWRk
            ciJdLnNwbGl0KCIvIilbMF0pCgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudjJfcG9ydF9l
            bmFibGU6CiAgICAgICAgICAgIGlmIHFfbGVhZGVyX2FkZHJ2ZWNbMF1bInR5cGUiXSA9PSAidjIi
            OgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVhZGVyX2FkZHJ2ZWNbMF1bImFkZHIiXQog
            ICAgICAgICAgICBlbGlmIGxlbihxX2xlYWRlcl9hZGRydmVjKSA+IDEgYW5kIHFfbGVhZGVyX2Fk
            ZHJ2ZWNbMV1bInR5cGUiXSA9PSAidjIiOgogICAgICAgICAgICAgICAgaXBfYWRkciA9IHFfbGVh
            ZGVyX2FkZHJ2ZWNbMV1bImFkZHIiXQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAg
            c3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAgICAiJ3YyJyBhZGRyZXNzIHR5cGUg
            bm90IHByZXNlbnQsIGFuZCAndjItcG9ydC1lbmFibGUnIGZsYWcgaXMgcHJvdmlkZWQiCiAgICAg
            ICAgICAgICAgICApCgogICAgICAgIHJldHVybiBmIntzdHIocV9sZWFkZXJfbmFtZSl9PXtpcF9h
            ZGRyfSIKCiAgICBkZWYgX2NvbnZlcnRfaG9zdG5hbWVfdG9faXAoc2VsZiwgaG9zdF9uYW1lLCBw
            b3J0LCBpcF90eXBlKToKICAgICAgICAjIGlmICdjbHVzdGVyJyBpbnN0YW5jZSBpcyBhIGR1bW15
            IHR5cGUsCiAgICAgICAgIyBjYWxsIHRoZSBkdW1teSBpbnN0YW5jZSdzICJjb252ZXJ0IiBtZXRo
            b2QKICAgICAgICBpZiBub3QgaG9zdF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKCJFbXB0eSBob3N0bmFtZSBwcm92aWRlZCIpCiAgICAgICAgaWYgaXNp
            bnN0YW5jZShzZWxmLmNsdXN0ZXIsIER1bW15UmFkb3MpOgogICAgICAgICAgICByZXR1cm4gc2Vs
            Zi5jbHVzdGVyLl9jb252ZXJ0X2hvc3RuYW1lX3RvX2lwKGhvc3RfbmFtZSkKCiAgICAgICAgaWYg
            aXBfdHlwZSA9PSAiRlFETiI6CiAgICAgICAgICAgICMgY2hlY2sgd2hpY2ggaXAgRlFETiBzaG91
            bGQgYmUgY29udmVydGVkIHRvLCBJUHY0IG9yIElQdjYKICAgICAgICAgICAgIyBjaGVjayB0aGUg
            aG9zdCBpcCwgdGhlIGVuZHBvaW50IGlwIHR5cGUgd291bGQgYmUgc2ltaWxhciB0byBob3N0IGlw
            CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAib3JjaCBob3N0IGxzIiwgImZvcm1h
            dCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAjIGlmIHRoZXJlIGlz
            IGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVy
            ZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ29yY2ggaG9zdCBscycgY29tbWFuZCBm
            YWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3Zh
            bCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgIGhvc3RfYWRkciA9IGpzb25fb3V0WzBdWyJhZGRyIl0KICAgICAgICAgICAgIyBh
            ZGQgOjgwIHNhbXBsZSBwb3J0IGluIGlwX3R5cGUsIGFzIF9pbnZhbGlkX2VuZHBvaW50IGFsc28g
            dmVyaWZ5IHBvcnQKICAgICAgICAgICAgaG9zdF9pcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRw
            b2ludChob3N0X2FkZHIgKyAiOjgwIikKICAgICAgICAgICAgaW1wb3J0IHNvY2tldAoKICAgICAg
            ICAgICAgaXAgPSBbXQogICAgICAgICAgICAjIGV4YW1wbGUgb3V0cHV0IFsoPEFkZHJlc3NGYW1p
            bHkuQUZfSU5FVDogMj4sIDxTb2NrZXRLaW5kLlNPQ0tfU1RSRUFNOiAxPiwgNiwgJycsICgnOTMu
            MTg0LjIxNi4zNCcsIDgwKSksIC4uLl0KICAgICAgICAgICAgIyB3ZSBuZWVkIHRvIGdldCA5My4x
            ODQuMjE2LjM0IHNvIGl0IHdvdWxkIGJlIGlwWzBdWzRdWzBdCiAgICAgICAgICAgIGlmIGhvc3Rf
            aXBfdHlwZSA9PSAiSVB2NiI6CiAgICAgICAgICAgICAgICBpcCA9IHNvY2tldC5nZXRhZGRyaW5m
            bygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZhbWlseT1zb2NrZXQuQUZf
            SU5FVDYsIHByb3RvPXNvY2tldC5JUFBST1RPX1RDUAogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbGlmIGhvc3RfaXBfdHlwZSA9PSAiSVB2NCI6CiAgICAgICAgICAgICAgICBpcCA9IHNv
            Y2tldC5nZXRhZGRyaW5mbygKICAgICAgICAgICAgICAgICAgICBob3N0X25hbWUsIHBvcnQsIGZh
            bWlseT1zb2NrZXQuQUZfSU5FVCwgcHJvdG89c29ja2V0LklQUFJPVE9fVENQCiAgICAgICAgICAg
            ICAgICApCiAgICAgICAgICAgIGRlbCBzb2NrZXQKICAgICAgICAgICAgcmV0dXJuIGlwWzBdWzRd
            WzBdCiAgICAgICAgcmV0dXJuIGhvc3RfbmFtZQoKICAgIGRlZiBnZXRfYWN0aXZlX2FuZF9zdGFu
            ZGJ5X21ncnMoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAg
            ICAgICAgICByZXR1cm4gIiIsIHNlbGYuZHJ5X3J1bigiY2VwaCBzdGF0dXMiKQogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYuX2FyZ19wYXJzZXIubW9uaXRvcmluZ19lbmRw
            b2ludF9wb3J0CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gc2VsZi5fYXJn
            X3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50CiAgICAgICAgc3RhbmRieV9tZ3JzID0gW10KICAg
            ICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0OgogICAgICAgICAgICBjbWRf
            anNvbiA9IHsicHJlZml4IjogInN0YXR1cyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgICAg
            IHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1w
            dCwKICAgICAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAg
            ICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICAgICAgIidtZ3Igc2VydmljZXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAg
            ICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlf
            T1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBtb25pdG9yaW5nX2Vu
            ZHBvaW50ID0gKAogICAgICAgICAgICAgICAganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0
            KCJzZXJ2aWNlcyIsIHt9KS5nZXQoInByb21ldGhldXMiLCAiIikKICAgICAgICAgICAgKQogICAg
            ICAgICAgICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludDoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgImNhbid0IGZp
            bmQgbW9uaXRvcmluZ19lbmRwb2ludCwgcHJvbWV0aGV1cyBtb2R1bGUgbWlnaHQgbm90IGJlIGVu
            YWJsZWQsICIKICAgICAgICAgICAgICAgICAgICAiZW5hYmxlIHRoZSBtb2R1bGUgYnkgcnVubmlu
            ZyAnY2VwaCBtZ3IgbW9kdWxlIGVuYWJsZSBwcm9tZXRoZXVzJyIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgIyBub3cgY2hlY2sgdGhlIHN0YW5kLWJ5IG1nci1zCiAgICAgICAgICAgIHN0
            YW5kYnlfYXJyID0ganNvbl9vdXQuZ2V0KCJtZ3JtYXAiLCB7fSkuZ2V0KCJzdGFuZGJ5cyIsIFtd
            KQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5IGluIHN0YW5kYnlfYXJyOgogICAgICAgICAg
            ICAgICAgaWYgIm5hbWUiIGluIGVhY2hfc3RhbmRieS5rZXlzKCk6CiAgICAgICAgICAgICAgICAg
            ICAgc3RhbmRieV9tZ3JzLmFwcGVuZChlYWNoX3N0YW5kYnlbIm5hbWUiXSkKICAgICAgICAgICAg
            dHJ5OgogICAgICAgICAgICAgICAgcGFyc2VkX2VuZHBvaW50ID0gdXJscGFyc2UobW9uaXRvcmlu
            Z19lbmRwb2ludCkKICAgICAgICAgICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAg
            ICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYi
            aW52YWxpZCBlbmRwb2ludDoge21vbml0b3JpbmdfZW5kcG9pbnR9IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBwYXJzZWRfZW5kcG9p
            bnQuaG9zdG5hbWUKICAgICAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoK
            ICAgICAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHN0cihwYXJzZWRfZW5k
            cG9pbnQucG9ydCkKCiAgICAgICAgIyBpZiBtb25pdG9yaW5nIGVuZHBvaW50IHBvcnQgaXMgbm90
            IHNldCwgcHV0IGEgZGVmYXVsdCBtb24gcG9ydAogICAgICAgIGlmIG5vdCBtb25pdG9yaW5nX2Vu
            ZHBvaW50X3BvcnQ6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCA9IHNlbGYu
            REVGQVVMVF9NT05JVE9SSU5HX0VORFBPSU5UX1BPUlQKCiAgICAgICAgIyB1c2VyIGNvdWxkIGdp
            dmUgY29tbWEgYW5kIHNwYWNlIHNlcGFyYXRlZCBpbnB1dHMgKGxpa2UgLS1tb25pdG9yaW5nLWVu
            ZHBvaW50PSI8aXAxPiwgPGlwMj4iKQogICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlz
            dCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdC5yZXBsYWNlKCIsIiwgIiAiKQogICAgICAg
            IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9zcGxpdCA9IG1vbml0b3JpbmdfZW5kcG9pbnRf
            aXBfbGlzdC5zcGxpdCgpCiAgICAgICAgIyBpZiBtb25pdG9yaW5nLWVuZHBvaW50IGNvdWxkIG5v
            dCBiZSBmb3VuZCwgcmFpc2UgYW4gZXJyb3IKICAgICAgICBpZiBsZW4obW9uaXRvcmluZ19lbmRw
            b2ludF9pcF9saXN0X3NwbGl0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKCJObyAnbW9uaXRvcmluZy1lbmRwb2ludCcgZm91bmQiKQogICAgICAgICMg
            Zmlyc3QgaXAgaXMgdHJlYXRlZCBhcyB0aGUgbWFpbiBtb25pdG9yaW5nLWVuZHBvaW50CiAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCA9IG1vbml0b3JpbmdfZW5kcG9pbnRfaXBfbGlzdF9z
            cGxpdFswXQogICAgICAgICMgcmVzdCBvZiB0aGUgaXAtcyBhcmUgYWRkZWQgdG8gdGhlICdzdGFu
            ZGJ5X21ncnMnIGxpc3QKICAgICAgICBzdGFuZGJ5X21ncnMuZXh0ZW5kKG1vbml0b3JpbmdfZW5k
            cG9pbnRfaXBfbGlzdF9zcGxpdFsxOl0pCiAgICAgICAgZmFpbGVkX2lwID0gbW9uaXRvcmluZ19l
            bmRwb2ludF9pcAoKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50ID0gIjoiLmpvaW4oCiAgICAg
            ICAgICAgIFttb25pdG9yaW5nX2VuZHBvaW50X2lwLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnRd
            CiAgICAgICAgKQogICAgICAgIGlwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KG1vbml0
            b3JpbmdfZW5kcG9pbnQpCiAgICAgICAgdHJ5OgogICAgICAgICAgICBtb25pdG9yaW5nX2VuZHBv
            aW50X2lwID0gc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydCwgaXBfdHlwZQog
            ICAgICAgICAgICApCiAgICAgICAgICAgICMgY29sbGVjdCBhbGwgdGhlICdzdGFuZC1ieScgbWdy
            IGlwcwogICAgICAgICAgICBtZ3JfaXBzID0gW10KICAgICAgICAgICAgZm9yIGVhY2hfc3RhbmRi
            eV9tZ3IgaW4gc3RhbmRieV9tZ3JzOgogICAgICAgICAgICAgICAgZmFpbGVkX2lwID0gZWFjaF9z
            dGFuZGJ5X21ncgogICAgICAgICAgICAgICAgbWdyX2lwcy5hcHBlbmQoCiAgICAgICAgICAgICAg
            ICAgICAgc2VsZi5fY29udmVydF9ob3N0bmFtZV90b19pcCgKICAgICAgICAgICAgICAgICAgICAg
            ICAgZWFjaF9zdGFuZGJ5X21nciwgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAg
            ICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiQ29udmVyc2lvbiBvZiBob3N0OiB7ZmFpbGVkX2lwfSB0byBJUCBmYWlsZWQuICIKICAgICAg
            ICAgICAgICAgICJQbGVhc2UgZW50ZXIgdGhlIElQIGFkZHJlc3NlcyBvZiBhbGwgdGhlIGNlcGgt
            bWdycyB3aXRoIHRoZSAnLS1tb25pdG9yaW5nLWVuZHBvaW50JyBmbGFnIgogICAgICAgICAgICAp
            CgogICAgICAgIF8sIF8sIGVyciA9IHNlbGYuZW5kcG9pbnRfZGlhbChtb25pdG9yaW5nX2VuZHBv
            aW50LCBpcF90eXBlKQogICAgICAgIGlmIGVyciA9PSAiLTEiOgogICAgICAgICAgICByYWlzZSBF
            eGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGVycikKICAgICAgICAjIGFkZCB0aGUgdmFsaWRhdGVk
            IGFjdGl2ZSBtZ3IgSVAgaW50byB0aGUgZmlyc3QgaW5kZXgKICAgICAgICBtZ3JfaXBzLmluc2Vy
            dCgwLCBtb25pdG9yaW5nX2VuZHBvaW50X2lwKQogICAgICAgIGFsbF9tZ3JfaXBzX3N0ciA9ICIs
            Ii5qb2luKG1ncl9pcHMpCiAgICAgICAgcmV0dXJuIGFsbF9tZ3JfaXBzX3N0ciwgbW9uaXRvcmlu
            Z19lbmRwb2ludF9wb3J0CgogICAgZGVmIGNoZWNrX3VzZXJfZXhpc3Qoc2VsZiwgdXNlcik6CiAg
            ICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJhdXRoIGdldCIsICJlbnRpdHkiOiBmInt1c2Vy
            fSIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIF8gPSBzZWxm
            Ll9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgIGlmIHJldF92YWwgIT0gMCBv
            ciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJldHVybiAiIgogICAgICAgIHJldHVy
            biBzdHIoanNvbl9vdXRbMF1bImtleSJdKQoKICAgIGRlZiBnZXRfY2VwaGZzX3Byb3Zpc2lvbmVy
            X2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1jZXBo
            ZnMtcHJvdmlzaW9uZXIiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAi
            YWxsb3cgcnciLAogICAgICAgICAgICAib3NkIjogImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRh
            dGE9KiIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQogICAgICAgIGlm
            IHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb246CiAgICAgICAgICAg
            IGs4c19jbHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICAgICAgaWYgazhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIms4c19jbHVz
            dGVyX25hbWUgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMtY2x1c3Rlci1uYW1lJyBm
            bGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBjZXBoZnNfZmlsZXN5c3RlbSA9IHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBpZiBjZXBo
            ZnNfZmlsZXN5c3RlbSA9PSAiIjoKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0eX0t
            e2s4c19jbHVzdGVyX25hbWV9IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZW50
            aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0te2NlcGhmc19maWxlc3lzdGVtfSIK
            ICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gZiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFk
            YXRhPXtjZXBoZnNfZmlsZXN5c3RlbX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAg
            ICBkZWYgZ2V0X2NlcGhmc19ub2RlX2NhcHNfYW5kX2VudGl0eShzZWxmKToKICAgICAgICBlbnRp
            dHkgPSAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIKICAgICAgICBjYXBzID0gewogICAgICAgICAg
            ICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAgICAg
            ICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAiYWxsb3cgcncgdGFn
            IGNlcGhmcyAqPSoiLAogICAgICAgICAgICAibWRzIjogImFsbG93IHJ3IiwKICAgICAgICB9CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAg
            ICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJf
            bmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAgICAgICAgICAg
            ICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAi
            azhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVy
            LW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGNlcGhmc19maWxlc3lz
            dGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAg
            IGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgogICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7
            ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVz
            eXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBo
            ZnMgKj17Y2VwaGZzX2ZpbGVzeXN0ZW19IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5Cgog
            ICAgZGVmIGdldF9lbnRpdHkoCiAgICAgICAgc2VsZiwKICAgICAgICBlbnRpdHksCiAgICAgICAg
            cmJkX3Bvb2xfbmFtZSwKICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLAogICAgICAgIGs4c19j
            bHVzdGVyX25hbWUsCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgKToKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHJiZF9wb29sX25hbWUuY291bnQoIi4iKSAhPSAwCiAgICAgICAgICAgIG9y
            IHJiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgIG9yIGFsaWFzX3JiZF9w
            b29sX25hbWUgIT0gIiIKICAgICAgICAgICAgIyBjaGVja2luZyBhbGlhc19yYmRfcG9vbF9uYW1l
            IGlzIG5vdCBlbXB0eSBhcyB0aGVyZSBtYXliZSBhIHNwZWNpYWwgY2hhcmFjdGVyIHVzZWQgb3Ro
            ZXIgdGhhbiAuIG9yIF8KICAgICAgICApOgogICAgICAgICAgICBpZiBhbGlhc19yYmRfcG9vbF9u
            YW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICAgICAicGxlYXNlIHNldCB0aGUgJy0tYWxpYXMtcmJkLWRhdGEt
            cG9vbC1uYW1lJyBmbGFnIGFzIHRoZSByYmQgZGF0YSBwb29sIG5hbWUgY29udGFpbnMgJy4nIG9y
            ICdfJyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAg
            YWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgICAgIG9yIGFs
            aWFzX3JiZF9wb29sX25hbWUuY291bnQoIl8iKSAhPSAwCiAgICAgICAgICAgICk6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICInLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgdmFsdWUgc2hvdWxkIG5vdCBj
            b250YWluICcuJyBvciAnXyciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVudGl0eSA9
            IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXthbGlhc19yYmRfcG9vbF9uYW1lfSIKICAg
            ICAgICBlbHNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfS17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZToKICAgICAg
            ICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17cmFkb3NfbmFtZXNwYWNlfSIKICAgICAgICByZXR1
            cm4gZW50aXR5CgogICAgZGVmIGdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNl
            bGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21tYW5k
            ICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cgcnciLAogICAgICAg
            ICAgICAib3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xf
            bmFtZQogICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNf
            Y2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJz
            ZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJi
            ZC1kYXRhLXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYg
            azhzX2NsdXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAg
            ICAgICAgZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAg
            ICAgIGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1l
            LAogICAgICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAg
            ICAgIGlmIHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFt
            ZX0gbmFtZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAg
            ICAgICBlbHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBv
            b2w9e3JiZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVm
            IGdldF9yYmRfbm9kZV9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gImNs
            aWVudC5jc2ktcmJkLW5vZGUiCiAgICAgICAgY2FwcyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJw
            cm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAi
            b3NkIjogInByb2ZpbGUgcmJkIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9
            IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGFsaWFzX3Ji
            ZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLmFsaWFzX3JiZF9kYXRhX3Bvb2xfbmFtZQog
            ICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rl
            cl9uYW1lCiAgICAgICAgICAgIHJhZG9zX25hbWVzcGFjZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFk
            b3NfbmFtZXNwYWNlCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJtYW5kYXRvcnkgZmxhZyBub3QgZm91bmQsIHBsZWFzZSBzZXQgdGhlICctLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgazhzX2Ns
            dXN0ZXJfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5vdCBmb3VuZCwg
            cGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgZW50aXR5ID0gc2VsZi5nZXRfZW50aXR5KAogICAgICAgICAgICAgICAg
            ZW50aXR5LAogICAgICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSwKICAgICAgICAgICAgICAgIGFs
            aWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgICAgICAgICBrOHNfY2x1c3Rlcl9uYW1lLAogICAg
            ICAgICAgICAgICAgcmFkb3NfbmFtZXNwYWNlLAogICAgICAgICAgICApCiAgICAgICAgICAgIGlm
            IHJhZG9zX25hbWVzcGFjZSAhPSAiIjoKICAgICAgICAgICAgICAgIGNhcHNbIm9zZCJdID0gKAog
            ICAgICAgICAgICAgICAgICAgIGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0gbmFt
            ZXNwYWNlPXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSBmInByb2ZpbGUgcmJkIHBvb2w9e3Ji
            ZF9wb29sX25hbWV9IgoKICAgICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9k
            ZWZhdWx0VXNlcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0gc2VsZi5y
            dW5fYXNfdXNlcgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLAogICAg
            ICAgICAgICAibWdyIjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwKICAgICAgICAgICAgIm9zZCI6
            IGYicHJvZmlsZSByYmQtcmVhZC1vbmx5LCBhbGxvdyByd3ggcG9vbD17c2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5tZXRhLCBhbGxvdyByIHBvb2w9LnJndy5yb290LCBhbGxv
            dyBydyBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2ws
            IGFsbG93IHJ4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubG9n
            LCBhbGxvdyB4IHBvb2w9e3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuYnVj
            a2V0cy5pbmRleCIsCiAgICAgICAgICAgICJtZHMiOiAiYWxsb3cgKiIsCiAgICAgICAgfQoKICAg
            ICAgICByZXR1cm4gY2FwcywgZW50aXR5CgogICAgZGVmIGdldF9jYXBzX2FuZF9lbnRpdHkoc2Vs
            ZiwgdXNlcl9uYW1lKToKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIi
            IGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lv
            bmVyIiAhPSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3Ry
            aWN0ZWRfYXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRf
            Y2VwaGZzX3Byb3Zpc2lvbmVyX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5j
            c2ktY2VwaGZzLW5vZGUiIGluIHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5jc2kt
            Y2VwaGZzLW5vZGUiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAgIHJldHVybiBz
            ZWxmLmdldF9jZXBoZnNfbm9kZV9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQu
            Y3NpLXJiZC1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50
            LmNzaS1yYmQtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24gPSBUcnVlCiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmdldF9yYmRfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KCkKICAgICAg
            ICBpZiAiY2xpZW50LmNzaS1yYmQtbm9kZSIgaW4gdXNlcl9uYW1lOgogICAgICAgICAgICBpZiAi
            Y2xpZW50LmNzaS1yYmQtbm9kZSIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAg
            cmV0dXJuIHNlbGYuZ2V0X3JiZF9ub2RlX2NhcHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNs
            aWVudC5oZWFsdGhjaGVja2VyIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQu
            aGVhbHRoY2hlY2tlciIgIT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJu
            IHNlbGYuZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2VudGl0eSgpCgogICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgIGYibm8gdXNlciBmb3VuZCB3aXRo
            IHVzZXJfbmFtZToge3VzZXJfbmFtZX0sICIKICAgICAgICAgICAgImdldF9jYXBzX2FuZF9lbnRp
            dHkgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2VwaENTSUtl
            eXJpbmdfdXNlcihzZWxmLCB1c2VyKToKICAgICAgICAiIiIKICAgICAgICBjb21tYW5kOiBjZXBo
            IGF1dGggZ2V0LW9yLWNyZWF0ZSBjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciBtb24gJ2Fs
            bG93IHInIG1nciAnYWxsb3cgcncnIG9zZCAnYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0q
            JwogICAgICAgICIiIgogICAgICAgIGNhcHMsIGVudGl0eSA9IHNlbGYuZ2V0X2NhcHNfYW5kX2Vu
            dGl0eSh1c2VyKQogICAgICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1
            dGggZ2V0LW9yLWNyZWF0ZSIsCiAgICAgICAgICAgICJlbnRpdHkiOiBlbnRpdHksCiAgICAgICAg
            ICAgICJjYXBzIjogW2NhcCBmb3IgY2FwX2xpc3QgaW4gbGlzdChjYXBzLml0ZW1zKCkpIGZvciBj
            YXAgaW4gY2FwX2xpc3RdLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0K
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4g
            KAogICAgICAgICAgICAgICAgc2VsZi5kcnlfcnVuKAogICAgICAgICAgICAgICAgICAgICJjZXBo
            ICIKICAgICAgICAgICAgICAgICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAg
            ICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50aXR5Il0KICAg
            ICAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgICAgICsgIiAiLmpvaW4oY21k
            X2pzb25bImNhcHMiXSkKICAgICAgICAgICAgICAgICksCiAgICAgICAgICAgICAgICAiIiwKICAg
            ICAgICAgICAgKQogICAgICAgICMgY2hlY2sgaWYgdXNlciBhbHJlYWR5IGV4aXN0CiAgICAgICAg
            dXNlcl9rZXkgPSBzZWxmLmNoZWNrX3VzZXJfZXhpc3QoZW50aXR5KQogICAgICAgIGlmIHVzZXJf
            a2V5ICE9ICIiOgogICAgICAgICAgICByZXR1cm4gdXNlcl9rZXksIGYie2VudGl0eS5zcGxpdCgn
            LicsIDEpWzFdfSIKICAgICAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4nLDEpWzFdIHRvIHJlbmFt
            ZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1ub2RlCgogICAgICAgIHJl
            dF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRf
            anNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAg
            ICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGdl
            dC1vci1jcmVhdGUge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAgICAgIGYi
            RXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJ
            U1R9IgogICAgICAgICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0p
            LCBmIntlbnRpdHkuc3BsaXQoJy4nLCAxKVsxXX0iCiAgICAgICAgIyBlbnRpdHkuc3BsaXQoJy4n
            LDEpWzFdIHRvIHJlbmFtZSBlbnRpdHkoY2xpZW50LmNzaS1yYmQtbm9kZSkgYXMgY3NpLXJiZC1u
            b2RlCgogICAgZGVmIGdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoc2VsZik6CiAgICAgICAg
            Y21kX2pzb24gPSB7InByZWZpeCI6ICJmcyBscyIsICJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAg
            aWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlf
            cnVuKCJjZXBoICIgKyBjbWRfanNvblsicHJlZml4Il0pCiAgICAgICAgcmV0X3ZhbCwganNvbl9v
            dXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAg
            ICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsIHJlcG9ydCBhbiBlcnJvcgog
            ICAgICAgIGlmIHJldF92YWwgIT0gMDoKICAgICAgICAgICAgIyBpZiBmcyBhbmQgZGF0YV9wb29s
            IGFyZ3VtZW50cyBhcmUgbm90IHNldCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSA9
            PSAiIgogICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9v
            bF9uYW1lID09ICIiCiAgICAgICAgICAgICk6CiAgICAgICAgICAgICAgICByZXR1cm4KICAgICAg
            ICAgICAgIyBpZiB1c2VyIGhhcyBwcm92aWRlZCBhbnkgb2YgdGhlCiAgICAgICAgICAgICMgJy0t
            Y2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgb3IgJy0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lJyBhcmd1
            bWVudHMsCiAgICAgICAgICAgICMgcmFpc2UgYW4gZXhjZXB0aW9uIGFzIHdlIGFyZSB1bmFibGUg
            dG8gdmVyaWZ5IHRoZSBhcmdzCiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNl
            cHRpb24oCiAgICAgICAgICAgICAgICBmIidmcyBscycgY2VwaCBjYWxsIGZhaWxlZCB3aXRoIGVy
            cm9yOiB7ZXJyX21zZ30iCiAgICAgICAgICAgICkKCiAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQg
            PSB7fQogICAgICAgICMgaWYgJy0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZScgYXJndW1lbnQgaXMg
            cHJvdmlkZWQsCiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBwcm92aWRlZCBmaWxlc3lzdGVt
            LW5hbWUgZXhpc3RzIG9yIG5vdAogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZToKICAgICAgICAgICAgIyBnZXQgdGhlIG1hdGNoaW5nIGxpc3QKICAgICAg
            ICAgICAgbWF0Y2hpbmdfanNvbl9vdXRfbGlzdCA9IFsKICAgICAgICAgICAgICAgIG1hdGNoZWQK
            ICAgICAgICAgICAgICAgIGZvciBtYXRjaGVkIGluIGpzb25fb3V0CiAgICAgICAgICAgICAgICBp
            ZiBzdHIobWF0Y2hlZFsibmFtZSJdKSA9PSBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lz
            dGVtX25hbWUKICAgICAgICAgICAgXQogICAgICAgICAgICAjIHVuYWJsZSB0byBmaW5kIGEgbWF0
            Y2hpbmcgZnMtbmFtZSwgcmFpc2UgYW4gZXJyb3IKICAgICAgICAgICAgaWYgbGVuKG1hdGNoaW5n
            X2pzb25fb3V0X2xpc3QpID09IDA6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiRmlsZXN5c3RlbSBwcm92aWRlZCwg
            J3tzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWV9JywgIgogICAgICAgICAg
            ICAgICAgICAgIGYiaXMgbm90IGZvdW5kIGluIHRoZSBmcy1saXN0OiB7W3N0cih4WyduYW1lJ10p
            IGZvciB4IGluIGpzb25fb3V0XX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1hdGNo
            aW5nX2pzb25fb3V0ID0gbWF0Y2hpbmdfanNvbl9vdXRfbGlzdFswXQogICAgICAgICMgaWYgY2Vw
            aGZzIGZpbGVzeXN0ZW0gbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgIyB0cnkgdG8gZ2V0
            IGEgZGVmYXVsdCBmcyBuYW1lIGJ5IGRvaW5nIHRoZSBmb2xsb3dpbmcKICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAjIGEuIGNoZWNrIGlmIHRoZXJlIGlzIG9ubHkgb25lIGZpbGVzeXN0ZW0gaXMg
            cHJlc2VudAogICAgICAgICAgICBpZiBsZW4oanNvbl9vdXQpID09IDE6CiAgICAgICAgICAgICAg
            ICBtYXRjaGluZ19qc29uX291dCA9IGpzb25fb3V0WzBdCiAgICAgICAgICAgICMgYi4gb3IgZWxz
            ZSwgY2hlY2sgaWYgZGF0YV9wb29sIG5hbWUgaXMgcHJvdmlkZWQKICAgICAgICAgICAgZWxpZiBz
            ZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFtZToKICAgICAgICAgICAgICAgICMg
            YW5kIGlmIHByZXNlbnQsIGNoZWNrIHdoZXRoZXIgdGhlcmUgZXhpc3RzIGEgZnMgd2hpY2ggaGFz
            IHRoZSBkYXRhX3Bvb2wKICAgICAgICAgICAgICAgIGZvciBlYWNoSiBpbiBqc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19kYXRhX3Bvb2xfbmFt
            ZSBpbiBlYWNoSlsiZGF0YV9wb29scyJdOgogICAgICAgICAgICAgICAgICAgICAgICBtYXRjaGlu
            Z19qc29uX291dCA9IGVhY2hKCiAgICAgICAgICAgICAgICAgICAgICAgIGJyZWFrCiAgICAgICAg
            ICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIG1hdGNoaW5nIGZzIGV4aXN0cywgdGhhdCBtZWFucyBw
            cm92aWRlZCBkYXRhX3Bvb2wgbmFtZSBpcyBpbnZhbGlkCiAgICAgICAgICAgICAgICBpZiBub3Qg
            bWF0Y2hpbmdfanNvbl9vdXQ6CiAgICAgICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFp
            bHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAgICAgZiJQcm92aWRlZCBkYXRhX3Bv
            b2wgbmFtZSwge3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfSwiCiAgICAg
            ICAgICAgICAgICAgICAgICAgICIgZG9lcyBub3QgZXhpc3RzIgogICAgICAgICAgICAgICAgICAg
            ICkKICAgICAgICAgICAgIyBjLiBpZiBub3RoaW5nIGlzIHNldCBhbmQgY291bGRuJ3QgZmluZCBh
            IGRlZmF1bHQsCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICAjIGp1c3QgcmV0dXJu
            IHNpbGVudGx5CiAgICAgICAgICAgICAgICByZXR1cm4KCiAgICAgICAgaWYgbWF0Y2hpbmdfanNv
            bl9vdXQ6CiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSA9IHN0cihtYXRjaGluZ19qc29uX291dFsibmFtZSJdKQogICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLmNlcGhmc19tZXRhZGF0YV9wb29sX25hbWUgPSBzdHIoCiAgICAgICAgICAgICAgICBt
            YXRjaGluZ19qc29uX291dFsibWV0YWRhdGFfcG9vbCJdCiAgICAgICAgICAgICkKCiAgICAgICAg
            aWYgaXNpbnN0YW5jZShtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdLCBsaXN0KToKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBoYXMgYWxyZWFkeSBwcm92aWRlZCBkYXRhLXBvb2wtbmFt
            ZSwKICAgICAgICAgICAgIyB0aHJvdWdoIC0tY2VwaGZzLWRhdGEtcG9vbC1uYW1lCiAgICAgICAg
            ICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAg
            ICAgICAgIyBpZiB0aGUgcHJvdmlkZWQgbmFtZSBpcyBub3QgbWF0Y2hpbmcgd2l0aCB0aGUgb25l
            IGluIHRoZSBsaXN0CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgICAgICAgICBu
            b3QgaW4gbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXQogICAgICAgICAgICAgICAgKToK
            ICAgICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAg
            ICAgICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lOiAne3NlbGYuX2Fy
            Z19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lfScsICIKICAgICAgICAgICAgICAgICAgICAg
            ICAgImRvZXNuJ3QgbWF0Y2ggZnJvbSB0aGUgZGF0YS1wb29scyBsaXN0OiAiCiAgICAgICAgICAg
            ICAgICAgICAgICAgIGYie1tzdHIoeCkgZm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFf
            cG9vbHMnXV19IgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiBkYXRhX3Bv
            b2wgbmFtZSBpcyBub3QgcHJvdmlkZWQsCiAgICAgICAgICAgICMgdGhlbiB0cnkgdG8gZmluZCBh
            IGRlZmF1bHQgZGF0YSBwb29sIG5hbWUKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAg
            ICMgaWYgbm8gZGF0YV9wb29scyBleGlzdCwgc2lsZW50bHkgcmV0dXJuCiAgICAgICAgICAgICAg
            ICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXSkgPT0gMDoKICAgICAgICAg
            ICAgICAgICAgICByZXR1cm4KICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lID0gc3RyKAogICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25f
            b3V0WyJkYXRhX3Bvb2xzIl1bMF0KICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgIyBpZiB0
            aGVyZSBhcmUgbW9yZSB0aGFuIG9uZSAnZGF0YV9wb29scycgZXhpc3QsCiAgICAgICAgICAgICMg
            dGhlbiB3YXJuIHRoZSB1c2VyIHRoYXQgd2UgYXJlIHVzaW5nIHRoZSBzZWxlY3RlZCBuYW1lCiAg
            ICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29uX291dFsiZGF0YV9wb29scyJdKSA+IDE6CiAg
            ICAgICAgICAgICAgICBwcmludCgKICAgICAgICAgICAgICAgICAgICAiV0FSTklORzogTXVsdGlw
            bGUgZGF0YSBwb29scyBkZXRlY3RlZDogIgogICAgICAgICAgICAgICAgICAgIGYie1tzdHIoeCkg
            Zm9yIHggaW4gbWF0Y2hpbmdfanNvbl9vdXRbJ2RhdGFfcG9vbHMnXV19XG4iCiAgICAgICAgICAg
            ICAgICAgICAgZiJVc2luZyB0aGUgZGF0YS1wb29sOiAne3NlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lfSdcbiIKICAgICAgICAgICAgICAgICkKCiAgICBkZWYgY3JlYXRlX2No
            ZWNrZXJLZXkoc2VsZiwgdXNlcik6CiAgICAgICAgY2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fw
            c19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21kX2pzb24gPSB7CiAgICAgICAgICAgICJwcmVm
            aXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAgICAgICAgICAgImVudGl0eSI6IGVudGl0eSwK
            ICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBjYXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMo
            KSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAgICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAg
            ICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAg
            IHJldHVybiBzZWxmLmRyeV9ydW4oCiAgICAgICAgICAgICAgICAiY2VwaCAiCiAgICAgICAgICAg
            ICAgICArIGNtZF9qc29uWyJwcmVmaXgiXQogICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICsgY21kX2pzb25bImVudGl0eSJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAg
            ICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAgICAgICApCiAgICAg
            ICAgIyBjaGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYu
            Y2hlY2tfdXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAg
            ICAgICAgIHJldHVybiB1c2VyX2tleQoKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21z
            ZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVy
            ZSBpcyBhbiB1bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3Ig
            bGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBnZXQtb3ItY3JlYXRlIHtzZWxmLnJ1bl9h
            c191c2VyfScgY29tbWFuZCBmYWlsZWRcbiIKICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJf
            bXNnIGlmIHJldF92YWwgIT0gMCBlbHNlIHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAg
            ICAgICApCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5Il0pCgogICAgZGVmIGdl
            dF9jZXBoX2Rhc2hib2FyZF9saW5rKHNlbGYpOgogICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgi
            OiAibWdyIHNlcnZpY2VzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiAr
            IGNtZF9qc29uWyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYu
            X2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1
            bnN1Y2Nlc3NmdWwgYXR0ZW1wdCwKICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25f
            b3V0KSA9PSAwOgogICAgICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIGlmICJkYXNoYm9hcmQi
            IG5vdCBpbiBqc29uX291dDoKICAgICAgICAgICAgcmV0dXJuIE5vbmUKICAgICAgICByZXR1cm4g
            anNvbl9vdXRbImRhc2hib2FyZCJdCgogICAgZGVmIGNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIo
            c2VsZik6CiAgICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAg
            ICAgICAgICJ1c2VyIiwKICAgICAgICAgICAgImNyZWF0ZSIsCiAgICAgICAgICAgICItLXVpZCIs
            CiAgICAgICAgICAgIHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAg
            ICAgICAgICItLWRpc3BsYXktbmFtZSIsCiAgICAgICAgICAgICJSb29rIFJHVyBBZG1pbiBPcHMg
            dXNlciIsCiAgICAgICAgICAgICItLWNhcHMiLAogICAgICAgICAgICAiYnVja2V0cz0qO3VzZXJz
            PSo7dXNhZ2U9cmVhZDttZXRhZGF0YT1yZWFkO3pvbmU9cmVhZCIsCiAgICAgICAgICAgICItLXJn
            dy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUsCiAg
            ICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAg
            ICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigi
            Y2VwaCAiICsgIiAiLmpvaW4oY21kKSkKICAgICAgICB0cnk6CiAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAg
            ICAgICAgICAgIyBpZiB0aGUgdXNlciBhbHJlYWR5IGV4aXN0cywgd2UganVzdCBxdWVyeSBpdAog
            ICAgICAgICAgICBpZiBleGVjRXJyLnJldHVybmNvZGUgPT0gZXJybm8uRUVYSVNUOgogICAgICAg
            ICAgICAgICAgY21kID0gWwogICAgICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAg
            ICAgICAgICAgICAgICAgICAidXNlciIsCiAgICAgICAgICAgICAgICAgICAgImluZm8iLAogICAg
            ICAgICAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5FWFRFUk5B
            TF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctcmVh
            bG0iLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUs
            CiAgICAgICAgICAgICAgICAgICAgIi0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUsCiAgICAgICAgICAgICAgICAg
            ICAgIi0tcmd3LXpvbmUiLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICAgICAgICAgIF0KICAgICAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgICAgICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVy
            cj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxs
            ZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgICAgICBlcnJfbXNnID0g
            KAogICAgICAgICAgICAgICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2Nt
            ZH0uIE91dHB1dDoge2V4ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgICAgICBm
            IkNvZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAg
            ICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShl
            cnJfbXNnKQogICAgICAgICAgICAgICAgICAgIHJldHVybiBOb25lLCBOb25lLCBGYWxzZSwgIi0x
            IgogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAg
            ICAgICAgICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4
            ZWNFcnIub3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0
            dXJuY29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAg
            ICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0
            dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCgogICAgICAgICMgc2VwYXJhdGVseSBhZGQgaW5m
            bz1yZWFkIGNhcHMgZm9yIHJndy1lbmRwb2ludCBpcCB2YWxpZGF0aW9uCiAgICAgICAgaW5mb19j
            YXBfc3VwcG9ydGVkID0gVHJ1ZQogICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgInJhZG9zZ3ct
            YWRtaW4iLAogICAgICAgICAgICAiY2FwcyIsCiAgICAgICAgICAgICJhZGQiLAogICAgICAgICAg
            ICAiLS11aWQiLAogICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAiLS1jYXBzIiwKICAgICAgICAgICAgImluZm89cmVhZCIsCiAgICAg
            ICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Jl
            YWxtX25hbWUsCiAgICAgICAgICAgICItLXJndy16b25lZ3JvdXAiLAogICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmUi
            LAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsCiAgICAgICAgXQog
            ICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQo
            Y21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNh
            bGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSBjZXBoIHZl
            cnNpb24gbm90IHN1cHBvcnRlZCBmb3IgYWRkaW5nIGBpbmZvPXJlYWRgIGNhcChyZ3dfdmFsaWRh
            dGlvbikKICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgImNvdWxkIG5vdCBhZGQgY2Fw
            czogdW5hYmxlIHRvIGFkZCBjYXBzOiBpbmZvPXJlYWRcbiIKICAgICAgICAgICAgICAgIGluIGV4
            ZWNFcnIuc3RkZXJyLmRlY29kZSgidXRmLTgiKQogICAgICAgICAgICAgICAgYW5kIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSAyNDQKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIGluZm9fY2Fw
            X3N1cHBvcnRlZCA9IEZhbHNlCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlcnJf
            bXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7
            Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgZiJD
            b2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVjRXJyLnN0ZGVycn0iCiAgICAg
            ICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAg
            ICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2UsICItMSIKCiAgICAgICAganNv
            bm91dHB1dCA9IGpzb24ubG9hZHMob3V0cHV0KQogICAgICAgIHJldHVybiAoCiAgICAgICAgICAg
            IGpzb25vdXRwdXRbImtleXMiXVswXVsiYWNjZXNzX2tleSJdLAogICAgICAgICAgICBqc29ub3V0
            cHV0WyJrZXlzIl1bMF1bInNlY3JldF9rZXkiXSwKICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9y
            dGVkLAogICAgICAgICAgICAiIiwKICAgICAgICApCgogICAgZGVmIHZhbGlkYXRlX3JiZF9wb29s
            KHNlbGYsIHBvb2xfbmFtZSk6CiAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0
            cyhwb29sX25hbWUpOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9u
            KAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3twb29sX25hbWV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICkKCiAgICBkZWYgaW5pdF9yYmRfcG9vbChzZWxmLCBy
            YmRfcG9vbF9uYW1lKToKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlS
            YWRvcyk6CiAgICAgICAgICAgIHJldHVybgogICAgICAgIGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9w
            ZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAg
            ICAgIHJiZF9pbnN0LnBvb2xfaW5pdChpb2N0eCwgVHJ1ZSkKCiAgICBkZWYgdmFsaWRhdGVfcmFk
            b3NfbmFtZXNwYWNlKHNlbGYpOgogICAgICAgIHJiZF9wb29sX25hbWUgPSBzZWxmLl9hcmdfcGFy
            c2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmJkX21l
            dGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQogICAgICAgIHJhZG9zX25hbWVzcGFj
            ZSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgaWYgcmFkb3NfbmFt
            ZXNwYWNlID09ICIiOgogICAgICAgICAgICByZXR1cm4KICAgICAgICBpZiByYWRvc19uYW1lc3Bh
            Y2UuaXNsb3dlcigpID09IEZhbHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJl
            RXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNl
            LCAne3JhZG9zX25hbWVzcGFjZX0nLCAiCiAgICAgICAgICAgICAgICBmImNvbnRhaW5zIHVwcGVy
            IGNhc2UiCiAgICAgICAgICAgICkKICAgICAgICByYmRfaW5zdCA9IHJiZC5SQkQoKQogICAgICAg
            IGlvY3R4ID0gc2VsZi5jbHVzdGVyLm9wZW5faW9jdHgocmJkX3Bvb2xfbmFtZSkKICAgICAgICBp
            ZiByYmRfaW5zdC5uYW1lc3BhY2VfZXhpc3RzKGlvY3R4LCByYWRvc19uYW1lc3BhY2UpIGlzIEZh
            bHNlOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgZiJUaGUgcHJvdmlkZWQgcmFkb3MgTmFtZXNwYWNlLCAne3JhZG9zX25hbWVzcGFj
            ZX0nLCAiCiAgICAgICAgICAgICAgICBmImlzIG5vdCBmb3VuZCBpbiB0aGUgcG9vbCAne3JiZF9w
            b29sX25hbWV9JyIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRfb3JfY3JlYXRlX3N1YnZvbHVt
            ZV9ncm91cChzZWxmLCBzdWJ2b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUpOgog
            ICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAg
            ICAgICAgICAic3Vidm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAiZ2V0cGF0aCIsCiAgICAgICAg
            ICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwK
            ICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19v
            dXRwdXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9j
            ZXNzLkNhbGxlZFByb2Nlc3NFcnJvcjoKICAgICAgICAgICAgY21kID0gWwogICAgICAgICAgICAg
            ICAgImNlcGgiLAogICAgICAgICAgICAgICAgImZzIiwKICAgICAgICAgICAgICAgICJzdWJ2b2x1
            bWVncm91cCIsCiAgICAgICAgICAgICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgICAgIGNlcGhm
            c19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgICAgICBzdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgIF0KICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgXyA9IHN1YnByb2Nlc3Mu
            Y2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAgICAgICAgICAgZXhj
            ZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICAgICAgcmFpc2Ug
            RXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICBmInN1YnZvbHVt
            ZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQgY3JlYXRlZCIKICAg
            ICAgICAgICAgICAgICkKCiAgICBkZWYgcGluX3N1YnZvbHVtZSgKICAgICAgICBzZWxmLCBzdWJ2
            b2x1bWVfZ3JvdXAsIGNlcGhmc19maWxlc3lzdGVtX25hbWUsIHBpbl90eXBlLCBwaW5fc2V0dGlu
            ZwogICAgKToKICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICJjZXBoIiwKICAgICAgICAgICAg
            ImZzIiwKICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAgInBpbiIsCiAg
            ICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtX25hbWUsCiAgICAgICAgICAgIHN1YnZvbHVtZV9n
            cm91cCwKICAgICAgICAgICAgcGluX3R5cGUsCiAgICAgICAgICAgIHBpbl9zZXR0aW5nLAogICAg
            ICAgIF0KICAgICAgICB0cnk6CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1
            dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3Mu
            Q2FsbGVkUHJvY2Vzc0Vycm9yOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhj
            ZXB0aW9uKAogICAgICAgICAgICAgICAgZiJzdWJ2b2x1bWUgZ3JvdXAge3N1YnZvbHVtZV9ncm91
            cH0gaXMgbm90IGFibGUgdG8gZ2V0IHBpbm5lZCIKICAgICAgICAgICAgKQoKICAgIGRlZiBnZXRf
            cmd3X2ZzaWQoc2VsZiwgYmFzZV91cmwsIHZlcmlmeSk6CiAgICAgICAgYWNjZXNzX2tleSA9IHNl
            bGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX0FDQ0VTU19LRVkiXQogICAgICAgIHNlY3Jl
            dF9rZXkgPSBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0KICAg
            ICAgICByZ3dfZW5kcG9pbnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAogICAgICAg
            IGJhc2VfdXJsID0gYmFzZV91cmwgKyAiOi8vIiArIHJnd19lbmRwb2ludCArICIvYWRtaW4vaW5m
            bz8iCiAgICAgICAgcGFyYW1zID0geyJmb3JtYXQiOiAianNvbiJ9CiAgICAgICAgcmVxdWVzdF91
            cmwgPSBiYXNlX3VybCArIHVybGVuY29kZShwYXJhbXMpCgogICAgICAgIHRyeToKICAgICAgICAg
            ICAgciA9IHJlcXVlc3RzLmdldCgKICAgICAgICAgICAgICAgIHJlcXVlc3RfdXJsLAogICAgICAg
            ICAgICAgICAgYXV0aD1TM0F1dGgoYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgcmd3X2VuZHBvaW50
            KSwKICAgICAgICAgICAgICAgIHZlcmlmeT12ZXJpZnksCiAgICAgICAgICAgICkKICAgICAgICBl
            eGNlcHQgcmVxdWVzdHMuZXhjZXB0aW9ucy5UaW1lb3V0OgogICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKAogICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50Oiwgbm90IGFibGUgdG8g
            Y2FsbCBhZG1pbi1vcHMgYXBpe3Jnd19lbmRwb2ludH0iCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgcmV0dXJuICIiLCAiLTEiCiAgICAgICAgcjEgPSByLmpzb24oKQogICAgICAgIGlmIHIxIGlz
            IE5vbmUgb3IgcjEuZ2V0KCJpbmZvIikgaXMgTm9uZToKICAgICAgICAgICAgc3lzLnN0ZGVyci53
            cml0ZSgKICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0nLCBpcyBpbnZhbGlkLiIKICAgICAgICAgICAgKQog
            ICAgICAgICAgICByZXR1cm4gKAogICAgICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICAgICAi
            LTEiLAogICAgICAgICAgICApCgogICAgICAgIHJldHVybiByMVsiaW5mbyJdWyJzdG9yYWdlX2Jh
            Y2tlbmRzIl1bMF1bImNsdXN0ZXJfaWQiXSwgIiIKCiAgICBkZWYgdmFsaWRhdGVfcmd3X2VuZHBv
            aW50KHNlbGYsIGluZm9fY2FwX3N1cHBvcnRlZCk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXIn
            IGluc3RhbmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91
            dCB0byB0aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVt
            bXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybgoKICAgICAgICByZ3dfZW5kcG9pbnQgPSBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd19lbmRwb2ludAoKICAgICAgICAjIHZhbGlkYXRlIHJndyBlbmRwb2lu
            dCBvbmx5IGlmIGlwIGFkZHJlc3MgaXMgcGFzc2VkCiAgICAgICAgaXBfdHlwZSA9IHNlbGYuX2lu
            dmFsaWRfZW5kcG9pbnQocmd3X2VuZHBvaW50KQoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cg
            ZW5kcG9pbnQgaXMgcmVhY2hhYmxlCiAgICAgICAgY2VydCA9IE5vbmUKICAgICAgICBpZiBub3Qg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfc2tpcF90bHMgYW5kIHNlbGYudmFsaWRhdGVfcmd3X2VuZHBv
            aW50X3Rsc19jZXJ0KCk6CiAgICAgICAgICAgIGNlcnQgPSBzZWxmLl9hcmdfcGFyc2VyLnJnd190
            bHNfY2VydF9wYXRoCiAgICAgICAgYmFzZV91cmwsIHZlcmlmeSwgZXJyID0gc2VsZi5lbmRwb2lu
            dF9kaWFsKHJnd19lbmRwb2ludCwgaXBfdHlwZSwgY2VydD1jZXJ0KQogICAgICAgIGlmIGVyciAh
            PSAiIjoKICAgICAgICAgICAgcmV0dXJuICItMSIKCiAgICAgICAgIyBjaGVjayBpZiB0aGUgcmd3
            IGVuZHBvaW50IGJlbG9uZ3MgdG8gdGhlIHNhbWUgY2x1c3RlcgogICAgICAgICMgb25seSBjaGVj
            ayBpZiBgaW5mb2AgY2FwIGlzIHN1cHBvcnRlZAogICAgICAgIGlmIGluZm9fY2FwX3N1cHBvcnRl
            ZDoKICAgICAgICAgICAgZnNpZCA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgICAgICByZ3dfZnNp
            ZCwgZXJyID0gc2VsZi5nZXRfcmd3X2ZzaWQoYmFzZV91cmwsIHZlcmlmeSkKICAgICAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgogICAgICAgICAgICBp
            ZiBmc2lkICE9IHJnd19mc2lkOgogICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAg
            ICAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCByZ3cgRW5kcG9pbnQsICd7c2VsZi5fYXJn
            X3BhcnNlci5yZ3dfZW5kcG9pbnR9JywgaXMgaW52YWxpZC4gV2UgYXJlIHZhbGlkYXRpbmcgYnkg
            Y2FsbGluZyB0aGUgYWRtaW5vcHMgYXBpIHRocm91Z2ggcmd3LWVuZHBvaW50IGFuZCB2YWxpZGF0
            aW5nIHRoZSBjbHVzdGVyX2lkICd7cmd3X2ZzaWR9JyBpcyBlcXVhbCB0byB0aGUgY2VwaCBjbHVz
            dGVyIGZzaWQgJ3tmc2lkfSciCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgcG9vbCBleGlzdAog
            ICAgICAgICMgb25seSB2YWxpZGF0ZSBpZiByZ3dfcG9vbF9wcmVmaXggaXMgcGFzc2VkIGVsc2Ug
            aXQgd2lsbCB0YWtlIGRlZmF1bHQgdmFsdWUgYW5kIHdlIGRvbid0IGNyZWF0ZSB0aGVzZSBkZWZh
            dWx0IHBvb2xzCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXggIT0g
            ImRlZmF1bHQiOgogICAgICAgICAgICByZ3dfcG9vbHNfdG9fdmFsaWRhdGUgPSBbCiAgICAgICAg
            ICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3Lm1ldGEiLAog
            ICAgICAgICAgICAgICAgIi5yZ3cucm9vdCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdf
            cGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3LmNvbnRyb2wiLAogICAgICAgICAgICAgICAgZiJ7
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2ciLAogICAgICAgICAgICBd
            CiAgICAgICAgICAgIGZvciBfcmd3X3Bvb2xfdG9fdmFsaWRhdGUgaW4gcmd3X3Bvb2xzX3RvX3Zh
            bGlkYXRlOgogICAgICAgICAgICAgICAgaWYgbm90IHNlbGYuY2x1c3Rlci5wb29sX2V4aXN0cyhf
            cmd3X3Bvb2xfdG9fdmFsaWRhdGUpOgogICAgICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3Jp
            dGUoCiAgICAgICAgICAgICAgICAgICAgICAgIGYiVGhlIHByb3ZpZGVkIHBvb2wsICd7X3Jnd19w
            b29sX3RvX3ZhbGlkYXRlfScsIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgICAgICAgICByZXR1cm4gIi0xIgoKICAgICAgICByZXR1cm4gIiIKCiAgICBk
            ZWYgdmFsaWRhdGVfcmd3X211bHRpc2l0ZShzZWxmLCByZ3dfbXVsdGlzaXRlX2NvbmZpZ19uYW1l
            LCByZ3dfbXVsdGlzaXRlX2NvbmZpZyk6CiAgICAgICAgaWYgcmd3X211bHRpc2l0ZV9jb25maWcg
            IT0gIiI6CiAgICAgICAgICAgIGNtZCA9IFsKICAgICAgICAgICAgICAgICJyYWRvc2d3LWFkbWlu
            IiwKICAgICAgICAgICAgICAgIHJnd19tdWx0aXNpdGVfY29uZmlnLAogICAgICAgICAgICAgICAg
            ImdldCIsCiAgICAgICAgICAgICAgICAiLS1yZ3ctIiArIHJnd19tdWx0aXNpdGVfY29uZmlnLAog
            ICAgICAgICAgICAgICAgcmd3X211bHRpc2l0ZV9jb25maWdfbmFtZSwKICAgICAgICAgICAgXQog
            ICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJv
            Y2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3IgYXMgZXhlY0VycjoKICAgICAgICAgICAgICAgIGVycl9t
            c2cgPSAoCiAgICAgICAgICAgICAgICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtj
            bWR9LiBPdXRwdXQ6IHtleGVjRXJyLm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNv
            ZGU6IHtleGVjRXJyLnJldHVybmNvZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAg
            ICAgICAgICAgICkKICAgICAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAg
            ICAgICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgcmV0dXJuICIiCgogICAgZGVmIGNvbnZl
            cnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYsIHZhbHVlKToKICAgICAgICByZXR1cm4g
            dmFsdWUuc3BsaXQoIiwiKQoKICAgIGRlZiByYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9sb2d5
            X2ZsYWdfaXNfbWlzc2luZyhzZWxmKToKICAgICAgICBpZiAoCiAgICAgICAgICAgICgKICAgICAg
            ICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgIT0gIiIKICAgICAgICAg
            ICAgICAgIGFuZCAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgICAgIG9yIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fdmFsdWVzID09ICIiCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICAgICAgb3IgKAogICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAg
            ICAgICAgICAgICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRv
            cG9sb2d5X3Bvb2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNl
            ci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdf
            cGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICAgICAg
            ICAgYW5kICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9n
            eV9mYWlsdXJlX2RvbWFpbl9sYWJlbCA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICApCiAgICAgICAgKToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgICAgICJwcm92aWRlIGFsbCB0aGUgdG9wb2xvZ3kgZmxhZ3MgLS10b3Bv
            bG9neS1wb29scywgLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCwgLS10b3BvbG9neS1m
            YWlsdXJlLWRvbWFpbi12YWx1ZXMiCiAgICAgICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfdG9w
            b2xvZ3lfdmFsdWVzKHNlbGYsIHRvcG9sb2d5X3Bvb2xzLCB0b3BvbG9neV9mZCk6CiAgICAgICAg
            aWYgbGVuKHRvcG9sb2d5X3Bvb2xzKSAhPSBsZW4odG9wb2xvZ3lfZmQpOgogICAgICAgICAgICBy
            YWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJUaGUgcHJv
            dmlkZWQgdG9wb2xvZ3kgcG9vbHMsICd7dG9wb2xvZ3lfcG9vbHN9JywgYW5kICIKICAgICAgICAg
            ICAgICAgIGYidG9wb2xvZ3kgZmFpbHVyZSBkb21haW4sICd7dG9wb2xvZ3lfZmR9JywiCiAgICAg
            ICAgICAgICAgICBmImFyZSBvZiBkaWZmZXJlbnQgbGVuZ3RoLCAne2xlbih0b3BvbG9neV9wb29s
            cyl9JyBhbmQgJ3tsZW4odG9wb2xvZ3lfZmQpfScgcmVzcGN0aXZlbHkiCiAgICAgICAgICAgICkK
            ICAgICAgICByZXR1cm4KCiAgICBkZWYgdmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYs
            IHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAgZm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bv
            b2xzOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3JiZF9wb29sKHBvb2wpCgogICAgZGVmIGlu
            aXRfdG9wb2xvZ3lfcmJkX3Bvb2xzKHNlbGYsIHRvcG9sb2d5X3JiZF9wb29scyk6CiAgICAgICAg
            Zm9yIHBvb2wgaW4gdG9wb2xvZ3lfcmJkX3Bvb2xzOgogICAgICAgICAgICBzZWxmLmluaXRfcmJk
            X3Bvb2wocG9vbCkKCiAgICAjIHRoaXMgd2lsbCByZXR1cm4gdGhlIGZpbmFsIGFyZ3MgdGhhdCBz
            Y3JpcHQgdXNlcyB0byBwcm9jZXNzCiAgICAjIHRoZSBwcmlvcml0eSB0byBzZXQgYSBwYXJ0aWN1
            bGFyIHZhbHVlIGlzLAogICAgIyBjb21tYW5kLWxpbmUtYXJncyA+IGNvbmZpZy5pbmkgZmlsZSB2
            YWx1ZXMgPiBkZWZhdWx0IHZhbHVlcwogICAgZGVmIGdldEZpbmFsVXNlZEFyZ3Moc2VsZik6CiAg
            ICAgICAgYXJndW1lbnQgPSBmIltDb25maWd1cmF0aW9uc11cbiIKICAgICAgICBmb3IgYXJnIGlu
            IHZhcnMoc2VsZi5fYXJnX3BhcnNlcik6CiAgICAgICAgICAgIGlmIHN0cihnZXRhdHRyKHNlbGYu
            X2FyZ19wYXJzZXIsIGFyZykpOgogICAgICAgICAgICAgICAgIyBweXRob24gdHJlYXRzIGZsYWct
            bmFtZSBhcyBmbGFnX25hbWUgaW50ZXJuYWxseSwgc28gY29udmVydGluZyBiYWNrIHRvIGZsYWct
            bmFtZSwKICAgICAgICAgICAgICAgICMgc28gd2UgY2FuIGdldCB0aG9zZSB2YWx1ZXMgZnJvbSBj
            b25maWcgZmlsZQogICAgICAgICAgICAgICAgYXJnVmFsdWUgPSBhcmcucmVwbGFjZSgiXyIsICIt
            IikKICAgICAgICAgICAgICAgICMgZG8gbm90IGFkZCB0aGUgY29uZmlnLWZpbGUgZmxhZyBhbmQg
            YWxzbyBub3QgYWRkIHRoZSBib29sZWFuIGZsYWdzIHdoaWNoIGFyZSBzZXQgdG8gRmFsc2UKICAg
            ICAgICAgICAgICAgICMgYmVjYXVzZSBjb25maWcuaW5pIGZpbGUgdHJlYXRzIGJvb2xlYW4gZmxh
            Z3MgYXMgVHJ1ZSBhbHdheXMKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAg
            ICBhcmdWYWx1ZSAhPSAiY29uZmlnLWZpbGUiCiAgICAgICAgICAgICAgICAgICAgYW5kIGdldGF0
            dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSAhPSBGYWxzZQogICAgICAgICAgICAgICAgKToKICAg
            ICAgICAgICAgICAgICAgICBhcmd1bWVudCArPSBmInthcmdWYWx1ZX0gPSB7c3RyKGdldGF0dHIo
            c2VsZi5fYXJnX3BhcnNlciwgYXJnKSl9XG4iCiAgICAgICAgcmV0dXJuIGFyZ3VtZW50CgogICAg
            ZGVmIF9nZW5fb3V0cHV0X21hcChzZWxmKToKICAgICAgICBpZiBzZWxmLm91dF9tYXA6CiAgICAg
            ICAgICAgIHJldHVybgogICAgICAgICMgc3VwcG9ydCBsZWdhY3kgZmxhZyB3aXRoIHVwZ3JhZGVz
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWU6CiAgICAgICAgICAgIHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIuY2x1c3Rl
            cl9uYW1lCiAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lID0gKAogICAg
            ICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUubG93ZXIoKQogICAgICAg
            ICkgICMgYWx3YXlzIGNvbnZlcnQgY2x1c3RlciBuYW1lIHRvIGxvd2VyY2FzZSBjaGFyYWN0ZXJz
            CiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRh
            X3Bvb2xfbmFtZSkKICAgICAgICBzZWxmLmluaXRfcmJkX3Bvb2woc2VsZi5fYXJnX3BhcnNlci5y
            YmRfZGF0YV9wb29sX25hbWUpCiAgICAgICAgc2VsZi52YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uo
            KQogICAgICAgIHNlbGYuX2V4Y2x1ZGVkX2tleXMuYWRkKCJLOFNfQ0xVU1RFUl9OQU1FIikKICAg
            ICAgICBzZWxmLmdldF9jZXBoZnNfZGF0YV9wb29sX2RldGFpbHMoKQogICAgICAgICMgZG91Ymxl
            IHN0cmluZyBuZWVkZWQgZm9yIHVwc3RyZWFtIGV4cG9ydHMgb2YgZmxhZ3MKICAgICAgICBzZWxm
            Lm91dF9tYXBbIkFSR1MiXSA9IGYnIntzZWxmLmdldEZpbmFsVXNlZEFyZ3MoKX0iJwogICAgICAg
            IHNlbGYub3V0X21hcFsiTkFNRVNQQUNFIl0gPSBzZWxmLl9hcmdfcGFyc2VyLm5hbWVzcGFjZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiSzhTX0NMVVNURVJfTkFNRSJdID0gc2VsZi5fYXJnX3BhcnNl
            ci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0ZT
            SUQiXSA9IHNlbGYuZ2V0X2ZzaWQoKQogICAgICAgIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5B
            TF9VU0VSTkFNRSJdID0gc2VsZi5ydW5fYXNfdXNlcgogICAgICAgIHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0gPSBzZWxmLmdldF9jZXBoX2V4dGVybmFsX21vbl9k
            YXRhKCkKICAgICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUl9TRUNSRVQiXSA9
            IHNlbGYuY3JlYXRlX2NoZWNrZXJLZXkoCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIKICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FS
            RF9MSU5LIl0gPSBzZWxmLmdldF9jZXBoX2Rhc2hib2FyZF9saW5rKCkKICAgICAgICAoCiAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICBz
            ZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJdLAogICAgICAgICkgPSBzZWxm
            LmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLXJiZC1ub2RlIikKICAgICAg
            ICAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQi
            XSwKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVu
            dC5jc2ktcmJkLXByb3Zpc2lvbmVyIikKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09M
            X05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lCiAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDRVBIRlNfTUVUQURBVEFfUE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAg
            IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRhX3Bvb2xfbmFtZQogICAgICAgICkKICAg
            ICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLmNl
            cGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVU
            SF9QRVJNSVNTSU9OIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3Rl
            ZF9hdXRoX3Blcm1pc3Npb24KICAgICAgICApCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQURPU19O
            QU1FU1BBQ0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmFkb3NfbmFtZXNwYWNlCiAgICAgICAgc2Vs
            Zi5vdXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXSA9IHNlbGYuX2FyZ19wYXJzZXIuc3Vidm9sdW1l
            X2dyb3VwCiAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0gPSAi
            IgogICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQiXSA9
            ICIiCiAgICAgICAgIyBjcmVhdGUgQ2VwaEZTIG5vZGUgYW5kIHByb3Zpc2lvbmVyIGtleXJpbmcg
            b25seSB3aGVuIE1EUyBleGlzdHMKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIkNFUEhGU19GU19O
            QU1FIl0gYW5kIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdOgogICAgICAgICAgICAo
            CiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSwK
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1F
            Il0sCiAgICAgICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGll
            bnQuY3NpLWNlcGhmcy1ub2RlIikKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAg
            c2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKCJjbGllbnQuY3NpLWNl
            cGhmcy1wcm92aXNpb25lciIpCiAgICAgICAgICAgICMgY3JlYXRlIHRoZSBkZWZhdWx0ICJjc2ki
            IHN1YnZvbHVtZWdyb3VwCiAgICAgICAgICAgIHNlbGYuZ2V0X29yX2NyZWF0ZV9zdWJ2b2x1bWVf
            Z3JvdXAoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmls
            ZXN5c3RlbV9uYW1lCiAgICAgICAgICAgICkKICAgICAgICAgICAgIyBwaW4gdGhlIGRlZmF1bHQg
            ImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAgICAgICAgc2VsZi5waW5fc3Vidm9sdW1lKAogICAg
            ICAgICAgICAgICAgImNzaSIsIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFt
            ZSwgImRpc3RyaWJ1dGVkIiwgIjEiCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJTVUJWT0xVTUVfR1JPVVAiXToKICAgICAgICAgICAgICAgIHNlbGYuZ2V0X29yX2Ny
            ZWF0ZV9zdWJ2b2x1bWVfZ3JvdXAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5j
            ZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAg
            c2VsZi5waW5fc3Vidm9sdW1lKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIu
            c3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAiZGlzdHJpYnV0ZWQiLAog
            ICAgICAgICAgICAgICAgICAgICIxIiwKICAgICAgICAgICAgICAgICkKICAgICAgICBzZWxmLm91
            dF9tYXBbIlJHV19UTFNfQ0VSVCJdID0gIiIKICAgICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJ
            TkdfRU5EUE9JTlQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBP
            SU5UX1BPUlQiXSA9ICIiCiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIuc2tpcF9tb25p
            dG9yaW5nX2VuZHBvaW50OgogICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9t
            YXBbIk1PTklUT1JJTkdfRU5EUE9JTlQiXSwKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsi
            TU9OSVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICkgPSBzZWxmLmdldF9hY3Rp
            dmVfYW5kX3N0YW5kYnlfbWdycygpCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1F
            Il0gPSBzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZQogICAgICAgIHNlbGYub3V0
            X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdID0gKAogICAgICAgICAgICBzZWxmLnZh
            bGlkYXRlX3JiZF9tZXRhZGF0YV9lY19wb29sX25hbWUoKQogICAgICAgICkKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlRPUE9MT0dZX1BPT0xTIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X3Bv
            b2xzCiAgICAgICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJd
            ID0gKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWlu
            X2xhYmVsCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9E
            T01BSU5fVkFMVUVTIl0gPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lf
            ZmFpbHVyZV9kb21haW5fdmFsdWVzCiAgICAgICAgKQogICAgICAgIGlmICgKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICBhbmQgc2Vs
            Zi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl9sYWJlbCAhPSAiIgogICAgICAg
            ICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMg
            IT0gIiIKICAgICAgICApOgogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3ZhbHVl
            cygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXko
            c2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKSwKICAgICAgICAgICAgICAgIHNlbGYuY29u
            dmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJyYXkoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXQogICAgICAgICAgICAgICAg
            KSwKICAgICAgICAgICAgKQogICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3RvcG9sb2d5X3JiZF9w
            b29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9fYXJy
            YXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIHNlbGYuaW5pdF90b3BvbG9neV9yYmRfcG9vbHMoCiAgICAgICAgICAgICAgICBzZWxmLmNv
            bnZlcnRfY29tbWFfc2VwYXJhdGVkX3RvX2FycmF5KHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSkKICAgICAgICAgICAgKQogICAgICAgIGVsc2U6CiAgICAgICAgICAgIHNlbGYucmFpc2Vf
            ZXhjZXB0aW9uX2lmX2FueV90b3BvbG9neV9mbGFnX2lzX21pc3NpbmcoKQoKICAgICAgICBzZWxm
            Lm91dF9tYXBbIlJHV19QT09MX1BSRUZJWCJdID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXgKICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdID0gIiIKICAgICAgICBp
            ZiBzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludDoKICAgICAgICAgICAgaWYgc2VsZi5fYXJn
            X3BhcnNlci5kcnlfcnVuOgogICAgICAgICAgICAgICAgc2VsZi5jcmVhdGVfcmd3X2FkbWluX29w
            c191c2VyKCkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGlmICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lICE9ICIiCiAgICAgICAg
            ICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lICE9ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVfbmFtZSAh
            PSAiIgogICAgICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZh
            bGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUsICJyZWFsbSIKICAgICAgICAgICAgICAgICAgICApCiAgICAg
            ICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSwgInpv
            bmVncm91cCIKICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAgICAgZXJyID0g
            c2VsZi52YWxpZGF0ZV9yZ3dfbXVsdGlzaXRlKAogICAgICAgICAgICAgICAgICAgICAgICBzZWxm
            Ll9hcmdfcGFyc2VyLnJnd196b25lX25hbWUsICJ6b25lIgogICAgICAgICAgICAgICAgICAgICkK
            CiAgICAgICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcmVhbG1fbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdf
            cGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgPT0gIiIKICAgICAgICAgICAgICAgICkgb3Ig
            KAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3JlYWxtX25hbWUgIT0g
            IiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZWdyb3Vw
            X25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5fYXJnX3BhcnNlci5yZ3df
            em9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgICgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJf
            QUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19B
            RE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgICAgIGluZm9f
            Y2FwX3N1cHBvcnRlZCwKICAgICAgICAgICAgICAgICAgICAgICAgZXJyLAogICAgICAgICAgICAg
            ICAgICAgICkgPSBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICAg
            ICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X2VuZHBvaW50KGluZm9fY2FwX3N1cHBvcnRl
            ZCkKICAgICAgICAgICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJnd190bHNfY2VydF9w
            YXRoOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJd
            ID0gKAogICAgICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9p
            bnRfdGxzX2NlcnQoKQogICAgICAgICAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICAg
            ICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciwgc2V0IHRoZSBSR1dfRU5EUE9JTlQKICAgICAgICAg
            ICAgICAgICAgICBpZiBlcnIgIT0gIi0xIjoKICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJSR1dfRU5EUE9JTlQiXSA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAg
            ICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIGVyciA9ICJQbGVhc2UgcHJv
            dmlkZSBhbGwgdGhlIFJHVyBtdWx0aXNpdGUgcGFyYW1ldGVycyBvciBub25lIG9mIHRoZW0iCiAg
            ICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnIpCgogICAgZGVmIGdlbl9zaGVs
            bF9vdXQoc2VsZik6CiAgICAgICAgc2VsZi5fZ2VuX291dHB1dF9tYXAoKQogICAgICAgIHNoT3V0
            SU8gPSBTdHJpbmdJTygpCiAgICAgICAgZm9yIGssIHYgaW4gc2VsZi5vdXRfbWFwLml0ZW1zKCk6
            CiAgICAgICAgICAgIGlmIHYgYW5kIGsgbm90IGluIHNlbGYuX2V4Y2x1ZGVkX2tleXM6CiAgICAg
            ICAgICAgICAgICBzaE91dElPLndyaXRlKGYiZXhwb3J0IHtrfT17dn17TElORVNFUH0iKQogICAg
            ICAgIHNoT3V0ID0gc2hPdXRJTy5nZXR2YWx1ZSgpCiAgICAgICAgc2hPdXRJTy5jbG9zZSgpCiAg
            ICAgICAgcmV0dXJuIHNoT3V0CgogICAgZGVmIGdlbl9qc29uX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gIiIKICAgICAgICBqc29uX291dCA9IFsKICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgIm5hbWUiOiAiZXh0ZXJuYWwtY2x1c3Rlci11c2VyLWNvbW1hbmQi
            LAogICAgICAgICAgICAgICAgImtpbmQiOiAiQ29uZmlnTWFwIiwKICAgICAgICAgICAgICAgICJk
            YXRhIjogewogICAgICAgICAgICAgICAgICAgICJhcmdzIjogc2VsZi5vdXRfbWFwWyJBUkdTIl0s
            CiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAg
            ICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uLWVuZHBvaW50cyIsCiAgICAgICAgICAgICAg
            ICAia2luZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAg
            ICAgICAgICAgICAgImRhdGEiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfQ0VQSF9NT05f
            REFUQSJdLAogICAgICAgICAgICAgICAgICAgICJtYXhNb25JZCI6ICIwIiwKICAgICAgICAgICAg
            ICAgICAgICAibWFwcGluZyI6ICJ7fSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICB9
            LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgtbW9uIiwK
            ICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAiYWRtaW4tc2VjcmV0IjogImFkbWluLXNlY3JldCIsCiAg
            ICAgICAgICAgICAgICAgICAgImZzaWQiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJ
            RCJdLAogICAgICAgICAgICAgICAgICAgICJtb24tc2VjcmV0IjogIm1vbi1zZWNyZXQiLAogICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgIm5hbWUiOiAicm9vay1jZXBoLW9wZXJhdG9yLWNyZWRzIiwKICAgICAgICAgICAgICAgICJr
            aW5kIjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAg
            ICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0sCiAg
            ICAgICAgICAgICAgICAgICAgInVzZXJLZXkiOiBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxf
            VVNFUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAg
            XQoKICAgICAgICAjIGlmICdNT05JVE9SSU5HX0VORFBPSU5UJyBleGlzdHMsIHRoZW4gb25seSBh
            ZGQgJ21vbml0b3JpbmctZW5kcG9pbnQnIHRvIENsdXN0ZXIKICAgICAgICBpZiAoCiAgICAgICAg
            ICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdCiAgICAgICAgICAgIGFuZCBz
            ZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogIm1vbml0b3JpbmctZW5kcG9pbnQiLAogICAgICAgICAgICAgICAgICAgICJr
            aW5kIjogIkNlcGhDbHVzdGVyIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAg
            ICAgICAgICAgICAgICAgICAgIk1vbml0b3JpbmdFbmRwb2ludCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAgICAgICAgICAiTW9uaXRvcmluZ1Bv
            cnQiOiBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdLAogICAgICAgICAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgIyBp
            ZiAnQ1NJX1JCRF9OT0RFX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1y
            YmQtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsi
            Q1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lf
            UkJEX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJLZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAg
            ICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX1JC
            RF9QUk9WSVNJT05FUl9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktcmJk
            LXByb3Zpc2lvbmVyJyBTZWNyZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21h
            cFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRf
            bWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNS
            RVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInVzZXJJRCI6
            IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JCRF9QUk9WSVNJ
            T05FUl9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQog
            ICAgICAgICAgICApCiAgICAgICAgIyBpZiAnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVQn
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1jc2ktY2VwaGZzLXByb3Zpc2lvbmVyJyBTZWNy
            ZXQKICAgICAgICBpZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZT
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291
            dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSdd
            fSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAgICAgICAgICAg
            ICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImFkbWluSUQiOiBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImFkbWluS2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lP
            TkVSX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAg
            ICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfQ0VQSEZTX05PREVfU0VDUkVUJyBleGlzdHMs
            IHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLWNlcGhmcy1ub2RlJyBTZWNyZXQKICAgICAgICBpZiAo
            CiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJdCiAg
            ICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewog
            ICAgICAgICAgICAgICAgICAgICJuYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBI
            RlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2Vj
            cmV0IiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAg
            ICAgImFkbWluSUQiOiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXkiOiBzZWxmLm91dF9tYXBbIkNTSV9D
            RVBIRlNfTk9ERV9TRUNSRVQiXSwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAg
            ICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRf
            TElOSycgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsnIFNl
            Y3JldAogICAgICAgIGlmIHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9EQVNIQk9BUkRfTElO
            SyJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAg
            ICAgICAgICAgICAgICAgIm5hbWUiOiAicm9vay1jZXBoLWRhc2hib2FyZC1saW5rIiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRh
            IjogewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogImNlcGgtZGFzaGJvYXJkLWxp
            bmsiLAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiUk9P
            S19FWFRFUk5BTF9EQVNIQk9BUkRfTElOSyJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAg
            ICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdSQURPU19OQU1FU1BBQ0Un
            IGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgIlJBRE9TX05BTUVTUEFDRSIgbmFtZXNwYWNlCiAg
            ICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdCiAg
            ICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlJFU1RSSUNURURfQVVUSF9QRVJNSVNTSU9OIl0K
            ICAgICAgICAgICAgYW5kIG5vdCBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJyYWRvcy1uYW1lc3BhY2UiLAogICAg
            ICAgICAgICAgICAgICAgICJraW5kIjogIkNlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSIsCiAg
            ICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJyYWRv
            c05hbWVzcGFjZU5hbWUiOiBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdLAogICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAg
            ICAgICAgICJuYW1lIjogImNlcGgtcmJkLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAg
            ICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6
            IHsKICAgICAgICAgICAgICAgICAgICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09M
            X05BTUUiXSwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92
            aXNpb25lci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJ
            U0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9y
            YWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5v
            dXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAg
            IGVsc2U6CiAgICAgICAgICAgIGlmIHNlbGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xf
            TkFNRSJdOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQiLAogICAgICAg
            ICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJkYXRhUG9vbCI6
            IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9NRVRBREFUQV9FQ19QT09MX05BTUUiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vcHJvdmlzaW9uZXIt
            c2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9T
            RUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL2NvbnRyb2xsZXItZXhwYW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9t
            YXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjog
            ZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAg
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICAgICAgfQogICAgICAgICAgICAg
            ICAgKQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAog
            ICAgICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2Vw
            aC1yYmQiLAogICAgICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAg
            ICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1l
            IjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUn
            XX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1zZWNyZXQtbmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAg
            ICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgICAgICkKCiAgICAg
            ICAgIyBpZiAnVE9QT0xPR1lfUE9PTFMnLCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fTEFCRUwn
            LCAnVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTJyAgZXhpc3RzLAogICAgICAgICMgdGhl
            biBvbmx5IGFkZCAndG9wb2xvZ3knIFN0b3JhZ2VDbGFzcwogICAgICAgIGlmICgKICAgICAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91
            dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0KICAgICAgICAgICAgYW5kIHNl
            bGYub3V0X21hcFsiVE9QT0xPR1lfRkFJTFVSRV9ET01BSU5fVkFMVUVTIl0KICAgICAgICApOgog
            ICAgICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAg
            ICAgICAgICAgIm5hbWUiOiAiY2VwaC1yYmQtdG9wb2xvZ3kiLAogICAgICAgICAgICAgICAgICAg
            ICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAg
            ICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5MYWJlbCI6IHNlbGYu
            b3V0X21hcFsKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RP
            TUFJTl9MQUJFTCIKICAgICAgICAgICAgICAgICAgICAgICAgXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInRvcG9sb2d5RmFpbHVyZURvbWFpblZhbHVlcyI6IHNlbGYub3V0X21hcFsKICAgICAg
            ICAgICAgICAgICAgICAgICAgICAgICJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiCiAg
            ICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ0b3BvbG9n
            eVBvb2xzIjogc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJy
            b29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhw
            YW5kLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgICMgaWYgJ0NFUEhGU19GU19OQU1FJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGhmcycgU3RvcmFnZUNsYXNzCiAgICAgICAgaWYgc2Vs
            Zi5vdXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaGZzIiwK
            ICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAg
            ICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiZnNOYW1lIjogc2VsZi5v
            dXRfbWFwWyJDRVBIRlNfRlNfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAicG9vbCI6
            IHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAg
            ICAiY3NpLnN0b3JhZ2UuazhzLmlvL3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtz
            ZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAg
            ICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL2NvbnRyb2xsZXItZXhwYW5k
            LXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9DRVBIRlNfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2Uu
            azhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJ
            X0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICB9LAogICAg
            ICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgIyBpZiAnUkdXX0VORFBPSU5UJyBl
            eGlzdHMsIHRoZW4gb25seSBhZGQgJ2NlcGgtcmd3JyBTdG9yYWdlQ2xhc3MKICAgICAgICBpZiBz
            ZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAiY2VwaC1yZ3ci
            LAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAg
            ICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJlbmRwb2ludCI6IHNl
            bGYub3V0X21hcFsiUkdXX0VORFBPSU5UIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            UHJlZml4Ijogc2VsZi5vdXRfbWFwWyJSR1dfUE9PTF9QUkVGSVgiXSwKICAgICAgICAgICAgICAg
            ICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyZ3ctYWRtaW4tb3BzLXVzZXIiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjogIlNlY3Jl
            dCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJhY2Nlc3NLZXkiOiBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZ
            Il0sCiAgICAgICAgICAgICAgICAgICAgICAgICJzZWNyZXRLZXkiOiBzZWxmLm91dF9tYXBbIlJH
            V19BRE1JTl9PUFNfVVNFUl9TRUNSRVRfS0VZIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAg
            ICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JHV19UTFNfQ0VSVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkIHRoZSAiY2VwaC1yZ3ctdGxzLWNlcnQiIHNlY3JldAogICAg
            ICAgIGlmIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl06CiAgICAgICAgICAgIGpzb25fb3V0
            LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6ICJj
            ZXBoLXJndy10bHMtY2VydCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImNl
            cnQiOiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdLAogICAgICAgICAgICAgICAgICAgIH0s
            CiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKCiAgICAgICAgcmV0dXJuIGpzb24uZHVt
            cHMoanNvbl9vdXQpICsgTElORVNFUAoKICAgIGRlZiB1cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25z
            KHNlbGYpOgogICAgICAgIHVzZXJzID0gWwogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMt
            bm9kZSIsCiAgICAgICAgICAgICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsCiAgICAg
            ICAgICAgICJjbGllbnQuY3NpLXJiZC1ub2RlIiwKICAgICAgICAgICAgImNsaWVudC5jc2ktcmJk
            LXByb3Zpc2lvbmVyIiwKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIiwKICAgICAg
            ICBdCiAgICAgICAgaWYgc2VsZi5ydW5fYXNfdXNlciAhPSAiIiBhbmQgc2VsZi5ydW5fYXNfdXNl
            ciBub3QgaW4gdXNlcnM6CiAgICAgICAgICAgIHVzZXJzLmFwcGVuZChzZWxmLnJ1bl9hc191c2Vy
            KQogICAgICAgIGZvciB1c2VyIGluIHVzZXJzOgogICAgICAgICAgICBzZWxmLnVwZ3JhZGVfdXNl
            cl9wZXJtaXNzaW9ucyh1c2VyKQoKICAgIGRlZiBnZXRfcmd3X3Bvb2xfbmFtZV9kdXJpbmdfdXBn
            cmFkZShzZWxmLCB1c2VyLCBjYXBzKToKICAgICAgICBpZiB1c2VyID09ICJjbGllbnQuaGVhbHRo
            Y2hlY2tlciI6CiAgICAgICAgICAgICMgd2hlbiBhZG1pbiBoYXMgbm90IHByb3ZpZGVkIHJndyBw
            b29sIG5hbWUgZHVyaW5nIHVwZ3JhZGUsCiAgICAgICAgICAgICMgZ2V0IHRoZSByZ3cgcG9vbCBu
            YW1lIGZyb20gY2xpZW50LmhlYWx0aGNoZWNrZXIgdXNlciB3aGljaCB3YXMgdXNlZCBkdXJpbmcg
            Y29ubmVjdGlvbgogICAgICAgICAgICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcG9vbF9w
            cmVmaXg6CiAgICAgICAgICAgICAgICAjIFRvIGdldCB2YWx1ZSAnZGVmYXVsdCcgd2hpY2ggaXMg
            cmd3IHBvb2wgbmFtZSBmcm9tICdhbGxvdyByd3ggcG9vbD1kZWZhdWx0LnJndy5tZXRhJwogICAg
            ICAgICAgICAgICAgcGF0dGVybiA9IHIicG9vbD0oLio/KVwucmd3XC5tZXRhIgogICAgICAgICAg
            ICAgICAgbWF0Y2ggPSByZS5zZWFyY2gocGF0dGVybiwgY2FwcykKICAgICAgICAgICAgICAgIGlm
            IG1hdGNoOgogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJl
            Zml4ID0gbWF0Y2guZ3JvdXAoMSkKICAgICAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAg
            ICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAg
            ICAgICAgImZhaWxlZCB0byBnZXQgcmd3IHBvb2wgbmFtZSBmb3IgdXBncmFkZSIKICAgICAgICAg
            ICAgICAgICAgICApCgogICAgZGVmIHVwZ3JhZGVfdXNlcl9wZXJtaXNzaW9ucyhzZWxmLCB1c2Vy
            KToKICAgICAgICAjIGNoZWNrIHdoZXRoZXIgdGhlIGdpdmVuIHVzZXIgZXhpc3RzIG9yIG5vdAog
            ICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAiYXV0aCBnZXQiLCAiZW50aXR5IjogZiJ7dXNl
            cn0iLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNn
            ID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBpZiByZXRfdmFs
            ICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICBwcmludChmInVzZXIge3Vz
            ZXJ9IG5vdCBmb3VuZCBmb3IgdXBncmFkaW5nLiIpCiAgICAgICAgICAgIHJldHVybgogICAgICAg
            IGV4aXN0aW5nX2NhcHMgPSBqc29uX291dFswXVsiY2FwcyJdCiAgICAgICAgc2VsZi5nZXRfcmd3
            X3Bvb2xfbmFtZV9kdXJpbmdfdXBncmFkZSh1c2VyLCBzdHIoZXhpc3RpbmdfY2FwcykpCiAgICAg
            ICAgbmV3X2NhcCwgXyA9IHNlbGYuZ2V0X2NhcHNfYW5kX2VudGl0eSh1c2VyKQogICAgICAgIGNh
            cF9rZXlzID0gWyJtb24iLCAibWdyIiwgIm9zZCIsICJtZHMiXQogICAgICAgIGNhcHMgPSBbXQog
            ICAgICAgIGZvciBlYWNoQ2FwIGluIGNhcF9rZXlzOgogICAgICAgICAgICBjdXJfY2FwX3ZhbHVl
            cyA9IGV4aXN0aW5nX2NhcHMuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBuZXdfY2FwX3Zh
            bHVlcyA9IG5ld19jYXAuZ2V0KGVhY2hDYXAsICIiKQogICAgICAgICAgICBjdXJfY2FwX3Blcm1f
            bGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBmb3IgeCBpbiBjdXJfY2FwX3ZhbHVl
            cy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAgICBdCiAgICAgICAgICAgIG5ld19j
            YXBfcGVybV9saXN0ID0gWwogICAgICAgICAgICAgICAgeC5zdHJpcCgpIGZvciB4IGluIG5ld19j
            YXBfdmFsdWVzLnNwbGl0KCIsIikgaWYgeC5zdHJpcCgpCiAgICAgICAgICAgIF0KICAgICAgICAg
            ICAgIyBhcHBlbmQgbmV3X2NhcF9saXN0IHRvIGN1cl9jYXBfbGlzdCB0byBtYWludGFpbiB0aGUg
            b3JkZXIgb2YgY2FwcwogICAgICAgICAgICBjdXJfY2FwX3Blcm1fbGlzdC5leHRlbmQobmV3X2Nh
            cF9wZXJtX2xpc3QpCiAgICAgICAgICAgICMgZWxpbWluYXRlIGR1cGxpY2F0ZXMgd2l0aG91dCB1
            c2luZyAnc2V0JwogICAgICAgICAgICAjIHNldCByZS1vcmRlcnMgaXRlbXMgaW4gdGhlIGxpc3Qg
            YW5kIHdlIGhhdmUgdG8ga2VlcCB0aGUgb3JkZXIKICAgICAgICAgICAgbmV3X2NhcF9saXN0ID0g
            W10KICAgICAgICAgICAgW25ld19jYXBfbGlzdC5hcHBlbmQoeCkgZm9yIHggaW4gY3VyX2NhcF9w
            ZXJtX2xpc3QgaWYgeCBub3QgaW4gbmV3X2NhcF9saXN0XQogICAgICAgICAgICBleGlzdGluZ19j
            YXBzW2VhY2hDYXBdID0gIiwgIi5qb2luKG5ld19jYXBfbGlzdCkKICAgICAgICAgICAgaWYgZXhp
            c3RpbmdfY2Fwc1tlYWNoQ2FwXToKICAgICAgICAgICAgICAgIGNhcHMuYXBwZW5kKGVhY2hDYXAp
            CiAgICAgICAgICAgICAgICBjYXBzLmFwcGVuZChleGlzdGluZ19jYXBzW2VhY2hDYXBdKQogICAg
            ICAgIGNtZF9qc29uID0gewogICAgICAgICAgICAicHJlZml4IjogImF1dGggY2FwcyIsCiAgICAg
            ICAgICAgICJlbnRpdHkiOiB1c2VyLAogICAgICAgICAgICAiY2FwcyI6IGNhcHMsCiAgICAgICAg
            ICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQogICAgICAgIHJldF92YWwsIGpzb25fb3V0
            LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICBp
            ZiByZXRfdmFsICE9IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oCiAgICAgICAgICAgICAgICBmIidhdXRoIGNhcHMge3VzZXJ9JyBjb21tYW5kIGZhaWxlZC5c
            biBFcnJvcjoge2Vycl9tc2d9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQoZiJVcGRhdGVk
            IHVzZXIge3VzZXJ9IHN1Y2Nlc3NmdWxseS4iKQoKICAgIGRlZiBtYWluKHNlbGYpOgogICAgICAg
            IGdlbmVyYXRlZF9vdXRwdXQgPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIudXBncmFk
            ZToKICAgICAgICAgICAgc2VsZi51cGdyYWRlX3VzZXJzX3Blcm1pc3Npb25zKCkKICAgICAgICBl
            bGlmIHNlbGYuX2FyZ19wYXJzZXIuZm9ybWF0ID09ICJqc29uIjoKICAgICAgICAgICAgZ2VuZXJh
            dGVkX291dHB1dCA9IHNlbGYuZ2VuX2pzb25fb3V0KCkKICAgICAgICBlbGlmIHNlbGYuX2FyZ19w
            YXJzZXIuZm9ybWF0ID09ICJiYXNoIjoKICAgICAgICAgICAgZ2VuZXJhdGVkX291dHB1dCA9IHNl
            bGYuZ2VuX3NoZWxsX291dCgpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgcmFpc2UgRXhlY3V0
            aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiVW5zdXBwb3J0ZWQgZm9ybWF0
            OiB7c2VsZi5fYXJnX3BhcnNlci5mb3JtYXR9IgogICAgICAgICAgICApCiAgICAgICAgcHJpbnQo
            Z2VuZXJhdGVkX291dHB1dCkKICAgICAgICBpZiBzZWxmLm91dHB1dF9maWxlIGFuZCBnZW5lcmF0
            ZWRfb3V0cHV0OgogICAgICAgICAgICBmT3V0ID0gb3BlbihzZWxmLm91dHB1dF9maWxlLCBtb2Rl
            PSJ3IiwgZW5jb2Rpbmc9IlVURi04IikKICAgICAgICAgICAgZk91dC53cml0ZShnZW5lcmF0ZWRf
            b3V0cHV0KQogICAgICAgICAgICBmT3V0LmNsb3NlKCkKCgojIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIE1BSU4gIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIwppZiBfX25hbWVfXyA9PSAiX19tYWluX18iOgogICAgcmpPYmogPSBSYWRvc0pT
            T04oKQogICAgdHJ5OgogICAgICAgIHJqT2JqLm1haW4oKQogICAgZXhjZXB0IEV4ZWN1dGlvbkZh
            aWx1cmVFeGNlcHRpb24gYXMgZXJyOgogICAgICAgIHByaW50KGYiRXhlY3V0aW9uIEZhaWxlZDog
            e2Vycn0iKQogICAgICAgIHJhaXNlIGVycgogICAgZXhjZXB0IEtleUVycm9yIGFzIGtFcnI6CiAg
            ICAgICAgcHJpbnQoZiJLZXlFcnJvcjoge2tFcnJ9IikKICAgIGV4Y2VwdCBPU0Vycm9yIGFzIG9z
            RXJyOgogICAgICAgIHByaW50KGYiRXJyb3Igd2hpbGUgdHJ5aW5nIHRvIG91dHB1dCB0aGUgZGF0
            YToge29zRXJyfSIpCiAgICBmaW5hbGx5OgogICAgICAgIHJqT2JqLnNodXRkb3duKCkK
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:7b7bea42852bc5b7645ce749f28fc8966d270363e9c1ecbe25da3e055be8ca14
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:94b1948eccdafc6dfa3acfd62823069974281b902f1dca2be7580ee6b7881c9e
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:35e88716ace3576683a9c4aa686df5f7a8b5bb6ee7a144fd2d87e3c9f5bf3961
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:06d00273e2e0de33ddaea55a013904cb09224c644c27d67ffd076e44b5786f47
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:72e2bde6be6198b8ca1f1dcb0ba75b9f8599674b113359ad5074a78e43a6e711
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:7c0c1ff6d677ce724d544bcf358278795b1d56659c53bb8cccd966b5cb7da286
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:2520ee9366fb54306b56c6d43ea969145defcccc4cfb1e50eae34e173dee1ca7
        version: 4.17.13-rhodf
      entries:
      - name: rook-ceph-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: rook-ceph-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: rook-ceph-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: rook-ceph-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: rook-ceph-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: rook-ceph-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: rook-ceph-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: rook-ceph-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: rook-ceph-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: rook-ceph-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: rook-ceph-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: rook-ceph-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: rook-ceph-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: rook-ceph-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: rook-ceph-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephCluster",
                "metadata": {
                  "name": "my-rook-ceph",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "cephVersion": {
                    "image": "registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9"
                  },
                  "dataDirHostPath": "/var/lib/rook",
                  "mon": {
                    "count": 3
                  },
                  "dashboard": {
                    "enabled": true
                  },
                  "network": {
                    "hostNetwork": false
                  },
                  "rbdMirroring": {
                    "workers": 0
                  },
                  "storage": {
                    "useAllNodes": true,
                    "useAllDevices": true
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephBlockPool",
                "metadata": {
                  "name": "replicapool",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "failureDomain": "host",
                  "replicated": {
                    "size": 3
                  },
                  "annotations": null
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStore",
                "metadata": {
                  "name": "my-store",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "metadataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "dataPool": {
                    "failureDomain": "host",
                    "replicated": {
                      "size": 3
                    }
                  },
                  "gateway": {
                    "type": "s3",
                    "sslCertificateRef": null,
                    "port": 8080,
                    "securePort": null,
                    "instances": 1,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephObjectStoreUser",
                "metadata": {
                  "name": "my-user",
                  "namespace": "my-rook-ceph"
                },
                "spec": {
                  "store": "my-store",
                  "displayName": "my display name"
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephNFS",
                "metadata": {
                  "name": "my-nfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "rados": {
                    "pool": "myfs-data0",
                    "namespace": "nfs-ns"
                  },
                  "server": {
                    "active": 3,
                    "placement": null,
                    "annotations": null,
                    "resources": null
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephClient",
                "metadata": {
                  "name": "cinder",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "caps": {
                    "mon": "profile rbd",
                    "osd": "profile rbd pool=volumes, profile rbd pool=vms, profile rbd-read-only pool=images"
                  }
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephFilesystem",
                "metadata": {
                  "name": "myfs",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "dataPools": [
                    {
                      "compressionMode": "",
                      "crushRoot": "",
                      "deviceClass": "",
                      "erasureCoded": {
                        "algorithm": "",
                        "codingChunks": 0,
                        "dataChunks": 0
                      },
                      "failureDomain": "host",
                      "replicated": {
                        "requireSafeReplicaSize": false,
                        "size": 1,
                        "targetSizeRatio": 0.5
                      }
                    }
                  ],
                  "metadataPool": {
                    "compressionMode": "",
                    "crushRoot": "",
                    "deviceClass": "",
                    "erasureCoded": {
                      "algorithm": "",
                      "codingChunks": 0,
                      "dataChunks": 0
                    },
                    "failureDomain": "",
                    "replicated": {
                      "requireSafeReplicaSize": false,
                      "size": 1,
                      "targetSizeRatio": 0
                    }
                  },
                  "metadataServer": {
                    "activeCount": 1,
                    "activeStandby": true,
                    "placement": {},
                    "resources": {}
                  },
                  "preservePoolsOnDelete": false,
                  "preserveFilesystemOnDelete": false
                }
              },
              {
                "apiVersion": "ceph.rook.io/v1",
                "kind": "CephRBDMirror",
                "metadata": {
                  "name": "my-rbd-mirror",
                  "namespace": "rook-ceph"
                },
                "spec": {
                  "annotations": null,
                  "count": 1,
                  "placement": {
                    "topologyKey": "kubernetes.io/hostname"
                  },
                  "resources": null
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
          externalClusterScript: |-
            IiIiCkNvcHlyaWdodCAyMDIwIFRoZSBSb29rIEF1dGhvcnMuIEFsbCByaWdodHMgcmVzZXJ2ZWQu
            CgpMaWNlbnNlZCB1bmRlciB0aGUgQXBhY2hlIExpY2Vuc2UsIFZlcnNpb24gMi4wICh0aGUgIkxp
            Y2Vuc2UiKTsKeW91IG1heSBub3QgdXNlIHRoaXMgZmlsZSBleGNlcHQgaW4gY29tcGxpYW5jZSB3
            aXRoIHRoZSBMaWNlbnNlLgpZb3UgbWF5IG9idGFpbiBhIGNvcHkgb2YgdGhlIExpY2Vuc2UgYXQK
            CglodHRwOi8vd3d3LmFwYWNoZS5vcmcvbGljZW5zZXMvTElDRU5TRS0yLjAKClVubGVzcyByZXF1
            aXJlZCBieSBhcHBsaWNhYmxlIGxhdyBvciBhZ3JlZWQgdG8gaW4gd3JpdGluZywgc29mdHdhcmUK
            ZGlzdHJpYnV0ZWQgdW5kZXIgdGhlIExpY2Vuc2UgaXMgZGlzdHJpYnV0ZWQgb24gYW4gIkFTIElT
            IiBCQVNJUywKV0lUSE9VVCBXQVJSQU5USUVTIE9SIENPTkRJVElPTlMgT0YgQU5ZIEtJTkQsIGVp
            dGhlciBleHByZXNzIG9yIGltcGxpZWQuClNlZSB0aGUgTGljZW5zZSBmb3IgdGhlIHNwZWNpZmlj
            IGxhbmd1YWdlIGdvdmVybmluZyBwZXJtaXNzaW9ucyBhbmQKbGltaXRhdGlvbnMgdW5kZXIgdGhl
            IExpY2Vuc2UuCiIiIgoKaW1wb3J0IGVycm5vCmltcG9ydCBzeXMKaW1wb3J0IGpzb24KaW1wb3J0
            IGFyZ3BhcnNlCmltcG9ydCByZQppbXBvcnQgc3VicHJvY2VzcwppbXBvcnQgaG1hYwppbXBvcnQg
            Y29uZmlncGFyc2VyCmZyb20gaGFzaGxpYiBpbXBvcnQgc2hhMSBhcyBzaGEKZnJvbSBvcyBpbXBv
            cnQgbGluZXNlcCBhcyBMSU5FU0VQCmZyb20gb3MgaW1wb3J0IHBhdGgKZnJvbSBlbWFpbC51dGls
            cyBpbXBvcnQgZm9ybWF0ZGF0ZQppbXBvcnQgcmVxdWVzdHMKZnJvbSByZXF1ZXN0cy5hdXRoIGlt
            cG9ydCBBdXRoQmFzZQpmcm9tIGlvIGltcG9ydCBTdHJpbmdJTwpmcm9tIHVybGxpYi5wYXJzZSBp
            bXBvcnQgdXJscGFyc2UKZnJvbSB1cmxsaWIucGFyc2UgaW1wb3J0IHVybGVuY29kZSBhcyB1cmxl
            bmNvZGUKZnJvbSBpcGFkZHJlc3MgaW1wb3J0IGlwX2FkZHJlc3MKZnJvbSBpcGFkZHJlc3MgaW1w
            b3J0IElQdjRBZGRyZXNzCmZyb20gYmFzZTY0IGltcG9ydCBlbmNvZGVieXRlcyBhcyBlbmNvZGVz
            dHJpbmcKCk1vZHVsZU5vdEZvdW5kRXJyb3IgPSBJbXBvcnRFcnJvcgoKdHJ5OgogICAgaW1wb3J0
            IHJhZG9zCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5vTW9kRXJyOgogICAgcHJpbnQo
            ZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0Li4uIikKICAgIHN5cy5leGl0
            KDEpCgp0cnk6CiAgICBpbXBvcnQgcmJkCmV4Y2VwdCBNb2R1bGVOb3RGb3VuZEVycm9yIGFzIG5v
            TW9kRXJyOgogICAgcHJpbnQoZiJFcnJvcjoge25vTW9kRXJyfVxuRXhpdGluZyB0aGUgc2NyaXB0
            Li4uIikKICAgIHN5cy5leGl0KDEpCgoKY2xhc3MgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihF
            eGNlcHRpb24pOgogICAgcGFzcwoKCiMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMgRHVtbXlSYWRvcyAjIyMjIyMjIyMjIyMj
            IyMjIyMKIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMg
            dGhpcyBpcyBtYWlubHkgZm9yIHRlc3RpbmcgYW5kIGNvdWxkIGJlIHVzZWQgd2hlcmUgJ3JhZG9z
            JyBpcyBub3QgYXZhaWxhYmxlCgoKY2xhc3MgRHVtbXlSYWRvcyhvYmplY3QpOgogICAgZGVmIF9f
            aW5pdF9fKHNlbGYpOgogICAgICAgIHNlbGYucmV0dXJuX3ZhbCA9IDAKICAgICAgICBzZWxmLmVy
            cl9tZXNzYWdlID0gIiIKICAgICAgICBzZWxmLnN0YXRlID0gImNvbm5lY3RlZCIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwID0ge30KICAgICAgICBzZWxmLmNtZF9uYW1lcyA9IHt9CiAgICAg
            ICAgc2VsZi5faW5pdF9jbWRfb3V0cHV0X21hcCgpCiAgICAgICAgc2VsZi5kdW1teV9ob3N0X2lw
            X21hcCA9IHt9CgogICAgZGVmIF9pbml0X2NtZF9vdXRwdXRfbWFwKHNlbGYpOgogICAgICAgIGpz
            b25fZmlsZV9uYW1lID0gInRlc3QtZGF0YS9jZXBoLXN0YXR1cy1vdXQiCiAgICAgICAgc2NyaXB0
            X2RpciA9IHBhdGguYWJzcGF0aChwYXRoLmRpcm5hbWUoX19maWxlX18pKQogICAgICAgIGNlcGhf
            c3RhdHVzX3N0ciA9ICIiCiAgICAgICAgd2l0aCBvcGVuKAogICAgICAgICAgICBwYXRoLmpvaW4o
            c2NyaXB0X2RpciwganNvbl9maWxlX25hbWUpLCBtb2RlPSJyIiwgZW5jb2Rpbmc9IlVURi04Igog
            ICAgICAgICkgYXMganNvbl9maWxlOgogICAgICAgICAgICBjZXBoX3N0YXR1c19zdHIgPSBqc29u
            X2ZpbGUucmVhZCgpCiAgICAgICAgc2VsZi5jbWRfbmFtZXNbImZzIGxzIl0gPSAiIiJ7ImZvcm1h
            dCI6ICJqc29uIiwgInByZWZpeCI6ICJmcyBscyJ9IiIiCiAgICAgICAgc2VsZi5jbWRfbmFtZXNb
            InF1b3J1bV9zdGF0dXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAicXVvcnVtX3N0YXR1cyJ9IiIiCiAgICAgICAgKQogICAgICAgIHNlbGYuY21kX25h
            bWVzWyJtZ3Igc2VydmljZXMiXSA9ICgKICAgICAgICAgICAgIiIieyJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAibWdyIHNlcnZpY2VzIn0iIiIKICAgICAgICApCiAgICAgICAgIyBhbGwgdGhl
            IGNvbW1hbmRzIGFuZCB0aGVpciBvdXRwdXQKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwW3Nl
            bGYuY21kX25hbWVzWyJmcyBscyJdXSA9ICgKICAgICAgICAgICAgIiIiW3sibmFtZSI6Im15ZnMi
            LCJtZXRhZGF0YV9wb29sIjoibXlmcy1tZXRhZGF0YSIsIm1ldGFkYXRhX3Bvb2xfaWQiOjIsImRh
            dGFfcG9vbF9pZHMiOlszXSwiZGF0YV9wb29scyI6WyJteWZzLXJlcGxpY2F0ZWQiXX1dIiIiCiAg
            ICAgICAgKQogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbc2VsZi5jbWRfbmFtZXNbInF1b3J1
            bV9zdGF0dXMiXV0gPSAoCiAgICAgICAgICAgICIiInsiZWxlY3Rpb25fZXBvY2giOjMsInF1b3J1
            bSI6WzBdLCJxdW9ydW1fbmFtZXMiOlsiYSJdLCJxdW9ydW1fbGVhZGVyX25hbWUiOiJhIiwicXVv
            cnVtX2FnZSI6MTQzODUsImZlYXR1cmVzIjp7InF1b3J1bV9jb24iOiI0NTQwMTM4MjkyODM2Njk2
            MDYzIiwicXVvcnVtX21vbiI6WyJrcmFrZW4iLCJsdW1pbm91cyIsIm1pbWljIiwib3NkbWFwLXBy
            dW5lIiwibmF1dGlsdXMiLCJvY3RvcHVzIl19LCJtb25tYXAiOnsiZXBvY2giOjEsImZzaWQiOiJh
            ZjRlMTY3My0wYjcyLTQwMmQtOTkwYS0yMmQyOTE5ZDBmMWMiLCJtb2RpZmllZCI6IjIwMjAtMDUt
            MDdUMDM6MzY6MzkuOTE4MDM1WiIsImNyZWF0ZWQiOiIyMDIwLTA1LTA3VDAzOjM2OjM5LjkxODAz
            NVoiLCJtaW5fbW9uX3JlbGVhc2UiOjE1LCJtaW5fbW9uX3JlbGVhc2VfbmFtZSI6Im9jdG9wdXMi
            LCJmZWF0dXJlcyI6eyJwZXJzaXN0ZW50IjpbImtyYWtlbiIsImx1bWlub3VzIiwibWltaWMiLCJv
            c2RtYXAtcHJ1bmUiLCJuYXV0aWx1cyIsIm9jdG9wdXMiXSwib3B0aW9uYWwiOltdfSwibW9ucyI6
            W3sicmFuayI6MCwibmFtZSI6ImEiLCJwdWJsaWNfYWRkcnMiOnsiYWRkcnZlYyI6W3sidHlwZSI6
            InYyIiwiYWRkciI6IjEwLjExMC4yMDUuMTc0OjMzMDAiLCJub25jZSI6MH0seyJ0eXBlIjoidjEi
            LCJhZGRyIjoiMTAuMTEwLjIwNS4xNzQ6Njc4OSIsIm5vbmNlIjowfV19LCJhZGRyIjoiMTAuMTEw
            LjIwNS4xNzQ6Njc4OS8wIiwicHVibGljX2FkZHIiOiIxMC4xMTAuMjA1LjE3NDo2Nzg5LzAiLCJw
            cmlvcml0eSI6MCwid2VpZ2h0IjowfV19fSIiIgogICAgICAgICkKICAgICAgICBzZWxmLmNtZF9v
            dXRwdXRfbWFwW3NlbGYuY21kX25hbWVzWyJtZ3Igc2VydmljZXMiXV0gPSAoCiAgICAgICAgICAg
            ICIiInsiZGFzaGJvYXJkIjoiaHR0cHM6Ly9jZXBoLWRhc2hib2FyZDo4NDQzLyIsInByb21ldGhl
            dXMiOiJodHRwOi8vY2VwaC1kYXNoYm9hcmQtZGI6OTI4My8ifSIiIgogICAgICAgICkKICAgICAg
            ICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJh
            bGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMiLCAib3NkIiwgInByb2ZpbGUgcmJk
            LXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29s
            PS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyB4IHBv
            b2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5
            IjogImNsaWVudC5oZWFsdGhjaGVja2VyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJh
            dXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQu
            aGVhbHRoY2hlY2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNs
            cmRBPT0iLCJjYXBzIjp7Im1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgcXVvcnVtX3N0YXR1
            cyIsIm9zZCI6InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJn
            dy5jb250cm9sLCBhbGxvdyB4IHBvb2w9ZGVmYXVsdC5yZ3cuYnVja2V0cy5pbmRleCIsIm1kcyI6
            ImFsbG93ICoifX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAg
            IiIieyJjYXBzIjogWyJtb24iLCAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9j
            a2xpc3QnIiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5IjogImNsaWVudC5jc2ktcmJk
            LW5vZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9
            IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktcmJkLW5vZGUiLCJrZXki
            OiJBUUJPZ3JOZUhiSzFBeEFBdWJZQmVWOFMxVS9HUHpxNVNWZXE2Zz09IiwiY2FwcyI6eyJtb24i
            OiJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJwcm9m
            aWxlIHJiZCJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAi
            IiJ7ImNhcHMiOiBbIm1vbiIsICJwcm9maWxlIHJiZCwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2Nr
            bGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIsICJwcm9maWxlIHJiZCJdLCAiZW50aXR5
            IjogImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsImtleSI6IkFRQk5nck5lMWdleUt4QUE4ZWtWaVJk
            RStoc3M1T3dlWUJrd05nPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoicHJvZmls
            ZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoicHJvZmlsZSByYmQi
            fX1dIiIiCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBz
            IjogWyJtb24iLCAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdy
            IiwgImFsbG93IHJ3IiwgIm9zZCIsICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsICJtZHMiLCAi
            YWxsb3cgcnciXSwgImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIiIgogICAgICAgIF0gPSAi
            IiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1ub2RlIiwia2V5IjoiQVFCT2dyTmVFTnVu
            S3hBQVBDbWdFN1I2RzhEY1huYUoxRjMycWc9PSIsImNhcHMiOnsibWRzIjoiYWxsb3cgcnciLCJt
            Z3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xp
            c3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyAqPSoifX1dIiIiCiAgICAgICAgc2VsZi5j
            bWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJjYXBzIjogWyJtb24iLCAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCAibWdyIiwgImFsbG93IHJ3IiwgIm9zZCIs
            ICJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCAibWRzIiwgImFsbG93ICoiXSwgImVu
            dGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQiOiAianNvbiIs
            ICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50
            aXR5IjoiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCJrZXkiOiJBUUJPZ3JOZUFGZ2NH
            QkFBdkdxS09BRDBEM3h4bVZZMFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1v
            biI6ImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cg
            cncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBz
            ZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxv
            dyByLCBhbGxvdyBjb21tYW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAi
            b3NkIiwgImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsICJtZHMiLCAiYWxsb3cgKiJd
            LCAiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9y
            YWdlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVhdGUifSIi
            IgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25l
            ci1vcGVuc2hpZnQtc3RvcmFnZSIsImtleSI6IkJRQk9nck5lQUZnY0dCQUF2R3FLT0FEMEQzeHht
            VlkwUjkxMmRnPT0iLCJjYXBzIjp7Im1nciI6ImFsbG93IHJ3IiwibW9uIjoiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLCJvc2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1l
            dGFkYXRhPSoiLCAibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsICJtZ3IiLCAiYWxsb3cgcnciLCAib3NkIiwgImFsbG93IHJ3
            IHRhZyBjZXBoZnMgbWV0YWRhdGE9bXlmcyIsICJtZHMiLCAiYWxsb3cgKiJdLCAiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0LW9yLWNyZWF0ZSJ9IiIiCiAgICAg
            ICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5z
            aGlmdC1zdG9yYWdlLW15ZnMiLCJrZXkiOiJDUUJPZ3JOZUFGZ2NHQkFBdkdxS09BRDBEM3h4bVZZ
            MFI5MTJkZz09IiwiY2FwcyI6eyJtZ3IiOiJhbGxvdyBydyIsIm1vbiI6ImFsbG93IHIsIGFsbG93
            IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwib3NkIjoiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRh
            ZGF0YT1teWZzIiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRf
            bWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1vbiIsICJhbGxvdyByLCBhbGxvdyBjb21t
            YW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4
            IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncg
            cG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywg
            YWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMuaW5kZXgiLCAibWRzIiwgImFsbG93ICoi
            XSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJw
            cmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIn0iIiIKICAgICAgICBdID0gIiIiW3siZW50aXR5
            IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0NWJGUkFBVG5kTE5VU0VL
            cnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwgYWxsb3cgY29tbWFuZCBx
            dW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdyIjogImFsbG93IGNvbW1h
            bmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHksIGFsbG93IHJ3eCBwb29s
            PWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9
            ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0LnJndy5sb2csIGFsbG93
            IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoiYWxsb3cgKiJ9fV0iIiIK
            ICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImNhcHMiOiBbIm1v
            biIsICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93IGNvbW1hbmQg
            dmVyc2lvbiIsICJtZ3IiLCAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIiwgInByb2ZpbGUg
            cmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwgYWxsb3cgciBw
            b29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9sLCBhbGxvdyBy
            eCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3LmJ1Y2tldHMu
            aW5kZXgiLCAibWRzIiwgImFsbG93ICoiXSwgImVudGl0eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tl
            ciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBjYXBzIn0iIiIKICAgICAgICBd
            ID0gIiIiW3siZW50aXR5IjoiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCJrZXkiOiJBUURGa2JOZWZ0
            NWJGUkFBVG5kTE5VU1JLcnVvenhpWmkzbHJkQT09IiwiY2FwcyI6eyJtb24iOiAiYWxsb3cgciwg
            YWxsb3cgY29tbWFuZCBxdW9ydW1fc3RhdHVzLCBhbGxvdyBjb21tYW5kIHZlcnNpb24iLCAibWdy
            IjogImFsbG93IGNvbW1hbmQgY29uZmlnIiwgIm9zZCI6ICJwcm9maWxlIHJiZC1yZWFkLW9ubHks
            IGFsbG93IHJ3eCBwb29sPWRlZmF1bHQucmd3Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3Qs
            IGFsbG93IHJ3IHBvb2w9ZGVmYXVsdC5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD1kZWZhdWx0
            LnJndy5sb2csIGFsbG93IHggcG9vbD1kZWZhdWx0LnJndy5idWNrZXRzLmluZGV4IiwibWRzIjoi
            YWxsb3cgKiJ9fV0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWyIiInsiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogIm1nciBzZXJ2aWNlcyJ9IiIiXSA9ICgKICAgICAgICAgICAgIiIi
            eyJkYXNoYm9hcmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcw
            MDAvIiwgInByb21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0
            am5sOjkyODMvIn0iIiIKICAgICAgICApCiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAg
            ICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiLCAiZm9ybWF0Ijog
            Impzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBdID0gIiIieyJkYXNoYm9h
            cmQiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjcwMDAvIiwgInBy
            b21ldGhldXMiOiAiaHR0cDovL3Jvb2stY2VwaC1tZ3ItYS01N2NmOWY4NGJjLWY0am5sOjkyODMv
            In0iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7ImVudGl0
            eSI6ICJjbGllbnQuaGVhbHRoY2hlY2tlciIsICJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgiOiAi
            YXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbeyJlbnRpdHkiOiJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsImtleSI6IkFRREZrYk5lZnQ1YkZSQUFUbmRMTlVTRUtydW96eGlaaTNscmRBPT0iLCJj
            YXBzIjp7Im1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9zdGF0dXMsIGFsbG93
            IGNvbW1hbmQgdmVyc2lvbiIsICJtZ3IiOiAiYWxsb3cgY29tbWFuZCBjb25maWciLCAib3NkIjog
            InByb2ZpbGUgcmJkLXJlYWQtb25seSwgYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5yZ3cubWV0YSwg
            YWxsb3cgciBwb29sPS5yZ3cucm9vdCwgYWxsb3cgcncgcG9vbD1kZWZhdWx0LnJndy5jb250cm9s
            LCBhbGxvdyByeCBwb29sPWRlZmF1bHQucmd3LmxvZywgYWxsb3cgeCBwb29sPWRlZmF1bHQucmd3
            LmJ1Y2tldHMuaW5kZXgiLCJtZHMiOiJhbGxvdyAqIn19XSIiIgogICAgICAgIHNlbGYuY21kX291
            dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNsaWVudC5jc2ktY2VwaGZzLW5v
            ZGUiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAgICAgICBd
            ID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAgICAiIiJ7
            ImVudGl0eSI6ICJjbGllbnQuY3NpLXJiZC1ub2RlIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAgICAgICAgc2VsZi5jbWRf
            b3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xpZW50LmNzaS1yYmQtcHJv
            dmlzaW9uZXIiLCAiZm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggZ2V0In0iIiIKICAg
            ICAgICBdID0gIiIiW10iIiIKICAgICAgICBzZWxmLmNtZF9vdXRwdXRfbWFwWwogICAgICAgICAg
            ICAiIiJ7ImVudGl0eSI6ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIsICJmb3JtYXQi
            OiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIiIgog
            ICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5IjogImNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlIiwgImZvcm1hdCI6
            ICJqc29uIiwgInByZWZpeCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIltdIiIiCiAg
            ICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsKICAgICAgICAgICAgIiIieyJlbnRpdHkiOiAiY2xp
            ZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXItb3BlbnNoaWZ0LXN0b3JhZ2UtbXlmcyIsICJmb3Jt
            YXQiOiAianNvbiIsICJwcmVmaXgiOiAiYXV0aCBnZXQifSIiIgogICAgICAgIF0gPSAiIiJbXSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiZW50aXR5Ijog
            ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwgImZvcm1hdCI6ICJqc29uIiwgInByZWZp
            eCI6ICJhdXRoIGdldCJ9IiIiCiAgICAgICAgXSA9ICIiIlt7ImVudGl0eSI6ImNsaWVudC5jc2kt
            Y2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmtiTmVmdDViRlJBQVRuZExOVVNFS3J1b3p4
            aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cgciIsICJtZ3IiOiJhbGxvdyBydyIsICJv
            c2QiOiJhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSoiLCJtZHMiOiJhbGxvdyAqIn19XSIi
            IgogICAgICAgIHNlbGYuY21kX291dHB1dF9tYXBbCiAgICAgICAgICAgICIiInsiY2FwcyI6IFsi
            bW9uIiwgImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciIsICJh
            bGxvdyBydyIsICJvc2QiLCAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwgIm1kcyIs
            ICJhbGxvdyAqIl0sICJlbnRpdHkiOiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiLCAi
            Zm9ybWF0IjogImpzb24iLCAicHJlZml4IjogImF1dGggY2FwcyJ9IiIiCiAgICAgICAgXSA9ICIi
            Ilt7ImVudGl0eSI6ImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIiwia2V5IjoiQVFERmti
            TmVmdDViRlJBQVRuZExOVVNFS3J1b3p4aVppM2xyZEE9PSIsImNhcHMiOnsibW9uIjoiYWxsb3cg
            ciwgIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwgIm1nciI6ImFsbG93IHJ3IiwgIm9z
            ZCI6ImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9KiIsIm1kcyI6ImFsbG93ICoifX1dIiIi
            CiAgICAgICAgc2VsZi5jbWRfb3V0cHV0X21hcFsneyJmb3JtYXQiOiAianNvbiIsICJwcmVmaXgi
            OiAic3RhdHVzIn0nXSA9IGNlcGhfc3RhdHVzX3N0cgoKICAgIGRlZiBzaHV0ZG93bihzZWxmKToK
            ICAgICAgICBwYXNzCgogICAgZGVmIGdldF9mc2lkKHNlbGYpOgogICAgICAgIHJldHVybiAiYWY0
            ZTE2NzMtMGI3Mi00MDJkLTk5MGEtMjJkMjkxOWQwZjFjIgoKICAgIGRlZiBjb25mX3JlYWRfZmls
            ZShzZWxmKToKICAgICAgICBwYXNzCgogICAgZGVmIGNvbm5lY3Qoc2VsZik6CiAgICAgICAgcGFz
            cwoKICAgIGRlZiBwb29sX2V4aXN0cyhzZWxmLCBwb29sX25hbWUpOgogICAgICAgIHJldHVybiBU
            cnVlCgogICAgZGVmIG1vbl9jb21tYW5kKHNlbGYsIGNtZCwgb3V0KToKICAgICAgICBqc29uX2Nt
            ZCA9IGpzb24ubG9hZHMoY21kKQogICAgICAgIGpzb25fY21kX3N0ciA9IGpzb24uZHVtcHMoanNv
            bl9jbWQsIHNvcnRfa2V5cz1UcnVlKQogICAgICAgIGNtZF9vdXRwdXQgPSBzZWxmLmNtZF9vdXRw
            dXRfbWFwW2pzb25fY21kX3N0cl0KICAgICAgICByZXR1cm4gc2VsZi5yZXR1cm5fdmFsLCBjbWRf
            b3V0cHV0LCBzdHIoc2VsZi5lcnJfbWVzc2FnZS5lbmNvZGUoInV0Zi04IikpCgogICAgZGVmIF9j
            b252ZXJ0X2hvc3RuYW1lX3RvX2lwKHNlbGYsIGhvc3RfbmFtZSk6CiAgICAgICAgaXBfcmVnX3gg
            PSByZS5jb21waWxlKHIiXGR7MSwzfS5cZHsxLDN9LlxkezEsM30uXGR7MSwzfSIpCiAgICAgICAg
            IyBpZiBwcm92aWRlZCBob3N0IGlzIGRpcmVjdGx5IGFuIElQIGFkZHJlc3MsIHJldHVybiB0aGUg
            c2FtZQogICAgICAgIGlmIGlwX3JlZ194Lm1hdGNoKGhvc3RfbmFtZSk6CiAgICAgICAgICAgIHJl
            dHVybiBob3N0X25hbWUKICAgICAgICBpbXBvcnQgcmFuZG9tCgogICAgICAgIGhvc3RfaXAgPSBz
            ZWxmLmR1bW15X2hvc3RfaXBfbWFwLmdldChob3N0X25hbWUsICIiKQogICAgICAgIGlmIG5vdCBo
            b3N0X2lwOgogICAgICAgICAgICBob3N0X2lwID0gZiIxNzIuOS57cmFuZG9tLnJhbmRpbnQoMCwg
            MjU0KX0ue3JhbmRvbS5yYW5kaW50KDAsIDI1NCl9IgogICAgICAgICAgICBzZWxmLmR1bW15X2hv
            c3RfaXBfbWFwW2hvc3RfbmFtZV0gPSBob3N0X2lwCiAgICAgICAgZGVsIHJhbmRvbQogICAgICAg
            IHJldHVybiBob3N0X2lwCgogICAgQGNsYXNzbWV0aG9kCiAgICBkZWYgUmFkb3MoY29uZmZpbGU9
            Tm9uZSk6CiAgICAgICAgcmV0dXJuIER1bW15UmFkb3MoKQoKCmNsYXNzIFMzQXV0aChBdXRoQmFz
            ZSk6CiAgICAiIiJBdHRhY2hlcyBBV1MgQXV0aGVudGljYXRpb24gdG8gdGhlIGdpdmVuIFJlcXVl
            c3Qgb2JqZWN0LiIiIgoKICAgIHNlcnZpY2VfYmFzZV91cmwgPSAiczMuYW1hem9uYXdzLmNvbSIK
            CiAgICBkZWYgX19pbml0X18oc2VsZiwgYWNjZXNzX2tleSwgc2VjcmV0X2tleSwgc2VydmljZV91
            cmw9Tm9uZSk6CiAgICAgICAgaWYgc2VydmljZV91cmw6CiAgICAgICAgICAgIHNlbGYuc2Vydmlj
            ZV9iYXNlX3VybCA9IHNlcnZpY2VfdXJsCiAgICAgICAgc2VsZi5hY2Nlc3Nfa2V5ID0gc3RyKGFj
            Y2Vzc19rZXkpCiAgICAgICAgc2VsZi5zZWNyZXRfa2V5ID0gc3RyKHNlY3JldF9rZXkpCgogICAg
            ZGVmIF9fY2FsbF9fKHNlbGYsIHIpOgogICAgICAgICMgQ3JlYXRlIGRhdGUgaGVhZGVyIGlmIGl0
            IGlzIG5vdCBjcmVhdGVkIHlldC4KICAgICAgICBpZiAiZGF0ZSIgbm90IGluIHIuaGVhZGVycyBh
            bmQgIngtYW16LWRhdGUiIG5vdCBpbiByLmhlYWRlcnM6CiAgICAgICAgICAgIHIuaGVhZGVyc1si
            ZGF0ZSJdID0gZm9ybWF0ZGF0ZSh0aW1ldmFsPU5vbmUsIGxvY2FsdGltZT1GYWxzZSwgdXNlZ210
            PVRydWUpCiAgICAgICAgc2lnbmF0dXJlID0gc2VsZi5nZXRfc2lnbmF0dXJlKHIpCiAgICAgICAg
            c2lnbmF0dXJlID0gc2lnbmF0dXJlLmRlY29kZSgidXRmLTgiKQogICAgICAgIHIuaGVhZGVyc1si
            QXV0aG9yaXphdGlvbiJdID0gZiJBV1Mge3NlbGYuYWNjZXNzX2tleX06e3NpZ25hdHVyZX0iCiAg
            ICAgICAgcmV0dXJuIHIKCiAgICBkZWYgZ2V0X3NpZ25hdHVyZShzZWxmLCByKToKICAgICAgICBj
            YW5vbmljYWxfc3RyaW5nID0gc2VsZi5nZXRfY2Fub25pY2FsX3N0cmluZyhyLnVybCwgci5oZWFk
            ZXJzLCByLm1ldGhvZCkKICAgICAgICBrZXkgPSBzZWxmLnNlY3JldF9rZXkuZW5jb2RlKCJ1dGYt
            OCIpCiAgICAgICAgbXNnID0gY2Fub25pY2FsX3N0cmluZy5lbmNvZGUoInV0Zi04IikKICAgICAg
            ICBoID0gaG1hYy5uZXcoa2V5LCBtc2csIGRpZ2VzdG1vZD1zaGEpCiAgICAgICAgcmV0dXJuIGVu
            Y29kZXN0cmluZyhoLmRpZ2VzdCgpKS5zdHJpcCgpCgogICAgZGVmIGdldF9jYW5vbmljYWxfc3Ry
            aW5nKHNlbGYsIHVybCwgaGVhZGVycywgbWV0aG9kKToKICAgICAgICBwYXJzZWR1cmwgPSB1cmxw
            YXJzZSh1cmwpCiAgICAgICAgb2JqZWN0a2V5ID0gcGFyc2VkdXJsLnBhdGhbMTpdCgogICAgICAg
            IGJ1Y2tldCA9IHBhcnNlZHVybC5uZXRsb2NbOiAtbGVuKHNlbGYuc2VydmljZV9iYXNlX3VybCld
            CiAgICAgICAgaWYgbGVuKGJ1Y2tldCkgPiAxOgogICAgICAgICAgICAjIHJlbW92ZSBsYXN0IGRv
            dAogICAgICAgICAgICBidWNrZXQgPSBidWNrZXRbOi0xXQoKICAgICAgICBpbnRlcmVzdGluZ19o
            ZWFkZXJzID0geyJjb250ZW50LW1kNSI6ICIiLCAiY29udGVudC10eXBlIjogIiIsICJkYXRlIjog
            IiJ9CiAgICAgICAgZm9yIGtleSBpbiBoZWFkZXJzOgogICAgICAgICAgICBsayA9IGtleS5sb3dl
            cigpCiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGxrID0gbGsuZGVjb2RlKCJ1dGYt
            OCIpCiAgICAgICAgICAgIGV4Y2VwdDoKICAgICAgICAgICAgICAgIHBhc3MKICAgICAgICAgICAg
            aWYgaGVhZGVyc1trZXldIGFuZCAoCiAgICAgICAgICAgICAgICBsayBpbiBpbnRlcmVzdGluZ19o
            ZWFkZXJzLmtleXMoKSBvciBsay5zdGFydHN3aXRoKCJ4LWFtei0iKQogICAgICAgICAgICApOgog
            ICAgICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1tsa10gPSBoZWFkZXJzW2tleV0uc3Ry
            aXAoKQoKICAgICAgICAjIElmIHgtYW16LWRhdGUgaXMgdXNlZCBpdCBzdXBlcnNlZGVzIHRoZSBk
            YXRlIGhlYWRlci4KICAgICAgICBpZiAieC1hbXotZGF0ZSIgaW4gaW50ZXJlc3RpbmdfaGVhZGVy
            czoKICAgICAgICAgICAgaW50ZXJlc3RpbmdfaGVhZGVyc1siZGF0ZSJdID0gIiIKCiAgICAgICAg
            YnVmID0gZiJ7bWV0aG9kfVxuIgogICAgICAgIGZvciBrZXkgaW4gc29ydGVkKGludGVyZXN0aW5n
            X2hlYWRlcnMua2V5cygpKToKICAgICAgICAgICAgdmFsID0gaW50ZXJlc3RpbmdfaGVhZGVyc1tr
            ZXldCiAgICAgICAgICAgIGlmIGtleS5zdGFydHN3aXRoKCJ4LWFtei0iKToKICAgICAgICAgICAg
            ICAgIGJ1ZiArPSBmIntrZXl9Ont2YWx9XG4iCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBidWYgKz0gZiJ7dmFsfVxuIgoKICAgICAgICAjIGFwcGVuZCB0aGUgYnVja2V0IGlmIGl0
            IGV4aXN0cwogICAgICAgIGlmIGJ1Y2tldCAhPSAiIjoKICAgICAgICAgICAgYnVmICs9IGYiL3ti
            dWNrZXR9IgoKICAgICAgICAjIGFkZCB0aGUgb2JqZWN0a2V5LiBldmVuIGlmIGl0IGRvZXNuJ3Qg
            ZXhpc3QsIGFkZCB0aGUgc2xhc2gKICAgICAgICBidWYgKz0gZiIve29iamVjdGtleX0iCgogICAg
            ICAgIHJldHVybiBidWYKCgpjbGFzcyBSYWRvc0pTT046CiAgICBFWFRFUk5BTF9VU0VSX05BTUUg
            PSAiY2xpZW50LmhlYWx0aGNoZWNrZXIiCiAgICBFWFRFUk5BTF9SR1dfQURNSU5fT1BTX1VTRVJf
            TkFNRSA9ICJyZ3ctYWRtaW4tb3BzLXVzZXIiCiAgICBFTVBUWV9PVVRQVVRfTElTVCA9ICJFbXB0
            eSBvdXRwdXQgbGlzdCIKICAgIERFRkFVTFRfUkdXX1BPT0xfUFJFRklYID0gImRlZmF1bHQiCiAg
            ICBERUZBVUxUX01PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCA9ICI5MjgzIgoKICAgIEBjbGFzc21l
            dGhvZAogICAgZGVmIGdlbl9hcmdfcGFyc2VyKGNscywgYXJnc190b19wYXJzZT1Ob25lKToKICAg
            ICAgICBhcmdQID0gYXJncGFyc2UuQXJndW1lbnRQYXJzZXIoKQoKICAgICAgICBjb21tb25fZ3Jv
            dXAgPSBhcmdQLmFkZF9hcmd1bWVudF9ncm91cCgiY29tbW9uIikKICAgICAgICBjb21tb25fZ3Jv
            dXAuYWRkX2FyZ3VtZW50KCItLXZlcmJvc2UiLCAiLXYiLCBhY3Rpb249InN0b3JlX3RydWUiLCBk
            ZWZhdWx0PUZhbHNlKQogICAgICAgIGNvbW1vbl9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLWNlcGgtY29uZiIsICItYyIsIGhlbHA9IlByb3ZpZGUgYSBjZXBoIGNvbmYgZmlsZS4i
            LCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRk
            X2FyZ3VtZW50KAogICAgICAgICAgICAiLS1rZXlyaW5nIiwgIi1rIiwgaGVscD0iUGF0aCB0byBj
            ZXBoIGtleXJpbmcgZmlsZS4iLCB0eXBlPXN0ciwgZGVmYXVsdD0iIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1ydW4tYXMtdXNlciIs
            CiAgICAgICAgICAgICItdSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHR5
            cGU9c3RyLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyBhIHVzZXIgbmFtZSB0byBjaGVjayB0
            aGUgY2x1c3RlcidzIGhlYWx0aCBzdGF0dXMsIG11c3QgYmUgcHJlZml4ZWQgYnkgJ2NsaWVudC4n
            IiwKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tY2x1c3Rlci1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAg
            aGVscD0iS3ViZXJuZXRlcyBjbHVzdGVyIG5hbWUobGVnYWN5IGZsYWcpLCBOb3RlOiBFaXRoZXIg
            dXNlIHRoaXMgb3IgLS1rOHMtY2x1c3Rlci1uYW1lIiwKICAgICAgICApCiAgICAgICAgY29tbW9u
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tazhzLWNsdXN0ZXItbmFtZSIsIGRl
            ZmF1bHQ9IiIsIGhlbHA9Ikt1YmVybmV0ZXMgY2x1c3RlciBuYW1lIgogICAgICAgICkKICAgICAg
            ICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1uYW1lc3BhY2UiLAog
            ICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICBoZWxwPSJOYW1lc3BhY2Ugd2hlcmUg
            Q2VwaENsdXN0ZXIgaXMgcnVubmluZyIsCiAgICAgICAgKQogICAgICAgIGNvbW1vbl9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1wb29sLXByZWZpeCIsIGRlZmF1bHQ9IiIs
            IGhlbHA9IlJHVyBQb29sIHByZWZpeCIKICAgICAgICApCiAgICAgICAgY29tbW9uX2dyb3VwLmFk
            ZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24iLAog
            ICAgICAgICAgICBkZWZhdWx0PUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSZXN0cmljdCBjZXBo
            Q1NJS2V5cmluZ3MgYXV0aCBwZXJtaXNzaW9ucyB0byBzcGVjaWZpYyBwb29scywgY2x1c3Rlci4i
            CiAgICAgICAgICAgICsgIk1hbmRhdG9yeSBmbGFncyB0aGF0IG5lZWQgdG8gYmUgc2V0IGFyZSAt
            LXJiZC1kYXRhLXBvb2wtbmFtZSwgYW5kIC0tazhzLWNsdXN0ZXItbmFtZS4iCiAgICAgICAgICAg
            ICsgIi0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBmbGFnIGNhbiBhbHNvIGJlIHBhc3NlZCBpbiBj
            YXNlIG9mIGNlcGhmcyB1c2VyIHJlc3RyaWN0aW9uLCBzbyBpdCBjYW4gcmVzdHJpY3QgdXNlciB0
            byBwYXJ0aWN1bGFyIGNlcGhmcyBmaWxlc3lzdGVtIgogICAgICAgICAgICArICJzYW1wbGUgcnVu
            OiBgcHl0aG9uMyAvZXRjL2NlcGgvY3JlYXRlLWV4dGVybmFsLWNsdXN0ZXItcmVzb3VyY2VzLnB5
            IC0tY2VwaGZzLWZpbGVzeXN0ZW0tbmFtZSBteWZzIC0tcmJkLWRhdGEtcG9vbC1uYW1lIHJlcGxp
            Y2Fwb29sIC0tazhzLWNsdXN0ZXItbmFtZSByb29rc3RvcmFnZSAtLXJlc3RyaWN0ZWQtYXV0aC1w
            ZXJtaXNzaW9uIHRydWVgIgogICAgICAgICAgICArICJOb3RlOiBSZXN0cmljdGluZyB0aGUgY3Np
            LXVzZXJzIHBlciBwb29sLCBhbmQgcGVyIGNsdXN0ZXIgd2lsbCByZXF1aXJlIGNyZWF0aW5nIG5l
            dyBjc2ktdXNlcnMgYW5kIG5ldyBzZWNyZXRzIGZvciB0aGF0IGNzaS11c2Vycy4iCiAgICAgICAg
            ICAgICsgIlNvIGFwcGx5IHRoZXNlIHNlY3JldHMgb25seSB0byBuZXcgYENvbnN1bWVyIGNsdXN0
            ZXJgIGRlcGxveW1lbnQgd2hpbGUgdXNpbmcgdGhlIHNhbWUgYFNvdXJjZSBjbHVzdGVyYC4iLAog
            ICAgICAgICkKICAgICAgICBjb21tb25fZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS12Mi1wb3J0LWVuYWJsZSIsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAg
            ICAgICAgIGRlZmF1bHQ9VHJ1ZSwgICMgZGVmYXVsdCBpcyBUcnVlIHRvIGVuYWJsZSB2MiBtb24g
            cG9ydCBpbiBkb3duc3RyZWFtCiAgICAgICAgICAgIGhlbHA9IkVuYWJsZSB2MiBtb24gcG9ydCgz
            MzAwKSBmb3IgbW9ucyIsCiAgICAgICAgKQoKICAgICAgICBvdXRwdXRfZ3JvdXAgPSBhcmdQLmFk
            ZF9hcmd1bWVudF9ncm91cCgib3V0cHV0IikKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3Vt
            ZW50KAogICAgICAgICAgICAiLS1mb3JtYXQiLAogICAgICAgICAgICAiLXQiLAogICAgICAgICAg
            ICBjaG9pY2VzPVsianNvbiIsICJiYXNoIl0sCiAgICAgICAgICAgIGRlZmF1bHQ9Impzb24iLAog
            ICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgb3V0cHV0IGZvcm1hdCAoanNvbiB8IGJhc2gp
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tb3V0cHV0IiwKICAgICAgICAgICAgIi1vIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwK
            ICAgICAgICAgICAgaGVscD0iT3V0cHV0IHdpbGwgYmUgc3RvcmVkIGludG8gdGhlIHByb3ZpZGVk
            IGZpbGUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0i
            IiwKICAgICAgICAgICAgaGVscD0iUHJvdmlkZXMgdGhlIG5hbWUgb2YgdGhlIENlcGggZmlsZXN5
            c3RlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAg
            ICAgICAgICItLWNlcGhmcy1tZXRhZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0
            PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZzIG1l
            dGFkYXRhIHBvb2wiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50
            KAogICAgICAgICAgICAiLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUiLAogICAgICAgICAgICBkZWZh
            dWx0PSIiLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgY2VwaGZz
            IGRhdGEgcG9vbCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQo
            CiAgICAgICAgICAgICItLXJiZC1kYXRhLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9
            IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRl
            cyB0aGUgbmFtZSBvZiB0aGUgUkJEIGRhdGFwb29sIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0
            X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tYWxpYXMtcmJkLWRhdGEtcG9vbC1u
            YW1lIiwKICAgICAgICAgICAgZGVmYXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2Us
            CiAgICAgICAgICAgIGhlbHA9IlByb3ZpZGVzIGFuIGFsaWFzIGZvciB0aGUgIFJCRCBkYXRhIHBv
            b2wgbmFtZSwgbmVjZXNzYXJ5IGlmIGEgc3BlY2lhbCBjaGFyYWN0ZXIgaXMgcHJlc2VudCBpbiB0
            aGUgcG9vbCBuYW1lIHN1Y2ggYXMgYSBwZXJpb2Qgb3IgdW5kZXJzY29yZSIsCiAgICAgICAgKQog
            ICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJndy1lbmRw
            b2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNl
            LAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50IChpbiBgPElQdjQ+OjxQ
            T1JUPmAgb3IgYDxbSVB2Nl0+OjxQT1JUPmAgb3IgYDxGUUROPjo8UE9SVD5gIGZvcm1hdCkiLAog
            ICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAi
            LS1yZ3ctdGxzLWNlcnQtcGF0aCIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAg
            IHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJSQURPUyBHYXRld2F5IGVuZHBvaW50
            IFRMUyBjZXJ0aWZpY2F0ZSAob3IgaW50ZXJtZWRpYXRlIHNpZ25pbmcgY2VydGlmaWNhdGUpIiwK
            ICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAg
            Ii0tcmd3LXNraXAtdGxzIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAg
            IGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9Iklnbm9yZSBUTFMgY2VydGlmaWNhdGlv
            biB2YWxpZGF0aW9uIHdoZW4gYSBzZWxmLXNpZ25lZCBjZXJ0aWZpY2F0ZSBpcyBwcm92aWRlZCAo
            Tk9UIFJFQ09NTUVOREVEIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1
            bWVudCgKICAgICAgICAgICAgIi0tbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRl
            ZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJD
            ZXBoIE1hbmFnZXIgcHJvbWV0aGV1cyBleHBvcnRlciBlbmRwb2ludHMgKGNvbW1hIHNlcGFyYXRl
            ZCBsaXN0IG9mIChmb3JtYXQgYDxJUHY0PmAgb3IgYDxbSVB2Nl0+YCBvciBgPEZRRE4+YCkgZW50
            cmllcyBvZiBhY3RpdmUgYW5kIHN0YW5kYnkgbWdycykiLAogICAgICAgICkKICAgICAgICBvdXRw
            dXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1tb25pdG9yaW5nLWVuZHBvaW50
            LXBvcnQiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxz
            ZSwKICAgICAgICAgICAgaGVscD0iQ2VwaCBNYW5hZ2VyIHByb21ldGhldXMgZXhwb3J0ZXIgcG9y
            dCIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAg
            ICAgICItLXNraXAtbW9uaXRvcmluZy1lbmRwb2ludCIsCiAgICAgICAgICAgIGRlZmF1bHQ9RmFs
            c2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1ZSIsCiAgICAgICAgICAgIGhlbHA9IkRv
            IG5vdCBjaGVjayBmb3IgYSBtb25pdG9yaW5nIGVuZHBvaW50IGZvciB0aGUgQ2VwaCBjbHVzdGVy
            IiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAg
            ICAgIi0tcmJkLW1ldGFkYXRhLWVjLXBvb2wtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJQcm92aWRlcyB0
            aGUgbmFtZSBvZiBlcmFzdXJlIGNvZGVkIFJCRCBtZXRhZGF0YSBwb29sIiwKICAgICAgICApCiAg
            ICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0tZHJ5LXJ1biIs
            CiAgICAgICAgICAgIGRlZmF1bHQ9RmFsc2UsCiAgICAgICAgICAgIGFjdGlvbj0ic3RvcmVfdHJ1
            ZSIsCiAgICAgICAgICAgIGhlbHA9IkRyeSBydW4gcHJpbnRzIHRoZSBleGVjdXRlZCBjb21tYW5k
            cyB3aXRob3V0IHJ1bm5pbmcgdGhlbSIsCiAgICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5h
            ZGRfYXJndW1lbnQoCiAgICAgICAgICAgICItLXJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAg
            IGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxw
            PSJEaXZpZGVzIGEgcG9vbCBpbnRvIHNlcGFyYXRlIGxvZ2ljYWwgbmFtZXNwYWNlcywgdXNlZCBm
            b3IgY3JlYXRpbmcgUkJEIFBWQyBpbiBhIENlcGhCbG9ja1Bvb2xSYWRvc05hbWVzcGFjZSAoc2hv
            dWxkIGJlIGxvd2VyIGNhc2UpIiwKICAgICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tc3Vidm9sdW1lLWdyb3VwIiwKICAgICAgICAgICAgZGVm
            YXVsdD0iIiwKICAgICAgICAgICAgcmVxdWlyZWQ9RmFsc2UsCiAgICAgICAgICAgIGhlbHA9InBy
            b3ZpZGVzIHRoZSBuYW1lIG9mIHRoZSBzdWJ2b2x1bWUgZ3JvdXAiLAogICAgICAgICkKICAgICAg
            ICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS1yZ3ctcmVhbG0tbmFt
            ZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAog
            ICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3LXJlYWxtIiwKICAg
            ICAgICApCiAgICAgICAgb3V0cHV0X2dyb3VwLmFkZF9hcmd1bWVudCgKICAgICAgICAgICAgIi0t
            cmd3LXpvbmUtbmFtZSIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVp
            cmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJwcm92aWRlcyB0aGUgbmFtZSBvZiB0aGUgcmd3
            LXpvbmUiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS1yZ3ctem9uZWdyb3VwLW5hbWUiLAogICAgICAgICAgICBkZWZhdWx0PSIiLAog
            ICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgaGVscD0icHJvdmlkZXMgdGhl
            IG5hbWUgb2YgdGhlIHJndy16b25lZ3JvdXAiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3Jv
            dXAuYWRkX2FyZ3VtZW50KAogICAgICAgICAgICAiLS10b3BvbG9neS1wb29scyIsCiAgICAgICAg
            ICAgIGRlZmF1bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBo
            ZWxwPSJjb21tYS1zZXBhcmF0ZWQgbGlzdCBvZiB0b3BvbG9neS1jb25zdHJhaW5lZCByYmQgcG9v
            bHMiLAogICAgICAgICkKICAgICAgICBvdXRwdXRfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAgICAg
            ICAgICAiLS10b3BvbG9neS1mYWlsdXJlLWRvbWFpbi1sYWJlbCIsCiAgICAgICAgICAgIGRlZmF1
            bHQ9IiIsCiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJrOHMg
            Y2x1c3RlciBmYWlsdXJlIGRvbWFpbiBsYWJlbCAoZXhhbXBsZTogem9uZSwgcmFjaywgb3IgaG9z
            dCkgZm9yIHRoZSB0b3BvbG9neS1wb29scyB0aGF0IG1hdGNoIHRoZSBjZXBoIGRvbWFpbiIsCiAg
            ICAgICAgKQogICAgICAgIG91dHB1dF9ncm91cC5hZGRfYXJndW1lbnQoCiAgICAgICAgICAgICIt
            LXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIsCiAgICAgICAgICAgIGRlZmF1bHQ9IiIs
            CiAgICAgICAgICAgIHJlcXVpcmVkPUZhbHNlLAogICAgICAgICAgICBoZWxwPSJjb21tYS1zZXBh
            cmF0ZWQgbGlzdCBvZiB0aGUgazhzIGNsdXN0ZXIgZmFpbHVyZSBkb21haW4gdmFsdWVzIGNvcnJl
            c3BvbmRpbmcgdG8gZWFjaCBvZiB0aGUgcG9vbHMgaW4gdGhlIGB0b3BvbG9neS1wb29sc2AgbGlz
            dCIsCiAgICAgICAgKQoKICAgICAgICB1cGdyYWRlX2dyb3VwID0gYXJnUC5hZGRfYXJndW1lbnRf
            Z3JvdXAoInVwZ3JhZGUiKQogICAgICAgIHVwZ3JhZGVfZ3JvdXAuYWRkX2FyZ3VtZW50KAogICAg
            ICAgICAgICAiLS11cGdyYWRlIiwKICAgICAgICAgICAgYWN0aW9uPSJzdG9yZV90cnVlIiwKICAg
            ICAgICAgICAgZGVmYXVsdD1GYWxzZSwKICAgICAgICAgICAgaGVscD0iVXBncmFkZXMgdGhlIGNl
            cGhDU0lLZXlyaW5ncyhGb3IgZXhhbXBsZTogY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIp
            IGFuZCBjbGllbnQuaGVhbHRoY2hlY2tlciBjZXBoIHVzZXJzIHdpdGggbmV3IHBlcm1pc3Npb25z
            IG5lZWRlZCBmb3IgdGhlIG5ldyBjbHVzdGVyIHZlcnNpb24gYW5kIG9sZGVyIHBlcm1pc3Npb24g
            d2lsbCBzdGlsbCBiZSBhcHBsaWVkLiIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhv
            bjMgL2V0Yy9jZXBoL2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3Jh
            ZGVgLCB0aGlzIHdpbGwgdXBncmFkZSBhbGwgdGhlIGRlZmF1bHQgY3NpIHVzZXJzKG5vbi1yZXN0
            cmljdGVkKSIKICAgICAgICAgICAgKyAiRm9yIHJlc3RyaWN0ZWQgdXNlcnMoRm9yIGV4YW1wbGU6
            IGNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyLW9wZW5zaGlmdC1zdG9yYWdlLW15ZnMpLCB1
            c2VycyBjcmVhdGVkIHVzaW5nIC0tcmVzdHJpY3RlZC1hdXRoLXBlcm1pc3Npb24gZmxhZyBuZWVk
            IHRvIHBhc3MgbWFuZGF0b3J5IGZsYWdzIgogICAgICAgICAgICArICJtYW5kYXRvcnkgZmxhZ3M6
            ICctLXJiZC1kYXRhLXBvb2wtbmFtZSwgLS1rOHMtY2x1c3Rlci1uYW1lIGFuZCAtLXJ1bi1hcy11
            c2VyJyBmbGFncyB3aGlsZSB1cGdyYWRpbmciCiAgICAgICAgICAgICsgImluIGNhc2Ugb2YgY2Vw
            aGZzIHVzZXJzIGlmIHlvdSBoYXZlIHBhc3NlZCAtLWNlcGhmcy1maWxlc3lzdGVtLW5hbWUgZmxh
            ZyB3aGlsZSBjcmVhdGluZyB1c2VyIHRoZW4gd2hpbGUgdXBncmFkaW5nIGl0IHdpbGwgYmUgbWFu
            ZGF0b3J5IHRvbyIKICAgICAgICAgICAgKyAiU2FtcGxlIHJ1bjogYHB5dGhvbjMgL2V0Yy9jZXBo
            L2NyZWF0ZS1leHRlcm5hbC1jbHVzdGVyLXJlc291cmNlcy5weSAtLXVwZ3JhZGUgLS1yYmQtZGF0
            YS1wb29sLW5hbWUgcmVwbGljYXBvb2wgLS1rOHMtY2x1c3Rlci1uYW1lIHJvb2tzdG9yYWdlICAt
            LXJ1bi1hcy11c2VyIGNsaWVudC5jc2ktcmJkLW5vZGUtcm9va3N0b3JhZ2UtcmVwbGljYXBvb2xg
            IgogICAgICAgICAgICArICJQUzogQW4gZXhpc3Rpbmcgbm9uLXJlc3RyaWN0ZWQgdXNlciBjYW5u
            b3QgYmUgY29udmVydGVkIHRvIGEgcmVzdHJpY3RlZCB1c2VyIGJ5IHVwZ3JhZGluZy4iCiAgICAg
            ICAgICAgICsgIlVwZ3JhZGUgZmxhZyBzaG91bGQgb25seSBiZSB1c2VkIHRvIGFwcGVuZCBuZXcg
            cGVybWlzc2lvbnMgdG8gdXNlcnMsIGl0IHNob3VsZG4ndCBiZSB1c2VkIGZvciBjaGFuZ2luZyB1
            c2VyIGFscmVhZHkgYXBwbGllZCBwZXJtaXNzaW9uLCBmb3IgZXhhbXBsZSB5b3Ugc2hvdWxkbid0
            IGNoYW5nZSBpbiB3aGljaCBwb29sIHVzZXIgaGFzIGFjY2VzcyIsCiAgICAgICAgKQoKICAgICAg
            ICAjIEFkZCBjb21tYW5kLWxpbmUgYXJndW1lbnRzCiAgICAgICAgY29uZmlnX2dyb3VwID0gYXJn
            UC5hZGRfYXJndW1lbnRfZ3JvdXAoImNvbmZpZyIpCiAgICAgICAgY29uZmlnX2dyb3VwLmFkZF9h
            cmd1bWVudCgKICAgICAgICAgICAgIi0tY29uZmlnLWZpbGUiLAogICAgICAgICAgICB0eXBlPXN0
            ciwKICAgICAgICAgICAgaGVscD0iUGF0aCB0byB0aGUgY29uZmlndXJhdGlvbiBmaWxlLCBQcmlv
            cml0eTogY29tbWFuZC1saW5lLWFyZ3MgPiBjb25maWcuaW5pIGZpbGUgdmFsdWVzID4gZGVmYXVs
            dCB2YWx1ZXMiLAogICAgICAgICAgICByZXF1aXJlZD1GYWxzZSwKICAgICAgICAgICAgZGVmYXVs
            dD0iIiwKICAgICAgICApCgogICAgICAgIGlmIGFyZ3NfdG9fcGFyc2U6CiAgICAgICAgICAgIGFz
            c2VydCAoCiAgICAgICAgICAgICAgICB0eXBlKGFyZ3NfdG9fcGFyc2UpID09IGxpc3QKICAgICAg
            ICAgICAgKSwgIkFyZ3VtZW50IHRvICdnZW5fYXJnX3BhcnNlcicgc2hvdWxkIGJlIGEgbGlzdCIK
            ICAgICAgICBlbHNlOgogICAgICAgICAgICBhcmdzX3RvX3BhcnNlID0gc3lzLmFyZ3ZbMTpdCiAg
            ICAgICAgcmV0dXJuIGFyZ1AucGFyc2VfYXJncyhhcmdzX3RvX3BhcnNlKQoKICAgIGRlZiB2YWxp
            ZGF0ZV9yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lKHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZToKICAgICAgICAgICAgcmJkX21ldGFk
            YXRhX2VjX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX21ldGFkYXRhX2VjX3Bvb2xf
            bmFtZQogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfZGF0
            YV9wb29sX25hbWUKCiAgICAgICAgICAgIGlmIHJiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAg
            ICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAg
            ICAgICJGbGFnICctLXJiZC1kYXRhLXBvb2wtbmFtZScgc2hvdWxkIG5vdCBiZSBlbXB0eSIKICAg
            ICAgICAgICAgICAgICkKCiAgICAgICAgICAgIGlmIHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUg
            PT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAog
            ICAgICAgICAgICAgICAgICAgICJGbGFnICctLXJiZC1tZXRhZGF0YS1lYy1wb29sLW5hbWUnIHNo
            b3VsZCBub3QgYmUgZW1wdHkiCiAgICAgICAgICAgICAgICApCgogICAgICAgICAgICBjbWRfanNv
            biA9IHsicHJlZml4IjogIm9zZCBkdW1wIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAg
            cmV0X3ZhbCwganNvbl9vdXQsIGVycl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNt
            ZF9qc29uKQogICAgICAgICAgICBpZiByZXRfdmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAw
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICBmIntjbWRfanNvblsncHJlZml4J119IGNvbW1hbmQgZmFpbGVkLlxuIgog
            ICAgICAgICAgICAgICAgICAgIGYiRXJyb3I6IHtlcnJfbXNnIGlmIHJldF92YWwgIT0gMCBlbHNl
            IHNlbGYuRU1QVFlfT1VUUFVUX0xJU1R9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBt
            ZXRhZGF0YV9wb29sX2V4aXN0LCBwb29sX2V4aXN0ID0gRmFsc2UsIEZhbHNlCgogICAgICAgICAg
            ICBmb3Iga2V5IGluIGpzb25fb3V0WyJwb29scyJdOgogICAgICAgICAgICAgICAgIyBpZiBlcmFz
            dXJlX2NvZGVfcHJvZmlsZSBpcyBlbXB0eSBhbmQgcG9vbCBuYW1lIGV4aXN0cyB0aGVuIGl0IHJl
            cGxpY2EgcG9vbAogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGtleVsi
            ZXJhc3VyZV9jb2RlX3Byb2ZpbGUiXSA9PSAiIgogICAgICAgICAgICAgICAgICAgIGFuZCBrZXlb
            InBvb2xfbmFtZSJdID09IHJiZF9tZXRhZGF0YV9lY19wb29sX25hbWUKICAgICAgICAgICAgICAg
            ICk6CiAgICAgICAgICAgICAgICAgICAgbWV0YWRhdGFfcG9vbF9leGlzdCA9IFRydWUKICAgICAg
            ICAgICAgICAgICMgaWYgZXJhc3VyZV9jb2RlX3Byb2ZpbGUgaXMgbm90IGVtcHR5IGFuZCBwb29s
            IG5hbWUgZXhpc3RzIHRoZW4gaXQgaXMgZWMgcG9vbAogICAgICAgICAgICAgICAgaWYga2V5WyJl
            cmFzdXJlX2NvZGVfcHJvZmlsZSJdIGFuZCBrZXlbInBvb2xfbmFtZSJdID09IHJiZF9wb29sX25h
            bWU6CiAgICAgICAgICAgICAgICAgICAgcG9vbF9leGlzdCA9IFRydWUKCiAgICAgICAgICAgIGlm
            IG5vdCBtZXRhZGF0YV9wb29sX2V4aXN0OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9u
            RmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiUHJvdmlkZWQgcmJkX2VjX21l
            dGFkYXRhX3Bvb2wgbmFtZSwiCiAgICAgICAgICAgICAgICAgICAgZiIge3JiZF9tZXRhZGF0YV9l
            Y19wb29sX25hbWV9LCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgaWYgbm90IHBvb2xfZXhpc3Q6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiUHJvdmlkZWQgcmJkX2RhdGFfcG9v
            bCBuYW1lLCB7cmJkX3Bvb2xfbmFtZX0sIGRvZXMgbm90IGV4aXN0IgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICByZXR1cm4gcmJkX21ldGFkYXRhX2VjX3Bvb2xfbmFtZQoKICAgIGRlZiBk
            cnlfcnVuKHNlbGYsIG1zZyk6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICBwcmludCgiRXhlY3V0ZTogIiArICInIiArIG1zZyArICInIikKCiAgICBkZWYg
            dmFsaWRhdGVfcmd3X2VuZHBvaW50X3Rsc19jZXJ0KHNlbGYpOgogICAgICAgIGlmIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgIHdpdGggb3BlbihzZWxmLl9h
            cmdfcGFyc2VyLnJnd190bHNfY2VydF9wYXRoLCBlbmNvZGluZz0idXRmOCIpIGFzIGY6CiAgICAg
            ICAgICAgICAgICBjb250ZW50cyA9IGYucmVhZCgpCiAgICAgICAgICAgICAgICByZXR1cm4gY29u
            dGVudHMucnN0cmlwKCkKCiAgICBkZWYgX2NoZWNrX2NvbmZsaWN0aW5nX29wdGlvbnMoc2VsZik6
            CiAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZSBhbmQgbm90IHNlbGYuX2Fy
            Z19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgIkVpdGhlciAnLS11cGdyYWRlJyBvciAn
            LS1yYmQtZGF0YS1wb29sLW5hbWUgPHBvb2xfbmFtZT4nIHNob3VsZCBiZSBzcGVjaWZpZWQiCiAg
            ICAgICAgICAgICkKCiAgICBkZWYgX2ludmFsaWRfZW5kcG9pbnQoc2VsZiwgZW5kcG9pbnRfc3Ry
            KToKICAgICAgICAjIGV4dHJhY3QgdGhlIHBvcnQgYnkgZ2V0dGluZyB0aGUgbGFzdCBzcGxpdCBv
            biBgOmAgZGVsaW1pdGVyCiAgICAgICAgdHJ5OgogICAgICAgICAgICBlbmRwb2ludF9zdHJfaXAs
            IHBvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICBleGNlcHQgVmFsdWVF
            cnJvcjoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbihmIk5vdCBh
            IHByb3BlciBlbmRwb2ludDoge2VuZHBvaW50X3N0cn0iKQoKICAgICAgICB0cnk6CiAgICAgICAg
            ICAgIGlmIGVuZHBvaW50X3N0cl9pcFswXSA9PSAiWyI6CiAgICAgICAgICAgICAgICBlbmRwb2lu
            dF9zdHJfaXAgPSBlbmRwb2ludF9zdHJfaXBbMSA6IGxlbihlbmRwb2ludF9zdHJfaXApIC0gMV0K
            ICAgICAgICAgICAgaXBfdHlwZSA9ICgKICAgICAgICAgICAgICAgICJJUHY0IiBpZiB0eXBlKGlw
            X2FkZHJlc3MoZW5kcG9pbnRfc3RyX2lwKSkgaXMgSVB2NEFkZHJlc3MgZWxzZSAiSVB2NiIKICAg
            ICAgICAgICAgKQogICAgICAgIGV4Y2VwdCBWYWx1ZUVycm9yOgogICAgICAgICAgICBpcF90eXBl
            ID0gIkZRRE4iCiAgICAgICAgaWYgbm90IHBvcnQuaXNkaWdpdCgpOgogICAgICAgICAgICByYWlz
            ZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKGYiUG9ydCBub3QgdmFsaWQ6IHtwb3J0fSIpCiAg
            ICAgICAgaW50UG9ydCA9IGludChwb3J0KQogICAgICAgIGlmIGludFBvcnQgPCAxIG9yIGludFBv
            cnQgPiAyKioxNiAtIDE6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZiJPdXQgb2YgcmFuZ2UgcG9ydCBudW1iZXI6IHtwb3J0fSIpCgogICAgICAgIHJldHVybiBp
            cF90eXBlCgogICAgZGVmIGVuZHBvaW50X2RpYWwoc2VsZiwgZW5kcG9pbnRfc3RyLCBpcF90eXBl
            LCB0aW1lb3V0PTMsIGNlcnQ9Tm9uZSk6CiAgICAgICAgIyBpZiB0aGUgJ2NsdXN0ZXInIGluc3Rh
            bmNlIGlzIGEgZHVtbXkgb25lLAogICAgICAgICMgZG9uJ3QgdHJ5IHRvIHJlYWNoIG91dCB0byB0
            aGUgZW5kcG9pbnQKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3RlciwgRHVtbXlSYWRv
            cyk6CiAgICAgICAgICAgIHJldHVybiAiIiwgIiIsICIiCiAgICAgICAgaWYgaXBfdHlwZSA9PSAi
            SVB2NiI6CiAgICAgICAgICAgIHRyeToKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCwg
            ZW5kcG9pbnRfc3RyX3BvcnQgPSBlbmRwb2ludF9zdHIucnNwbGl0KCI6IiwgMSkKICAgICAgICAg
            ICAgZXhjZXB0IFZhbHVlRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgIGYiTm90IGEgcHJvcGVyIGVuZHBvaW50
            OiB7ZW5kcG9pbnRfc3RyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgZW5kcG9p
            bnRfc3RyX2lwWzBdICE9ICJbIjoKICAgICAgICAgICAgICAgIGVuZHBvaW50X3N0cl9pcCA9ICJb
            IiArIGVuZHBvaW50X3N0cl9pcCArICJdIgogICAgICAgICAgICBlbmRwb2ludF9zdHIgPSAiOiIu
            am9pbihbZW5kcG9pbnRfc3RyX2lwLCBlbmRwb2ludF9zdHJfcG9ydF0pCgogICAgICAgIHByb3Rv
            Y29scyA9IFsiaHR0cCIsICJodHRwcyJdCiAgICAgICAgcmVzcG9uc2VfZXJyb3IgPSBOb25lCiAg
            ICAgICAgZm9yIHByZWZpeCBpbiBwcm90b2NvbHM6CiAgICAgICAgICAgIHRyeToKICAgICAgICAg
            ICAgICAgIGVwID0gZiJ7cHJlZml4fTovL3tlbmRwb2ludF9zdHJ9IgogICAgICAgICAgICAgICAg
            dmVyaWZ5ID0gTm9uZQogICAgICAgICAgICAgICAgIyBJZiB2ZXJpZnkgaXMgc2V0IHRvIGEgcGF0
            aCB0byBhIGRpcmVjdG9yeSwKICAgICAgICAgICAgICAgICMgdGhlIGRpcmVjdG9yeSBtdXN0IGhh
            dmUgYmVlbiBwcm9jZXNzZWQgdXNpbmcgdGhlIGNfcmVoYXNoIHV0aWxpdHkgc3VwcGxpZWQgd2l0
            aCBPcGVuU1NMLgogICAgICAgICAgICAgICAgaWYgcHJlZml4ID09ICJodHRwcyIgYW5kIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3NraXBfdGxzOgogICAgICAgICAgICAgICAgICAgIHZlcmlmeSA9IEZh
            bHNlCiAgICAgICAgICAgICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGlt
            ZW91dCwgdmVyaWZ5PUZhbHNlKQogICAgICAgICAgICAgICAgZWxpZiBwcmVmaXggPT0gImh0dHBz
            IiBhbmQgY2VydDoKICAgICAgICAgICAgICAgICAgICB2ZXJpZnkgPSBjZXJ0CiAgICAgICAgICAg
            ICAgICAgICAgciA9IHJlcXVlc3RzLmhlYWQoZXAsIHRpbWVvdXQ9dGltZW91dCwgdmVyaWZ5PWNl
            cnQpCiAgICAgICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgICAgIHIgPSByZXF1ZXN0
            cy5oZWFkKGVwLCB0aW1lb3V0PXRpbWVvdXQpCiAgICAgICAgICAgICAgICBpZiByLnN0YXR1c19j
            b2RlID09IDIwMDoKICAgICAgICAgICAgICAgICAgICByZXR1cm4gcHJlZml4LCB2ZXJpZnksICIi
            CiAgICAgICAgICAgIGV4Y2VwdCBFeGNlcHRpb24gYXMgZXJyOgogICAgICAgICAgICAgICAgcmVz
            cG9uc2VfZXJyb3IgPSBlcnIKICAgICAgICAgICAgICAgIGNvbnRpbnVlCiAgICAgICAgc3lzLnN0
            ZGVyci53cml0ZSgKICAgICAgICAgICAgZiJ1bmFibGUgdG8gY29ubmVjdCB0byBlbmRwb2ludDog
            e2VuZHBvaW50X3N0cn0sIGZhaWxlZCBlcnJvcjoge3Jlc3BvbnNlX2Vycm9yfSIKICAgICAgICAp
            CiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAgIiIsCiAgICAgICAgICAgICIiLAogICAgICAg
            ICAgICAoIi0xIiksCiAgICAgICAgKQoKICAgIGRlZiBwYXJzZV9jb25maWdfZmlsZShzZWxmLCBj
            b25maWdfZmlsZSk6CiAgICAgICAgY29uZmlnID0gY29uZmlncGFyc2VyLkNvbmZpZ1BhcnNlcigp
            CiAgICAgICAgY29uZmlnLnJlYWQoY29uZmlnX2ZpbGUpCiAgICAgICAgZm9yIGFyZyBpbiBsaXN0
            KHZhcnMoc2VsZi5fYXJnX3BhcnNlcikpOgogICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxh
            Zy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRlcm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxh
            Zy1uYW1lLAogICAgICAgICAgICAjIHNvIHdlIGNhbiBnZXQgdGhvc2UgdmFsdWVzIGZyb20gY29u
            ZmlnIGZpbGUKICAgICAgICAgICAgYXJndW1lbnQgPSBhcmcucmVwbGFjZSgiXyIsICItIikKICAg
            ICAgICAgICAgYXJndW1lbnRWYWx1ZSA9IHN0cihnZXRhdHRyKHNlbGYuX2FyZ19wYXJzZXIsIGFy
            ZykpCiAgICAgICAgICAgIGNvbmZpZ192YWx1ZSA9IGNvbmZpZy5nZXQoIkNvbmZpZ3VyYXRpb25z
            IiwgYXJndW1lbnQsIGZhbGxiYWNrPU5vbmUpCiAgICAgICAgICAgICMgZ2l2ZSBwcmlvcml0eSB0
            byBjb21tYW5kIGxpbmUgYXJndW1lbnQsIGlmIGNvbW1hbmQgbGluZSBhcmd1bWVudCBpcyBub3Qg
            cHJlc2VudCB1c2UgY29uZmlnLmluaSB2YWx1ZSwKICAgICAgICAgICAgaWYgKHN0cihzeXMuYXJn
            dikuZmluZChhcmd1bWVudCkgPT0gLTEpIGFuZCAoCiAgICAgICAgICAgICAgICAoY29uZmlnX3Zh
            bHVlICE9IE5vbmUpIGFuZCAoY29uZmlnX3ZhbHVlICE9ICIiKQogICAgICAgICAgICApOgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5fX3NldGF0dHJfXyhhcmcsIGNvbmZpZ192YWx1
            ZSkKCiAgICAgICAgcmV0dXJuIGNvbmZpZwoKICAgIGRlZiBfX2luaXRfXyhzZWxmLCBhcmdfbGlz
            dD1Ob25lKToKICAgICAgICBzZWxmLm91dF9tYXAgPSB7fQogICAgICAgIHNlbGYuX2V4Y2x1ZGVk
            X2tleXMgPSBzZXQoKQogICAgICAgIHNlbGYuX2FyZ19wYXJzZXIgPSBzZWxmLmdlbl9hcmdfcGFy
            c2VyKGFyZ3NfdG9fcGFyc2U9YXJnX2xpc3QpCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5j
            b25maWdfZmlsZToKICAgICAgICAgICAgc2VsZi5jb25maWcgPSBzZWxmLnBhcnNlX2NvbmZpZ19m
            aWxlKHNlbGYuX2FyZ19wYXJzZXIuY29uZmlnX2ZpbGUpCiAgICAgICAgc2VsZi5fY2hlY2tfY29u
            ZmxpY3Rpbmdfb3B0aW9ucygpCiAgICAgICAgc2VsZi5ydW5fYXNfdXNlciA9IHNlbGYuX2FyZ19w
            YXJzZXIucnVuX2FzX3VzZXIKICAgICAgICBzZWxmLm91dHB1dF9maWxlID0gc2VsZi5fYXJnX3Bh
            cnNlci5vdXRwdXQKICAgICAgICBzZWxmLmNlcGhfY29uZiA9IHNlbGYuX2FyZ19wYXJzZXIuY2Vw
            aF9jb25mCiAgICAgICAgc2VsZi5jZXBoX2tleXJpbmcgPSBzZWxmLl9hcmdfcGFyc2VyLmtleXJp
            bmcKICAgICAgICAjIGlmIHVzZXIgbm90IHByb3ZpZGVkLCBnaXZlIGEgZGVmYXVsdCB1c2VyCiAg
            ICAgICAgaWYgbm90IHNlbGYucnVuX2FzX3VzZXIgYW5kIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnVw
            Z3JhZGU6CiAgICAgICAgICAgIHNlbGYucnVuX2FzX3VzZXIgPSBzZWxmLkVYVEVSTkFMX1VTRVJf
            TkFNRQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCBhbmQg
            bm90IHNlbGYuX2FyZ19wYXJzZXIudXBncmFkZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNl
            ci5yZ3dfcG9vbF9wcmVmaXggPSBzZWxmLkRFRkFVTFRfUkdXX1BPT0xfUFJFRklYCiAgICAgICAg
            aWYgc2VsZi5jZXBoX2NvbmY6CiAgICAgICAgICAgIGt3YXJncyA9IHt9CiAgICAgICAgICAgIGlm
            IHNlbGYuY2VwaF9rZXlyaW5nOgogICAgICAgICAgICAgICAga3dhcmdzWyJjb25mIl0gPSB7Imtl
            eXJpbmciOiBzZWxmLmNlcGhfa2V5cmluZ30KICAgICAgICAgICAgc2VsZi5jbHVzdGVyID0gcmFk
            b3MuUmFkb3MoY29uZmZpbGU9c2VsZi5jZXBoX2NvbmYsICoqa3dhcmdzKQogICAgICAgIGVsc2U6
            CiAgICAgICAgICAgIHNlbGYuY2x1c3RlciA9IHJhZG9zLlJhZG9zKCkKICAgICAgICAgICAgc2Vs
            Zi5jbHVzdGVyLmNvbmZfcmVhZF9maWxlKCkKICAgICAgICBzZWxmLmNsdXN0ZXIuY29ubmVjdCgp
            CgogICAgZGVmIHNodXRkb3duKHNlbGYpOgogICAgICAgIGlmIHNlbGYuY2x1c3Rlci5zdGF0ZSA9
            PSAiY29ubmVjdGVkIjoKICAgICAgICAgICAgc2VsZi5jbHVzdGVyLnNodXRkb3duKCkKCiAgICBk
            ZWYgZ2V0X2ZzaWQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVuOgog
            ICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoIGZzaWQiKQogICAgICAgIHJldHVy
            biBzdHIoc2VsZi5jbHVzdGVyLmdldF9mc2lkKCkpCgogICAgZGVmIF9jb21tb25fY21kX2pzb25f
            Z2VuKHNlbGYsIGNtZF9qc29uKToKICAgICAgICBjbWQgPSBqc29uLmR1bXBzKGNtZF9qc29uLCBz
            b3J0X2tleXM9VHJ1ZSkKICAgICAgICByZXRfdmFsLCBjbWRfb3V0LCBlcnJfbXNnID0gc2VsZi5j
            bHVzdGVyLm1vbl9jb21tYW5kKGNtZCwgYiIiKQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIu
            dmVyYm9zZToKICAgICAgICAgICAgcHJpbnQoZiJDb21tYW5kIElucHV0OiB7Y21kfSIpCiAgICAg
            ICAgICAgIHByaW50KAogICAgICAgICAgICAgICAgZiJSZXR1cm4gVmFsOiB7cmV0X3ZhbH1cbkNv
            bW1hbmQgT3V0cHV0OiB7Y21kX291dH1cbiIKICAgICAgICAgICAgICAgIGYiRXJyb3IgTWVzc2Fn
            ZToge2Vycl9tc2d9XG4tLS0tLS0tLS0tXG4iCiAgICAgICAgICAgICkKICAgICAgICBqc29uX291
            dCA9IHt9CiAgICAgICAgIyBpZiB0aGVyZSBpcyBubyBlcnJvciAoaS5lOyByZXRfdmFsIGlzIFpF
            Uk8pIGFuZCAnY21kX291dCcgaXMgbm90IGVtcHR5CiAgICAgICAgIyB0aGVuIGNvbnZlcnQgJ2Nt
            ZF9vdXQnIHRvIGEganNvbiBvdXRwdXQKICAgICAgICBpZiByZXRfdmFsID09IDAgYW5kIGNtZF9v
            dXQ6CiAgICAgICAgICAgIGpzb25fb3V0ID0ganNvbi5sb2FkcyhjbWRfb3V0KQogICAgICAgIHJl
            dHVybiByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZwoKICAgIGRlZiBnZXRfY2VwaF9leHRlcm5h
            bF9tb25fZGF0YShzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogInF1b3J1bV9z
            dGF0dXMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5
            X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25b
            InByZWZpeCJdKQogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29t
            bW9uX2NtZF9qc29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3Vj
            Y2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQp
            ID09IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAiJ3F1b3J1bV9zdGF0dXMnIGNvbW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAg
            ICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9P
            VVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICBxX2xlYWRlcl9uYW1lID0ganNvbl9v
            dXRbInF1b3J1bV9sZWFkZXJfbmFtZSJdCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHt9CiAg
            ICAgICAgcV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCA9IFsKICAgICAgICAgICAgbCBmb3IgbCBpbiBq
            c29uX291dFsibW9ubWFwIl1bIm1vbnMiXSBpZiBsWyJuYW1lIl0gPT0gcV9sZWFkZXJfbmFtZQog
            ICAgICAgIF0KICAgICAgICBpZiBsZW4ocV9sZWFkZXJfbWF0Y2hpbmdfbGlzdCkgPT0gMDoKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigiTm8gbWF0Y2hpbmcgJ21v
            bicgZGV0YWlscyBmb3VuZCIpCiAgICAgICAgcV9sZWFkZXJfZGV0YWlscyA9IHFfbGVhZGVyX21h
            dGNoaW5nX2xpc3RbMF0KICAgICAgICAjIGdldCB0aGUgYWRkcmVzcyB2ZWN0b3Igb2YgdGhlIHF1
            b3J1bS1sZWFkZXIKICAgICAgICBxX2xlYWRlcl9hZGRydmVjID0gcV9sZWFkZXJfZGV0YWlscy5n
            ZXQoInB1YmxpY19hZGRycyIsIHt9KS5nZXQoImFkZHJ2ZWMiLCBbXSkKICAgICAgICBpcF9hZGRy
            ID0gc3RyKHFfbGVhZGVyX2RldGFpbHNbInB1YmxpY19hZGRyIl0uc3BsaXQoIi8iKVswXSkKCiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci52Ml9wb3J0X2VuYWJsZToKICAgICAgICAgICAgaWYg
            cV9sZWFkZXJfYWRkcnZlY1swXVsidHlwZSJdID09ICJ2MiI6CiAgICAgICAgICAgICAgICBpcF9h
            ZGRyID0gcV9sZWFkZXJfYWRkcnZlY1swXVsiYWRkciJdCiAgICAgICAgICAgIGVsaWYgbGVuKHFf
            bGVhZGVyX2FkZHJ2ZWMpID4gMSBhbmQgcV9sZWFkZXJfYWRkcnZlY1sxXVsidHlwZSJdID09ICJ2
            MiI6CiAgICAgICAgICAgICAgICBpcF9hZGRyID0gcV9sZWFkZXJfYWRkcnZlY1sxXVsiYWRkciJd
            CiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAg
            ICAgICAgICAgICAgICAgICIndjInIGFkZHJlc3MgdHlwZSBub3QgcHJlc2VudCwgYW5kICd2Mi1w
            b3J0LWVuYWJsZScgZmxhZyBpcyBwcm92aWRlZCIKICAgICAgICAgICAgICAgICkKCiAgICAgICAg
            cmV0dXJuIGYie3N0cihxX2xlYWRlcl9uYW1lKX09e2lwX2FkZHJ9IgoKICAgIGRlZiBfY29udmVy
            dF9ob3N0bmFtZV90b19pcChzZWxmLCBob3N0X25hbWUsIHBvcnQsIGlwX3R5cGUpOgogICAgICAg
            ICMgaWYgJ2NsdXN0ZXInIGluc3RhbmNlIGlzIGEgZHVtbXkgdHlwZSwKICAgICAgICAjIGNhbGwg
            dGhlIGR1bW15IGluc3RhbmNlJ3MgImNvbnZlcnQiIG1ldGhvZAogICAgICAgIGlmIG5vdCBob3N0
            X25hbWU6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIkVtcHR5
            IGhvc3RuYW1lIHByb3ZpZGVkIikKICAgICAgICBpZiBpc2luc3RhbmNlKHNlbGYuY2x1c3Rlciwg
            RHVtbXlSYWRvcyk6CiAgICAgICAgICAgIHJldHVybiBzZWxmLmNsdXN0ZXIuX2NvbnZlcnRfaG9z
            dG5hbWVfdG9faXAoaG9zdF9uYW1lKQoKICAgICAgICBpZiBpcF90eXBlID09ICJGUUROIjoKICAg
            ICAgICAgICAgIyBjaGVjayB3aGljaCBpcCBGUUROIHNob3VsZCBiZSBjb252ZXJ0ZWQgdG8sIElQ
            djQgb3IgSVB2NgogICAgICAgICAgICAjIGNoZWNrIHRoZSBob3N0IGlwLCB0aGUgZW5kcG9pbnQg
            aXAgdHlwZSB3b3VsZCBiZSBzaW1pbGFyIHRvIGhvc3QgaXAKICAgICAgICAgICAgY21kX2pzb24g
            PSB7InByZWZpeCI6ICJvcmNoIGhvc3QgbHMiLCAiZm9ybWF0IjogImpzb24ifQogICAgICAgICAg
            ICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgICAgICMgaWYgdGhlcmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVt
            cHQsCiAgICAgICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAg
            ICAgICAgICAgICInb3JjaCBob3N0IGxzJyBjb21tYW5kIGZhaWxlZC5cbiIKICAgICAgICAgICAg
            ICAgICAgICBmIkVycm9yOiB7ZXJyX21zZyBpZiByZXRfdmFsICE9IDAgZWxzZSBzZWxmLkVNUFRZ
            X09VVFBVVF9MSVNUfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgaG9zdF9hZGRyID0g
            anNvbl9vdXRbMF1bImFkZHIiXQogICAgICAgICAgICAjIGFkZCA6ODAgc2FtcGxlIHBvcnQgaW4g
            aXBfdHlwZSwgYXMgX2ludmFsaWRfZW5kcG9pbnQgYWxzbyB2ZXJpZnkgcG9ydAogICAgICAgICAg
            ICBob3N0X2lwX3R5cGUgPSBzZWxmLl9pbnZhbGlkX2VuZHBvaW50KGhvc3RfYWRkciArICI6ODAi
            KQogICAgICAgICAgICBpbXBvcnQgc29ja2V0CgogICAgICAgICAgICBpcCA9IFtdCiAgICAgICAg
            ICAgICMgZXhhbXBsZSBvdXRwdXQgWyg8QWRkcmVzc0ZhbWlseS5BRl9JTkVUOiAyPiwgPFNvY2tl
            dEtpbmQuU09DS19TVFJFQU06IDE+LCA2LCAnJywgKCc5My4xODQuMjE2LjM0JywgODApKSwgLi4u
            XQogICAgICAgICAgICAjIHdlIG5lZWQgdG8gZ2V0IDkzLjE4NC4yMTYuMzQgc28gaXQgd291bGQg
            YmUgaXBbMF1bNF1bMF0KICAgICAgICAgICAgaWYgaG9zdF9pcF90eXBlID09ICJJUHY2IjoKICAg
            ICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAgICAgICAgICAgICAgICAg
            IGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVUNiwgcHJvdG89c29ja2V0LklQ
            UFJPVE9fVENQCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsaWYgaG9zdF9pcF90eXBl
            ID09ICJJUHY0IjoKICAgICAgICAgICAgICAgIGlwID0gc29ja2V0LmdldGFkZHJpbmZvKAogICAg
            ICAgICAgICAgICAgICAgIGhvc3RfbmFtZSwgcG9ydCwgZmFtaWx5PXNvY2tldC5BRl9JTkVULCBw
            cm90bz1zb2NrZXQuSVBQUk9UT19UQ1AKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZGVs
            IHNvY2tldAogICAgICAgICAgICByZXR1cm4gaXBbMF1bNF1bMF0KICAgICAgICByZXR1cm4gaG9z
            dF9uYW1lCgogICAgZGVmIGdldF9hY3RpdmVfYW5kX3N0YW5kYnlfbWdycyhzZWxmKToKICAgICAg
            ICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAiIiwgc2Vs
            Zi5kcnlfcnVuKCJjZXBoIHN0YXR1cyIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0
            ID0gc2VsZi5fYXJnX3BhcnNlci5tb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKICAgICAgICBtb25p
            dG9yaW5nX2VuZHBvaW50X2lwX2xpc3QgPSBzZWxmLl9hcmdfcGFyc2VyLm1vbml0b3JpbmdfZW5k
            cG9pbnQKICAgICAgICBzdGFuZGJ5X21ncnMgPSBbXQogICAgICAgIGlmIG5vdCBtb25pdG9yaW5n
            X2VuZHBvaW50X2lwX2xpc3Q6CiAgICAgICAgICAgIGNtZF9qc29uID0geyJwcmVmaXgiOiAic3Rh
            dHVzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVy
            cl9tc2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICAgICAj
            IGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAogICAgICAgICAgICBpZiByZXRf
            dmFsICE9IDAgb3IgbGVuKGpzb25fb3V0KSA9PSAwOgogICAgICAgICAgICAgICAgcmFpc2UgRXhl
            Y3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAgICAgICAiJ21nciBzZXJ2aWNl
            cycgY29tbWFuZCBmYWlsZWQuXG4iCiAgICAgICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9t
            c2cgaWYgcmV0X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAg
            ICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnQgPSAoCiAgICAgICAgICAg
            ICAgICBqc29uX291dC5nZXQoIm1ncm1hcCIsIHt9KS5nZXQoInNlcnZpY2VzIiwge30pLmdldCgi
            cHJvbWV0aGV1cyIsICIiKQogICAgICAgICAgICApCiAgICAgICAgICAgIGlmIG5vdCBtb25pdG9y
            aW5nX2VuZHBvaW50OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiY2FuJ3QgZmluZCBtb25pdG9yaW5nX2VuZHBvaW50
            LCBwcm9tZXRoZXVzIG1vZHVsZSBtaWdodCBub3QgYmUgZW5hYmxlZCwgIgogICAgICAgICAgICAg
            ICAgICAgICJlbmFibGUgdGhlIG1vZHVsZSBieSBydW5uaW5nICdjZXBoIG1nciBtb2R1bGUgZW5h
            YmxlIHByb21ldGhldXMnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIG5vdyBjaGVj
            ayB0aGUgc3RhbmQtYnkgbWdyLXMKICAgICAgICAgICAgc3RhbmRieV9hcnIgPSBqc29uX291dC5n
            ZXQoIm1ncm1hcCIsIHt9KS5nZXQoInN0YW5kYnlzIiwgW10pCiAgICAgICAgICAgIGZvciBlYWNo
            X3N0YW5kYnkgaW4gc3RhbmRieV9hcnI6CiAgICAgICAgICAgICAgICBpZiAibmFtZSIgaW4gZWFj
            aF9zdGFuZGJ5LmtleXMoKToKICAgICAgICAgICAgICAgICAgICBzdGFuZGJ5X21ncnMuYXBwZW5k
            KGVhY2hfc3RhbmRieVsibmFtZSJdKQogICAgICAgICAgICB0cnk6CiAgICAgICAgICAgICAgICBw
            YXJzZWRfZW5kcG9pbnQgPSB1cmxwYXJzZShtb25pdG9yaW5nX2VuZHBvaW50KQogICAgICAgICAg
            ICBleGNlcHQgVmFsdWVFcnJvcjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1
            cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgZiJpbnZhbGlkIGVuZHBvaW50OiB7bW9u
            aXRvcmluZ19lbmRwb2ludH0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIG1vbml0b3Jp
            bmdfZW5kcG9pbnRfaXBfbGlzdCA9IHBhcnNlZF9lbmRwb2ludC5ob3N0bmFtZQogICAgICAgICAg
            ICBpZiBub3QgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0OgogICAgICAgICAgICAgICAgbW9uaXRv
            cmluZ19lbmRwb2ludF9wb3J0ID0gc3RyKHBhcnNlZF9lbmRwb2ludC5wb3J0KQoKICAgICAgICAj
            IGlmIG1vbml0b3JpbmcgZW5kcG9pbnQgcG9ydCBpcyBub3Qgc2V0LCBwdXQgYSBkZWZhdWx0IG1v
            biBwb3J0CiAgICAgICAgaWYgbm90IG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydDoKICAgICAgICAg
            ICAgbW9uaXRvcmluZ19lbmRwb2ludF9wb3J0ID0gc2VsZi5ERUZBVUxUX01PTklUT1JJTkdfRU5E
            UE9JTlRfUE9SVAoKICAgICAgICAjIHVzZXIgY291bGQgZ2l2ZSBjb21tYSBhbmQgc3BhY2Ugc2Vw
            YXJhdGVkIGlucHV0cyAobGlrZSAtLW1vbml0b3JpbmctZW5kcG9pbnQ9IjxpcDE+LCA8aXAyPiIp
            CiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0ID0gbW9uaXRvcmluZ19lbmRwb2lu
            dF9pcF9saXN0LnJlcGxhY2UoIiwiLCAiICIpCiAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9p
            cF9saXN0X3NwbGl0ID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0LnNwbGl0KCkKICAgICAg
            ICAjIGlmIG1vbml0b3JpbmctZW5kcG9pbnQgY291bGQgbm90IGJlIGZvdW5kLCByYWlzZSBhbiBl
            cnJvcgogICAgICAgIGlmIGxlbihtb25pdG9yaW5nX2VuZHBvaW50X2lwX2xpc3Rfc3BsaXQpID09
            IDA6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oIk5vICdtb25p
            dG9yaW5nLWVuZHBvaW50JyBmb3VuZCIpCiAgICAgICAgIyBmaXJzdCBpcCBpcyB0cmVhdGVkIGFz
            IHRoZSBtYWluIG1vbml0b3JpbmctZW5kcG9pbnQKICAgICAgICBtb25pdG9yaW5nX2VuZHBvaW50
            X2lwID0gbW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzBdCiAgICAgICAgIyByZXN0
            IG9mIHRoZSBpcC1zIGFyZSBhZGRlZCB0byB0aGUgJ3N0YW5kYnlfbWdycycgbGlzdAogICAgICAg
            IHN0YW5kYnlfbWdycy5leHRlbmQobW9uaXRvcmluZ19lbmRwb2ludF9pcF9saXN0X3NwbGl0WzE6
            XSkKICAgICAgICBmYWlsZWRfaXAgPSBtb25pdG9yaW5nX2VuZHBvaW50X2lwCgogICAgICAgIG1v
            bml0b3JpbmdfZW5kcG9pbnQgPSAiOiIuam9pbigKICAgICAgICAgICAgW21vbml0b3JpbmdfZW5k
            cG9pbnRfaXAsIG1vbml0b3JpbmdfZW5kcG9pbnRfcG9ydF0KICAgICAgICApCiAgICAgICAgaXBf
            dHlwZSA9IHNlbGYuX2ludmFsaWRfZW5kcG9pbnQobW9uaXRvcmluZ19lbmRwb2ludCkKICAgICAg
            ICB0cnk6CiAgICAgICAgICAgIG1vbml0b3JpbmdfZW5kcG9pbnRfaXAgPSBzZWxmLl9jb252ZXJ0
            X2hvc3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgbW9uaXRvcmluZ19lbmRwb2ludF9pcCwg
            bW9uaXRvcmluZ19lbmRwb2ludF9wb3J0LCBpcF90eXBlCiAgICAgICAgICAgICkKICAgICAgICAg
            ICAgIyBjb2xsZWN0IGFsbCB0aGUgJ3N0YW5kLWJ5JyBtZ3IgaXBzCiAgICAgICAgICAgIG1ncl9p
            cHMgPSBbXQogICAgICAgICAgICBmb3IgZWFjaF9zdGFuZGJ5X21nciBpbiBzdGFuZGJ5X21ncnM6
            CiAgICAgICAgICAgICAgICBmYWlsZWRfaXAgPSBlYWNoX3N0YW5kYnlfbWdyCiAgICAgICAgICAg
            ICAgICBtZ3JfaXBzLmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9jb252ZXJ0X2hv
            c3RuYW1lX3RvX2lwKAogICAgICAgICAgICAgICAgICAgICAgICBlYWNoX3N0YW5kYnlfbWdyLCBt
            b25pdG9yaW5nX2VuZHBvaW50X3BvcnQsIGlwX3R5cGUKICAgICAgICAgICAgICAgICAgICApCiAg
            ICAgICAgICAgICAgICApCiAgICAgICAgZXhjZXB0OgogICAgICAgICAgICByYWlzZSBFeGVjdXRp
            b25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiJDb252ZXJzaW9uIG9mIGhvc3Q6
            IHtmYWlsZWRfaXB9IHRvIElQIGZhaWxlZC4gIgogICAgICAgICAgICAgICAgIlBsZWFzZSBlbnRl
            ciB0aGUgSVAgYWRkcmVzc2VzIG9mIGFsbCB0aGUgY2VwaC1tZ3JzIHdpdGggdGhlICctLW1vbml0
            b3JpbmctZW5kcG9pbnQnIGZsYWciCiAgICAgICAgICAgICkKCiAgICAgICAgXywgXywgZXJyID0g
            c2VsZi5lbmRwb2ludF9kaWFsKG1vbml0b3JpbmdfZW5kcG9pbnQsIGlwX3R5cGUpCiAgICAgICAg
            aWYgZXJyID09ICItMSI6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRp
            b24oZXJyKQogICAgICAgICMgYWRkIHRoZSB2YWxpZGF0ZWQgYWN0aXZlIG1nciBJUCBpbnRvIHRo
            ZSBmaXJzdCBpbmRleAogICAgICAgIG1ncl9pcHMuaW5zZXJ0KDAsIG1vbml0b3JpbmdfZW5kcG9p
            bnRfaXApCiAgICAgICAgYWxsX21ncl9pcHNfc3RyID0gIiwiLmpvaW4obWdyX2lwcykKICAgICAg
            ICByZXR1cm4gYWxsX21ncl9pcHNfc3RyLCBtb25pdG9yaW5nX2VuZHBvaW50X3BvcnQKCiAgICBk
            ZWYgY2hlY2tfdXNlcl9leGlzdChzZWxmLCB1c2VyKToKICAgICAgICBjbWRfanNvbiA9IHsicHJl
            Zml4IjogImF1dGggZ2V0IiwgImVudGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0K
            ICAgICAgICByZXRfdmFsLCBqc29uX291dCwgXyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4o
            Y21kX2pzb24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoK
            ICAgICAgICAgICAgcmV0dXJuICIiCiAgICAgICAgcmV0dXJuIHN0cihqc29uX291dFswXVsia2V5
            Il0pCgogICAgZGVmIGdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fwc19hbmRfZW50aXR5KHNlbGYp
            OgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIKICAgICAg
            ICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogImFsbG93IHIsIGFsbG93IGNvbW1hbmQgJ29z
            ZCBibG9ja2xpc3QnIiwKICAgICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAg
            ICJvc2QiOiAiYWxsb3cgcncgdGFnIGNlcGhmcyBtZXRhZGF0YT0qIiwKICAgICAgICAgICAgIm1k
            cyI6ICJhbGxvdyAqIiwKICAgICAgICB9CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yZXN0
            cmljdGVkX2F1dGhfcGVybWlzc2lvbjoKICAgICAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuazhzX2NsdXN0ZXJfbmFtZQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rl
            cl9uYW1lID09ICIiOgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2Vw
            dGlvbigKICAgICAgICAgICAgICAgICAgICAiazhzX2NsdXN0ZXJfbmFtZSBub3QgZm91bmQsIHBs
            ZWFzZSBzZXQgdGhlICctLWs4cy1jbHVzdGVyLW5hbWUnIGZsYWciCiAgICAgICAgICAgICAgICAp
            CiAgICAgICAgICAgIGNlcGhmc19maWxlc3lzdGVtID0gc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNf
            ZmlsZXN5c3RlbV9uYW1lCiAgICAgICAgICAgIGlmIGNlcGhmc19maWxlc3lzdGVtID09ICIiOgog
            ICAgICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0ZXJfbmFtZX0iCiAg
            ICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNf
            Y2x1c3Rlcl9uYW1lfS17Y2VwaGZzX2ZpbGVzeXN0ZW19IgogICAgICAgICAgICAgICAgY2Fwc1si
            b3NkIl0gPSBmImFsbG93IHJ3IHRhZyBjZXBoZnMgbWV0YWRhdGE9e2NlcGhmc19maWxlc3lzdGVt
            fSIKCiAgICAgICAgcmV0dXJuIGNhcHMsIGVudGl0eQoKICAgIGRlZiBnZXRfY2VwaGZzX25vZGVf
            Y2Fwc19hbmRfZW50aXR5KHNlbGYpOgogICAgICAgIGVudGl0eSA9ICJjbGllbnQuY3NpLWNlcGhm
            cy1ub2RlIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAgICJtb24iOiAiYWxsb3cgciwgYWxs
            b3cgY29tbWFuZCAnb3NkIGJsb2NrbGlzdCciLAogICAgICAgICAgICAibWdyIjogImFsbG93IHJ3
            IiwKICAgICAgICAgICAgIm9zZCI6ICJhbGxvdyBydyB0YWcgY2VwaGZzICo9KiIsCiAgICAgICAg
            ICAgICJtZHMiOiAiYWxsb3cgcnciLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFy
            c2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJtaXNzaW9uOgogICAgICAgICAgICBrOHNfY2x1c3Rlcl9u
            YW1lID0gc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1c3Rlcl9uYW1lCiAgICAgICAgICAgIGlmIGs4
            c19jbHVzdGVyX25hbWUgPT0gIiI6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWls
            dXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAgICJrOHNfY2x1c3Rlcl9uYW1lIG5vdCBm
            b3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tazhzLWNsdXN0ZXItbmFtZScgZmxhZyIKICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW0gPSBzZWxmLl9hcmdfcGFyc2Vy
            LmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAgICAgaWYgY2VwaGZzX2ZpbGVzeXN0ZW0g
            PT0gIiI6CiAgICAgICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRpdHl9LXtrOHNfY2x1c3Rlcl9u
            YW1lfSIKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVudGl0eSA9IGYie2VudGl0
            eX0te2s4c19jbHVzdGVyX25hbWV9LXtjZXBoZnNfZmlsZXN5c3RlbX0iCiAgICAgICAgICAgICAg
            ICBjYXBzWyJvc2QiXSA9IGYiYWxsb3cgcncgdGFnIGNlcGhmcyAqPXtjZXBoZnNfZmlsZXN5c3Rl
            bX0iCgogICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2VudGl0eSgKICAg
            ICAgICBzZWxmLAogICAgICAgIGVudGl0eSwKICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAg
            IGFsaWFzX3JiZF9wb29sX25hbWUsCiAgICAgICAgazhzX2NsdXN0ZXJfbmFtZSwKICAgICAgICBy
            YWRvc19uYW1lc3BhY2UsCiAgICApOgogICAgICAgIGlmICgKICAgICAgICAgICAgcmJkX3Bvb2xf
            bmFtZS5jb3VudCgiLiIpICE9IDAKICAgICAgICAgICAgb3IgcmJkX3Bvb2xfbmFtZS5jb3VudCgi
            XyIpICE9IDAKICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZSAhPSAiIgogICAgICAg
            ICAgICAjIGNoZWNraW5nIGFsaWFzX3JiZF9wb29sX25hbWUgaXMgbm90IGVtcHR5IGFzIHRoZXJl
            IG1heWJlIGEgc3BlY2lhbCBjaGFyYWN0ZXIgdXNlZCBvdGhlciB0aGFuIC4gb3IgXwogICAgICAg
            ICk6CiAgICAgICAgICAgIGlmIGFsaWFzX3JiZF9wb29sX25hbWUgPT0gIiI6CiAgICAgICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgICAg
            ICJwbGVhc2Ugc2V0IHRoZSAnLS1hbGlhcy1yYmQtZGF0YS1wb29sLW5hbWUnIGZsYWcgYXMgdGhl
            IHJiZCBkYXRhIHBvb2wgbmFtZSBjb250YWlucyAnLicgb3IgJ18nIgogICAgICAgICAgICAgICAg
            KQogICAgICAgICAgICBpZiAoCiAgICAgICAgICAgICAgICBhbGlhc19yYmRfcG9vbF9uYW1lLmNv
            dW50KCIuIikgIT0gMAogICAgICAgICAgICAgICAgb3IgYWxpYXNfcmJkX3Bvb2xfbmFtZS5jb3Vu
            dCgiXyIpICE9IDAKICAgICAgICAgICAgKToKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIictLWFsaWFzLXJiZC1kYXRh
            LXBvb2wtbmFtZScgZmxhZyB2YWx1ZSBzaG91bGQgbm90IGNvbnRhaW4gJy4nIG9yICdfJyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgZW50aXR5ID0gZiJ7ZW50aXR5fS17azhzX2NsdXN0
            ZXJfbmFtZX0te2FsaWFzX3JiZF9wb29sX25hbWV9IgogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IGVudGl0eSA9IGYie2VudGl0eX0te2s4c19jbHVzdGVyX25hbWV9LXtyYmRfcG9vbF9uYW1lfSIK
            CiAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlOgogICAgICAgICAgICBlbnRpdHkgPSBmIntlbnRp
            dHl9LXtyYWRvc19uYW1lc3BhY2V9IgogICAgICAgIHJldHVybiBlbnRpdHkKCiAgICBkZWYgZ2V0
            X3JiZF9wcm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoc2VsZik6CiAgICAgICAgZW50aXR5ID0g
            ImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIgogICAgICAgIGNhcHMgPSB7CiAgICAgICAgICAg
            ICJtb24iOiAicHJvZmlsZSByYmQsIGFsbG93IGNvbW1hbmQgJ29zZCBibG9ja2xpc3QnIiwKICAg
            ICAgICAgICAgIm1nciI6ICJhbGxvdyBydyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSBy
            YmQiLAogICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0
            aF9wZXJtaXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNl
            ci5yYmRfZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNl
            bGYuX2FyZ19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19j
            bHVzdGVyX25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAg
            ICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAg
            ICAgICAgICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4
            ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBm
            bGFnIG5vdCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFn
            IgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIi
            OgogICAgICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAg
            ICAgICAgICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAn
            LS1rOHMtY2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBl
            bnRpdHkgPSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAg
            ICAgICAgICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFt
            ZSwKICAgICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRv
            c19uYW1lc3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNl
            ICE9ICIiOgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgZiJwcm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25h
            bWVzcGFjZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAg
            ICAgICBjYXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgog
            ICAgICAgIHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X3JiZF9ub2RlX2NhcHNfYW5k
            X2VudGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSAiY2xpZW50LmNzaS1yYmQtbm9kZSIKICAg
            ICAgICBjYXBzID0gewogICAgICAgICAgICAibW9uIjogInByb2ZpbGUgcmJkLCBhbGxvdyBjb21t
            YW5kICdvc2QgYmxvY2tsaXN0JyIsCiAgICAgICAgICAgICJvc2QiOiAicHJvZmlsZSByYmQiLAog
            ICAgICAgIH0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9wZXJt
            aXNzaW9uOgogICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRf
            ZGF0YV9wb29sX25hbWUKICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2Fy
            Z19wYXJzZXIuYWxpYXNfcmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgICAgIGs4c19jbHVzdGVy
            X25hbWUgPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAgICAgICAgICAgcmFk
            b3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICAg
            ICAgaWYgcmJkX3Bvb2xfbmFtZSA9PSAiIjoKICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlv
            bkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgIm1hbmRhdG9yeSBmbGFnIG5v
            dCBmb3VuZCwgcGxlYXNlIHNldCB0aGUgJy0tcmJkLWRhdGEtcG9vbC1uYW1lJyBmbGFnIgogICAg
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBrOHNfY2x1c3Rlcl9uYW1lID09ICIiOgogICAg
            ICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAg
            ICAgICAgICAibWFuZGF0b3J5IGZsYWcgbm90IGZvdW5kLCBwbGVhc2Ugc2V0IHRoZSAnLS1rOHMt
            Y2x1c3Rlci1uYW1lJyBmbGFnIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICBlbnRpdHkg
            PSBzZWxmLmdldF9lbnRpdHkoCiAgICAgICAgICAgICAgICBlbnRpdHksCiAgICAgICAgICAgICAg
            ICByYmRfcG9vbF9uYW1lLAogICAgICAgICAgICAgICAgYWxpYXNfcmJkX3Bvb2xfbmFtZSwKICAg
            ICAgICAgICAgICAgIGs4c19jbHVzdGVyX25hbWUsCiAgICAgICAgICAgICAgICByYWRvc19uYW1l
            c3BhY2UsCiAgICAgICAgICAgICkKICAgICAgICAgICAgaWYgcmFkb3NfbmFtZXNwYWNlICE9ICIi
            OgogICAgICAgICAgICAgICAgY2Fwc1sib3NkIl0gPSAoCiAgICAgICAgICAgICAgICAgICAgZiJw
            cm9maWxlIHJiZCBwb29sPXtyYmRfcG9vbF9uYW1lfSBuYW1lc3BhY2U9e3JhZG9zX25hbWVzcGFj
            ZX0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgIGVsc2U6CiAgICAgICAgICAgICAgICBj
            YXBzWyJvc2QiXSA9IGYicHJvZmlsZSByYmQgcG9vbD17cmJkX3Bvb2xfbmFtZX0iCgogICAgICAg
            IHJldHVybiBjYXBzLCBlbnRpdHkKCiAgICBkZWYgZ2V0X2RlZmF1bHRVc2VyX2NhcHNfYW5kX2Vu
            dGl0eShzZWxmKToKICAgICAgICBlbnRpdHkgPSBzZWxmLnJ1bl9hc191c2VyCiAgICAgICAgY2Fw
            cyA9IHsKICAgICAgICAgICAgIm1vbiI6ICJhbGxvdyByLCBhbGxvdyBjb21tYW5kIHF1b3J1bV9z
            dGF0dXMsIGFsbG93IGNvbW1hbmQgdmVyc2lvbiIsCiAgICAgICAgICAgICJtZ3IiOiAiYWxsb3cg
            Y29tbWFuZCBjb25maWciLAogICAgICAgICAgICAib3NkIjogZiJwcm9maWxlIHJiZC1yZWFkLW9u
            bHksIGFsbG93IHJ3eCBwb29sPXtzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeH0ucmd3
            Lm1ldGEsIGFsbG93IHIgcG9vbD0ucmd3LnJvb3QsIGFsbG93IHJ3IHBvb2w9e3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cuY29udHJvbCwgYWxsb3cgcnggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5sb2csIGFsbG93IHggcG9vbD17c2VsZi5f
            YXJnX3BhcnNlci5yZ3dfcG9vbF9wcmVmaXh9LnJndy5idWNrZXRzLmluZGV4IiwKICAgICAgICAg
            ICAgIm1kcyI6ICJhbGxvdyAqIiwKICAgICAgICB9CgogICAgICAgIHJldHVybiBjYXBzLCBlbnRp
            dHkKCiAgICBkZWYgZ2V0X2NhcHNfYW5kX2VudGl0eShzZWxmLCB1c2VyX25hbWUpOgogICAgICAg
            IGlmICJjbGllbnQuY3NpLWNlcGhmcy1wcm92aXNpb25lciIgaW4gdXNlcl9uYW1lOgogICAgICAg
            ICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlzaW9uZXIiICE9IHVzZXJfbmFtZToKICAg
            ICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmVzdHJpY3RlZF9hdXRoX3Blcm1pc3Npb24g
            PSBUcnVlCiAgICAgICAgICAgIHJldHVybiBzZWxmLmdldF9jZXBoZnNfcHJvdmlzaW9uZXJfY2Fw
            c19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgaW4gdXNl
            cl9uYW1lOgogICAgICAgICAgICBpZiAiY2xpZW50LmNzaS1jZXBoZnMtbm9kZSIgIT0gdXNlcl9u
            YW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVy
            bWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X2NlcGhmc19ub2RlX2Nh
            cHNfYW5kX2VudGl0eSgpCiAgICAgICAgaWYgImNsaWVudC5jc2ktcmJkLXByb3Zpc2lvbmVyIiBp
            biB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIg
            IT0gdXNlcl9uYW1lOgogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVk
            X2F1dGhfcGVybWlzc2lvbiA9IFRydWUKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZ2V0X3JiZF9w
            cm92aXNpb25lcl9jYXBzX2FuZF9lbnRpdHkoKQogICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1u
            b2RlIiBpbiB1c2VyX25hbWU6CiAgICAgICAgICAgIGlmICJjbGllbnQuY3NpLXJiZC1ub2RlIiAh
            PSB1c2VyX25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRf
            YXV0aF9wZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfcmJkX25v
            ZGVfY2Fwc19hbmRfZW50aXR5KCkKICAgICAgICBpZiAiY2xpZW50LmhlYWx0aGNoZWNrZXIiIGlu
            IHVzZXJfbmFtZToKICAgICAgICAgICAgaWYgImNsaWVudC5oZWFsdGhjaGVja2VyIiAhPSB1c2Vy
            X25hbWU6CiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJlc3RyaWN0ZWRfYXV0aF9w
            ZXJtaXNzaW9uID0gVHJ1ZQogICAgICAgICAgICByZXR1cm4gc2VsZi5nZXRfZGVmYXVsdFVzZXJf
            Y2Fwc19hbmRfZW50aXR5KCkKCiAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlv
            bigKICAgICAgICAgICAgZiJubyB1c2VyIGZvdW5kIHdpdGggdXNlcl9uYW1lOiB7dXNlcl9uYW1l
            fSwgIgogICAgICAgICAgICAiZ2V0X2NhcHNfYW5kX2VudGl0eSBjb21tYW5kIGZhaWxlZC5cbiIK
            ICAgICAgICApCgogICAgZGVmIGNyZWF0ZV9jZXBoQ1NJS2V5cmluZ191c2VyKHNlbGYsIHVzZXIp
            OgogICAgICAgICIiIgogICAgICAgIGNvbW1hbmQ6IGNlcGggYXV0aCBnZXQtb3ItY3JlYXRlIGNs
            aWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIG1vbiAnYWxsb3cgcicgbWdyICdhbGxvdyBydycg
            b3NkICdhbGxvdyBydyB0YWcgY2VwaGZzIG1ldGFkYXRhPSonCiAgICAgICAgIiIiCiAgICAgICAg
            Y2FwcywgZW50aXR5ID0gc2VsZi5nZXRfY2Fwc19hbmRfZW50aXR5KHVzZXIpCiAgICAgICAgY21k
            X2pzb24gPSB7CiAgICAgICAgICAgICJwcmVmaXgiOiAiYXV0aCBnZXQtb3ItY3JlYXRlIiwKICAg
            ICAgICAgICAgImVudGl0eSI6IGVudGl0eSwKICAgICAgICAgICAgImNhcHMiOiBbY2FwIGZvciBj
            YXBfbGlzdCBpbiBsaXN0KGNhcHMuaXRlbXMoKSkgZm9yIGNhcCBpbiBjYXBfbGlzdF0sCiAgICAg
            ICAgICAgICJmb3JtYXQiOiAianNvbiIsCiAgICAgICAgfQoKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAoCiAgICAgICAgICAgICAgICBzZWxm
            LmRyeV9ydW4oCiAgICAgICAgICAgICAgICAgICAgImNlcGggIgogICAgICAgICAgICAgICAgICAg
            ICsgY21kX2pzb25bInByZWZpeCJdCiAgICAgICAgICAgICAgICAgICAgKyAiICIKICAgICAgICAg
            ICAgICAgICAgICArIGNtZF9qc29uWyJlbnRpdHkiXQogICAgICAgICAgICAgICAgICAgICsgIiAi
            CiAgICAgICAgICAgICAgICAgICAgKyAiICIuam9pbihjbWRfanNvblsiY2FwcyJdKQogICAgICAg
            ICAgICAgICAgKSwKICAgICAgICAgICAgICAgICIiLAogICAgICAgICAgICApCiAgICAgICAgIyBj
            aGVjayBpZiB1c2VyIGFscmVhZHkgZXhpc3QKICAgICAgICB1c2VyX2tleSA9IHNlbGYuY2hlY2tf
            dXNlcl9leGlzdChlbnRpdHkpCiAgICAgICAgaWYgdXNlcl9rZXkgIT0gIiI6CiAgICAgICAgICAg
            IHJldHVybiB1c2VyX2tleSwgZiJ7ZW50aXR5LnNwbGl0KCcuJywgMSlbMV19IgogICAgICAgICAg
            ICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0eShjbGllbnQuY3NpLXJi
            ZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICAgICAgcmV0X3ZhbCwganNvbl9vdXQsIGVycl9t
            c2cgPSBzZWxmLl9jb21tb25fY21kX2pzb25fZ2VuKGNtZF9qc29uKQogICAgICAgICMgaWYgdGhl
            cmUgaXMgYW4gdW5zdWNjZXNzZnVsIGF0dGVtcHQsCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwIG9y
            IGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4
            Y2VwdGlvbigKICAgICAgICAgICAgICAgIGYiJ2F1dGggZ2V0LW9yLWNyZWF0ZSB7dXNlcn0nIGNv
            bW1hbmQgZmFpbGVkLlxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0
            X3ZhbCAhPSAwIGVsc2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAg
            ICAgICByZXR1cm4gc3RyKGpzb25fb3V0WzBdWyJrZXkiXSksIGYie2VudGl0eS5zcGxpdCgnLics
            IDEpWzFdfSIKICAgICAgICAjIGVudGl0eS5zcGxpdCgnLicsMSlbMV0gdG8gcmVuYW1lIGVudGl0
            eShjbGllbnQuY3NpLXJiZC1ub2RlKSBhcyBjc2ktcmJkLW5vZGUKCiAgICBkZWYgZ2V0X2NlcGhm
            c19kYXRhX3Bvb2xfZGV0YWlscyhzZWxmKToKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4Ijog
            ImZzIGxzIiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRy
            eV9ydW46CiAgICAgICAgICAgIHJldHVybiBzZWxmLmRyeV9ydW4oImNlcGggIiArIGNtZF9qc29u
            WyJwcmVmaXgiXSkKICAgICAgICByZXRfdmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2Nv
            bW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAgICAgIyBpZiB0aGVyZSBpcyBhbiB1bnN1
            Y2Nlc3NmdWwgYXR0ZW1wdCwgcmVwb3J0IGFuIGVycm9yCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAw
            OgogICAgICAgICAgICAjIGlmIGZzIGFuZCBkYXRhX3Bvb2wgYXJndW1lbnRzIGFyZSBub3Qgc2V0
            LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgc2VsZi5f
            YXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID09ICIiCiAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPT0gIiIKICAgICAgICAg
            ICAgKToKICAgICAgICAgICAgICAgIHJldHVybgogICAgICAgICAgICAjIGlmIHVzZXIgaGFzIHBy
            b3ZpZGVkIGFueSBvZiB0aGUKICAgICAgICAgICAgIyAnLS1jZXBoZnMtZmlsZXN5c3RlbS1uYW1l
            JyBvciAnLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUnIGFyZ3VtZW50cywKICAgICAgICAgICAgIyBy
            YWlzZSBhbiBleGNlcHRpb24gYXMgd2UgYXJlIHVuYWJsZSB0byB2ZXJpZnkgdGhlIGFyZ3MKICAg
            ICAgICAgICAgcmFpc2UgRXhlY3V0aW9uRmFpbHVyZUV4Y2VwdGlvbigKICAgICAgICAgICAgICAg
            IGYiJ2ZzIGxzJyBjZXBoIGNhbGwgZmFpbGVkIHdpdGggZXJyb3I6IHtlcnJfbXNnfSIKICAgICAg
            ICAgICAgKQoKICAgICAgICBtYXRjaGluZ19qc29uX291dCA9IHt9CiAgICAgICAgIyBpZiAnLS1j
            ZXBoZnMtZmlsZXN5c3RlbS1uYW1lJyBhcmd1bWVudCBpcyBwcm92aWRlZCwKICAgICAgICAjIGNo
            ZWNrIHdoZXRoZXIgdGhlIHByb3ZpZGVkIGZpbGVzeXN0ZW0tbmFtZSBleGlzdHMgb3Igbm90CiAg
            ICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lOgogICAgICAg
            ICAgICAjIGdldCB0aGUgbWF0Y2hpbmcgbGlzdAogICAgICAgICAgICBtYXRjaGluZ19qc29uX291
            dF9saXN0ID0gWwogICAgICAgICAgICAgICAgbWF0Y2hlZAogICAgICAgICAgICAgICAgZm9yIG1h
            dGNoZWQgaW4ganNvbl9vdXQKICAgICAgICAgICAgICAgIGlmIHN0cihtYXRjaGVkWyJuYW1lIl0p
            ID09IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQogICAgICAgICAgICBd
            CiAgICAgICAgICAgICMgdW5hYmxlIHRvIGZpbmQgYSBtYXRjaGluZyBmcy1uYW1lLCByYWlzZSBh
            biBlcnJvcgogICAgICAgICAgICBpZiBsZW4obWF0Y2hpbmdfanNvbl9vdXRfbGlzdCkgPT0gMDoK
            ICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAg
            ICAgICAgICAgICAgZiJGaWxlc3lzdGVtIHByb3ZpZGVkLCAne3NlbGYuX2FyZ19wYXJzZXIuY2Vw
            aGZzX2ZpbGVzeXN0ZW1fbmFtZX0nLCAiCiAgICAgICAgICAgICAgICAgICAgZiJpcyBub3QgZm91
            bmQgaW4gdGhlIGZzLWxpc3Q6IHtbc3RyKHhbJ25hbWUnXSkgZm9yIHggaW4ganNvbl9vdXRdfSIK
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXQgPSBtYXRjaGlu
            Z19qc29uX291dF9saXN0WzBdCiAgICAgICAgIyBpZiBjZXBoZnMgZmlsZXN5c3RlbSBuYW1lIGlz
            IG5vdCBwcm92aWRlZCwKICAgICAgICAjIHRyeSB0byBnZXQgYSBkZWZhdWx0IGZzIG5hbWUgYnkg
            ZG9pbmcgdGhlIGZvbGxvd2luZwogICAgICAgIGVsc2U6CiAgICAgICAgICAgICMgYS4gY2hlY2sg
            aWYgdGhlcmUgaXMgb25seSBvbmUgZmlsZXN5c3RlbSBpcyBwcmVzZW50CiAgICAgICAgICAgIGlm
            IGxlbihqc29uX291dCkgPT0gMToKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0g
            anNvbl9vdXRbMF0KICAgICAgICAgICAgIyBiLiBvciBlbHNlLCBjaGVjayBpZiBkYXRhX3Bvb2wg
            bmFtZSBpcyBwcm92aWRlZAogICAgICAgICAgICBlbGlmIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZz
            X2RhdGFfcG9vbF9uYW1lOgogICAgICAgICAgICAgICAgIyBhbmQgaWYgcHJlc2VudCwgY2hlY2sg
            d2hldGhlciB0aGVyZSBleGlzdHMgYSBmcyB3aGljaCBoYXMgdGhlIGRhdGFfcG9vbAogICAgICAg
            ICAgICAgICAgZm9yIGVhY2hKIGluIGpzb25fb3V0OgogICAgICAgICAgICAgICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuY2VwaGZzX2RhdGFfcG9vbF9uYW1lIGluIGVhY2hKWyJkYXRhX3Bvb2xz
            Il06CiAgICAgICAgICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0ID0gZWFjaEoKICAg
            ICAgICAgICAgICAgICAgICAgICAgYnJlYWsKICAgICAgICAgICAgICAgICMgaWYgdGhlcmUgaXMg
            bm8gbWF0Y2hpbmcgZnMgZXhpc3RzLCB0aGF0IG1lYW5zIHByb3ZpZGVkIGRhdGFfcG9vbCBuYW1l
            IGlzIGludmFsaWQKICAgICAgICAgICAgICAgIGlmIG5vdCBtYXRjaGluZ19qc29uX291dDoKICAg
            ICAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAg
            ICAgICAgICAgICAgICAgICBmIlByb3ZpZGVkIGRhdGFfcG9vbCBuYW1lLCB7c2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9LCIKICAgICAgICAgICAgICAgICAgICAgICAgIiBk
            b2VzIG5vdCBleGlzdHMiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGMuIGlm
            IG5vdGhpbmcgaXMgc2V0IGFuZCBjb3VsZG4ndCBmaW5kIGEgZGVmYXVsdCwKICAgICAgICAgICAg
            ZWxzZToKICAgICAgICAgICAgICAgICMganVzdCByZXR1cm4gc2lsZW50bHkKICAgICAgICAgICAg
            ICAgIHJldHVybgoKICAgICAgICBpZiBtYXRjaGluZ19qc29uX291dDoKICAgICAgICAgICAgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lID0gc3RyKG1hdGNoaW5nX2pzb25f
            b3V0WyJuYW1lIl0pCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX21ldGFkYXRh
            X3Bvb2xfbmFtZSA9IHN0cigKICAgICAgICAgICAgICAgIG1hdGNoaW5nX2pzb25fb3V0WyJtZXRh
            ZGF0YV9wb29sIl0KICAgICAgICAgICAgKQoKICAgICAgICBpZiBpc2luc3RhbmNlKG1hdGNoaW5n
            X2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0sIGxpc3QpOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGhhcyBhbHJlYWR5IHByb3ZpZGVkIGRhdGEtcG9vbC1uYW1lLAogICAgICAgICAgICAjIHRocm91
            Z2ggLS1jZXBoZnMtZGF0YS1wb29sLW5hbWUKICAgICAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNl
            ci5jZXBoZnNfZGF0YV9wb29sX25hbWU6CiAgICAgICAgICAgICAgICAjIGlmIHRoZSBwcm92aWRl
            ZCBuYW1lIGlzIG5vdCBtYXRjaGluZyB3aXRoIHRoZSBvbmUgaW4gdGhlIGxpc3QKICAgICAgICAg
            ICAgICAgIGlmICgKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19k
            YXRhX3Bvb2xfbmFtZQogICAgICAgICAgICAgICAgICAgIG5vdCBpbiBtYXRjaGluZ19qc29uX291
            dFsiZGF0YV9wb29scyJdCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIHJh
            aXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICAgICAgICAgIGYi
            UHJvdmlkZWQgZGF0YS1wb29sLW5hbWU6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9w
            b29sX25hbWV9JywgIgogICAgICAgICAgICAgICAgICAgICAgICAiZG9lc24ndCBtYXRjaCBmcm9t
            IHRoZSBkYXRhLXBvb2xzIGxpc3Q6ICIKICAgICAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4
            KSBmb3IgeCBpbiBtYXRjaGluZ19qc29uX291dFsnZGF0YV9wb29scyddXX0iCiAgICAgICAgICAg
            ICAgICAgICAgKQogICAgICAgICAgICAjIGlmIGRhdGFfcG9vbCBuYW1lIGlzIG5vdCBwcm92aWRl
            ZCwKICAgICAgICAgICAgIyB0aGVuIHRyeSB0byBmaW5kIGEgZGVmYXVsdCBkYXRhIHBvb2wgbmFt
            ZQogICAgICAgICAgICBlbHNlOgogICAgICAgICAgICAgICAgIyBpZiBubyBkYXRhX3Bvb2xzIGV4
            aXN0LCBzaWxlbnRseSByZXR1cm4KICAgICAgICAgICAgICAgIGlmIGxlbihtYXRjaGluZ19qc29u
            X291dFsiZGF0YV9wb29scyJdKSA9PSAwOgogICAgICAgICAgICAgICAgICAgIHJldHVybgogICAg
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUgPSBzdHIo
            CiAgICAgICAgICAgICAgICAgICAgbWF0Y2hpbmdfanNvbl9vdXRbImRhdGFfcG9vbHMiXVswXQog
            ICAgICAgICAgICAgICAgKQogICAgICAgICAgICAjIGlmIHRoZXJlIGFyZSBtb3JlIHRoYW4gb25l
            ICdkYXRhX3Bvb2xzJyBleGlzdCwKICAgICAgICAgICAgIyB0aGVuIHdhcm4gdGhlIHVzZXIgdGhh
            dCB3ZSBhcmUgdXNpbmcgdGhlIHNlbGVjdGVkIG5hbWUKICAgICAgICAgICAgaWYgbGVuKG1hdGNo
            aW5nX2pzb25fb3V0WyJkYXRhX3Bvb2xzIl0pID4gMToKICAgICAgICAgICAgICAgIHByaW50KAog
            ICAgICAgICAgICAgICAgICAgICJXQVJOSU5HOiBNdWx0aXBsZSBkYXRhIHBvb2xzIGRldGVjdGVk
            OiAiCiAgICAgICAgICAgICAgICAgICAgZiJ7W3N0cih4KSBmb3IgeCBpbiBtYXRjaGluZ19qc29u
            X291dFsnZGF0YV9wb29scyddXX1cbiIKICAgICAgICAgICAgICAgICAgICBmIlVzaW5nIHRoZSBk
            YXRhLXBvb2w6ICd7c2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWV9J1xuIgog
            ICAgICAgICAgICAgICAgKQoKICAgIGRlZiBjcmVhdGVfY2hlY2tlcktleShzZWxmLCB1c2VyKToK
            ICAgICAgICBjYXBzLCBlbnRpdHkgPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNlcikKICAg
            ICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRoIGdldC1vci1jcmVh
            dGUiLAogICAgICAgICAgICAiZW50aXR5IjogZW50aXR5LAogICAgICAgICAgICAiY2FwcyI6IFtj
            YXAgZm9yIGNhcF9saXN0IGluIGxpc3QoY2Fwcy5pdGVtcygpKSBmb3IgY2FwIGluIGNhcF9saXN0
            XSwKICAgICAgICAgICAgImZvcm1hdCI6ICJqc29uIiwKICAgICAgICB9CgogICAgICAgIGlmIHNl
            bGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAgICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigK
            ICAgICAgICAgICAgICAgICJjZXBoICIKICAgICAgICAgICAgICAgICsgY21kX2pzb25bInByZWZp
            eCJdCiAgICAgICAgICAgICAgICArICIgIgogICAgICAgICAgICAgICAgKyBjbWRfanNvblsiZW50
            aXR5Il0KICAgICAgICAgICAgICAgICsgIiAiCiAgICAgICAgICAgICAgICArICIgIi5qb2luKGNt
            ZF9qc29uWyJjYXBzIl0pCiAgICAgICAgICAgICkKICAgICAgICAjIGNoZWNrIGlmIHVzZXIgYWxy
            ZWFkeSBleGlzdAogICAgICAgIHVzZXJfa2V5ID0gc2VsZi5jaGVja191c2VyX2V4aXN0KGVudGl0
            eSkKICAgICAgICBpZiB1c2VyX2tleSAhPSAiIjoKICAgICAgICAgICAgcmV0dXJuIHVzZXJfa2V5
            CgogICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBlcnJfbXNnID0gc2VsZi5fY29tbW9uX2NtZF9q
            c29uX2dlbihjbWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBh
            dHRlbXB0LAogICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmIidhdXRoIGdldC1vci1jcmVhdGUge3NlbGYucnVuX2FzX3VzZXJ9JyBjb21tYW5kIGZhaWxl
            ZFxuIgogICAgICAgICAgICAgICAgZiJFcnJvcjoge2Vycl9tc2cgaWYgcmV0X3ZhbCAhPSAwIGVs
            c2Ugc2VsZi5FTVBUWV9PVVRQVVRfTElTVH0iCiAgICAgICAgICAgICkKICAgICAgICByZXR1cm4g
            c3RyKGpzb25fb3V0WzBdWyJrZXkiXSkKCiAgICBkZWYgZ2V0X2NlcGhfZGFzaGJvYXJkX2xpbmso
            c2VsZik6CiAgICAgICAgY21kX2pzb24gPSB7InByZWZpeCI6ICJtZ3Igc2VydmljZXMiLCAiZm9y
            bWF0IjogImpzb24ifQogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIuZHJ5X3J1bjoKICAgICAg
            ICAgICAgcmV0dXJuIHNlbGYuZHJ5X3J1bigiY2VwaCAiICsgY21kX2pzb25bInByZWZpeCJdKQog
            ICAgICAgIHJldF92YWwsIGpzb25fb3V0LCBfID0gc2VsZi5fY29tbW9uX2NtZF9qc29uX2dlbihj
            bWRfanNvbikKICAgICAgICAjIGlmIHRoZXJlIGlzIGFuIHVuc3VjY2Vzc2Z1bCBhdHRlbXB0LAog
            ICAgICAgIGlmIHJldF92YWwgIT0gMCBvciBsZW4oanNvbl9vdXQpID09IDA6CiAgICAgICAgICAg
            IHJldHVybiBOb25lCiAgICAgICAgaWYgImRhc2hib2FyZCIgbm90IGluIGpzb25fb3V0OgogICAg
            ICAgICAgICByZXR1cm4gTm9uZQogICAgICAgIHJldHVybiBqc29uX291dFsiZGFzaGJvYXJkIl0K
            CiAgICBkZWYgY3JlYXRlX3Jnd19hZG1pbl9vcHNfdXNlcihzZWxmKToKICAgICAgICBjbWQgPSBb
            CiAgICAgICAgICAgICJyYWRvc2d3LWFkbWluIiwKICAgICAgICAgICAgInVzZXIiLAogICAgICAg
            ICAgICAiY3JlYXRlIiwKICAgICAgICAgICAgIi0tdWlkIiwKICAgICAgICAgICAgc2VsZi5FWFRF
            Uk5BTF9SR1dfQURNSU5fT1BTX1VTRVJfTkFNRSwKICAgICAgICAgICAgIi0tZGlzcGxheS1uYW1l
            IiwKICAgICAgICAgICAgIlJvb2sgUkdXIEFkbWluIE9wcyB1c2VyIiwKICAgICAgICAgICAgIi0t
            Y2FwcyIsCiAgICAgICAgICAgICJidWNrZXRzPSo7dXNlcnM9Kjt1c2FnZT1yZWFkO21ldGFkYXRh
            PXJlYWQ7em9uZT1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwKICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgIi0tcmd3LXpvbmVn
            cm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLAog
            ICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5kcnlfcnVu
            OgogICAgICAgICAgICByZXR1cm4gc2VsZi5kcnlfcnVuKCJjZXBoICIgKyAiICIuam9pbihjbWQp
            KQogICAgICAgIHRyeToKICAgICAgICAgICAgb3V0cHV0ID0gc3VicHJvY2Vzcy5jaGVja19vdXRw
            dXQoY21kLCBzdGRlcnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNz
            LkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVjRXJyOgogICAgICAgICAgICAjIGlmIHRoZSB1c2Vy
            IGFscmVhZHkgZXhpc3RzLCB3ZSBqdXN0IHF1ZXJ5IGl0CiAgICAgICAgICAgIGlmIGV4ZWNFcnIu
            cmV0dXJuY29kZSA9PSBlcnJuby5FRVhJU1Q6CiAgICAgICAgICAgICAgICBjbWQgPSBbCiAgICAg
            ICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgICAgICJ1c2Vy
            IiwKICAgICAgICAgICAgICAgICAgICAiaW5mbyIsCiAgICAgICAgICAgICAgICAgICAgIi0tdWlk
            IiwKICAgICAgICAgICAgICAgICAgICBzZWxmLkVYVEVSTkFMX1JHV19BRE1JTl9PUFNfVVNFUl9O
            QU1FLAogICAgICAgICAgICAgICAgICAgICItLXJndy1yZWFsbSIsCiAgICAgICAgICAgICAgICAg
            ICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAgICAgICAgICAi
            LS1yZ3ctem9uZWdyb3VwIiwKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJn
            d196b25lZ3JvdXBfbmFtZSwKICAgICAgICAgICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lLAogICAgICAgICAg
            ICAgICAgXQogICAgICAgICAgICAgICAgdHJ5OgogICAgICAgICAgICAgICAgICAgIG91dHB1dCA9
            IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nlc3MuUElQRSkKICAg
            ICAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJvciBhcyBleGVj
            RXJyOgogICAgICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAgICAgICAg
            ICAgIGYiZmFpbGVkIHRvIGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5v
            dXRwdXR9LiAiCiAgICAgICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJu
            Y29kZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgICAgICkKICAg
            ICAgICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAg
            ICAgICAgcmV0dXJuIE5vbmUsIE5vbmUsIEZhbHNlLCAiLTEiCiAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICBlcnJfbXNnID0gKAogICAgICAgICAgICAgICAgICAgIGYiZmFpbGVkIHRv
            IGV4ZWN1dGUgY29tbWFuZCB7Y21kfS4gT3V0cHV0OiB7ZXhlY0Vyci5vdXRwdXR9LiAiCiAgICAg
            ICAgICAgICAgICAgICAgZiJDb2RlOiB7ZXhlY0Vyci5yZXR1cm5jb2RlfS4gRXJyb3I6IHtleGVj
            RXJyLnN0ZGVycn0iCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzeXMuc3RkZXJy
            LndyaXRlKGVycl9tc2cpCiAgICAgICAgICAgICAgICByZXR1cm4gTm9uZSwgTm9uZSwgRmFsc2Us
            ICItMSIKCiAgICAgICAgIyBzZXBhcmF0ZWx5IGFkZCBpbmZvPXJlYWQgY2FwcyBmb3Igcmd3LWVu
            ZHBvaW50IGlwIHZhbGlkYXRpb24KICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQgPSBUcnVlCiAg
            ICAgICAgY21kID0gWwogICAgICAgICAgICAicmFkb3Nndy1hZG1pbiIsCiAgICAgICAgICAgICJj
            YXBzIiwKICAgICAgICAgICAgImFkZCIsCiAgICAgICAgICAgICItLXVpZCIsCiAgICAgICAgICAg
            IHNlbGYuRVhURVJOQUxfUkdXX0FETUlOX09QU19VU0VSX05BTUUsCiAgICAgICAgICAgICItLWNh
            cHMiLAogICAgICAgICAgICAiaW5mbz1yZWFkIiwKICAgICAgICAgICAgIi0tcmd3LXJlYWxtIiwK
            ICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwKICAgICAgICAgICAg
            Ii0tcmd3LXpvbmVncm91cCIsCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVn
            cm91cF9uYW1lLAogICAgICAgICAgICAiLS1yZ3ctem9uZSIsCiAgICAgICAgICAgIHNlbGYuX2Fy
            Z19wYXJzZXIucmd3X3pvbmVfbmFtZSwKICAgICAgICBdCiAgICAgICAgdHJ5OgogICAgICAgICAg
            ICBvdXRwdXQgPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9jZXNz
            LlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9yIGFzIGV4
            ZWNFcnI6CiAgICAgICAgICAgICMgaWYgdGhlIGNlcGggdmVyc2lvbiBub3Qgc3VwcG9ydGVkIGZv
            ciBhZGRpbmcgYGluZm89cmVhZGAgY2FwKHJnd192YWxpZGF0aW9uKQogICAgICAgICAgICBpZiAo
            CiAgICAgICAgICAgICAgICAiY291bGQgbm90IGFkZCBjYXBzOiB1bmFibGUgdG8gYWRkIGNhcHM6
            IGluZm89cmVhZFxuIgogICAgICAgICAgICAgICAgaW4gZXhlY0Vyci5zdGRlcnIuZGVjb2RlKCJ1
            dGYtOCIpCiAgICAgICAgICAgICAgICBhbmQgZXhlY0Vyci5yZXR1cm5jb2RlID09IDI0NAogICAg
            ICAgICAgICApOgogICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkID0gRmFsc2UKICAg
            ICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGVycl9tc2cgPSAoCiAgICAgICAgICAgICAg
            ICAgICAgZiJmYWlsZWQgdG8gZXhlY3V0ZSBjb21tYW5kIHtjbWR9LiBPdXRwdXQ6IHtleGVjRXJy
            Lm91dHB1dH0uICIKICAgICAgICAgICAgICAgICAgICBmIkNvZGU6IHtleGVjRXJyLnJldHVybmNv
            ZGV9LiBFcnJvcjoge2V4ZWNFcnIuc3RkZXJyfSIKICAgICAgICAgICAgICAgICkKICAgICAgICAg
            ICAgICAgIHN5cy5zdGRlcnIud3JpdGUoZXJyX21zZykKICAgICAgICAgICAgICAgIHJldHVybiBO
            b25lLCBOb25lLCBGYWxzZSwgIi0xIgoKICAgICAgICBqc29ub3V0cHV0ID0ganNvbi5sb2Fkcyhv
            dXRwdXQpCiAgICAgICAgcmV0dXJuICgKICAgICAgICAgICAganNvbm91dHB1dFsia2V5cyJdWzBd
            WyJhY2Nlc3Nfa2V5Il0sCiAgICAgICAgICAgIGpzb25vdXRwdXRbImtleXMiXVswXVsic2VjcmV0
            X2tleSJdLAogICAgICAgICAgICBpbmZvX2NhcF9zdXBwb3J0ZWQsCiAgICAgICAgICAgICIiLAog
            ICAgICAgICkKCiAgICBkZWYgdmFsaWRhdGVfcmJkX3Bvb2woc2VsZiwgcG9vbF9uYW1lKToKICAg
            ICAgICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKHBvb2xfbmFtZSk6CiAgICAgICAg
            ICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRo
            ZSBwcm92aWRlZCBwb29sLCAne3Bvb2xfbmFtZX0nLCBkb2VzIG5vdCBleGlzdCIKICAgICAgICAg
            ICAgKQoKICAgIGRlZiBpbml0X3JiZF9wb29sKHNlbGYsIHJiZF9wb29sX25hbWUpOgogICAgICAg
            IGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAgcmV0
            dXJuCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIub3Blbl9pb2N0eChyYmRfcG9vbF9uYW1l
            KQogICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgcmJkX2luc3QucG9vbF9pbml0
            KGlvY3R4LCBUcnVlKQoKICAgIGRlZiB2YWxpZGF0ZV9yYWRvc19uYW1lc3BhY2Uoc2VsZik6CiAg
            ICAgICAgcmJkX3Bvb2xfbmFtZSA9IHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1l
            CiAgICAgICAgaWYgc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFfZWNfcG9vbF9uYW1lOgog
            ICAgICAgICAgICByYmRfcG9vbF9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5yYmRfbWV0YWRhdGFf
            ZWNfcG9vbF9uYW1lCiAgICAgICAgcmFkb3NfbmFtZXNwYWNlID0gc2VsZi5fYXJnX3BhcnNlci5y
            YWRvc19uYW1lc3BhY2UKICAgICAgICBpZiByYWRvc19uYW1lc3BhY2UgPT0gIiI6CiAgICAgICAg
            ICAgIHJldHVybgogICAgICAgIGlmIHJhZG9zX25hbWVzcGFjZS5pc2xvd2VyKCkgPT0gRmFsc2U6
            CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAg
            ICAgICBmIlRoZSBwcm92aWRlZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScs
            ICIKICAgICAgICAgICAgICAgIGYiY29udGFpbnMgdXBwZXIgY2FzZSIKICAgICAgICAgICAgKQog
            ICAgICAgIHJiZF9pbnN0ID0gcmJkLlJCRCgpCiAgICAgICAgaW9jdHggPSBzZWxmLmNsdXN0ZXIu
            b3Blbl9pb2N0eChyYmRfcG9vbF9uYW1lKQogICAgICAgIGlmIHJiZF9pbnN0Lm5hbWVzcGFjZV9l
            eGlzdHMoaW9jdHgsIHJhZG9zX25hbWVzcGFjZSkgaXMgRmFsc2U6CiAgICAgICAgICAgIHJhaXNl
            IEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRl
            ZCByYWRvcyBOYW1lc3BhY2UsICd7cmFkb3NfbmFtZXNwYWNlfScsICIKICAgICAgICAgICAgICAg
            IGYiaXMgbm90IGZvdW5kIGluIHRoZSBwb29sICd7cmJkX3Bvb2xfbmFtZX0nIgogICAgICAgICAg
            ICApCgogICAgZGVmIGdldF9vcl9jcmVhdGVfc3Vidm9sdW1lX2dyb3VwKHNlbGYsIHN1YnZvbHVt
            ZV9ncm91cCwgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSk6CiAgICAgICAgY21kID0gWwogICAgICAg
            ICAgICAiY2VwaCIsCiAgICAgICAgICAgICJmcyIsCiAgICAgICAgICAgICJzdWJ2b2x1bWVncm91
            cCIsCiAgICAgICAgICAgICJnZXRwYXRoIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1f
            bmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgIF0KICAgICAgICB0cnk6
            CiAgICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJw
            cm9jZXNzLlBJUEUpCiAgICAgICAgZXhjZXB0IHN1YnByb2Nlc3MuQ2FsbGVkUHJvY2Vzc0Vycm9y
            OgogICAgICAgICAgICBjbWQgPSBbCiAgICAgICAgICAgICAgICAiY2VwaCIsCiAgICAgICAgICAg
            ICAgICAiZnMiLAogICAgICAgICAgICAgICAgInN1YnZvbHVtZWdyb3VwIiwKICAgICAgICAgICAg
            ICAgICJjcmVhdGUiLAogICAgICAgICAgICAgICAgY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZSwKICAg
            ICAgICAgICAgICAgIHN1YnZvbHVtZV9ncm91cCwKICAgICAgICAgICAgXQogICAgICAgICAgICB0
            cnk6CiAgICAgICAgICAgICAgICBfID0gc3VicHJvY2Vzcy5jaGVja19vdXRwdXQoY21kLCBzdGRl
            cnI9c3VicHJvY2Vzcy5QSVBFKQogICAgICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQ
            cm9jZXNzRXJyb3I6CiAgICAgICAgICAgICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0
            aW9uKAogICAgICAgICAgICAgICAgICAgIGYic3Vidm9sdW1lIGdyb3VwIHtzdWJ2b2x1bWVfZ3Jv
            dXB9IGlzIG5vdCBhYmxlIHRvIGdldCBjcmVhdGVkIgogICAgICAgICAgICAgICAgKQoKICAgIGRl
            ZiBwaW5fc3Vidm9sdW1lKAogICAgICAgIHNlbGYsIHN1YnZvbHVtZV9ncm91cCwgY2VwaGZzX2Zp
            bGVzeXN0ZW1fbmFtZSwgcGluX3R5cGUsIHBpbl9zZXR0aW5nCiAgICApOgogICAgICAgIGNtZCA9
            IFsKICAgICAgICAgICAgImNlcGgiLAogICAgICAgICAgICAiZnMiLAogICAgICAgICAgICAic3Vi
            dm9sdW1lZ3JvdXAiLAogICAgICAgICAgICAicGluIiwKICAgICAgICAgICAgY2VwaGZzX2ZpbGVz
            eXN0ZW1fbmFtZSwKICAgICAgICAgICAgc3Vidm9sdW1lX2dyb3VwLAogICAgICAgICAgICBwaW5f
            dHlwZSwKICAgICAgICAgICAgcGluX3NldHRpbmcsCiAgICAgICAgXQogICAgICAgIHRyeToKICAg
            ICAgICAgICAgXyA9IHN1YnByb2Nlc3MuY2hlY2tfb3V0cHV0KGNtZCwgc3RkZXJyPXN1YnByb2Nl
            c3MuUElQRSkKICAgICAgICBleGNlcHQgc3VicHJvY2Vzcy5DYWxsZWRQcm9jZXNzRXJyb3I6CiAg
            ICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAg
            ICBmInN1YnZvbHVtZSBncm91cCB7c3Vidm9sdW1lX2dyb3VwfSBpcyBub3QgYWJsZSB0byBnZXQg
            cGlubmVkIgogICAgICAgICAgICApCgogICAgZGVmIGdldF9yZ3dfZnNpZChzZWxmLCBiYXNlX3Vy
            bCwgdmVyaWZ5KToKICAgICAgICBhY2Nlc3Nfa2V5ID0gc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5f
            T1BTX1VTRVJfQUNDRVNTX0tFWSJdCiAgICAgICAgc2VjcmV0X2tleSA9IHNlbGYub3V0X21hcFsi
            UkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9LRVkiXQogICAgICAgIHJnd19lbmRwb2ludCA9IHNl
            bGYuX2FyZ19wYXJzZXIucmd3X2VuZHBvaW50CiAgICAgICAgYmFzZV91cmwgPSBiYXNlX3VybCAr
            ICI6Ly8iICsgcmd3X2VuZHBvaW50ICsgIi9hZG1pbi9pbmZvPyIKICAgICAgICBwYXJhbXMgPSB7
            ImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXF1ZXN0X3VybCA9IGJhc2VfdXJsICsgdXJsZW5j
            b2RlKHBhcmFtcykKCiAgICAgICAgdHJ5OgogICAgICAgICAgICByID0gcmVxdWVzdHMuZ2V0KAog
            ICAgICAgICAgICAgICAgcmVxdWVzdF91cmwsCiAgICAgICAgICAgICAgICBhdXRoPVMzQXV0aChh
            Y2Nlc3Nfa2V5LCBzZWNyZXRfa2V5LCByZ3dfZW5kcG9pbnQpLAogICAgICAgICAgICAgICAgdmVy
            aWZ5PXZlcmlmeSwKICAgICAgICAgICAgKQogICAgICAgIGV4Y2VwdCByZXF1ZXN0cy5leGNlcHRp
            b25zLlRpbWVvdXQ6CiAgICAgICAgICAgIHN5cy5zdGRlcnIud3JpdGUoCiAgICAgICAgICAgICAg
            ICBmImludmFsaWQgZW5kcG9pbnQ6LCBub3QgYWJsZSB0byBjYWxsIGFkbWluLW9wcyBhcGl7cmd3
            X2VuZHBvaW50fSIKICAgICAgICAgICAgKQogICAgICAgICAgICByZXR1cm4gIiIsICItMSIKICAg
            ICAgICByMSA9IHIuanNvbigpCiAgICAgICAgaWYgcjEgaXMgTm9uZSBvciByMS5nZXQoImluZm8i
            KSBpcyBOb25lOgogICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAg
            ZiJUaGUgcHJvdmlkZWQgcmd3IEVuZHBvaW50LCAne3NlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50fScsIGlzIGludmFsaWQuIgogICAgICAgICAgICApCiAgICAgICAgICAgIHJldHVybiAoCiAg
            ICAgICAgICAgICAgICAiIiwKICAgICAgICAgICAgICAgICItMSIsCiAgICAgICAgICAgICkKCiAg
            ICAgICAgcmV0dXJuIHIxWyJpbmZvIl1bInN0b3JhZ2VfYmFja2VuZHMiXVswXVsiY2x1c3Rlcl9p
            ZCJdLCAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfZW5kcG9pbnQoc2VsZiwgaW5mb19jYXBfc3Vw
            cG9ydGVkKToKICAgICAgICAjIGlmIHRoZSAnY2x1c3RlcicgaW5zdGFuY2UgaXMgYSBkdW1teSBv
            bmUsCiAgICAgICAgIyBkb24ndCB0cnkgdG8gcmVhY2ggb3V0IHRvIHRoZSBlbmRwb2ludAogICAg
            ICAgIGlmIGlzaW5zdGFuY2Uoc2VsZi5jbHVzdGVyLCBEdW1teVJhZG9zKToKICAgICAgICAgICAg
            cmV0dXJuCgogICAgICAgIHJnd19lbmRwb2ludCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X2VuZHBv
            aW50CgogICAgICAgICMgdmFsaWRhdGUgcmd3IGVuZHBvaW50IG9ubHkgaWYgaXAgYWRkcmVzcyBp
            cyBwYXNzZWQKICAgICAgICBpcF90eXBlID0gc2VsZi5faW52YWxpZF9lbmRwb2ludChyZ3dfZW5k
            cG9pbnQpCgogICAgICAgICMgY2hlY2sgaWYgdGhlIHJndyBlbmRwb2ludCBpcyByZWFjaGFibGUK
            ICAgICAgICBjZXJ0ID0gTm9uZQogICAgICAgIGlmIG5vdCBzZWxmLl9hcmdfcGFyc2VyLnJnd19z
            a2lwX3RscyBhbmQgc2VsZi52YWxpZGF0ZV9yZ3dfZW5kcG9pbnRfdGxzX2NlcnQoKToKICAgICAg
            ICAgICAgY2VydCA9IHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGgKICAgICAgICBi
            YXNlX3VybCwgdmVyaWZ5LCBlcnIgPSBzZWxmLmVuZHBvaW50X2RpYWwocmd3X2VuZHBvaW50LCBp
            cF90eXBlLCBjZXJ0PWNlcnQpCiAgICAgICAgaWYgZXJyICE9ICIiOgogICAgICAgICAgICByZXR1
            cm4gIi0xIgoKICAgICAgICAjIGNoZWNrIGlmIHRoZSByZ3cgZW5kcG9pbnQgYmVsb25ncyB0byB0
            aGUgc2FtZSBjbHVzdGVyCiAgICAgICAgIyBvbmx5IGNoZWNrIGlmIGBpbmZvYCBjYXAgaXMgc3Vw
            cG9ydGVkCiAgICAgICAgaWYgaW5mb19jYXBfc3VwcG9ydGVkOgogICAgICAgICAgICBmc2lkID0g
            c2VsZi5nZXRfZnNpZCgpCiAgICAgICAgICAgIHJnd19mc2lkLCBlcnIgPSBzZWxmLmdldF9yZ3df
            ZnNpZChiYXNlX3VybCwgdmVyaWZ5KQogICAgICAgICAgICBpZiBlcnIgPT0gIi0xIjoKICAgICAg
            ICAgICAgICAgIHJldHVybiAiLTEiCiAgICAgICAgICAgIGlmIGZzaWQgIT0gcmd3X2ZzaWQ6CiAg
            ICAgICAgICAgICAgICBzeXMuc3RkZXJyLndyaXRlKAogICAgICAgICAgICAgICAgICAgIGYiVGhl
            IHByb3ZpZGVkIHJndyBFbmRwb2ludCwgJ3tzZWxmLl9hcmdfcGFyc2VyLnJnd19lbmRwb2ludH0n
            LCBpcyBpbnZhbGlkLiBXZSBhcmUgdmFsaWRhdGluZyBieSBjYWxsaW5nIHRoZSBhZG1pbm9wcyBh
            cGkgdGhyb3VnaCByZ3ctZW5kcG9pbnQgYW5kIHZhbGlkYXRpbmcgdGhlIGNsdXN0ZXJfaWQgJ3ty
            Z3dfZnNpZH0nIGlzIGVxdWFsIHRvIHRoZSBjZXBoIGNsdXN0ZXIgZnNpZCAne2ZzaWR9JyIKICAg
            ICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgIHJldHVybiAiLTEiCgogICAgICAgICMgY2hl
            Y2sgaWYgdGhlIHJndyBlbmRwb2ludCBwb29sIGV4aXN0CiAgICAgICAgIyBvbmx5IHZhbGlkYXRl
            IGlmIHJnd19wb29sX3ByZWZpeCBpcyBwYXNzZWQgZWxzZSBpdCB3aWxsIHRha2UgZGVmYXVsdCB2
            YWx1ZSBhbmQgd2UgZG9uJ3QgY3JlYXRlIHRoZXNlIGRlZmF1bHQgcG9vbHMKICAgICAgICBpZiBz
            ZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeCAhPSAiZGVmYXVsdCI6CiAgICAgICAgICAg
            IHJnd19wb29sc190b192YWxpZGF0ZSA9IFsKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19w
            YXJzZXIucmd3X3Bvb2xfcHJlZml4fS5yZ3cubWV0YSIsCiAgICAgICAgICAgICAgICAiLnJndy5y
            b290IiwKICAgICAgICAgICAgICAgIGYie3NlbGYuX2FyZ19wYXJzZXIucmd3X3Bvb2xfcHJlZml4
            fS5yZ3cuY29udHJvbCIsCiAgICAgICAgICAgICAgICBmIntzZWxmLl9hcmdfcGFyc2VyLnJnd19w
            b29sX3ByZWZpeH0ucmd3LmxvZyIsCiAgICAgICAgICAgIF0KICAgICAgICAgICAgZm9yIF9yZ3df
            cG9vbF90b192YWxpZGF0ZSBpbiByZ3dfcG9vbHNfdG9fdmFsaWRhdGU6CiAgICAgICAgICAgICAg
            ICBpZiBub3Qgc2VsZi5jbHVzdGVyLnBvb2xfZXhpc3RzKF9yZ3dfcG9vbF90b192YWxpZGF0ZSk6
            CiAgICAgICAgICAgICAgICAgICAgc3lzLnN0ZGVyci53cml0ZSgKICAgICAgICAgICAgICAgICAg
            ICAgICAgZiJUaGUgcHJvdmlkZWQgcG9vbCwgJ3tfcmd3X3Bvb2xfdG9fdmFsaWRhdGV9JywgZG9l
            cyBub3QgZXhpc3QiCiAgICAgICAgICAgICAgICAgICAgKQogICAgICAgICAgICAgICAgICAgIHJl
            dHVybiAiLTEiCgogICAgICAgIHJldHVybiAiIgoKICAgIGRlZiB2YWxpZGF0ZV9yZ3dfbXVsdGlz
            aXRlKHNlbGYsIHJnd19tdWx0aXNpdGVfY29uZmlnX25hbWUsIHJnd19tdWx0aXNpdGVfY29uZmln
            KToKICAgICAgICBpZiByZ3dfbXVsdGlzaXRlX2NvbmZpZyAhPSAiIjoKICAgICAgICAgICAgY21k
            ID0gWwogICAgICAgICAgICAgICAgInJhZG9zZ3ctYWRtaW4iLAogICAgICAgICAgICAgICAgcmd3
            X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICAiZ2V0IiwKICAgICAgICAgICAgICAg
            ICItLXJndy0iICsgcmd3X211bHRpc2l0ZV9jb25maWcsCiAgICAgICAgICAgICAgICByZ3dfbXVs
            dGlzaXRlX2NvbmZpZ19uYW1lLAogICAgICAgICAgICBdCiAgICAgICAgICAgIHRyeToKICAgICAg
            ICAgICAgICAgIF8gPSBzdWJwcm9jZXNzLmNoZWNrX291dHB1dChjbWQsIHN0ZGVycj1zdWJwcm9j
            ZXNzLlBJUEUpCiAgICAgICAgICAgIGV4Y2VwdCBzdWJwcm9jZXNzLkNhbGxlZFByb2Nlc3NFcnJv
            ciBhcyBleGVjRXJyOgogICAgICAgICAgICAgICAgZXJyX21zZyA9ICgKICAgICAgICAgICAgICAg
            ICAgICBmImZhaWxlZCB0byBleGVjdXRlIGNvbW1hbmQge2NtZH0uIE91dHB1dDoge2V4ZWNFcnIu
            b3V0cHV0fS4gIgogICAgICAgICAgICAgICAgICAgIGYiQ29kZToge2V4ZWNFcnIucmV0dXJuY29k
            ZX0uIEVycm9yOiB7ZXhlY0Vyci5zdGRlcnJ9IgogICAgICAgICAgICAgICAgKQogICAgICAgICAg
            ICAgICAgc3lzLnN0ZGVyci53cml0ZShlcnJfbXNnKQogICAgICAgICAgICAgICAgcmV0dXJuICIt
            MSIKICAgICAgICByZXR1cm4gIiIKCiAgICBkZWYgY29udmVydF9jb21tYV9zZXBhcmF0ZWRfdG9f
            YXJyYXkoc2VsZiwgdmFsdWUpOgogICAgICAgIHJldHVybiB2YWx1ZS5zcGxpdCgiLCIpCgogICAg
            ZGVmIHJhaXNlX2V4Y2VwdGlvbl9pZl9hbnlfdG9wb2xvZ3lfZmxhZ19pc19taXNzaW5nKHNlbGYp
            OgogICAgICAgIGlmICgKICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgc2VsZi5fYXJnX3Bh
            cnNlci50b3BvbG9neV9wb29scyAhPSAiIgogICAgICAgICAgICAgICAgYW5kICgKICAgICAgICAg
            ICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICAgICAgb3Igc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9m
            YWlsdXJlX2RvbWFpbl92YWx1ZXMgPT0gIiIKICAgICAgICAgICAgICAgICkKICAgICAgICAgICAg
            KQogICAgICAgICAgICBvciAoCiAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAg
            ICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAg
            ICAgICAgICAgICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9t
            YWluX3ZhbHVlcyA9PSAiIgogICAgICAgICAgICAgICAgKQogICAgICAgICAgICApCiAgICAgICAg
            ICAgIG9yICgKICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVy
            ZV9kb21haW5fdmFsdWVzICE9ICIiCiAgICAgICAgICAgICAgICBhbmQgKAogICAgICAgICAgICAg
            ICAgICAgIHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMgPT0gIiIKICAgICAgICAgICAg
            ICAgICAgICBvciBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX2xhYmVs
            ID09ICIiCiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICkKICAgICAgICApOgogICAgICAg
            ICAgICByYWlzZSBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgInBy
            b3ZpZGUgYWxsIHRoZSB0b3BvbG9neSBmbGFncyAtLXRvcG9sb2d5LXBvb2xzLCAtLXRvcG9sb2d5
            LWZhaWx1cmUtZG9tYWluLWxhYmVsLCAtLXRvcG9sb2d5LWZhaWx1cmUtZG9tYWluLXZhbHVlcyIK
            ICAgICAgICAgICAgKQoKICAgIGRlZiB2YWxpZGF0ZV90b3BvbG9neV92YWx1ZXMoc2VsZiwgdG9w
            b2xvZ3lfcG9vbHMsIHRvcG9sb2d5X2ZkKToKICAgICAgICBpZiBsZW4odG9wb2xvZ3lfcG9vbHMp
            ICE9IGxlbih0b3BvbG9neV9mZCk6CiAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVF
            eGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlRoZSBwcm92aWRlZCB0b3BvbG9neSBwb29scywg
            J3t0b3BvbG9neV9wb29sc30nLCBhbmQgIgogICAgICAgICAgICAgICAgZiJ0b3BvbG9neSBmYWls
            dXJlIGRvbWFpbiwgJ3t0b3BvbG9neV9mZH0nLCIKICAgICAgICAgICAgICAgIGYiYXJlIG9mIGRp
            ZmZlcmVudCBsZW5ndGgsICd7bGVuKHRvcG9sb2d5X3Bvb2xzKX0nIGFuZCAne2xlbih0b3BvbG9n
            eV9mZCl9JyByZXNwY3RpdmVseSIKICAgICAgICAgICAgKQogICAgICAgIHJldHVybgoKICAgIGRl
            ZiB2YWxpZGF0ZV90b3BvbG9neV9yYmRfcG9vbHMoc2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToK
            ICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9yYmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYu
            dmFsaWRhdGVfcmJkX3Bvb2wocG9vbCkKCiAgICBkZWYgaW5pdF90b3BvbG9neV9yYmRfcG9vbHMo
            c2VsZiwgdG9wb2xvZ3lfcmJkX3Bvb2xzKToKICAgICAgICBmb3IgcG9vbCBpbiB0b3BvbG9neV9y
            YmRfcG9vbHM6CiAgICAgICAgICAgIHNlbGYuaW5pdF9yYmRfcG9vbChwb29sKQoKICAgICMgdGhp
            cyB3aWxsIHJldHVybiB0aGUgZmluYWwgYXJncyB0aGF0IHNjcmlwdCB1c2VzIHRvIHByb2Nlc3MK
            ICAgICMgdGhlIHByaW9yaXR5IHRvIHNldCBhIHBhcnRpY3VsYXIgdmFsdWUgaXMsCiAgICAjIGNv
            bW1hbmQtbGluZS1hcmdzID4gY29uZmlnLmluaSBmaWxlIHZhbHVlcyA+IGRlZmF1bHQgdmFsdWVz
            CiAgICBkZWYgZ2V0RmluYWxVc2VkQXJncyhzZWxmKToKICAgICAgICBhcmd1bWVudCA9IGYiW0Nv
            bmZpZ3VyYXRpb25zXVxuIgogICAgICAgIGZvciBhcmcgaW4gdmFycyhzZWxmLl9hcmdfcGFyc2Vy
            KToKICAgICAgICAgICAgaWYgc3RyKGdldGF0dHIoc2VsZi5fYXJnX3BhcnNlciwgYXJnKSk6CiAg
            ICAgICAgICAgICAgICAjIHB5dGhvbiB0cmVhdHMgZmxhZy1uYW1lIGFzIGZsYWdfbmFtZSBpbnRl
            cm5hbGx5LCBzbyBjb252ZXJ0aW5nIGJhY2sgdG8gZmxhZy1uYW1lLAogICAgICAgICAgICAgICAg
            IyBzbyB3ZSBjYW4gZ2V0IHRob3NlIHZhbHVlcyBmcm9tIGNvbmZpZyBmaWxlCiAgICAgICAgICAg
            ICAgICBhcmdWYWx1ZSA9IGFyZy5yZXBsYWNlKCJfIiwgIi0iKQogICAgICAgICAgICAgICAgIyBk
            byBub3QgYWRkIHRoZSBjb25maWctZmlsZSBmbGFnIGFuZCBhbHNvIG5vdCBhZGQgdGhlIGJvb2xl
            YW4gZmxhZ3Mgd2hpY2ggYXJlIHNldCB0byBGYWxzZQogICAgICAgICAgICAgICAgIyBiZWNhdXNl
            IGNvbmZpZy5pbmkgZmlsZSB0cmVhdHMgYm9vbGVhbiBmbGFncyBhcyBUcnVlIGFsd2F5cwogICAg
            ICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIGFyZ1ZhbHVlICE9ICJjb25maWct
            ZmlsZSIKICAgICAgICAgICAgICAgICAgICBhbmQgZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBh
            cmcpICE9IEZhbHNlCiAgICAgICAgICAgICAgICApOgogICAgICAgICAgICAgICAgICAgIGFyZ3Vt
            ZW50ICs9IGYie2FyZ1ZhbHVlfSA9IHtzdHIoZ2V0YXR0cihzZWxmLl9hcmdfcGFyc2VyLCBhcmcp
            KX1cbiIKICAgICAgICByZXR1cm4gYXJndW1lbnQKCiAgICBkZWYgX2dlbl9vdXRwdXRfbWFwKHNl
            bGYpOgogICAgICAgIGlmIHNlbGYub3V0X21hcDoKICAgICAgICAgICAgcmV0dXJuCiAgICAgICAg
            IyBzdXBwb3J0IGxlZ2FjeSBmbGFnIHdpdGggdXBncmFkZXMKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLmNsdXN0ZXJfbmFtZToKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5rOHNfY2x1
            c3Rlcl9uYW1lID0gc2VsZi5fYXJnX3BhcnNlci5jbHVzdGVyX25hbWUKICAgICAgICBzZWxmLl9h
            cmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUgPSAoCiAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJz
            ZXIuazhzX2NsdXN0ZXJfbmFtZS5sb3dlcigpCiAgICAgICAgKSAgIyBhbHdheXMgY29udmVydCBj
            bHVzdGVyIG5hbWUgdG8gbG93ZXJjYXNlIGNoYXJhY3RlcnMKICAgICAgICBzZWxmLnZhbGlkYXRl
            X3JiZF9wb29sKHNlbGYuX2FyZ19wYXJzZXIucmJkX2RhdGFfcG9vbF9uYW1lKQogICAgICAgIHNl
            bGYuaW5pdF9yYmRfcG9vbChzZWxmLl9hcmdfcGFyc2VyLnJiZF9kYXRhX3Bvb2xfbmFtZSkKICAg
            ICAgICBzZWxmLnZhbGlkYXRlX3JhZG9zX25hbWVzcGFjZSgpCiAgICAgICAgc2VsZi5fZXhjbHVk
            ZWRfa2V5cy5hZGQoIks4U19DTFVTVEVSX05BTUUiKQogICAgICAgIHNlbGYuZ2V0X2NlcGhmc19k
            YXRhX3Bvb2xfZGV0YWlscygpCiAgICAgICAgIyBkb3VibGUgc3RyaW5nIG5lZWRlZCBmb3IgdXBz
            dHJlYW0gZXhwb3J0cyBvZiBmbGFncwogICAgICAgIHNlbGYub3V0X21hcFsiQVJHUyJdID0gZici
            e3NlbGYuZ2V0RmluYWxVc2VkQXJncygpfSInCiAgICAgICAgc2VsZi5vdXRfbWFwWyJOQU1FU1BB
            Q0UiXSA9IHNlbGYuX2FyZ19wYXJzZXIubmFtZXNwYWNlCiAgICAgICAgc2VsZi5vdXRfbWFwWyJL
            OFNfQ0xVU1RFUl9OQU1FIl0gPSBzZWxmLl9hcmdfcGFyc2VyLms4c19jbHVzdGVyX25hbWUKICAg
            ICAgICBzZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfRlNJRCJdID0gc2VsZi5nZXRfZnNpZCgp
            CiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX1VTRVJOQU1FIl0gPSBzZWxmLnJ1
            bl9hc191c2VyCiAgICAgICAgc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0NFUEhfTU9OX0RB
            VEEiXSA9IHNlbGYuZ2V0X2NlcGhfZXh0ZXJuYWxfbW9uX2RhdGEoKQogICAgICAgIHNlbGYub3V0
            X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdID0gc2VsZi5jcmVhdGVfY2hlY2tlcktl
            eSgKICAgICAgICAgICAgImNsaWVudC5oZWFsdGhjaGVja2VyIgogICAgICAgICkKICAgICAgICBz
            ZWxmLm91dF9tYXBbIlJPT0tfRVhURVJOQUxfREFTSEJPQVJEX0xJTksiXSA9IHNlbGYuZ2V0X2Nl
            cGhfZGFzaGJvYXJkX2xpbmsoKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJD
            U0lfUkJEX05PREVfU0VDUkVUIl0sCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiQ1NJX1JCRF9O
            T0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgKSA9IHNlbGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5n
            X3VzZXIoImNsaWVudC5jc2ktcmJkLW5vZGUiKQogICAgICAgICgKICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICApID0gc2Vs
            Zi5jcmVhdGVfY2VwaENTSUtleXJpbmdfdXNlcigiY2xpZW50LmNzaS1yYmQtcHJvdmlzaW9uZXIi
            KQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQSEZTX1BPT0xfTkFNRSJdID0gc2VsZi5fYXJnX3Bh
            cnNlci5jZXBoZnNfZGF0YV9wb29sX25hbWUKICAgICAgICBzZWxmLm91dF9tYXBbIkNFUEhGU19N
            RVRBREFUQV9QT09MX05BTUUiXSA9ICgKICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBo
            ZnNfbWV0YWRhdGFfcG9vbF9uYW1lCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiQ0VQ
            SEZTX0ZTX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIuY2VwaGZzX2ZpbGVzeXN0ZW1fbmFtZQog
            ICAgICAgIHNlbGYub3V0X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZXN0cmljdGVkX2F1dGhfcGVybWlzc2lvbgogICAg
            ICAgICkKICAgICAgICBzZWxmLm91dF9tYXBbIlJBRE9TX05BTUVTUEFDRSJdID0gc2VsZi5fYXJn
            X3BhcnNlci5yYWRvc19uYW1lc3BhY2UKICAgICAgICBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9H
            Uk9VUCJdID0gc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAKICAgICAgICBzZWxmLm91
            dF9tYXBbIkNTSV9DRVBIRlNfTk9ERV9TRUNSRVQiXSA9ICIiCiAgICAgICAgc2VsZi5vdXRfbWFw
            WyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdID0gIiIKICAgICAgICAjIGNyZWF0ZSBD
            ZXBoRlMgbm9kZSBhbmQgcHJvdmlzaW9uZXIga2V5cmluZyBvbmx5IHdoZW4gTURTIGV4aXN0cwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSBhbmQgc2VsZi5vdXRfbWFw
            WyJDRVBIRlNfUE9PTF9OQU1FIl06CiAgICAgICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYu
            b3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgc2VsZi5v
            dXRfbWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNl
            bGYuY3JlYXRlX2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLW5vZGUiKQog
            ICAgICAgICAgICAoCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJP
            VklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9DRVBI
            RlNfUFJPVklTSU9ORVJfU0VDUkVUX05BTUUiXSwKICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X2NlcGhDU0lLZXlyaW5nX3VzZXIoImNsaWVudC5jc2ktY2VwaGZzLXByb3Zpc2lvbmVyIikKICAg
            ICAgICAgICAgIyBjcmVhdGUgdGhlIGRlZmF1bHQgImNzaSIgc3Vidm9sdW1lZ3JvdXAKICAgICAg
            ICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgKICAgICAgICAgICAgICAg
            ICJjc2kiLCBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUKICAgICAgICAg
            ICAgKQogICAgICAgICAgICAjIHBpbiB0aGUgZGVmYXVsdCAiY3NpIiBzdWJ2b2x1bWVncm91cAog
            ICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAgICAgICAgICAgICAgICAiY3NpIiwgc2Vs
            Zi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLCAiZGlzdHJpYnV0ZWQiLCAiMSIK
            ICAgICAgICAgICAgKQogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlNVQlZPTFVNRV9HUk9V
            UCJdOgogICAgICAgICAgICAgICAgc2VsZi5nZXRfb3JfY3JlYXRlX3N1YnZvbHVtZV9ncm91cCgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnN1YnZvbHVtZV9ncm91cCwKICAg
            ICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLmNlcGhmc19maWxlc3lzdGVtX25hbWUs
            CiAgICAgICAgICAgICAgICApCiAgICAgICAgICAgICAgICBzZWxmLnBpbl9zdWJ2b2x1bWUoCiAg
            ICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5zdWJ2b2x1bWVfZ3JvdXAsCiAgICAg
            ICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5jZXBoZnNfZmlsZXN5c3RlbV9uYW1lLAog
            ICAgICAgICAgICAgICAgICAgICJkaXN0cmlidXRlZCIsCiAgICAgICAgICAgICAgICAgICAgIjEi
            LAogICAgICAgICAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0g
            PSAiIgogICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lOVCJdID0gIiIKICAg
            ICAgICBzZWxmLm91dF9tYXBbIk1PTklUT1JJTkdfRU5EUE9JTlRfUE9SVCJdID0gIiIKICAgICAg
            ICBpZiBub3Qgc2VsZi5fYXJnX3BhcnNlci5za2lwX21vbml0b3JpbmdfZW5kcG9pbnQ6CiAgICAg
            ICAgICAgICgKICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiTU9OSVRPUklOR19FTkRQT0lO
            VCJdLAogICAgICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UX1BP
            UlQiXSwKICAgICAgICAgICAgKSA9IHNlbGYuZ2V0X2FjdGl2ZV9hbmRfc3RhbmRieV9tZ3JzKCkK
            ICAgICAgICBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSA9IHNlbGYuX2FyZ19wYXJzZXIu
            cmJkX2RhdGFfcG9vbF9uYW1lCiAgICAgICAgc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNf
            UE9PTF9OQU1FIl0gPSAoCiAgICAgICAgICAgIHNlbGYudmFsaWRhdGVfcmJkX21ldGFkYXRhX2Vj
            X3Bvb2xfbmFtZSgpCiAgICAgICAgKQogICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9P
            TFMiXSA9IHNlbGYuX2FyZ19wYXJzZXIudG9wb2xvZ3lfcG9vbHMKICAgICAgICBzZWxmLm91dF9t
            YXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX0xBQkVMIl0gPSAoCiAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIudG9wb2xvZ3lfZmFpbHVyZV9kb21haW5fbGFiZWwKICAgICAgICApCiAgICAg
            ICAgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9WQUxVRVMiXSA9ICgKICAg
            ICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci50b3BvbG9neV9mYWlsdXJlX2RvbWFpbl92YWx1ZXMK
            ICAgICAgICApCiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnRvcG9s
            b2d5X3Bvb2xzICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnRvcG9sb2d5
            X2ZhaWx1cmVfZG9tYWluX2xhYmVsICE9ICIiCiAgICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFy
            c2VyLnRvcG9sb2d5X2ZhaWx1cmVfZG9tYWluX3ZhbHVlcyAhPSAiIgogICAgICAgICk6CiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfdmFsdWVzKAogICAgICAgICAgICAgICAgc2Vs
            Zi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9MT0dZ
            X1BPT0xTIl0pLAogICAgICAgICAgICAgICAgc2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90
            b19hcnJheSgKICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxV
            UkVfRE9NQUlOX1ZBTFVFUyJdCiAgICAgICAgICAgICAgICApLAogICAgICAgICAgICApCiAgICAg
            ICAgICAgIHNlbGYudmFsaWRhdGVfdG9wb2xvZ3lfcmJkX3Bvb2xzKAogICAgICAgICAgICAgICAg
            c2VsZi5jb252ZXJ0X2NvbW1hX3NlcGFyYXRlZF90b19hcnJheShzZWxmLm91dF9tYXBbIlRPUE9M
            T0dZX1BPT0xTIl0pCiAgICAgICAgICAgICkKICAgICAgICAgICAgc2VsZi5pbml0X3RvcG9sb2d5
            X3JiZF9wb29scygKICAgICAgICAgICAgICAgIHNlbGYuY29udmVydF9jb21tYV9zZXBhcmF0ZWRf
            dG9fYXJyYXkoc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9QT09MUyJdKQogICAgICAgICAgICApCiAg
            ICAgICAgZWxzZToKICAgICAgICAgICAgc2VsZi5yYWlzZV9leGNlcHRpb25faWZfYW55X3RvcG9s
            b2d5X2ZsYWdfaXNfbWlzc2luZygpCgogICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJF
            RklYIl0gPSBzZWxmLl9hcmdfcGFyc2VyLnJnd19wb29sX3ByZWZpeAogICAgICAgIHNlbGYub3V0
            X21hcFsiUkdXX0VORFBPSU5UIl0gPSAiIgogICAgICAgIGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3
            X2VuZHBvaW50OgogICAgICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAg
            ICAgICAgICAgICBzZWxmLmNyZWF0ZV9yZ3dfYWRtaW5fb3BzX3VzZXIoKQogICAgICAgICAgICBl
            bHNlOgogICAgICAgICAgICAgICAgaWYgKAogICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19w
            YXJzZXIucmd3X3JlYWxtX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBhbmQgc2VsZi5f
            YXJnX3BhcnNlci5yZ3dfem9uZWdyb3VwX25hbWUgIT0gIiIKICAgICAgICAgICAgICAgICAgICBh
            bmQgc2VsZi5fYXJnX3BhcnNlci5yZ3dfem9uZV9uYW1lICE9ICIiCiAgICAgICAgICAgICAgICAp
            OgogICAgICAgICAgICAgICAgICAgIGVyciA9IHNlbGYudmFsaWRhdGVfcmd3X211bHRpc2l0ZSgK
            ICAgICAgICAgICAgICAgICAgICAgICAgc2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSwg
            InJlYWxtIgogICAgICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBz
            ZWxmLnZhbGlkYXRlX3Jnd19tdWx0aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYu
            X2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9uYW1lLCAiem9uZWdyb3VwIgogICAgICAgICAgICAg
            ICAgICAgICkKICAgICAgICAgICAgICAgICAgICBlcnIgPSBzZWxmLnZhbGlkYXRlX3Jnd19tdWx0
            aXNpdGUoCiAgICAgICAgICAgICAgICAgICAgICAgIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVf
            bmFtZSwgInpvbmUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgICAgICAgICAgICAgIGlmICgK
            ICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2VyLnJnd19yZWFsbV9uYW1lID09ICIi
            CiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pvbmVncm91cF9u
            YW1lID09ICIiCiAgICAgICAgICAgICAgICAgICAgYW5kIHNlbGYuX2FyZ19wYXJzZXIucmd3X3pv
            bmVfbmFtZSA9PSAiIgogICAgICAgICAgICAgICAgKSBvciAoCiAgICAgICAgICAgICAgICAgICAg
            c2VsZi5fYXJnX3BhcnNlci5yZ3dfcmVhbG1fbmFtZSAhPSAiIgogICAgICAgICAgICAgICAgICAg
            IGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lZ3JvdXBfbmFtZSAhPSAiIgogICAgICAgICAg
            ICAgICAgICAgIGFuZCBzZWxmLl9hcmdfcGFyc2VyLnJnd196b25lX25hbWUgIT0gIiIKICAgICAg
            ICAgICAgICAgICk6CiAgICAgICAgICAgICAgICAgICAgKAogICAgICAgICAgICAgICAgICAgICAg
            ICBzZWxmLm91dF9tYXBbIlJHV19BRE1JTl9PUFNfVVNFUl9BQ0NFU1NfS0VZIl0sCiAgICAgICAg
            ICAgICAgICAgICAgICAgIHNlbGYub3V0X21hcFsiUkdXX0FETUlOX09QU19VU0VSX1NFQ1JFVF9L
            RVkiXSwKICAgICAgICAgICAgICAgICAgICAgICAgaW5mb19jYXBfc3VwcG9ydGVkLAogICAgICAg
            ICAgICAgICAgICAgICAgICBlcnIsCiAgICAgICAgICAgICAgICAgICAgKSA9IHNlbGYuY3JlYXRl
            X3Jnd19hZG1pbl9vcHNfdXNlcigpCiAgICAgICAgICAgICAgICAgICAgZXJyID0gc2VsZi52YWxp
            ZGF0ZV9yZ3dfZW5kcG9pbnQoaW5mb19jYXBfc3VwcG9ydGVkKQogICAgICAgICAgICAgICAgICAg
            IGlmIHNlbGYuX2FyZ19wYXJzZXIucmd3X3Rsc19jZXJ0X3BhdGg6CiAgICAgICAgICAgICAgICAg
            ICAgICAgIHNlbGYub3V0X21hcFsiUkdXX1RMU19DRVJUIl0gPSAoCiAgICAgICAgICAgICAgICAg
            ICAgICAgICAgICBzZWxmLnZhbGlkYXRlX3Jnd19lbmRwb2ludF90bHNfY2VydCgpCiAgICAgICAg
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgICAgICAgICAjIGlmIHRoZXJlIGlzIG5vIGVy
            cm9yLCBzZXQgdGhlIFJHV19FTkRQT0lOVAogICAgICAgICAgICAgICAgICAgIGlmIGVyciAhPSAi
            LTEiOgogICAgICAgICAgICAgICAgICAgICAgICBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJd
            ID0gc2VsZi5fYXJnX3BhcnNlci5yZ3dfZW5kcG9pbnQKICAgICAgICAgICAgICAgIGVsc2U6CiAg
            ICAgICAgICAgICAgICAgICAgZXJyID0gIlBsZWFzZSBwcm92aWRlIGFsbCB0aGUgUkdXIG11bHRp
            c2l0ZSBwYXJhbWV0ZXJzIG9yIG5vbmUgb2YgdGhlbSIKICAgICAgICAgICAgICAgICAgICBzeXMu
            c3RkZXJyLndyaXRlKGVycikKCiAgICBkZWYgZ2VuX3NoZWxsX291dChzZWxmKToKICAgICAgICBz
            ZWxmLl9nZW5fb3V0cHV0X21hcCgpCiAgICAgICAgc2hPdXRJTyA9IFN0cmluZ0lPKCkKICAgICAg
            ICBmb3IgaywgdiBpbiBzZWxmLm91dF9tYXAuaXRlbXMoKToKICAgICAgICAgICAgaWYgdiBhbmQg
            ayBub3QgaW4gc2VsZi5fZXhjbHVkZWRfa2V5czoKICAgICAgICAgICAgICAgIHNoT3V0SU8ud3Jp
            dGUoZiJleHBvcnQge2t9PXt2fXtMSU5FU0VQfSIpCiAgICAgICAgc2hPdXQgPSBzaE91dElPLmdl
            dHZhbHVlKCkKICAgICAgICBzaE91dElPLmNsb3NlKCkKICAgICAgICByZXR1cm4gc2hPdXQKCiAg
            ICBkZWYgZ2VuX2pzb25fb3V0KHNlbGYpOgogICAgICAgIHNlbGYuX2dlbl9vdXRwdXRfbWFwKCkK
            ICAgICAgICBpZiBzZWxmLl9hcmdfcGFyc2VyLmRyeV9ydW46CiAgICAgICAgICAgIHJldHVybiAi
            IgogICAgICAgIGpzb25fb3V0ID0gWwogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFt
            ZSI6ICJleHRlcm5hbC1jbHVzdGVyLXVzZXItY29tbWFuZCIsCiAgICAgICAgICAgICAgICAia2lu
            ZCI6ICJDb25maWdNYXAiLAogICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgImFyZ3MiOiBzZWxmLm91dF9tYXBbIkFSR1MiXSwKICAgICAgICAgICAgICAgIH0sCiAg
            ICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICJuYW1lIjogInJvb2st
            Y2VwaC1tb24tZW5kcG9pbnRzIiwKICAgICAgICAgICAgICAgICJraW5kIjogIkNvbmZpZ01hcCIs
            CiAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHNl
            bGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9DRVBIX01PTl9EQVRBIl0sCiAgICAgICAgICAgICAg
            ICAgICAgIm1heE1vbklkIjogIjAiLAogICAgICAgICAgICAgICAgICAgICJtYXBwaW5nIjogInt9
            IiwKICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgIH0sCiAgICAgICAgICAgIHsKICAgICAg
            ICAgICAgICAgICJuYW1lIjogInJvb2stY2VwaC1tb24iLAogICAgICAgICAgICAgICAgImtpbmQi
            OiAiU2VjcmV0IiwKICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAg
            ICJhZG1pbi1zZWNyZXQiOiAiYWRtaW4tc2VjcmV0IiwKICAgICAgICAgICAgICAgICAgICAiZnNp
            ZCI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9GU0lEIl0sCiAgICAgICAgICAgICAgICAg
            ICAgIm1vbi1zZWNyZXQiOiAibW9uLXNlY3JldCIsCiAgICAgICAgICAgICAgICB9LAogICAgICAg
            ICAgICB9LAogICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAibmFtZSI6ICJyb29rLWNlcGgt
            b3BlcmF0b3ItY3JlZHMiLAogICAgICAgICAgICAgICAgImtpbmQiOiAiU2VjcmV0IiwKICAgICAg
            ICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICJ1c2VySUQiOiBzZWxmLm91
            dF9tYXBbIlJPT0tfRVhURVJOQUxfVVNFUk5BTUUiXSwKICAgICAgICAgICAgICAgICAgICAidXNl
            cktleSI6IHNlbGYub3V0X21hcFsiUk9PS19FWFRFUk5BTF9VU0VSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgfSwKICAgICAgICAgICAgfSwKICAgICAgICBdCgogICAgICAgICMgaWYgJ01PTklU
            T1JJTkdfRU5EUE9JTlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAnbW9uaXRvcmluZy1lbmRwb2lu
            dCcgdG8gQ2x1c3RlcgogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJNT05J
            VE9SSU5HX0VORFBPSU5UIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21hcFsiTU9OSVRPUklO
            R19FTkRQT0lOVF9QT1JUIl0KICAgICAgICApOgogICAgICAgICAgICBqc29uX291dC5hcHBlbmQo
            CiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiAibW9uaXRvcmlu
            Zy1lbmRwb2ludCIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiQ2VwaENsdXN0ZXIiLAog
            ICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiTW9u
            aXRvcmluZ0VuZHBvaW50Ijogc2VsZi5vdXRfbWFwWyJNT05JVE9SSU5HX0VORFBPSU5UIl0sCiAg
            ICAgICAgICAgICAgICAgICAgICAgICJNb25pdG9yaW5nUG9ydCI6IHNlbGYub3V0X21hcFsiTU9O
            SVRPUklOR19FTkRQT0lOVF9QT1JUIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAg
            ICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDU0lfUkJEX05PREVfU0VDUkVU
            JyBleGlzdHMsIHRoZW4gb25seSBhZGQgJ3Jvb2stY3NpLXJiZC1wcm92aXNpb25lcicgU2VjcmV0
            CiAgICAgICAgaWYgKAogICAgICAgICAgICBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNS
            RVQiXQogICAgICAgICAgICBhbmQgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfUkJEX05PREVfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiBzZWxmLm91dF9tYXBbIkNTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSJd
            LAogICAgICAgICAgICAgICAgICAgICAgICAidXNlcktleSI6IHNlbGYub3V0X21hcFsiQ1NJX1JC
            RF9OT0RFX1NFQ1JFVCJdLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9
            CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCcg
            ZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdyb29rLWNzaS1yYmQtcHJvdmlzaW9uZXInIFNlY3JldAog
            ICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVCJdCiAgICAgICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9SQkRfUFJPVklTSU9O
            RVJfU0VDUkVUX05BTUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgK
            ICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2Vs
            Zi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAg
            ICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjog
            ewogICAgICAgICAgICAgICAgICAgICAgICAidXNlcklEIjogc2VsZi5vdXRfbWFwWyJDU0lfUkJE
            X1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJ1c2Vy
            S2V5Ijogc2VsZi5vdXRfbWFwWyJDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVCJdLAogICAgICAg
            ICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAj
            IGlmICdDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            ICdyb29rLWNzaS1jZXBoZnMtcHJvdmlzaW9uZXInIFNlY3JldAogICAgICAgIGlmICgKICAgICAg
            ICAgICAgc2VsZi5vdXRfbWFwWyJDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUX05B
            TUUiXQogICAgICAgICk6CiAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAgICAgICAgICAg
            ICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydD
            U0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAg
            ICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAg
            ICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19QUk9W
            SVNJT05FUl9TRUNSRVRfTkFNRSJdLAogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5LZXki
            OiBzZWxmLm91dF9tYXBbIkNTSV9DRVBIRlNfUFJPVklTSU9ORVJfU0VDUkVUIl0sCiAgICAgICAg
            ICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMg
            aWYgJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCAncm9vay1j
            c2ktY2VwaGZzLW5vZGUnIFNlY3JldAogICAgICAgIGlmICgKICAgICAgICAgICAgc2VsZi5vdXRf
            bWFwWyJDU0lfQ0VQSEZTX05PREVfU0VDUkVUIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0X21h
            cFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0KICAgICAgICApOgogICAgICAgICAgICBq
            c29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5h
            bWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FJ119
            IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAg
            ICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAiYWRtaW5JRCI6IHNlbGYub3V0
            X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJhZG1pbktleSI6IHNlbGYub3V0X21hcFsiQ1NJX0NFUEhGU19OT0RFX1NFQ1JFVCJdLAog
            ICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAg
            ICAgICAjIGlmICdST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LJyBleGlzdHMsIHRoZW4gb25s
            eSBhZGQgJ3Jvb2stY2VwaC1kYXNoYm9hcmQtbGluaycgU2VjcmV0CiAgICAgICAgaWYgc2VsZi5v
            dXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9MSU5LIl06CiAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAibmFtZSI6
            ICJyb29rLWNlcGgtZGFzaGJvYXJkLWxpbmsiLAogICAgICAgICAgICAgICAgICAgICJraW5kIjog
            IlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VySUQiOiAiY2VwaC1kYXNoYm9hcmQtbGluayIsCiAgICAgICAgICAgICAgICAg
            ICAgICAgICJ1c2VyS2V5Ijogc2VsZi5vdXRfbWFwWyJST09LX0VYVEVSTkFMX0RBU0hCT0FSRF9M
            SU5LIl0sCiAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAg
            ICAgKQogICAgICAgICMgaWYgJ1JBRE9TX05BTUVTUEFDRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRk
            IHRoZSAiUkFET1NfTkFNRVNQQUNFIiBuYW1lc3BhY2UKICAgICAgICBpZiAoCiAgICAgICAgICAg
            IHNlbGYub3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0KICAgICAgICAgICAgYW5kIHNlbGYub3V0
            X21hcFsiUkVTVFJJQ1RFRF9BVVRIX1BFUk1JU1NJT04iXQogICAgICAgICAgICBhbmQgbm90IHNl
            bGYub3V0X21hcFsiUkJEX01FVEFEQVRBX0VDX1BPT0xfTkFNRSJdCiAgICAgICAgKToKICAgICAg
            ICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAg
            ICAgICJuYW1lIjogInJhZG9zLW5hbWVzcGFjZSIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQi
            OiAiQ2VwaEJsb2NrUG9vbFJhZG9zTmFtZXNwYWNlIiwKICAgICAgICAgICAgICAgICAgICAiZGF0
            YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgInJhZG9zTmFtZXNwYWNlTmFtZSI6IHNlbGYu
            b3V0X21hcFsiUkFET1NfTkFNRVNQQUNFIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJwb29s
            Ijogc2VsZi5vdXRfbWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgfSwK
            ICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICAgICBqc29uX291dC5hcHBl
            bmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgIm5hbWUiOiBmImNlcGgt
            cmJkLXJhZG9zLW5hbWVzcGFjZS17c2VsZi5vdXRfbWFwWydSQURPU19OQU1FU1BBQ0UnXX0iLAog
            ICAgICAgICAgICAgICAgICAgICJraW5kIjogIlN0b3JhZ2VDbGFzcyIsCiAgICAgICAgICAgICAg
            ICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRf
            bWFwWyJSQkRfUE9PTF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vcHJvdmlzaW9uZXItc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0X21hcFsn
            Q1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAgICAg
            ICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBm
            InJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIs
            CiAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAg
            ICAgICkKICAgICAgICBlbHNlOgogICAgICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJCRF9NRVRB
            REFUQV9FQ19QT09MX05BTUUiXToKICAgICAgICAgICAgICAgIGpzb25fb3V0LmFwcGVuZCgKICAg
            ICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgt
            cmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAgICAgICAgICAgICAg
            ICAiZGF0YVBvb2wiOiBzZWxmLm91dF9tYXBbIlJCRF9QT09MX05BTUUiXSwKICAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICJwb29sIjogc2VsZi5vdXRfbWFwWyJSQkRfTUVUQURBVEFfRUNfUE9P
            TF9OQU1FIl0sCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlv
            L3Byb3Zpc2lvbmVyLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxmLm91dF9tYXBbJ0NTSV9SQkRf
            UFJPVklTSU9ORVJfU0VDUkVUX05BTUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9jb250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9v
            ay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFnZS5rOHMuaW8vbm9kZS1zdGFnZS1z
            ZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX05PREVfU0VDUkVUX05B
            TUUnXX0iLAogICAgICAgICAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgICAgIH0K
            ICAgICAgICAgICAgICAgICkKICAgICAgICAgICAgZWxzZToKICAgICAgICAgICAgICAgIGpzb25f
            b3V0LmFwcGVuZCgKICAgICAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAgICAgICAg
            ICJuYW1lIjogImNlcGgtcmJkIiwKICAgICAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3Rv
            cmFnZUNsYXNzIiwKICAgICAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAg
            ICAgICAgICAgICAgICAgICAicG9vbCI6IHNlbGYub3V0X21hcFsiUkJEX1BPT0xfTkFNRSJdLAog
            ICAgICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25l
            ci1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVS
            X1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgICAgICJjc2kuc3RvcmFn
            ZS5rOHMuaW8vY29udHJvbGxlci1leHBhbmQtc2VjcmV0LW5hbWUiOiBmInJvb2ste3NlbGYub3V0
            X21hcFsnQ1NJX1JCRF9QUk9WSVNJT05FUl9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgICAgICAgICAiY3NpLnN0b3JhZ2UuazhzLmlvL25vZGUtc3RhZ2Utc2VjcmV0LW5hbWUi
            OiBmInJvb2ste3NlbGYub3V0X21hcFsnQ1NJX1JCRF9OT0RFX1NFQ1JFVF9OQU1FJ119IiwKICAg
            ICAgICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgICAgICB9CiAgICAgICAgICAg
            ICAgICApCgogICAgICAgICMgaWYgJ1RPUE9MT0dZX1BPT0xTJywgJ1RPUE9MT0dZX0ZBSUxVUkVf
            RE9NQUlOX0xBQkVMJywgJ1RPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUycgIGV4aXN0cywK
            ICAgICAgICAjIHRoZW4gb25seSBhZGQgJ3RvcG9sb2d5JyBTdG9yYWdlQ2xhc3MKICAgICAgICBp
            ZiAoCiAgICAgICAgICAgIHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXQogICAgICAgICAg
            ICBhbmQgc2VsZi5vdXRfbWFwWyJUT1BPTE9HWV9GQUlMVVJFX0RPTUFJTl9MQUJFTCJdCiAgICAg
            ICAgICAgIGFuZCBzZWxmLm91dF9tYXBbIlRPUE9MT0dZX0ZBSUxVUkVfRE9NQUlOX1ZBTFVFUyJd
            CiAgICAgICAgKToKICAgICAgICAgICAganNvbl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAg
            ewogICAgICAgICAgICAgICAgICAgICJuYW1lIjogImNlcGgtcmJkLXRvcG9sb2d5IiwKICAgICAg
            ICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3MiLAogICAgICAgICAgICAgICAgICAg
            ICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAidG9wb2xvZ3lGYWlsdXJlRG9tYWlu
            TGFiZWwiOiBzZWxmLm91dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xP
            R1lfRkFJTFVSRV9ET01BSU5fTEFCRUwiCiAgICAgICAgICAgICAgICAgICAgICAgIF0sCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJ0b3BvbG9neUZhaWx1cmVEb21haW5WYWx1ZXMiOiBzZWxmLm91
            dF9tYXBbCiAgICAgICAgICAgICAgICAgICAgICAgICAgICAiVE9QT0xPR1lfRkFJTFVSRV9ET01B
            SU5fVkFMVUVTIgogICAgICAgICAgICAgICAgICAgICAgICBdLAogICAgICAgICAgICAgICAgICAg
            ICAgICAidG9wb2xvZ3lQb29scyI6IHNlbGYub3V0X21hcFsiVE9QT0xPR1lfUE9PTFMiXSwKICAg
            ICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNy
            ZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfUkJEX1BST1ZJU0lPTkVSX1NFQ1JF
            VF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9j
            b250cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lf
            UkJEX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9SQkRfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAgICAg
            ICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQoKICAgICAgICAjIGlmICdDRVBI
            RlNfRlNfTkFNRScgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoZnMnIFN0b3JhZ2VDbGFzcwog
            ICAgICAgIGlmIHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGhmcyIsCiAgICAgICAgICAgICAgICAgICAgImtpbmQiOiAiU3RvcmFnZUNsYXNzIiwK
            ICAgICAgICAgICAgICAgICAgICAiZGF0YSI6IHsKICAgICAgICAgICAgICAgICAgICAgICAgImZz
            TmFtZSI6IHNlbGYub3V0X21hcFsiQ0VQSEZTX0ZTX05BTUUiXSwKICAgICAgICAgICAgICAgICAg
            ICAgICAgInBvb2wiOiBzZWxmLm91dF9tYXBbIkNFUEhGU19QT09MX05BTUUiXSwKICAgICAgICAg
            ICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9wcm92aXNpb25lci1zZWNyZXQtbmFt
            ZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQSEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9O
            QU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAgImNzaS5zdG9yYWdlLms4cy5pby9jb250
            cm9sbGVyLWV4cGFuZC1zZWNyZXQtbmFtZSI6IGYicm9vay17c2VsZi5vdXRfbWFwWydDU0lfQ0VQ
            SEZTX1BST1ZJU0lPTkVSX1NFQ1JFVF9OQU1FJ119IiwKICAgICAgICAgICAgICAgICAgICAgICAg
            ImNzaS5zdG9yYWdlLms4cy5pby9ub2RlLXN0YWdlLXNlY3JldC1uYW1lIjogZiJyb29rLXtzZWxm
            Lm91dF9tYXBbJ0NTSV9DRVBIRlNfTk9ERV9TRUNSRVRfTkFNRSddfSIsCiAgICAgICAgICAgICAg
            ICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAgICAgICMgaWYgJ1JH
            V19FTkRQT0lOVCcgZXhpc3RzLCB0aGVuIG9ubHkgYWRkICdjZXBoLXJndycgU3RvcmFnZUNsYXNz
            CiAgICAgICAgaWYgc2VsZi5vdXRfbWFwWyJSR1dfRU5EUE9JTlQiXToKICAgICAgICAgICAganNv
            bl9vdXQuYXBwZW5kKAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJuYW1l
            IjogImNlcGgtcmd3IiwKICAgICAgICAgICAgICAgICAgICAia2luZCI6ICJTdG9yYWdlQ2xhc3Mi
            LAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAgICAgICAgICAgICAgICAi
            ZW5kcG9pbnQiOiBzZWxmLm91dF9tYXBbIlJHV19FTkRQT0lOVCJdLAogICAgICAgICAgICAgICAg
            ICAgICAgICAicG9vbFByZWZpeCI6IHNlbGYub3V0X21hcFsiUkdXX1BPT0xfUFJFRklYIl0sCiAg
            ICAgICAgICAgICAgICAgICAgfSwKICAgICAgICAgICAgICAgIH0KICAgICAgICAgICAgKQogICAg
            ICAgICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAg
            ICAgICAgIm5hbWUiOiAicmd3LWFkbWluLW9wcy11c2VyIiwKICAgICAgICAgICAgICAgICAgICAi
            a2luZCI6ICJTZWNyZXQiLAogICAgICAgICAgICAgICAgICAgICJkYXRhIjogewogICAgICAgICAg
            ICAgICAgICAgICAgICAiYWNjZXNzS2V5Ijogc2VsZi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VT
            RVJfQUNDRVNTX0tFWSJdLAogICAgICAgICAgICAgICAgICAgICAgICAic2VjcmV0S2V5Ijogc2Vs
            Zi5vdXRfbWFwWyJSR1dfQURNSU5fT1BTX1VTRVJfU0VDUkVUX0tFWSJdLAogICAgICAgICAgICAg
            ICAgICAgIH0sCiAgICAgICAgICAgICAgICB9CiAgICAgICAgICAgICkKICAgICAgICAjIGlmICdS
            R1dfVExTX0NFUlQnIGV4aXN0cywgdGhlbiBvbmx5IGFkZCB0aGUgImNlcGgtcmd3LXRscy1jZXJ0
            IiBzZWNyZXQKICAgICAgICBpZiBzZWxmLm91dF9tYXBbIlJHV19UTFNfQ0VSVCJdOgogICAgICAg
            ICAgICBqc29uX291dC5hcHBlbmQoCiAgICAgICAgICAgICAgICB7CiAgICAgICAgICAgICAgICAg
            ICAgIm5hbWUiOiAiY2VwaC1yZ3ctdGxzLWNlcnQiLAogICAgICAgICAgICAgICAgICAgICJraW5k
            IjogIlNlY3JldCIsCiAgICAgICAgICAgICAgICAgICAgImRhdGEiOiB7CiAgICAgICAgICAgICAg
            ICAgICAgICAgICJjZXJ0Ijogc2VsZi5vdXRfbWFwWyJSR1dfVExTX0NFUlQiXSwKICAgICAgICAg
            ICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgfQogICAgICAgICAgICApCgogICAgICAgIHJl
            dHVybiBqc29uLmR1bXBzKGpzb25fb3V0KSArIExJTkVTRVAKCiAgICBkZWYgdXBncmFkZV91c2Vy
            c19wZXJtaXNzaW9ucyhzZWxmKToKICAgICAgICB1c2VycyA9IFsKICAgICAgICAgICAgImNsaWVu
            dC5jc2ktY2VwaGZzLW5vZGUiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1jZXBoZnMtcHJvdmlz
            aW9uZXIiLAogICAgICAgICAgICAiY2xpZW50LmNzaS1yYmQtbm9kZSIsCiAgICAgICAgICAgICJj
            bGllbnQuY3NpLXJiZC1wcm92aXNpb25lciIsCiAgICAgICAgICAgICJjbGllbnQuaGVhbHRoY2hl
            Y2tlciIsCiAgICAgICAgXQogICAgICAgIGlmIHNlbGYucnVuX2FzX3VzZXIgIT0gIiIgYW5kIHNl
            bGYucnVuX2FzX3VzZXIgbm90IGluIHVzZXJzOgogICAgICAgICAgICB1c2Vycy5hcHBlbmQoc2Vs
            Zi5ydW5fYXNfdXNlcikKICAgICAgICBmb3IgdXNlciBpbiB1c2VyczoKICAgICAgICAgICAgc2Vs
            Zi51cGdyYWRlX3VzZXJfcGVybWlzc2lvbnModXNlcikKCiAgICBkZWYgZ2V0X3Jnd19wb29sX25h
            bWVfZHVyaW5nX3VwZ3JhZGUoc2VsZiwgdXNlciwgY2Fwcyk6CiAgICAgICAgaWYgdXNlciA9PSAi
            Y2xpZW50LmhlYWx0aGNoZWNrZXIiOgogICAgICAgICAgICAjIHdoZW4gYWRtaW4gaGFzIG5vdCBw
            cm92aWRlZCByZ3cgcG9vbCBuYW1lIGR1cmluZyB1cGdyYWRlLAogICAgICAgICAgICAjIGdldCB0
            aGUgcmd3IHBvb2wgbmFtZSBmcm9tIGNsaWVudC5oZWFsdGhjaGVja2VyIHVzZXIgd2hpY2ggd2Fz
            IHVzZWQgZHVyaW5nIGNvbm5lY3Rpb24KICAgICAgICAgICAgaWYgbm90IHNlbGYuX2FyZ19wYXJz
            ZXIucmd3X3Bvb2xfcHJlZml4OgogICAgICAgICAgICAgICAgIyBUbyBnZXQgdmFsdWUgJ2RlZmF1
            bHQnIHdoaWNoIGlzIHJndyBwb29sIG5hbWUgZnJvbSAnYWxsb3cgcnd4IHBvb2w9ZGVmYXVsdC5y
            Z3cubWV0YScKICAgICAgICAgICAgICAgIHBhdHRlcm4gPSByInBvb2w9KC4qPylcLnJnd1wubWV0
            YSIKICAgICAgICAgICAgICAgIG1hdGNoID0gcmUuc2VhcmNoKHBhdHRlcm4sIGNhcHMpCiAgICAg
            ICAgICAgICAgICBpZiBtYXRjaDoKICAgICAgICAgICAgICAgICAgICBzZWxmLl9hcmdfcGFyc2Vy
            LnJnd19wb29sX3ByZWZpeCA9IG1hdGNoLmdyb3VwKDEpCiAgICAgICAgICAgICAgICBlbHNlOgog
            ICAgICAgICAgICAgICAgICAgIHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAg
            ICAgICAgICAgICAgICAgICAgICJmYWlsZWQgdG8gZ2V0IHJndyBwb29sIG5hbWUgZm9yIHVwZ3Jh
            ZGUiCiAgICAgICAgICAgICAgICAgICAgKQoKICAgIGRlZiB1cGdyYWRlX3VzZXJfcGVybWlzc2lv
            bnMoc2VsZiwgdXNlcik6CiAgICAgICAgIyBjaGVjayB3aGV0aGVyIHRoZSBnaXZlbiB1c2VyIGV4
            aXN0cyBvciBub3QKICAgICAgICBjbWRfanNvbiA9IHsicHJlZml4IjogImF1dGggZ2V0IiwgImVu
            dGl0eSI6IGYie3VzZXJ9IiwgImZvcm1hdCI6ICJqc29uIn0KICAgICAgICByZXRfdmFsLCBqc29u
            X291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pzb24pCiAgICAg
            ICAgaWYgcmV0X3ZhbCAhPSAwIG9yIGxlbihqc29uX291dCkgPT0gMDoKICAgICAgICAgICAgcHJp
            bnQoZiJ1c2VyIHt1c2VyfSBub3QgZm91bmQgZm9yIHVwZ3JhZGluZy4iKQogICAgICAgICAgICBy
            ZXR1cm4KICAgICAgICBleGlzdGluZ19jYXBzID0ganNvbl9vdXRbMF1bImNhcHMiXQogICAgICAg
            IHNlbGYuZ2V0X3Jnd19wb29sX25hbWVfZHVyaW5nX3VwZ3JhZGUodXNlciwgc3RyKGV4aXN0aW5n
            X2NhcHMpKQogICAgICAgIG5ld19jYXAsIF8gPSBzZWxmLmdldF9jYXBzX2FuZF9lbnRpdHkodXNl
            cikKICAgICAgICBjYXBfa2V5cyA9IFsibW9uIiwgIm1nciIsICJvc2QiLCAibWRzIl0KICAgICAg
            ICBjYXBzID0gW10KICAgICAgICBmb3IgZWFjaENhcCBpbiBjYXBfa2V5czoKICAgICAgICAgICAg
            Y3VyX2NhcF92YWx1ZXMgPSBleGlzdGluZ19jYXBzLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAg
            ICAgbmV3X2NhcF92YWx1ZXMgPSBuZXdfY2FwLmdldChlYWNoQ2FwLCAiIikKICAgICAgICAgICAg
            Y3VyX2NhcF9wZXJtX2xpc3QgPSBbCiAgICAgICAgICAgICAgICB4LnN0cmlwKCkgZm9yIHggaW4g
            Y3VyX2NhcF92YWx1ZXMuc3BsaXQoIiwiKSBpZiB4LnN0cmlwKCkKICAgICAgICAgICAgXQogICAg
            ICAgICAgICBuZXdfY2FwX3Blcm1fbGlzdCA9IFsKICAgICAgICAgICAgICAgIHguc3RyaXAoKSBm
            b3IgeCBpbiBuZXdfY2FwX3ZhbHVlcy5zcGxpdCgiLCIpIGlmIHguc3RyaXAoKQogICAgICAgICAg
            ICBdCiAgICAgICAgICAgICMgYXBwZW5kIG5ld19jYXBfbGlzdCB0byBjdXJfY2FwX2xpc3QgdG8g
            bWFpbnRhaW4gdGhlIG9yZGVyIG9mIGNhcHMKICAgICAgICAgICAgY3VyX2NhcF9wZXJtX2xpc3Qu
            ZXh0ZW5kKG5ld19jYXBfcGVybV9saXN0KQogICAgICAgICAgICAjIGVsaW1pbmF0ZSBkdXBsaWNh
            dGVzIHdpdGhvdXQgdXNpbmcgJ3NldCcKICAgICAgICAgICAgIyBzZXQgcmUtb3JkZXJzIGl0ZW1z
            IGluIHRoZSBsaXN0IGFuZCB3ZSBoYXZlIHRvIGtlZXAgdGhlIG9yZGVyCiAgICAgICAgICAgIG5l
            d19jYXBfbGlzdCA9IFtdCiAgICAgICAgICAgIFtuZXdfY2FwX2xpc3QuYXBwZW5kKHgpIGZvciB4
            IGluIGN1cl9jYXBfcGVybV9saXN0IGlmIHggbm90IGluIG5ld19jYXBfbGlzdF0KICAgICAgICAg
            ICAgZXhpc3RpbmdfY2Fwc1tlYWNoQ2FwXSA9ICIsICIuam9pbihuZXdfY2FwX2xpc3QpCiAgICAg
            ICAgICAgIGlmIGV4aXN0aW5nX2NhcHNbZWFjaENhcF06CiAgICAgICAgICAgICAgICBjYXBzLmFw
            cGVuZChlYWNoQ2FwKQogICAgICAgICAgICAgICAgY2Fwcy5hcHBlbmQoZXhpc3RpbmdfY2Fwc1tl
            YWNoQ2FwXSkKICAgICAgICBjbWRfanNvbiA9IHsKICAgICAgICAgICAgInByZWZpeCI6ICJhdXRo
            IGNhcHMiLAogICAgICAgICAgICAiZW50aXR5IjogdXNlciwKICAgICAgICAgICAgImNhcHMiOiBj
            YXBzLAogICAgICAgICAgICAiZm9ybWF0IjogImpzb24iLAogICAgICAgIH0KICAgICAgICByZXRf
            dmFsLCBqc29uX291dCwgZXJyX21zZyA9IHNlbGYuX2NvbW1vbl9jbWRfanNvbl9nZW4oY21kX2pz
            b24pCiAgICAgICAgaWYgcmV0X3ZhbCAhPSAwOgogICAgICAgICAgICByYWlzZSBFeGVjdXRpb25G
            YWlsdXJlRXhjZXB0aW9uKAogICAgICAgICAgICAgICAgZiInYXV0aCBjYXBzIHt1c2VyfScgY29t
            bWFuZCBmYWlsZWQuXG4gRXJyb3I6IHtlcnJfbXNnfSIKICAgICAgICAgICAgKQogICAgICAgIHBy
            aW50KGYiVXBkYXRlZCB1c2VyIHt1c2VyfSBzdWNjZXNzZnVsbHkuIikKCiAgICBkZWYgbWFpbihz
            ZWxmKToKICAgICAgICBnZW5lcmF0ZWRfb3V0cHV0ID0gIiIKICAgICAgICBpZiBzZWxmLl9hcmdf
            cGFyc2VyLnVwZ3JhZGU6CiAgICAgICAgICAgIHNlbGYudXBncmFkZV91c2Vyc19wZXJtaXNzaW9u
            cygpCiAgICAgICAgZWxpZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAianNvbiI6CiAgICAg
            ICAgICAgIGdlbmVyYXRlZF9vdXRwdXQgPSBzZWxmLmdlbl9qc29uX291dCgpCiAgICAgICAgZWxp
            ZiBzZWxmLl9hcmdfcGFyc2VyLmZvcm1hdCA9PSAiYmFzaCI6CiAgICAgICAgICAgIGdlbmVyYXRl
            ZF9vdXRwdXQgPSBzZWxmLmdlbl9zaGVsbF9vdXQoKQogICAgICAgIGVsc2U6CiAgICAgICAgICAg
            IHJhaXNlIEV4ZWN1dGlvbkZhaWx1cmVFeGNlcHRpb24oCiAgICAgICAgICAgICAgICBmIlVuc3Vw
            cG9ydGVkIGZvcm1hdDoge3NlbGYuX2FyZ19wYXJzZXIuZm9ybWF0fSIKICAgICAgICAgICAgKQog
            ICAgICAgIHByaW50KGdlbmVyYXRlZF9vdXRwdXQpCiAgICAgICAgaWYgc2VsZi5vdXRwdXRfZmls
            ZSBhbmQgZ2VuZXJhdGVkX291dHB1dDoKICAgICAgICAgICAgZk91dCA9IG9wZW4oc2VsZi5vdXRw
            dXRfZmlsZSwgbW9kZT0idyIsIGVuY29kaW5nPSJVVEYtOCIpCiAgICAgICAgICAgIGZPdXQud3Jp
            dGUoZ2VuZXJhdGVkX291dHB1dCkKICAgICAgICAgICAgZk91dC5jbG9zZSgpCgoKIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjCiMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyBNQUlOICMjIyMjIyMjIyMjIyMjIyMjIyMjIwojIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMj
            IyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIHJq
            T2JqID0gUmFkb3NKU09OKCkKICAgIHRyeToKICAgICAgICByak9iai5tYWluKCkKICAgIGV4Y2Vw
            dCBFeGVjdXRpb25GYWlsdXJlRXhjZXB0aW9uIGFzIGVycjoKICAgICAgICBwcmludChmIkV4ZWN1
            dGlvbiBGYWlsZWQ6IHtlcnJ9IikKICAgICAgICByYWlzZSBlcnIKICAgIGV4Y2VwdCBLZXlFcnJv
            ciBhcyBrRXJyOgogICAgICAgIHByaW50KGYiS2V5RXJyb3I6IHtrRXJyfSIpCiAgICBleGNlcHQg
            T1NFcnJvciBhcyBvc0VycjoKICAgICAgICBwcmludChmIkVycm9yIHdoaWxlIHRyeWluZyB0byBv
            dXRwdXQgdGhlIGRhdGE6IHtvc0Vycn0iKQogICAgZmluYWxseToKICAgICAgICByak9iai5zaHV0
            ZG93bigpCg==
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/red-hat-storage/rook
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Ceph cluster.
            displayName: Ceph Cluster
            kind: CephCluster
            name: cephclusters.ceph.rook.io
            version: v1
          - description: Represents a Ceph Block Pool.
            displayName: Ceph Block Pool
            kind: CephBlockPool
            name: cephblockpools.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store.
            displayName: Ceph Object Store
            kind: CephObjectStore
            name: cephobjectstores.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store User.
            displayName: Ceph Object Store User
            kind: CephObjectStoreUser
            name: cephobjectstoreusers.ceph.rook.io
            version: v1
          - description: Represents a cluster of Ceph NFS ganesha gateways.
            displayName: Ceph NFS
            kind: CephNFS
            name: cephnfses.ceph.rook.io
            version: v1
          - description: Represents a Ceph User.
            displayName: Ceph Client
            kind: CephClient
            name: cephclients.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem.
            displayName: Ceph Filesystem
            kind: CephFilesystem
            name: cephfilesystems.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem Mirror.
            displayName: Ceph Filesystem Mirror
            kind: CephFilesystemMirror
            name: cephfilesystemmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph RBD Mirror.
            displayName: Ceph RBD Mirror
            kind: CephRBDMirror
            name: cephrbdmirrors.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Realm.
            displayName: Ceph Object Store Realm
            kind: CephObjectRealm
            name: cephobjectrealms.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone Group.
            displayName: Ceph Object Store Zone Group
            kind: CephObjectZoneGroup
            name: cephobjectzonegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph Object Store Zone.
            displayName: Ceph Object Store Zone
            kind: CephObjectZone
            name: cephobjectzones.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Notification.
            displayName: Ceph Bucket Notification
            kind: CephBucketNotification
            name: cephbucketnotifications.ceph.rook.io
            version: v1
          - description: Represents a Ceph Bucket Topic.
            displayName: Ceph Bucket Topic
            kind: CephBucketTopic
            name: cephbuckettopics.ceph.rook.io
            version: v1
          - description: Represents a Ceph Filesystem SubVolumeGroup.
            displayName: Ceph Filesystem SubVolumeGroup
            kind: CephFilesystemSubVolumeGroup
            name: cephfilesystemsubvolumegroups.ceph.rook.io
            version: v1
          - description: Represents a Ceph BlockPool Rados Namespace.
            displayName: Ceph BlockPool Rados Namespace
            kind: CephBlockPoolRadosNamespace
            name: cephblockpoolradosnamespaces.ceph.rook.io
            version: v1
          - description: Represents a Ceph COSI Driver.
            displayName: Ceph COSI Driver
            kind: CephCOSIDriver
            name: cephcosidrivers.ceph.rook.io
            version: v1
        description: |2

          The Rook-Ceph storage operator packages, deploys, manages, upgrades and scales Ceph storage for providing persistent storage to infrastructure services (Logging, Metrics, Registry) as well as stateful applications in Kubernetes clusters.

          ## Rook-Ceph Storage Operator

          Rook runs as a cloud-native service in Kubernetes clusters for optimal integration with applications in need of storage, and handles the heavy-lifting behind the scenes such as provisioning and management.
          Rook orchestrates battle-tested open-source storage technology Ceph, which has years of production deployments and runs some of the worlds largest clusters.

          Ceph is a massively scalable, software-defined, cloud native storage platform that offers block, file and object storage services.
          Ceph can be used to back a wide variety of applications including relational databases, NoSQL databases, CI/CD tool-sets, messaging, AI/ML and analytics applications.
          Ceph is a proven storage platform that backs some of the world's largest storage deployments and has a large vibrant open source community backing the project.

          ## Supported features
          * **High Availability and resiliency** - Ceph has no single point of failures (SPOF) and all its components work natively in a highly available fashion
          * **Data Protection** - Ceph periodically scrub for inconsistent objects and repair them if necessary, making sure your replicas are always coherent
          * **Consistent storage platform across hybrid cloud** - Ceph can be deployed anywhere (on-premise or bare metal) and thus offers a similar experience regardless
          * **Block, File & Object storage service** - Ceph can expose your data through several storage interfaces, solving all the application use cases
          * **Scale up/down** - addition and removal of storage is fully covered by the operator.
          * **Dashboard** - The Operator deploys a dashboard for monitoring and introspecting your cluster.

          ## Before you start
          https://rook.io/docs/rook/v1.0/k8s-pre-reqs.html
        displayName: Rook-Ceph
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - rook
        - ceph
        - storage
        - object storage
        - open source
        - block storage
        - shared filesystem
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/rook
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:c4367e141c1e9030d71d014616d4d6670e1394811d5e3790b8c19b6c493379e7
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:42cdce7884931c2e19018e581abf0bcde4710a05ea897668ca9d2b690faec936
        - registry.redhat.io/odf4/rook-ceph-operator-bundle@sha256:62cd98508ab41f9526f10c55f6c8054475a1b029230f255fe7461fb750fd1def
        - registry.redhat.io/odf4/cephcsi-rhel9@sha256:96c3fcee0755a7dced20bcacea4ceb3ff70b06fd073430cd5b428a70975c9170
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:c69ce4abba781f2c1cdb2ddfb5ddd3650e6a2f2aca62eb9864a64115ff3ed8b8
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:193a216cd1c4c48f544b6c204cb57e28eb2c8d162942742b3d585bf24ca0b4fb
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:0e26d59e58c16e076d8e8e5dc81b5a77165461291866cfb9588f8381a602239a
        version: 4.18.11-rhodf
      entries:
      - name: rook-ceph-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: rook-ceph-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: rook-ceph-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: rook-ceph-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: rook-ceph-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: rook-ceph-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: rook-ceph-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: rook-ceph-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: rook-ceph-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: rook-ceph-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: rook-ceph-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: rook-ceph-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: rook-ceph-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-console
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-console.v7.10.3-0.1655306919.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.10",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.10.3'
          olm.substitutesFor: fuse-console.v7.10.3-0.1652362834.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.10/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:bbfddd7cb4a6cb7a131a0cef174054890fd39956d63e737c92fe8233ad8f98e4
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:37dc38375f5d553ac1d9f20f66ce005ed9495354b81d7915ae3bd7dd6af132f3
        version: 7.10.3+0.1655306919.p
      entries:
      - name: fuse-console.v7.10.3
        version: 7.10.3
      - name: fuse-console.v7.10.3-0.1652362834.p
        version: 7.10.3+0.1652362834.p
      - name: fuse-console.v7.10.3-0.1655306919.p
        version: 7.10.3+0.1655306919.p
      name: 7.10.x
    - currentCSV: fuse-console.v7.11.1-0.1681224695.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.11",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.11.1'
          olm.substitutesFor: fuse-console.v7.11.1-0.1675972857.p
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.11/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5a35593357fbf280561a1d4a64ecc0d5c2044409acff3ce6cb490a4a61462a4e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:518b5ace9178e2aafd7b6f247fd41d9392abd1a7bccb4cb53783531820611193
        version: 7.11.1+0.1681224695.p
      entries:
      - name: fuse-console.v7.11.1
        version: 7.11.1
      - name: fuse-console.v7.11.1-0.1669922044.p
        version: 7.11.1+0.1669922044.p
      - name: fuse-console.v7.11.1-0.1675972857.p
        version: 7.11.1+0.1675972857.p
      - name: fuse-console.v7.11.1-0.1681224695.p
        version: 7.11.1+0.1681224695.p
      name: 7.11.x
    - currentCSV: fuse-console.v7.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.12",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.12.1'
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.12/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:791ac512370af5cde1e089de3def760e7dbce69badc03137051520ce316d63e2
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:4e6889bf76d323956e0b594db7381e8162fea6b108f53f9e6dd0b38cf3596c8d
        version: 7.12.1
      entries:
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.12.x
    - currentCSV: fuse-console.v7.13.0-0.1757604404.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "auth": {
                    "clientCertCheckSchedule": "* */12 * * *",
                    "clientCertExpirationPeriod": 24
                  },
                  "replicas": 1,
                  "version": "1.13",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "512Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
          createdAt: "2022-09-07T14:03:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-console.v7.13.0-0.1754559584.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.13/html/managing_fuse_on_openshift
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:5dc66b7c486d1604c4bbb04c85544147c43264840e597958e6c5ca83e7e34c0e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:72d98254a821bce237b3dde896f2825628cb7b6df564e23949337590586e8bea
        version: 7.13.0+0.1757604404.p
      entries:
      - name: fuse-console.v7.13.0-0.1744281421.p
        version: 7.13.0+0.1744281421.p
      - name: fuse-console.v7.13.0-0.1723709432.p
        version: 7.13.0+0.1723709432.p
      - name: fuse-console.v7.13.0-0.1757604404.p
        version: 7.13.0+0.1757604404.p
      - name: fuse-console.v7.13.0-0.1721224332.p
        version: 7.13.0+0.1721224332.p
      - name: fuse-console.v7.13.0-0.1744737026.p
        version: 7.13.0+0.1744737026.p
      - name: fuse-console.v7.13.0-0.1724659726.p
        version: 7.13.0+0.1724659726.p
      - name: fuse-console.v7.13.0-0.1725026290.p
        version: 7.13.0+0.1725026290.p
      - name: fuse-console.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-console.v7.13.0-0.1738056254.p
        version: 7.13.0+0.1738056254.p
      - name: fuse-console.v7.13.0-0.1739885034.p
        version: 7.13.0+0.1739885034.p
      - name: fuse-console.v7.13.0-0.1741875974.p
        version: 7.13.0+0.1741875974.p
      - name: fuse-console.v7.13.0-0.1741884704.p
        version: 7.13.0+0.1741884704.p
      - name: fuse-console.v7.13.0-0.1719822976.p
        version: 7.13.0+0.1719822976.p
      - name: fuse-console.v7.13.0
        version: 7.13.0
      - name: fuse-console.v7.13.0-0.1731460809.p
        version: 7.13.0+0.1731460809.p
      - name: fuse-console.v7.13.0-0.1749640277.p
        version: 7.13.0+0.1749640277.p
      - name: fuse-console.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-console.v7.13.0-0.1753702606.p
        version: 7.13.0+0.1753702606.p
      - name: fuse-console.v7.13.0-0.1753778626.p
        version: 7.13.0+0.1753778626.p
      - name: fuse-console.v7.13.0-0.1753785080.p
        version: 7.13.0+0.1753785080.p
      - name: fuse-console.v7.13.0-0.1753860418.p
        version: 7.13.0+0.1753860418.p
      - name: fuse-console.v7.13.0-0.1754303647.p
        version: 7.13.0+0.1754303647.p
      - name: fuse-console.v7.13.0-0.1754559584.p
        version: 7.13.0+0.1754559584.p
      - name: fuse-console.v7.13.0-0.1754567932.p
        version: 7.13.0+0.1754567932.p
      - name: fuse-console.v7.12.1
        version: 7.12.1
      name: 7.13.x
    - currentCSV: fuse-console.v7.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hawt.io/v1alpha1",
                "kind": "Hawtio",
                "metadata": {
                  "name": "fuse-console"
                },
                "spec": {
                  "replicas": 1,
                  "version": "1.9",
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "memory": "100Mi"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "32Mi"
                    }
                  },
                  "type": "Namespace"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
          createdAt: "2020-09-14T10:59:00Z"
          description: The Fuse Console eases the discovery and management of Fuse
            applications deployed on OpenShift.
          olm.skipRange: '>=7.7.0 <7.9.0'
          repository: https://github.com/hawtio/hawtio-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Fuse Console
            displayName: Hawtio
            kind: Hawtio
            name: hawtios.hawt.io
            version: v1alpha1
        description: |
          The Fuse Console eases the discovery and management of Fuse applications deployed on OpenShift.

          To secure the communication between the Fuse Console and the Jolokia agents, a client certificate is automatically generated.

          Note that the client certificate CN (defaults to `fuse-console.fuse.svc`) must be trusted by the Jolokia agents, for which client certification authentication is enabled. See the `clientPrincipal` parameter from the [Jolokia agent configuration options](https://jolokia.org/reference/html/agents.html#agent-jvm-config).
        displayName: Red Hat Integration - Fuse Console
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Fuse
        - Hawtio
        - Java
        - Management
        - Console
        links:
        - name: Red Hat Fuse Console Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_fuse/7.9/html/managing_fuse
        - name: Upstream Hawtio Web site
          url: https://hawt.io
        - name: Upstream Hawtio operator
          url: https://github.com/hawtio/hawtio-operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator-bundle@sha256:ddfdf850c9238a6988889e1840efe3175f03dad2489a5651e21cac62ac44e23e
        - registry.redhat.io/fuse7/fuse-console-rhel8-operator@sha256:846d57f3fd9f307a98b0169b6733c26299209d99ed189f219c48f36ba771b198
        version: 7.9.0
      entries:
      - name: fuse-console.v7.9.0
        version: 7.9.0
      name: 7.9.x
    defaultChannel: 7.13.x
    packageName: fuse-console
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: application-services-metering-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: application-services-metering-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "applicationservices.redhat.com/v1",
                "kind": "Meter",
                "metadata": {
                  "name": "app-services-metrics-example"
                },
                "spec": {
                  "includeInfrastructure": false,
                  "meterCollectionEnabled": true
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Application Runtime
          certified: "false"
          containerImage: quay.io/redhat-developer/application-services-metering-operator:0.6.1
          createdAt: "2021-09-28T13:00:00Z"
          description: Collect the core usage of products from the Application Services
            portfolio into a single metric labeled by Runtimes, Integration, and Automation.
          repository: https://github.com/redhat-developer/application-services-metering-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Meter
            displayName: Meter
            kind: Meter
            name: meters.applicationservices.redhat.com
            version: v1
        description: |-
          ## About the managed application

          There is no managed application created with this operator.

          ## About this Operator

          The metering operator watches pods for predetermined labels, across a cluster or namespace(s).
          The label schema for Red Hat Application Services is in the format:

          * `com.company=Red_Hat`
          * `rht.comp=PAM`
          * `rht.comp_ver=7.10`
          * `rht.prod_name=Red_Hat_Process_Automation`
          * `rht.prod_ver=7.11`
          * `rht.subcomp=rhpam-kogito-runtime`
          * `rht.subcomp_t=application`

          It's important not to remove labels as the operator will cease monitoring those pods.

          The operator exposes a metric named: `appsvcs_cpu_usage_cores`
          The metric has a label for product category, these are Runtimes, Integration and Automation.
          There is a prometheus rule using the metric to group by product name, `appsvcs_cores_by_product:sum`.

          The operator utilises the OpenShift Monitoring Stack to provide usage observability via cloud.redhat.com

          ## Prerequisites for enabling this Operator

          None.
        displayName: Application Services Metering Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - metering
        - operator
        links:
        - name: Red Hat Application Services portfolio
          url: https://www.redhat.com/en/products/middleware
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/application-services-metering-operator@sha256:c3600cc8cd789df1f3c8f48bb3b1bc4c9cbf0285b6f6ce06f1d563c52f50915e
        - quay.io/redhat-developer/application-services-metering-operator:0.6.1
        version: 0.6.1
      entries:
      - name: application-services-metering-operator.v0.6.1
        version: 0.6.1
      - name: application-services-metering-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: application-services-metering-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CNCF
      provider-url: ""
    name: etcd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: etcdoperator.v0.6.1
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          description: etcd is a distributed key value store providing a reliable
            way to store data across a cluster of machines.
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
        description: |
          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader. Your applications can read and write data into etcd.
          A simple use-case is to store database connection details or feature flags within etcd as key value pairs. These values can be watched, allowing your app to reconfigure itself when they change. Advanced uses take advantage of the consistency guarantees to implement database leader elections or do distributed locking across a cluster of workers.

          _The etcd Open Cloud Service is Public Alpha. The goal before Beta is to fully implement backup features._

          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` or with the API using the automatically generated Kubernetes Service.

          [Read the complete guide to using the etcd Open Cloud Service](https://coreos.com/tectonic/docs/latest/alm/etcd-ocs.html)

          ### Supported Features
          **High availability**
          Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.
          **Automated updates**
          Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.
          **Backups included**
          Coming soon, the ability to schedule backups to happen on or off cluster.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: support@coreos.com
          name: CoreOS, Inc
        maturity: alpha
        provider:
          name: CoreOS, Inc
        relatedImages:
        - quay.io/openshift-community-operators/etcd@sha256:e42581eacbc8349f40c6a4ed6d7d82d6168518e15b1a24cba36817205bd8688f
        - quay.io/coreos/etcd-operator@sha256:bd944a211eaf8f31da5e6d69e8541e7cada8f16a9f7a5a570b22478997819943
        version: 0.6.1
      entries:
      - name: etcdoperator.v0.6.1
        version: 0.6.1
      name: alpha
    - currentCSV: etcdoperator.v0.9.4-clusterwide
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example",
                  "annotations": {
                    "etcd.database.coreos.com/scope": "clusterwide"
                  }
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:1d308347e5e290ac4bb7dd7aab0a1c0659d661153541d81dd30733b52d1fbf83
        version: 0.9.4-clusterwide
      entries:
      - name: etcdoperator.v0.9.4-clusterwide
        version: 0.9.4-clusterwide
      - name: etcdoperator.v0.9.2-clusterwide
        version: 0.9.2-clusterwide
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: clusterwide-alpha
    - currentCSV: etcdoperator.v0.9.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "size": 3,
                  "version": "3.2.13"
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdRestore",
                "metadata": {
                  "name": "example-etcd-cluster-restore"
                },
                "spec": {
                  "etcdCluster": {
                    "name": "example-etcd-cluster"
                  },
                  "backupStorageType": "S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              },
              {
                "apiVersion": "etcd.database.coreos.com/v1beta2",
                "kind": "EtcdBackup",
                "metadata": {
                  "name": "example-etcd-cluster-backup"
                },
                "spec": {
                  "etcdEndpoints": ["<etcd-cluster-endpoints>"],
                  "storageType":"S3",
                  "s3": {
                    "path": "<full-s3-path>",
                    "awsSecret": "<aws-secret>"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
          createdAt: "2019-02-28 01:03:00"
          description: Create and maintain highly-available etcd clusters on Kubernetes
          repository: https://github.com/coreos/etcd-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a cluster of etcd nodes.
            displayName: etcd Cluster
            kind: EtcdCluster
            name: etcdclusters.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to backup an etcd cluster.
            displayName: etcd Backup
            kind: EtcdBackup
            name: etcdbackups.etcd.database.coreos.com
            version: v1beta2
          - description: Represents the intent to restore an etcd cluster from a backup.
            displayName: etcd Restore
            kind: EtcdRestore
            name: etcdrestores.etcd.database.coreos.com
            version: v1beta2
        description: |
          The etcd Operater creates and maintains highly-available etcd clusters on Kubernetes, allowing engineers to easily deploy and manage etcd clusters for their applications.

          etcd is a distributed key value store that provides a reliable way to store data across a cluster of machines. Itâ€™s open-source and available on GitHub. etcd gracefully handles leader elections during network partitions and will tolerate machine failure, including the leader.


          ### Reading and writing to etcd

          Communicate with etcd though its command line utility `etcdctl` via port forwarding:

              $ kubectl --namespace default port-forward service/example-client 2379:2379
              $ etcdctl --endpoints http://127.0.0.1:2379 get /

          Or directly to the API using the automatically generated Kubernetes Service:

              $ etcdctl --endpoints http://example-client.default.svc:2379 get /

          Be sure to secure your etcd cluster (see Common Configurations) before exposing it outside of the namespace or cluster.


          ### Supported Features

          * **High availability** - Multiple instances of etcd are networked together and secured. Individual failures or networking issues are transparently handled to keep your cluster up and running.

          * **Automated updates** - Rolling out a new etcd version works like all Kubernetes rolling updates. Simply declare the desired version, and the etcd service starts a safe rolling update to the new version automatically.

          * **Backups included** - Create etcd backups and restore them through the etcd Operator.

          ### Common Configurations

          * **Configure TLS** - Specify [static TLS certs](https://github.com/coreos/etcd-operator/blob/master/doc/user/cluster_tls.md) as Kubernetes secrets.

          * **Set Node Selector and Affinity** - [Spread your etcd Pods](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-node-selector-and-anti-affinity-across-nodes) across Nodes and availability zones.

          * **Set Resource Limits** - [Set the Kubernetes limit and request](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#three-member-cluster-with-resource-requirement) values for your etcd Pods.

          * **Customize Storage** - [Set a custom StorageClass](https://github.com/coreos/etcd-operator/blob/master/doc/user/spec_examples.md#custom-persistentvolumeclaim-definition) that you would like to use.
        displayName: etcd
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - etcd
        - key value
        - database
        - coreos
        - open source
        links:
        - name: Blog
          url: https://coreos.com/etcd
        - name: Documentation
          url: https://coreos.com/operators/etcd/docs/latest/
        - name: etcd Operator Source Code
          url: https://github.com/coreos/etcd-operator
        maintainers:
        - email: etcd-dev@googlegroups.com
          name: etcd Community
        maturity: alpha
        provider:
          name: CNCF
        relatedImages:
        - quay.io/coreos/etcd-operator@sha256:66a37fd61a06a43969854ee6d3e21087a98b93838e284a6086b13917f96b0d9b
        - quay.io/openshift-community-operators/etcd@sha256:f1757deff8c4048ae42cb033f3ddd7686dec2968e5a4cd61ff8f74fda69a8a09
        version: 0.9.4
      entries:
      - name: etcdoperator.v0.9.4
        version: 0.9.4
      - name: etcdoperator.v0.9.2
        version: 0.9.2
      - name: etcdoperator.v0.9.0
        version: 0.9.0
      name: singlenamespace-alpha
    defaultChannel: singlenamespace-alpha
    packageName: etcd
    provider:
      name: CNCF
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Tempo Operator SIG
      provider-url: ""
    name: tempo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          containerImage: ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
          createdAt: "2025-09-17T14:23:28Z"
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: tempo-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Grafana Tempo Operator SIG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Tempo is an open source, easy-to-use, and high-scale distributed tracing backend.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.
          Please refer to the [Tempo documentation](https://grafana.com/docs/tempo/latest/) for more information about Tempo.

          The Community Tempo Operator can manage Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Prerequisites
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that an object storage solution is available and configured.
        displayName: Community Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: ruben.vp8510@gmail.com
          name: Ruben Vargas
        - email: p.loffay@gmail.com
          name: Pavol Loffay
        - email: bongartz@klimlive.de
          name: Benedikt Bongartz
        - email: andreas@gerstmayr.me
          name: Andreas Gerstmayr
        - email: iblancas@redhat.com
          name: Israel Blancas Alvarez
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Grafana Tempo Operator SIG
        relatedImages:
        - quay.io/observatorium/opa-openshift:main-2025-06-16-ecdeca0
        - quay.io/openshift/origin-oauth-proxy:4.14
        - docker.io/grafana/tempo-query:2.8.2
        - docker.io/grafana/tempo:2.8.2
        - docker.io/jaegertracing/jaeger-query:1.68.0
        - ghcr.io/grafana/tempo-operator/tempo-operator:v0.18.0
        - quay.io/community-operator-pipeline-prod/tempo-operator@sha256:14282a0ca95401ab3e4c11dc2b611a6f57c00c461254f52ddf03ef74c725767b
        - quay.io/observatorium/api:main-2025-08-25-7489eb0
        version: 0.18.0
      entries:
      - name: tempo-operator.v0.18.0
        version: 0.18.0
      - name: tempo-operator.v0.17.1
        version: 0.17.1
      - name: tempo-operator.v0.15.3
        version: 0.15.3
      - name: tempo-operator.v0.15.2
        version: 0.15.2
      - name: tempo-operator.v0.15.1
        version: 0.15.1
      - name: tempo-operator.v0.15.0
        version: 0.15.0
      - name: tempo-operator.v0.14.2
        version: 0.14.2
      - name: tempo-operator.v0.14.1
        version: 0.14.1
      - name: tempo-operator.v0.14.0
        version: 0.14.0
      - name: tempo-operator.v0.13.0
        version: 0.13.0
      - name: tempo-operator.v0.12.0
        version: 0.12.0
      - name: tempo-operator.v0.11.1
        version: 0.11.1
      - name: tempo-operator.v0.11.0
        version: 0.11.0
      - name: tempo-operator.v0.10.0
        version: 0.10.0
      - name: tempo-operator.v0.9.0
        version: 0.9.0
      - name: tempo-operator.v0.8.0
        version: 0.8.0
      - name: tempo-operator.v0.7.0
        version: 0.7.0
      - name: tempo-operator.v0.6.0
        version: 0.6.0
      - name: tempo-operator.v0.5.0
        version: 0.5.0
      - name: tempo-operator.v0.4.0
        version: 0.4.0
      - name: tempo-operator.v0.3.0
        version: 0.3.0
      - name: tempo-operator.v0.2.0
        version: 0.2.0
      - name: tempo-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: tempo-operator
    provider:
      name: Grafana Tempo Operator SIG
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: vault-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vault-config-operator.v0.8.34
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AuthEngineMount",
                "metadata": {
                  "name": "authenginemount-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample",
                  "type": "kubernetes"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "maxRetries": 3,
                  "maxRetryDelay": 60,
                  "path": "azure/test",
                  "resource": "https://management.azure.com/",
                  "retryDelay": 4,
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azureauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azureauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azureauthenginerole-sample"
                },
                "spec": {
                  "BoundResourceGroups": [
                    "resourcegroup1",
                    "resourcegroup2"
                  ],
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundGroupIDs": [
                    "group1",
                    "group2"
                  ],
                  "boundLocations": [
                    "location1",
                    "location2"
                  ],
                  "boundScaleSets": [
                    "scaleset1",
                    "scaleset1"
                  ],
                  "boundServicePrincipalIDs": [
                    "sp1",
                    "sp2"
                  ],
                  "boundSubscriptionIDs": [
                    "subscription1",
                    "subscription2"
                  ],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "dev-role",
                  "path": "azure",
                  "policies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenBoundCIDRs": [
                    "10.20.102.13",
                    "10.20.102.15"
                  ],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [
                    "policy1",
                    "policy2"
                  ],
                  "tokenTTL": "",
                  "tokenType": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureCredentials": {
                    "passwordKey": "clientsecret",
                    "secret": {
                      "name": "aad-credentials"
                    },
                    "usernameKey": "clientid"
                  },
                  "clientID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "environment": "AzurePublicCloud",
                  "passwordPolicy": "",
                  "path": "azure",
                  "rootPasswordTTL": "128d",
                  "subscriptionID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx",
                  "tenantID": "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxx"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "AzureSecretEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "azuresecretenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "azuresecretenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "azuresecretenginerole-sample"
                },
                "spec": {
                  "TTL": "",
                  "applicationObjectID": "",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "azureGroups": "",
                  "azureRoles": "",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxTTL": "",
                  "name": "azure-role",
                  "path": "azure",
                  "permanentlyDelete": "",
                  "persistApp": "",
                  "signInAudience": "",
                  "tags": ""
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test",
                  "path": "cert"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "CertAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "certauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "certauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "certauthenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "certificate": "-----BEGIN CERTIFICATE-----\nMIIDJTCCAg2gAwIBAgIUF119kmCNlnMwK6yOOpfRs4JGhAAwDQYJKoZIhvcNAQEL\nBQAwFjEUMBIGA1UEAwwLZXhhbXBsZS5jb20wHhcNMjUwMTIxMTM0MjExWhcNMjYw\nMTIxMTM0MjExWjAWMRQwEgYDVQQDDAtleGFtcGxlLmNvbTCCASIwDQYJKoZIhvcN\nAQEBBQADggEPADCCAQoCggEBALad4Sx97VyFSfOepJoXUGYljg7hL+z20pVL2/M8\n8HXAQl4qf/i45b//Qz8VogMxdg3Y9J7O/M0n9hxbL18i1gm270VjEFyzjJCyUq0R\n72WMM4Sr3OjQPnVtJoQkWnH7wR3zsWK2N4NuIPdE3Ofb48Olyachffv7xCI+MLIa\n/adcyRggM4Z5f2nBCMcXL8D/eY03qCgd/U1aaPJ0WyJTZHLL4euL0ewKlKtB1JnU\n4MrRa5mbPQ3B+OhxcaD/7BNV6hPXAe2zftEv8LcimhbVoj3XmLJ+om58YIaEMadk\nJj61xnXbf+F/iV41onzwvTPNWCSWiLowDl8P3zkabuS583sCAwEAAaNrMGkwHQYD\nVR0OBBYEFH971CwrqkTvUz2qrUncWMS5ULwKMB8GA1UdIwQYMBaAFH971CwrqkTv\nUz2qrUncWMS5ULwKMA8GA1UdEwEB/wQFMAMBAf8wFgYDVR0RBA8wDYILZXhhbXBs\nZS5jb20wDQYJKoZIhvcNAQELBQADggEBAIQzT113TQC9BqitOCOoDFCJqcM+BHmf\nITEKlEhlqubPvA40lD4wkz67RP2SdQbF1glzZW5sDTxMjpHL+8nhdmo65CJpsQGP\nyMPwnOvx7W5JWQJa1Yt+48JNhgIymDMun+xwDsjNnlBqp0Kr7HkOxPimMPWN4R6B\nydG96puDGBYqxB1ujx5seOGfGvxPBBr3h8kNg054rtIIDrUG2iUTfpNIILJrd8Q2\nshop9//3VcIpUJO7Z1brvK/Hp4DrcKBZVo7vd8n+5uCx3vBA0tjqNAXYtFnSRvDZ\n6XV+rCZecDBFkaLuJ5YgYEK15gGe6lQAez6DTVLXMAKhZ52kceefnwM=\n-----END CERTIFICATE-----\n",
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "name": "test-role",
                  "path": "cert/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineConfig",
                "metadata": {
                  "name": "my-postgresql-database"
                },
                "spec": {
                  "allowedRoles": [
                    "read-only"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "connectionURL": "postgresql://{{username}}:{{password}}@my-postgresql-database.test-vault-config-operator.svc:5432",
                  "path": "test-vault-config-operator/database",
                  "pluginName": "postgresql-database-plugin",
                  "rootCredentials": {
                    "passwordKey": "postgresql-password",
                    "secret": {
                      "name": "postgresql-admin-password"
                    }
                  },
                  "username": "postgres"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineRole",
                "metadata": {
                  "name": "read-only"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "creationStatements": [
                    "CREATE ROLE \"{{name}}\" WITH LOGIN PASSWORD '{{password}}' VALID UNTIL '{{expiration}}'; GRANT SELECT ON ALL TABLES IN SCHEMA public TO \"{{name}}\";"
                  ],
                  "dBName": "my-postgresql-database",
                  "defaultTTL": "1h",
                  "maxTTL": "24h",
                  "path": "test-vault-config-operator/database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "DatabaseSecretEngineStaticRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "databasesecretenginestaticrole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "databasesecretenginestaticrole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "databasesecretenginestaticrole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "dBName": "db-name",
                  "path": "test-vault-config-operator/database",
                  "rotationPeriod": "24h",
                  "rotationStatements": [
                    "ALTER ROLE \"{{name}}\" WITH PASSWORD '{{password}}';git add"
                  ],
                  "username": "db-user"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthengineconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthengineconfig",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthengineconfig-sample"
                },
                "spec": {
                  "GCEalias": "role_id",
                  "GCEmetadata": "default",
                  "GCPCredentials": {
                    "passwordKey": "credentials",
                    "secret": {
                      "name": "gcp-serviceaccount-credentials"
                    },
                    "usernameKey": "serviceaccount"
                  },
                  "IAMalias": "default",
                  "IAMmetadata": "default",
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "customEndpoint": {},
                  "path": "azure",
                  "serviceAccount": "vault-sa"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GCPAuthEngineRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "gcpauthenginerole-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "gcpauthenginerole",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "gcpauthenginerole-sample"
                },
                "spec": {
                  "addGroupAliases": false,
                  "allowGCEInference": false,
                  "authentication": {
                    "path": "vault-admin",
                    "role": "vault-admin",
                    "serviceAccount": {
                      "name": "vault"
                    }
                  },
                  "boundInstanceGroups": [],
                  "boundLabels": [],
                  "boundProjects": {},
                  "boundRegions": [],
                  "boundServiceAccounts": {},
                  "boundZones": [],
                  "connection": {
                    "address": "https://vault.example.com"
                  },
                  "maxJWTExp": "",
                  "name": "gcp-iam-role",
                  "path": "gcp",
                  "policies": [],
                  "tokenBoundCIDRs": [],
                  "tokenExplicitMaxTTL": "",
                  "tokenMaxTTL": "",
                  "tokenNoDefaultPolicy": false,
                  "tokenNumUses": 0,
                  "tokenPeriod": 0,
                  "tokenPolicies": [],
                  "tokenTTL": "",
                  "tokenType": "",
                  "type": "iam"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineConfig",
                "metadata": {
                  "name": "githubsecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GitHubSecretEngineRole",
                "metadata": {
                  "name": "githubsecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "group-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "group",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "group-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "metadata": {
                    "team": "team-abc"
                  },
                  "policies": [
                    "team-abc-access"
                  ],
                  "type": "external"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupAlias",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "vault-config-operator",
                    "app.kubernetes.io/instance": "groupalias-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "groupalias",
                    "app.kubernetes.io/part-of": "vault-config-operator"
                  },
                  "name": "groupalias-sample"
                },
                "spec": {
                  "authEngineMountPath": "kubernetes",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupName": "group-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineConfig",
                "metadata": {
                  "name": "azure-oidc-config"
                },
                "spec": {
                  "OIDCClientID": "000000000-0000-0000-0000-000000000",
                  "OIDCCredentials": {
                    "passwordKey": "oidc_client_secret",
                    "secret": {
                      "name": "oidccredentials"
                    },
                    "usernameKey": "oidc_client_id"
                  },
                  "OIDCDiscoveryURL": "https://login.microsoftonline.com/000000000-0000-0000-0000-000000000/v2.0",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "oidc/azuread-oidc/"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "JWTOIDCAuthEngineRole",
                "metadata": {
                  "name": "azure-oidc-dev-role"
                },
                "spec": {
                  "OIDCScopes": [
                    "https://graph.microsoft.com/.default"
                  ],
                  "allowedRedirectURIs": [
                    "http://localhost:8250/oidc/callback",
                    "http://localhost:8200/ui/vault/auth/oidc/azuread-oidc/oidc/callback"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "groupsClaim": "groups",
                  "name": "dev-role",
                  "path": "oidc-aad/azuread-oidc",
                  "roleType": "oidc",
                  "tokenPolicies": [
                    "dev",
                    "prod"
                  ],
                  "userClaim": "email"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineConfig",
                "metadata": {
                  "name": "kubernetesauthengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kube-authengine-mount-sample"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesAuthEngineRole",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "path": "kubernetes",
                  "policies": [
                    "database-engine-admin"
                  ],
                  "targetNamespaces": {
                    "targetNamespaceSelector": {
                      "matchLabels": {
                        "database-engine-admin": "true"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineConfig",
                "metadata": {
                  "name": "kubernetessecretengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "KubernetesSecretEngineRole",
                "metadata": {
                  "name": "kubernetessecretenginerole-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineConfig",
                "metadata": {
                  "name": "ldapauthengineconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "LDAPAuthEngineGroup",
                "metadata": {
                  "name": "ldapauthenginegroup-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin",
                    "serviceAccount": {
                      "name": "default"
                    }
                  },
                  "name": "test-group",
                  "path": "ldap/test"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineConfig",
                "metadata": {
                  "name": "pkisecretengineconfig-sample"
                },
                "spec": {
                  "TTL": "8760h",
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "commonName": "pki-vault-demo.internal.io",
                  "crlDistributionPoints": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/crl\""
                  ],
                  "issuingCertificates": [
                    "https://vault-internal.vault.svc:8200/v1/test-vault-config-operator/pki/ca"
                  ],
                  "path": "test-vault-config-operator/pki",
                  "privateKeyType": "internal",
                  "type": "root"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PKISecretEngineRole",
                "metadata": {
                  "name": "pkisecretenginerole-sample"
                },
                "spec": {
                  "allowGlobDomains": true,
                  "allowSubdomains": true,
                  "allowedDomains": [
                    "internal.io",
                    "pki-vault-demo.svc",
                    "example.com"
                  ],
                  "allowedOtherSans": "*",
                  "allowedURISans": [
                    "*-pki-vault-demo.apps.example.com"
                  ],
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "maxTTL": "8760h",
                  "path": "test-vault-config-operator/pki"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "PasswordPolicy",
                "metadata": {
                  "name": "simple-password-policy"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "passwordPolicy": "length = 20\nrule \"charset\" {\n  charset = \"abcdefghijklmnopqrstuvwxyz\"\n}"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "database-engine-admin"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "policy-admin"
                  },
                  "policy": "# mount database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n  allowed_parameters = {\n    \"type\" = [\"database\"]\n    \"*\"   = []\n  }\n}\n\n# tune database secrets engines\npath \"/sys/mounts/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/tune\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database secrets engines\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/config/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n\n# Configure database roles\npath \"/{{identity.entity.aliases.${auth/kubernetes/@accessor}.metadata.service_account_namespace}}/database/roles/+\" {\n  capabilities = [ \"create\", \"read\", \"update\", \"delete\"]\n}\n"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineConfig",
                "metadata": {
                  "name": "quaysecretengineconfig-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineRole",
                "metadata": {
                  "name": "quaysecretenginerole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "QuaySecretEngineStaticRole",
                "metadata": {
                  "name": "quaysecretenginestaticrole-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineConfig",
                "metadata": {
                  "name": "rabbitmqsecretengineconfig-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitMQ-engine-admin"
                  },
                  "connectionURI": "https://test.com",
                  "leaseMaxTTL": 86400,
                  "leaseTTL": 86400,
                  "path": "vault-config-operator/rabbitmq"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RabbitMQSecretEngineRole",
                "metadata": {
                  "name": "rabbitmqsecretenginerole-sample"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "rabbitmq-engine-admin"
                  },
                  "path": "vault-config-operator/rabbitmq",
                  "tags": "administrator",
                  "vhostTopics": [
                    {
                      "topics": [
                        {
                          "permissions": {
                            "configure": ".*",
                            "read": ".*",
                            "write": ".*"
                          },
                          "topicName": "my-topic"
                        },
                        {
                          "permissions": {
                            "read": ".*"
                          },
                          "topicName": "my-read-topic"
                        }
                      ],
                      "vhostName": "/"
                    }
                  ],
                  "vhosts": [
                    {
                      "permissions": {
                        "configure": ".*",
                        "read": ".*",
                        "write": ".*"
                      },
                      "vhostName": "/"
                    },
                    {
                      "permissions": {
                        "read": "my-queue",
                        "write": "my-exchange"
                      },
                      "vhostName": "my-vhost"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "RandomSecret",
                "metadata": {
                  "name": "randomsecret-password"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "secret-writer"
                  },
                  "path": "test-vault-config-operator/kv",
                  "secretFormat": {
                    "passwordPolicyName": "simple-password-policy"
                  },
                  "secretKey": "password"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "SecretEngineMount",
                "metadata": {
                  "name": "database"
                },
                "spec": {
                  "authentication": {
                    "path": "kubernetes",
                    "role": "database-engine-admin"
                  },
                  "path": "test-vault-config-operator",
                  "type": "database"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "VaultSecret",
                "metadata": {
                  "name": "vaultsecret-sample"
                },
                "spec": {
                  "output": {
                    "annotations": {
                      "refresh": "every-minute"
                    },
                    "labels": {
                      "app": "test-vault-config-operator"
                    },
                    "name": "randomsecret",
                    "stringData": {
                      "anotherpassword": "{{ .anotherrandomsecret.password }}",
                      "password": "{{ .randomsecret.password }}"
                    },
                    "type": "Opaque"
                  },
                  "refreshPeriod": "3m0s",
                  "vaultSecretDefinitions": [
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "randomsecret",
                      "path": "test-vault-config-operator/kv/randomsecret-password"
                    },
                    {
                      "authentication": {
                        "path": "kubernetes",
                        "role": "secret-reader",
                        "serviceAccount": {
                          "name": "default"
                        }
                      },
                      "name": "anotherrandomsecret",
                      "path": "test-vault-config-operator/kv/another-password"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
          createdAt: "2025-09-18T18:56:36Z"
          description: This operator provides primitives to declaratively configure
            Hashicorp Vault.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: vault-config-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/vault-config-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthEngineMount is the Schema for the authenginemounts API
            displayName: Auth Engine Mount
            kind: AuthEngineMount
            name: authenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineConfig is the Schema for the azureauthengineconfigs
              API
            displayName: Azure Auth Engine Config
            kind: AzureAuthEngineConfig
            name: azureauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureAuthEngineRole is the Schema for the azureauthengineroles
              API
            displayName: Azure Auth Engine Role
            kind: AzureAuthEngineRole
            name: azureauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - kind: AzureSecretEngineConfig
            name: azuresecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: AzureSecretEngineRole is the Schema for the azuresecretengineroles
              API
            displayName: Azure Secret Engine Role
            kind: AzureSecretEngineRole
            name: azuresecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineConfig is the Schema for the certauthengineconfigs
              API
            displayName: Cert Auth Engine Config
            kind: CertAuthEngineConfig
            name: certauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: CertAuthEngineRole is the Schema for the certauthengineroles
              API
            displayName: Cert Auth Engine Role
            kind: CertAuthEngineRole
            name: certauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineConfig is the Schema for the databasesecretengineconfigs
              API
            displayName: Database Secret Engine Config
            kind: DatabaseSecretEngineConfig
            name: databasesecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineRole is the Schema for the databasesecretengineroles
              API
            displayName: Database Secret Engine Role
            kind: DatabaseSecretEngineRole
            name: databasesecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: DatabaseSecretEngineStaticRole is the Schema for the databasesecretenginestaticroles
              API
            displayName: Database Secret Engine Static Role
            kind: DatabaseSecretEngineStaticRole
            name: databasesecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineConfig is the Schema for the gcpauthengineconfigs
              API
            displayName: GCPAuth Engine Config
            kind: GCPAuthEngineConfig
            name: gcpauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GCPAuthEngineRole is the Schema for the gcpauthengineroles
              API
            displayName: GCPAuth Engine Role
            kind: GCPAuthEngineRole
            name: gcpauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineConfig is the Schema for the githubsecretengineconfigs
              API
            displayName: Git Hub Secret Engine Config
            kind: GitHubSecretEngineConfig
            name: githubsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: GitHubSecretEngineRole is the Schema for the githubsecretengineroles
              API
            displayName: Git Hub Secret Engine Role
            kind: GitHubSecretEngineRole
            name: githubsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: GroupAlias is the Schema for the groupaliases API
            displayName: Group Alias
            kind: GroupAlias
            name: groupaliases.redhatcop.redhat.io
            version: v1alpha1
          - description: Group is the Schema for the groups API
            displayName: Group
            kind: Group
            name: groups.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineConfig is the Schema for the jwtoidcauthengineconfigs
              API
            displayName: JWTOIDCAuth Engine Config
            kind: JWTOIDCAuthEngineConfig
            name: jwtoidcauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: JWTOIDCAuthEngineRole is the Schema for the jwtoidcauthengineroles
              API
            displayName: JWTOIDCAuth Engine Role
            kind: JWTOIDCAuthEngineRole
            name: jwtoidcauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineConfig is the Schema for the kubernetesauthengineconfigs
              API
            displayName: Kubernetes Auth Engine Config
            kind: KubernetesAuthEngineConfig
            name: kubernetesauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesAuthEngineRole can be used to define a KubernetesAuthEngineRole
              for the kube-auth authentication method
            displayName: Kubernetes Auth Engine Role
            kind: KubernetesAuthEngineRole
            name: kubernetesauthengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineConfig is the Schema for the kubernetessecretengineconfigs
              API
            displayName: Kubernetes Secret Engine Config
            kind: KubernetesSecretEngineConfig
            name: kubernetessecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: KubernetesSecretEngineRole is the Schema for the kubernetessecretengineroles
              API
            displayName: Kubernetes Secret Engine Role
            kind: KubernetesSecretEngineRole
            name: kubernetessecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineConfig is the Schema for the ldapauthengineconfigs
              API
            displayName: LDAPAuth Engine Config
            kind: LDAPAuthEngineConfig
            name: ldapauthengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: LDAPAuthEngineGroup is the Schema for the ldapauthenginegroups
              API
            displayName: LDAPAuth Engine Group
            kind: LDAPAuthEngineGroup
            name: ldapauthenginegroups.redhatcop.redhat.io
            version: v1alpha1
          - description: PasswordPolicy is the Schema for the passwordpolicies API
            displayName: Password Policy
            kind: PasswordPolicy
            name: passwordpolicies.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineConfig is the Schema for the pkisecretengineconfigs
              API
            displayName: PKISecret Engine Config
            kind: PKISecretEngineConfig
            name: pkisecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: PKISecretEngineRole is the Schema for the pkisecretengineroles
              API
            displayName: PKISecret Engine Role
            kind: PKISecretEngineRole
            name: pkisecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: Policy is the Schema for the policies API
            displayName: Policy
            kind: Policy
            name: policies.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineConfig is the Schema for the quaysecretengineconfigs
              API
            displayName: Quay Secret Engine Config
            kind: QuaySecretEngineConfig
            name: quaysecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineRole is the Schema for the quaysecretengineroles
              API
            displayName: Quay Secret Engine Role
            kind: QuaySecretEngineRole
            name: quaysecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: QuaySecretEngineStaticRole is the Schema for the quaysecretenginestaticroles
              API
            displayName: Quay Secret Engine Static Role
            kind: QuaySecretEngineStaticRole
            name: quaysecretenginestaticroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineConfig is the Schema for the rabbitmqsecretengineconfigs
              API
            displayName: Rabbit MQSecret Engine Config
            kind: RabbitMQSecretEngineConfig
            name: rabbitmqsecretengineconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: RabbitMQSecretEngineRole is the Schema for the rabbitmqsecretengineroles
              API
            displayName: Rabbit MQSecret Engine Role
            kind: RabbitMQSecretEngineRole
            name: rabbitmqsecretengineroles.redhatcop.redhat.io
            version: v1alpha1
          - description: RandomSecret is the Schema for the randomsecrets API
            displayName: Random Secret
            kind: RandomSecret
            name: randomsecrets.redhatcop.redhat.io
            version: v1alpha1
          - description: SecretEngineMount is the Schema for the secretenginemounts
              API
            displayName: Secret Engine Mount
            kind: SecretEngineMount
            name: secretenginemounts.redhatcop.redhat.io
            version: v1alpha1
          - description: VaultSecret is the Schema for the vaultsecrets API
            displayName: Vault Secret
            kind: VaultSecret
            name: vaultsecrets.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator helps set up Vault Configurations. The main intent is to do so such that subsequently pods can consume the secrets made available.
          There are two main principles through all of the capabilities of this operator:

          1. high-fidelity API. The CRD exposed by this operator reflect field by field the Vault APIs. This is because we don't want to make any assumption on the kinds of configuration workflow that user will set up. That being said the Vault API is very extensive and we are starting with enough API coverage to support, we think, some simple and very common configuration workflows.
          2. attention to security (after all we are integrating with a security tool). To prevent credential leaks we give no permissions to the operator itself against Vault. All APIs exposed by this operator contains enough information to authenticate to Vault using a local service account (local to the namespace where the API exist). In other word for a namespace user to be abel to successfully configure Vault, a service account in that namespace must have been previously given the needed Vault permissions.

          Currently this operator supports the following CRDs:

          1. [Policy](https://github.com/redhat-cop/vault-config-operator#policy) Configures Vault [Policies](https://www.vaultproject.io/docs/concepts/policies)
          2. [KubernetesAuthEngineRole](https://github.com/redhat-cop/vault-config-operator#KubernetesAuthEngineRole) Configures a Vault [Kubernetes Authentication](https://www.vaultproject.io/docs/auth/kubernetes) Role
          3. [SecretEngineMount](https://github.com/redhat-cop/vault-config-operator#SecretEngineMount) Configures a Mount point for a [SecretEngine](https://www.vaultproject.io/docs/secrets)
          4. [DatabaseSecretEngineConfig](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineConfig) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Connection
          5. [DatabaseSecretEngineRole](https://github.com/redhat-cop/vault-config-operator#DatabaseSecretEngineRole) Configures a [Database Secret Engine](https://www.vaultproject.io/docs/secrets/databases) Role
          6. [RandomSecret](https://github.com/redhat-cop/vault-config-operator#RandomSecret) Creates a random secret in a vault [kv Secret Engine](https://www.vaultproject.io/docs/secrets/kv) with one password field generated using a [PasswordPolicy](https://www.vaultproject.io/docs/concepts/password-policies)vault-config-operator
        displayName: Vault Config Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - secrets
        - gitops
        - policy
        - credentials
        links:
        - name: Vault Config Operator
          url: https://github.com/redhat-cop/vault-config-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/vault-config-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/vault-config-operator@sha256:696bfe3769605870286e228f463dc2582202d6aeb8f199d78d87f153c68d7401
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/vault-config-operator@sha256:490852417fac37f3003ab7c1ee797f9035fd2905c354cc46eca5e33d2ad780ce
        version: 0.8.34
      entries:
      - name: vault-config-operator.v0.8.34
        version: 0.8.34
      - name: vault-config-operator.v0.8.33
        version: 0.8.33
      - name: vault-config-operator.v0.8.32
        version: 0.8.32
      - name: vault-config-operator.v0.8.31
        version: 0.8.31
      - name: vault-config-operator.v0.8.30
        version: 0.8.30
      - name: vault-config-operator.v0.8.29
        version: 0.8.29
      - name: vault-config-operator.v0.8.28
        version: 0.8.28
      - name: vault-config-operator.v0.8.27
        version: 0.8.27
      - name: vault-config-operator.v0.8.25
        version: 0.8.25
      - name: vault-config-operator.v0.8.24
        version: 0.8.24
      - name: vault-config-operator.v0.8.23
        version: 0.8.23
      - name: vault-config-operator.v0.8.22
        version: 0.8.22
      - name: vault-config-operator.v0.8.21
        version: 0.8.21
      - name: vault-config-operator.v0.8.20
        version: 0.8.20
      - name: vault-config-operator.v0.8.19
        version: 0.8.19
      - name: vault-config-operator.v0.8.18
        version: 0.8.18
      - name: vault-config-operator.v0.8.17
        version: 0.8.17
      - name: vault-config-operator.v0.8.16
        version: 0.8.16
      - name: vault-config-operator.v0.8.15
        version: 0.8.15
      - name: vault-config-operator.v0.8.14
        version: 0.8.14
      - name: vault-config-operator.v0.8.13
        version: 0.8.13
      - name: vault-config-operator.v0.8.12
        version: 0.8.12
      - name: vault-config-operator.v0.8.11
        version: 0.8.11
      - name: vault-config-operator.v0.8.10
        version: 0.8.10
      - name: vault-config-operator.v0.8.9
        version: 0.8.9
      - name: vault-config-operator.v0.8.7
        version: 0.8.7
      - name: vault-config-operator.v0.8.5
        version: 0.8.5
      - name: vault-config-operator.v0.8.4
        version: 0.8.4
      - name: vault-config-operator.v0.8.3
        version: 0.8.3
      - name: vault-config-operator.v0.8.2
        version: 0.8.2
      - name: vault-config-operator.v0.8.0
        version: 0.8.0
      - name: vault-config-operator.v0.7.1
        version: 0.7.1
      - name: vault-config-operator.v0.6.5
        version: 0.6.5
      - name: vault-config-operator.v0.6.4
        version: 0.6.4
      - name: vault-config-operator.v0.6.3
        version: 0.6.3
      - name: vault-config-operator.v0.6.2
        version: 0.6.2
      - name: vault-config-operator.v0.6.1
        version: 0.6.1
      - name: vault-config-operator.v0.6.0
        version: 0.6.0
      - name: vault-config-operator.v0.5.1
        version: 0.5.1
      - name: vault-config-operator.v0.5.0
        version: 0.5.0
      - name: vault-config-operator.v0.4.1
        version: 0.4.1
      - name: vault-config-operator.v0.4.0
        version: 0.4.0
      - name: vault-config-operator.v0.3.1
        version: 0.3.1
      - name: vault-config-operator.v0.3.0
        version: 0.3.0
      - name: vault-config-operator.v0.2.0
        version: 0.2.0
      - name: vault-config-operator.v0.1.2
        version: 0.1.2
      - name: vault-config-operator.v0.1.1
        version: 0.1.1
      - name: vault-config-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: vault-config-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-emrcontainers-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-emrcontainers-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "emrcontainers.services.k8s.aws/v1alpha1",
                "kind": "VirtualCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
          createdAt: "2025-09-19T19:36:26Z"
          description: AWS EMR on EKS controller is a service controller for managing
            EMR on EKS resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: JobRun represents the state of an AWS emrcontainers JobRun
              resource.
            displayName: JobRun
            kind: JobRun
            name: jobruns.emrcontainers.services.k8s.aws
            version: v1alpha1
          - description: VirtualCluster represents the state of an AWS emrcontainers
              VirtualCluster resource.
            displayName: VirtualCluster
            kind: VirtualCluster
            name: virtualclusters.emrcontainers.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EMR on EKS resources in AWS from within your Kubernetes cluster.

          **About Amazon EMR on EKS**

          Amazon EMR on EKS provides a deployment option for Amazon EMR that allows you to run open-source big data frameworks on Amazon Elastic Kubernetes Service (Amazon EKS). With this deployment option, you can focus on running analytics workloads while Amazon EMR on EKS builds, configures, and manages containers for open-source applications.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EMR on EKS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - emrcontainers
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EMR on EKS Developer Resources
          url: https://docs.aws.amazon.com/emr/latest/EMR-on-EKS-DevelopmentGuide/emr-eks.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: emr on eks maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/emrcontainers-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-emrcontainers-controller@sha256:cde2c51ca18d0969be0549e37ecd03cc430be3b4746d53b8744cb552a2c475df
        version: 1.1.1
      entries:
      - name: ack-emrcontainers-controller.v1.1.1
        version: 1.1.1
      - name: ack-emrcontainers-controller.v1.1.0
        version: 1.1.0
      - name: ack-emrcontainers-controller.v1.0.27
        version: 1.0.27
      - name: ack-emrcontainers-controller.v1.0.25
        version: 1.0.25
      - name: ack-emrcontainers-controller.v1.0.24
        version: 1.0.24
      - name: ack-emrcontainers-controller.v1.0.23
        version: 1.0.23
      - name: ack-emrcontainers-controller.v1.0.22
        version: 1.0.22
      - name: ack-emrcontainers-controller.v1.0.21
        version: 1.0.21
      - name: ack-emrcontainers-controller.v1.0.18
        version: 1.0.18
      - name: ack-emrcontainers-controller.v1.0.16
        version: 1.0.16
      - name: ack-emrcontainers-controller.v1.0.15
        version: 1.0.15
      - name: ack-emrcontainers-controller.v1.0.14
        version: 1.0.14
      - name: ack-emrcontainers-controller.v1.0.13
        version: 1.0.13
      - name: ack-emrcontainers-controller.v1.0.12
        version: 1.0.12
      - name: ack-emrcontainers-controller.v1.0.11
        version: 1.0.11
      - name: ack-emrcontainers-controller.v1.0.9
        version: 1.0.9
      - name: ack-emrcontainers-controller.v1.0.8
        version: 1.0.8
      - name: ack-emrcontainers-controller.v1.0.4
        version: 1.0.4
      - name: ack-emrcontainers-controller.v1.0.3
        version: 1.0.3
      - name: ack-emrcontainers-controller.v1.0.1
        version: 1.0.1
      - name: ack-emrcontainers-controller.v1.0.0
        version: 1.0.0
      - name: ack-emrcontainers-controller.v0.1.2
        version: 0.1.2
      - name: ack-emrcontainers-controller.v0.1.0
        version: 0.1.0
      - name: ack-emrcontainers-controller.v0.0.10
        version: 0.0.10
      - name: ack-emrcontainers-controller.v0.0.9
        version: 0.0.9
      - name: ack-emrcontainers-controller.v0.0.8
        version: 0.0.8
      - name: ack-emrcontainers-controller.v0.0.7
        version: 0.0.7
      - name: ack-emrcontainers-controller.v0.0.6
        version: 0.0.6
      - name: ack-emrcontainers-controller.v0.0.5
        version: 0.0.5
      - name: ack-emrcontainers-controller.v0.0.4
        version: 0.0.4
      - name: ack-emrcontainers-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-emrcontainers-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ocs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocs-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storageprofiles.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageProfile is the Schema for the storageprofiles API
            displayName: Storage Profile
            kind: StorageProfile
            name: storageprofiles.ocs.openshift.io
            version: v1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:29ec87ce725d2c277824312607915bb7de92e18015c388543173dbdadd5834a5
        - registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:b0d8139b4aff524d12a77962d189113e6ffa28486be35391a85f75ad967b299d
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a148fb746cbe799e8aa4a2eb5ded57d96e37f0dd9cfe7ae3a13e0ffe2f3833e9
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:c906c822fbab57c1eb95880b0dfb9cd04100c86222a5d5cfe6342298e3bb1b9e
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:66d47fe893b697a681b015b2dc15f50f50b8447bd400e77ca65b606a6f9c7985
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:551dc7c894d2d6f403729b2f2ca01624debd09d8e18448168f7d025072b78f5b
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:e26bc4b183f69a448552f9fe740292d4e803165d341d7d66047688d01fcd060c
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:3e47dfc8a5a136f492c1aea72e8741b5ebb3cae775d53bcf9ee484950c992bd2
        version: 4.17.13-rhodf
      entries:
      - name: ocs-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: ocs-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: ocs-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: ocs-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: ocs-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: ocs-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: ocs-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: ocs-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: ocs-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: ocs-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: ocs-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: ocs-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: ocs-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: ocs-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: ocs-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |2-

            [
                {
                    "apiVersion": "ocs.openshift.io/v1",
                    "kind": "StorageCluster",
                    "metadata": {
                        "name": "example-storagecluster",
                        "namespace": "openshift-storage"
                    },
                    "spec": {
                        "manageNodes": false,
                        "monPVCTemplate": {
                            "spec": {
                                "accessModes": [
                                    "ReadWriteOnce"
                                ],
                                "resources": {
                                    "requests": {
                                        "storage": "10Gi"
                                    }
                                },
                                "storageClassName": "gp2-csi"
                            }
                        },
                        "storageDeviceSets": [
                            {
                                "count": 3,
                                "dataPVCTemplate": {
                                    "spec": {
                                        "accessModes": [
                                            "ReadWriteOnce"
                                        ],
                                        "resources": {
                                            "requests": {
                                                "storage": "1Ti"
                                            }
                                        },
                                        "storageClassName": "gp2-csi",
                                        "volumeMode": "Block"
                                    }
                                },
                                "name": "example-deviceset",
                                "placement": {},
                                "portable": true,
                                "resources": {}
                            }
                        ]
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Storage
          containerImage: registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
          description: Red Hat OpenShift Container Storage provides hyperconverged
            storage for applications within an OpenShift cluster.
          external.features.ocs.openshift.io/supported-platforms: '["BareMetal", "None",
            "VSphere", "OpenStack", "oVirt"]'
          external.features.ocs.openshift.io/validation: '{"secrets":["rook-ceph-operator-creds",
            "rook-csi-rbd-node", "rook-csi-rbd-provisioner"], "configMaps": ["rook-ceph-mon-endpoints",
            "rook-ceph-mon"], "storageClasses": ["ceph-rbd"], "cephClusters": ["monitoring-endpoint"]}'
          features.ocs.openshift.io/disabled: '["ss-list", "install-wizard", "block-pool",
            "mcg-resource", "odf-dashboard", "common", "storage-provider", "storage-provisioner",
            "dashboard-resources", "csv-actions", "inventory-item", "alert-actions"]'
          features.ocs.openshift.io/enabled: '["kms", "arbiter", "flexible-scaling",
            "multus", "pool-management", "namespace-store", "mcg-standalone", "taint-nodes",
            "vault-sa-kms", "hpcs-kms"]'
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/initialization-resource: "\n    {\n        \"apiVersion\":
            \"ocs.openshift.io/v1\",\n        \"kind\": \"StorageCluster\",\n        \"metadata\":
            {\n            \"name\": \"example-storagecluster\",\n            \"namespace\":
            \"openshift-storage\"\n        },\n        \"spec\": {\n            \"manageNodes\":
            false,\n            \"monPVCTemplate\": {\n                \"spec\": {\n
            \                   \"accessModes\": [\n                        \"ReadWriteOnce\"\n
            \                   ],\n                    \"resources\": {\n                        \"requests\":
            {\n                            \"storage\": \"10Gi\"\n                        }\n
            \                   },\n                    \"storageClassName\": \"gp2-csi\"\n
            \               }\n            },\n            \"storageDeviceSets\":
            [\n                {\n                    \"count\": 3,\n                    \"dataPVCTemplate\":
            {\n                        \"spec\": {\n                            \"accessModes\":
            [\n                                \"ReadWriteOnce\"\n                            ],\n
            \                           \"resources\": {\n                                \"requests\":
            {\n                                    \"storage\": \"1Ti\"\n                                }\n
            \                           },\n                            \"storageClassName\":
            \"gp2-csi\",\n                            \"volumeMode\": \"Block\"\n
            \                       }\n                    },\n                    \"name\":
            \"example-deviceset\",\n                    \"placement\": {},\n                    \"portable\":
            true,\n                    \"resources\": {}\n                }\n            ]\n
            \       }\n    }\n\t"
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.4
          operators.operatorframework.io/internal-objects: '["ocsinitializations.ocs.openshift.io","storageclusterpeers.ocs.openshift.io","storageconsumers.ocs.openshift.io","storagerequests.ocs.openshift.io"]'
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/ocs-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OCSInitialization represents the initial data to be created
              when the operator is installed.
            displayName: OCS Initialization
            kind: OCSInitialization
            name: ocsinitializations.ocs.openshift.io
            version: v1
          - description: StorageClusterPeer is the Schema for the storageclusterpeers
              API
            displayName: Storage Cluster Peer
            kind: StorageClusterPeer
            name: storageclusterpeers.ocs.openshift.io
            version: v1
          - description: StorageCluster represents a cluster including Ceph Cluster,
              NooBaa and all the storage and compute resources required.
            displayName: Storage Cluster
            kind: StorageCluster
            name: storageclusters.ocs.openshift.io
            version: v1
          - description: StorageConsumer is the Schema for the storageconsumers API
            displayName: Storage Consumer
            kind: StorageConsumer
            name: storageconsumers.ocs.openshift.io
            version: v1alpha1
          - description: StorageRequest is the Schema for the StorageRequests API
            displayName: Storage Request
            kind: StorageRequest
            name: storagerequests.ocs.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Container Storage** deploys three operators.

          ### OpenShift Container Storage operator

          The OpenShift Container Storage operator is the primary operator for OpenShift Container Storage. It serves to facilitate the other operators in OpenShift Container Storage by performing administrative tasks outside their scope as well as watching and configuring their CustomResources.

          ### Rook

          [Rook][1] deploys and manages Ceph on OpenShift, which provides block and file storage.

          # Core Capabilities

          * **Self-managing service:** No matter which supported storage technologies you choose, OpenShift Container Storage ensures that resources can be deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Container Storage you can either build dedicated storage clusters or hyper-converged clusters where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Container Storage:** OpenShift Container Storage integrates Ceph with multiple storage presentations including object storage (compatible with S3), block storage, and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Container Storage efficiently distributes and replicates your data across your cluster to minimize the risk of data loss. With snapshots, cloning, and versioning, no more losing sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data storage for your workloads, using a single namespace.

          [1]: https://rook.io
        displayName: OpenShift Container Storage
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - storage
        - rook
        - ceph
        - block storage
        - shared filesystem
        - object storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ocs-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/ocs-metrics-exporter-rhel9@sha256:a4b5197db1ccda9934f30bfd63e41b0bef430ea9bd224cd3dc022466742b82b5
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/odf4/ocs-operator-bundle@sha256:0fcc80541d8966f3ea3502da1c4fdda24de277d8d73d0752b46e7d69fbcf7038
        - registry.redhat.io/odf4/rook-ceph-rhel9-operator@sha256:45c04107453f0b4c13c542627a74c38a4400791c21a2f0b96a8a6ae9290cfbc1
        - registry.redhat.io/odf4/odf-must-gather-rhel9@sha256:269b30202e6168ab2bf2cc7b13db433e71d8dbbd927375c05b8a1cfa49b089c7
        - registry.redhat.io/rhceph/rhceph-8-rhel9@sha256:df9b26d716e75e8680200cb6ed5c5ee51c24852afb4e5691c8a92775a8dfe2b9
        - registry.redhat.io/odf4/mcg-core-rhel9@sha256:5cb7b974d33e4be08003bcdd67ac93d54c57485220b9a94503043c6b4ca4bbb0
        - registry.redhat.io/odf4/ocs-rhel9-operator@sha256:a433bba1511c9efd80153df6e18a78a2390b6e3a7b82a5e195838bee4dde1dbe
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: ocs-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: ocs-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: ocs-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: ocs-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: ocs-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: ocs-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: ocs-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: ocs-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: ocs-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: ocs-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: ocs-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: ocs-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: ocs-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhsso-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhsso-operator.7.5.1-opr-005
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          olm.skipRange: <7.5.0
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:614a9b1d7975769e8559e0b8c1578d16f6e3f7fee0778283797f4d62f5fc2100
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:9fcdfa4615b91a5bb919aff71fc4df046365b2ee6d540b5f5f47f9ccaeb9877c
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:66577ec32a9a975f304e52b8ab49b9f7602158dcab1ef5d8c28a5d3455491336
        - registry.redhat.io/rh-sso-7/sso75-openj9-openshift-rhel8@sha256:2d4ea7226f43e7e8e8d3fcef6d425eb5104a15acf07791355033300b80db2901
        - registry.redhat.io/rh-sso-7/sso75-openshift-rhel8@sha256:5f0840a69642d13016430151892ec987fd67e711563b1dfb1a57353037c3bf79
        version: 7.5.1-opr-005
      entries:
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: alpha
    - currentCSV: rhsso-operator.7.6.12-opr-003
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "example-keycloak",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "instances": 1,
                  "externalAccess": {
                    "enabled": true
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "example-keycloakrealm"
                },
                "spec": {
                  "realm": {
                    "id": "basic",
                    "realm": "basic",
                    "enabled": true,
                    "displayName": "Basic Realm"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakBackup",
                "metadata": {
                  "name": "example-keycloakbackup"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "client-secret",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  },
                  "client": {
                    "clientId": "client-secret",
                    "secret": "client-secret",
                    "clientAuthenticatorType": "client-secret"
                  }
                }
              },
              {
                "apiVersion": "keycloak.org/v1alpha1",
                "kind": "KeycloakUser",
                "metadata": {
                  "name": "example-realm-user",
                  "labels": {
                    "app": "sso"
                  }
                },
                "spec": {
                  "user": {
                    "username": "realm_user",
                    "firstName": "John",
                    "lastName": "Doe",
                    "email": "user@example.com",
                    "enabled": true,
                    "emailVerified": false
                  },
                  "realmSelector": {
                    "matchLabels": {
                      "app": "sso"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "False"
          containerImage: registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
          createdAt: "2021-08-09 00:00:00"
          description: An Operator for installing and managing Red Hat Single Sign-On
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/keycloak/keycloak-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Red Hat Single Sign-On Instance
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Realm
            displayName: KeycloakRealm
            kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Backup
            displayName: KeycloakBackup
            kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On Client
            displayName: KeycloakClient
            kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - description: Represents a Red Hat Single Sign-On User
            displayName: KeycloakUser
            kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: |
          A Kubernetes Operator based on the Operator SDK for installing and managing Red Hat Single Sign-On.

          Red Hat Single Sign-On lets you add authentication to applications and secure services with minimum fuss. No need to deal with storing users or authenticating users. It's all available out of the box.

          The operator can deploy and manage Keycloak instances on Kubernetes and OpenShift.
          The following features are supported:

          * Install Keycloak to a namespace
          * Import Keycloak Realms
          * Import Keycloak Clients
          * Import Keycloak Users
          * Create scheduled backups of the database
        displayName: Red Hat Single Sign-On Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Keycloak
        - Identity
        - Access
        links:
        - name: Documentation
          url: https://access.redhat.com/products/red-hat-single-sign-on
        - name: Keycloak
          url: https://www.keycloak.org/
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat Customer Service
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator@sha256:5f157f534f9c14be7359489af775f48890d897c8444053e469feb6613fdef21b
        - registry.redhat.io/rh-sso-7/sso76-openshift-rhel8@sha256:ee2202b2d975c0692c06b9d18f705d6f45b43660b0a7927f50d4441b41ccdb0e
        - registry.redhat.io/rh-sso-7/sso7-rhel8-init-container@sha256:9f5b63888cbf4b1c90e3803d0017a01ad9c0c22693b978c9ee8299bfc5337832
        - registry.redhat.io/rh-sso-7/sso7-rhel8-operator-bundle@sha256:212ab9249707ca8ff835186d88e7888af18118fcbe98d948a3156b9700307a79
        version: 7.6.12-opr-003
      entries:
      - name: rhsso-operator.7.6.12-opr-003
        version: 7.6.12-opr-003
      - name: rhsso-operator.7.6.12-opr-002
        version: 7.6.12-opr-002
      - name: rhsso-operator.7.6.12-opr-001
        version: 7.6.12-opr-001
      - name: rhsso-operator.7.6.11-opr-006
        version: 7.6.11-opr-006
      - name: rhsso-operator.7.6.11-opr-005
        version: 7.6.11-opr-005
      - name: rhsso-operator.7.6.11-opr-004
        version: 7.6.11-opr-004
      - name: rhsso-operator.7.6.11-opr-003
        version: 7.6.11-opr-003
      - name: rhsso-operator.7.6.11-opr-002
        version: 7.6.11-opr-002
      - name: rhsso-operator.7.6.11-opr-001
        version: 7.6.11-opr-001
      - name: rhsso-operator.7.6.10-opr-001
        version: 7.6.10-opr-001
      - name: rhsso-operator.7.6.9-opr-003
        version: 7.6.9-opr-003
      - name: rhsso-operator.7.6.9-opr-002
        version: 7.6.9-opr-002
      - name: rhsso-operator.7.6.9-opr-001
        version: 7.6.9-opr-001
      - name: rhsso-operator.7.6.8-opr-002
        version: 7.6.8-opr-002
      - name: rhsso-operator.7.6.8-opr-001
        version: 7.6.8-opr-001
      - name: rhsso-operator.7.6.7-opr-002
        version: 7.6.7-opr-002
      - name: rhsso-operator.7.6.7-opr-001
        version: 7.6.7-opr-001
      - name: rhsso-operator.7.6.6-opr-004
        version: 7.6.6-opr-004
      - name: rhsso-operator.7.6.6-opr-003
        version: 7.6.6-opr-003
      - name: rhsso-operator.7.6.6-opr-002
        version: 7.6.6-opr-002
      - name: rhsso-operator.7.6.6-opr-001
        version: 7.6.6-opr-001
      - name: rhsso-operator.7.6.5-opr-004
        version: 7.6.5-opr-004
      - name: rhsso-operator.7.6.5-opr-003
        version: 7.6.5-opr-003
      - name: rhsso-operator.7.6.5-opr-002
        version: 7.6.5-opr-002
      - name: rhsso-operator.7.6.5-opr-001
        version: 7.6.5-opr-001
      - name: rhsso-operator.7.6.4-opr-003
        version: 7.6.4-opr-003
      - name: rhsso-operator.7.6.4-opr-002
        version: 7.6.4-opr-002
      - name: rhsso-operator.7.6.4-opr-001
        version: 7.6.4-opr-001
      - name: rhsso-operator.7.6.3-opr-002
        version: 7.6.3-opr-002
      - name: rhsso-operator.7.6.3-opr-001
        version: 7.6.3-opr-001
      - name: rhsso-operator.7.6.2-opr-003
        version: 7.6.2-opr-003
      - name: rhsso-operator.7.6.2-opr-002
        version: 7.6.2-opr-002
      - name: rhsso-operator.7.6.2-opr-001
        version: 7.6.2-opr-001
      - name: rhsso-operator.7.6.1-opr-005
        version: 7.6.1-opr-005
      - name: rhsso-operator.7.6.1-opr-004
        version: 7.6.1-opr-004
      - name: rhsso-operator.7.6.1-opr-003
        version: 7.6.1-opr-003
      - name: rhsso-operator.7.6.1-opr-002
        version: 7.6.1-opr-002
      - name: rhsso-operator.7.6.1-opr-001
        version: 7.6.1-opr-001
      - name: rhsso-operator.7.6.0-opr-005
        version: 7.6.0-opr-005
      - name: rhsso-operator.7.6.0-opr-004
        version: 7.6.0-opr-004
      - name: rhsso-operator.7.6.0-opr-003
        version: 7.6.0-opr-003
      - name: rhsso-operator.7.6.0-opr-002
        version: 7.6.0-opr-002
      - name: rhsso-operator.7.6.0-opr-001
        version: 7.6.0-opr-001
      - name: rhsso-operator.7.5.2-opr-003
        version: 7.5.2-opr-003
      - name: rhsso-operator.7.5.2-opr-002
        version: 7.5.2-opr-002
      - name: rhsso-operator.7.5.2-opr-001
        version: 7.5.2-opr-001
      - name: rhsso-operator.7.5.1-opr-011
        version: 7.5.1-opr-011
      - name: rhsso-operator.7.5.1-opr-010
        version: 7.5.1-opr-010
      - name: rhsso-operator.7.5.1-opr-009
        version: 7.5.1-opr-009
      - name: rhsso-operator.7.5.1-opr-009-0.1646836600.p
        version: 7.5.1-opr-009+0.1646836600.p
      - name: rhsso-operator.7.5.1-opr-008
        version: 7.5.1-opr-008
      - name: rhsso-operator.7.5.1-opr-007
        version: 7.5.1-opr-007
      - name: rhsso-operator.7.5.1-opr-006
        version: 7.5.1-opr-006
      - name: rhsso-operator.7.5.1-opr-005
        version: 7.5.1-opr-005
      - name: rhsso-operator.7.5.1-opr-004
        version: 7.5.1-opr-004
      - name: rhsso-operator.7.5.1-opr-003
        version: 7.5.1-opr-003
      - name: rhsso-operator.7.5.1-opr-002
        version: 7.5.1-opr-002
      - name: rhsso-operator.7.5.1-opr-001
        version: 7.5.1-opr-001
      - name: rhsso-operator.7.5.0
        version: 7.5.0
      name: stable
    defaultChannel: stable
    packageName: rhsso-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://www.intel.com
    name: ipm-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipm-controller.v25.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ipm.intel.com/v1",
                "kind": "CPUPowerProv",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ipm-controller",
                    "node-label": "ipm-provisioning"
                  },
                  "name": "cpupowerprov-sample"
                },
                "spec": {
                  "cstates": {
                    "cores": {
                      "5": {
                        "C1": true,
                        "C1E": true,
                        "C6": false
                      },
                      "9,12": {
                        "C1": true,
                        "C6": true
                      }
                    },
                    "idleGovernor": "ladder"
                  },
                  "pstates": {
                    "cores": {
                      "5": {
                        "energyPerformanceBias": "normal",
                        "energyPerformancePreference": "performance",
                        "freqGovernor": "performance",
                        "maxFreqMHz": "3500",
                        "minFreqMHz": "800"
                      },
                      "9,12": {
                        "freqGovernor": "powersave",
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    },
                    "global": {
                      "conservative": {
                        "downThreshold": 20,
                        "freqStep": 5,
                        "ignoreNiceLoad": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "maxPerfPct": 95,
                      "minPerfPct": 45,
                      "ondemand": {
                        "ignoreNiceLoad": 0,
                        "ioIsBusy": 1,
                        "powersaveBias": 0,
                        "samplingDownFactor": 1,
                        "samplingRate": 8888,
                        "upThreshold": 88
                      },
                      "schedutil": {
                        "rateLimitUs": 5000
                      },
                      "turbo": true
                    }
                  },
                  "uncoreFreqCPU": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  },
                  "uncoreFreqIO": {
                    "cpu0": [
                      {
                        "maxFreqMHz": "1600",
                        "minFreqMHz": "800"
                      }
                    ],
                    "cpu1": [
                      {
                        "maxFreqMHz": "1800",
                        "minFreqMHz": "1000"
                      }
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Monitoring
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
          createdAt: "2025-08-07T13:13:25Z"
          description: IPM Controller manages resources via REST API interactions.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.operatorframework.io/required-sccs: '["privileged"]'
          olm.securityContextConstraints.required: privileged
          operators.openshift.io/infrastructure-features: RestrictedNetwork
          operators.openshift.io/valid-subscription: Intel OBL Limited Distribution
            (Commercial Use) License Agreement [v2022.12.20]
          operators.operatorframework.io.bundle.package.v1: ipm-controller
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Intel® Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CPUPowerProv is the Schema for the cpupowerprovs API
            displayName: CPUPower Prov
            kind: CPUPowerProv
            name: cpupowerprovs.ipm.intel.com
            version: v1
        description: Infrastructure Power Manager (IPM) governs platform power technologies
          based on input from application workloads, the platform itself and customer
          input
        displayName: Intel® Infrastructure Power Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - intel
        - infrastructure
        - power
        - manager
        - management
        links:
        - name: IPM
          url: https://builders.intel.com/intel-technologies/power-management
        maintainers:
        - email: infrastructure.power.manager.operator@intel.com
          name: Infrastructure Power Manager Operator
        maturity: stable
        minKubeVersion: 1.29.0
        provider:
          name: Intel® Corporation
          url: https://www.intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/ipm-bundle@sha256:68c38ac59cf0d6cc145eb694e5b5a724da733caddb82c5585ef132e880dde544
        - registry.connect.redhat.com/intel/ipm-controller@sha256:32135c888f86a2c725d093ada9a59d222229bee5640d37f6e19a9e9812b4b67a
        - registry.connect.redhat.com/intel/ipm-watcher@sha256:51eb773d5a4d1f401d4cbf2b42b8df4d0359beb85d5e274b8392dd7c5165e6ff
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.13.0@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.connect.redhat.com/intel/ipm-agent@sha256:54de45cdc2e34768f6cdbf1cdee219406684143dddc4d236877709dc2536fe9c
        version: 25.7.0
      entries:
      - name: ipm-controller.v25.7.0
        version: 25.7.0
      name: stable
    defaultChannel: stable
    packageName: ipm-controller
    provider:
      name: Intel® Corporation
      url: https://www.intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-cpd-wspipelines
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-cpd-wspipelines.v10.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wspipelines.cpd.ibm.com/v1beta1",
                "kind": "WSPipelines",
                "metadata": {
                  "name": "wspipelines-cr",
                  "labels": {
                    "app.kubernetes.io/name": "wspipelines",
                    "app.kubernetes.io/instance": "wspipelines.wspipelines.cpd.ibm.com",
                    "app.kubernetes.io/managed-by": "wspipelines-operator",
                    "wspipelinesBuildNumber": "2017"
                }
                },
                "spec": {
                  "version": "5.1.3",
                  "scaleConfig": "small",
                  "license": {
                    "accept": false
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, Application Runtime
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
          createdAt: "2025-02-11"
          description: WS Pipelines CSV file
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-cpd-wspipelines
          olm.skipRange: '>=1.0.0 <10.3.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM watsonx.ai"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: 'Documentation For additional details regarding install parameters
              check: https://ibm.biz/BdaaL9 or https://ibm.biz/BdaaLC. License By
              installing this product you accept the license terms https://ibm.biz/BdaaL9
              or https://ibm.biz/BdaaLC.'
            kind: WSPipelines
            name: wspipelines.wspipelines.cpd.ibm.com
            version: v1beta1
        description: |-
          # Name
          IBM&reg; Watson Pipelines
          # Introduction
          ## Summary
          The Watson Pipelines editor provides a graphical interface for orchestrating an end-to-end flow of assets from creation through deployment. It enables to automate an end-to-end flow to prepare data, then create, train, deploy, monitor and update machine learning models and Python scripts. To design a pipeline you drag nodes onto the canvas, specify objects and parameters, then run and monitor the pipeline.
          ## Features
          Watson Pipelines are built off of Kubeflow pipelines on Tekton runtime and are fully integrated into the Watson Studio platform, allowing users to combine tools including:
          - Notebooks
          - Data refinery flows
          - AutoAI experiments
          - Web service / online deployments
          - Batch deployments
          - Import and export of Project / Space assets
          into repeatable and automated pipelines.
          # Details
          ## Prerequisites
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing-pre-installation-tasks)
          ### Resources Required
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-system-requirements)
          # Installing
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Configuration
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=installing)
          ## Storage
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=planning-storage-considerations)
          ## Limitations
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=issues-watson-studio)
          ## Documentation
          See [Details](https://www.ibm.com/docs/en/cloud-paks/cp-data/4.0?topic=overview)
          ## PodSecurityPolicy Requirements
          Custom PodSecurityPolicy definition:
          ```
          Not Applicable
          ```
          ## SecurityContextConstraints Requirements
          Custom SecurityContextConstraints definition:
          ```
          apiVersion: security.openshift.io/v1
          kind: SecurityContextConstraints
          allowHostDirVolumePlugin: false
          allowHostIPC: false
          allowHostNetwork: false
          allowHostPID: false
          allowHostPorts: false
          allowPrivilegeEscalation: false
          allowPrivilegedContainer: false
          allowedCapabilities: null
          defaultAddCapabilities: null
          fsGroup:
            type: MustRunAs
          metadata:
            annotations:
              include.release.openshift.io/self-managed-high-availability: "true"
              kubernetes.io/description: restrict denies access to all host features and requires pods to be run with a UID, and SELinux context that are allocated to the namespace.  This is the most restrictive SCC and it is used by default for authenticated users.
              release.openshift.io/create-only: "true"
            name: restrict
          priority: null
          readOnlyRootFilesystem: false
          requiredDropCapabilities:
          - ALL
          runAsUser:
            type: MustRunAsRange
          seLinuxContext:
            type: MustRunAs
          supplementalGroups:
            type: RunAsAny
          users: []
          volumes:
          - configMap
          - downwardAPI
          - emptyDir
          - persistentVolumeClaim
          - projected
          - secret
          ```
        displayName: Watson Pipelines
        installModes:
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: OwnNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wspipelines
        - WSPipelines
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: v1
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cp/cpd/wsp-ts@sha256:8d7625a0ba656425f601b6ec9a398a6d30511fe1de119cf0e29f680c704ea225
        - registry.marketplace.redhat.com/cp/cpd/rbs-ext@sha256:f0c68715f78bd20c352c97dd48ac5a912cad58e19a00159b5e6878f6beeb00fb
        - registry.marketplace.redhat.com/cp/cpd/artifact-store@sha256:00f8bcee4ba7e0c5cf5f0c52e9e7707df367041c9f4dd5508dbefe0dbdba8877
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:48b7cfa31b9a185176c20afc119e37346a6f9380d74000edf6c729b4677ca44e
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-entrypoint@sha256:e4888fadfadb3462e3a9ad0395a526a2631872f70caf291dc18b60269ded199f
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-worker@sha256:a24f4cb635e8b5759a5e2bda20ddefc6d5ea9b16fb1081592885c54aefb5480c
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-catalog@sha256:99d0b3192e6e46167423505f48655e7d49af1a461ae00b95c9c7c9d4209e5b65
        - registry.marketplace.redhat.com/cp/cpd/tekton-runtime-controller@sha256:6fc976bc622c2165dd8bf93c9244e1c29b7c973a4da7a5ce4d0a097a1c1ea4fb
        - registry.marketplace.redhat.com/cp/cpd/ubi-minimal-cpd@sha256:0164d8fbea591bbd9d67d84603162c3781b08551eddcc2a8a842aafc828af18c
        - registry.marketplace.redhat.com/cp/cpd/cpdctl@sha256:ad2215adfb8dd2dc600bf5097b7f88c257d5fbaf541409c01b36cdb67c130236
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:29cd5c9cd60bbcfe211337786616622b911e2319db9870eb9d38d4f630c03e75
        - registry.marketplace.redhat.com/cp/cpd/send-email@sha256:a477d47eaf8f8e569601d3c23c1def66f31c098ac3f5a1ae350618c460917dd8
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:f9f616d17f816468957b69bbb09e6c92d678acb4ba7a40f1cb493d08fb98deb8
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-nop@sha256:25a4bb9cea1c3bdbeeeb205622512bcba98163f9629eba35625756346ce9d652
        - registry.marketplace.redhat.com/cp/cpd/opa@sha256:ebcfc8941330da6991480dc18ac34d6df4fe879e562ee537e5c36df73f4bc536
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-controller@sha256:19bb32e0389361e8314e627bcad2e9bd93984289485cf632466834353d9c3449
        - registry.marketplace.redhat.com/cp/cpd/tekton-extensions-webhook@sha256:e29885c1ee083a73b952b0de8d1cd675b2b2830abf44d69a8577f56972a601fb
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:61155acdb8982648ae5748be9f5944ba913db1590ec3e773206d6fbb1d8952cb
        - registry.connect.redhat.com/cpopen/ibm-cpd-wspipelines-operator@sha256:19b163f055da43946643057c07500784210d02db868d2aa10f25a6501d767d3d
        - registry.marketplace.redhat.com/cp/cpd/tekton-pipelines-controller@sha256:a2949c3868c2c91bf35e50c64b51fcc24372986744b55520e21e9ce31fde59f0
        - registry.marketplace.redhat.com/cpopen/ibm-cpd-wspipelines-operator-bundle@sha256:842f2ac568db4db79cf0450b413b4b09b5f2180ef3541385b46fb650ac7066c0
        - registry.marketplace.redhat.com/cp/cpd/portal-pipelines@sha256:164f6063f3a50a36e28a240ae7c060c4f90266b7ed799bfdc075f90628bea7ff
        version: 10.3.0
      entries:
      - name: ibm-cpd-wspipelines.v10.3.0
        version: 10.3.0
      name: v10.3
    defaultChannel: v10.3
    packageName: ibm-cpd-wspipelines
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-common-service-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-common-service-operator.v4.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v3",
                "kind": "CommonService",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-common-service-operator",
                    "app.kubernetes.io/managed-by": "ibm-common-service-operator",
                    "app.kubernetes.io/name": "ibm-common-service-operator"
                  },
                  "name": "example-commonservice"
                },
                "spec": {
                  "size": "starterset"
                }
              }
            ]
          capabilities: Seamless Upgrades
          cloudPakThemesVersion: styles4100.css
          containerImage: registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
          createdAt: "2025-03-18T21:27:31Z"
          description: The IBM Cloud Pak foundational services operator is used to
            deploy IBM foundational services.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-common-service-operator
          nss.operator.ibm.com/managed-webhooks: ""
          olm.skipRange: '>=3.3.0 <4.12.0'
          operatorChannel: v4.12
          operatorVersion: 4.12.0
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/ibm-common-service-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CommonService is the Schema for the commonservices API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: CommonService
            kind: CommonService
            name: commonservices.operator.ibm.com
            version: v3
        description: |-
          # Introduction

          The `IBM Cloud Pak foundational services` is a bridge to connect IBM Cloud Paks and Operand Deployment Lifecycle Manager (ODLM) with IBM Cloud Pak foundational services. You can also install the `IBM Cloud Pak foundational services` in stand-alone mode.

          When you install this operator, the operator completes the following tasks:

            - Installs ODLM operator namespace
            - Creates the Common Services `OperandRegistry` and `OperandConfig` in the service namespace

          # Details
          For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). If you are using this operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          For more information about the available IBM Cloud Pak foundational services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

          Red Hat OpenShift Container Platform 4.5 or newer installed on one of the following platforms:

           - Linux x86_64
           - Linux on Power (ppc64le)
           - Linux on IBM Z and LinuxONE

          ## Operator versions

           - 4.12.0

          ## Prerequisites

          Before you install this operator, you need to first install the operator prerequisites:

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) | Disk (GB) | Nodes  |
          | -------------------------- | ----------- | ----------- | --------- | ------ |
          | ibm common service operator | 200          | 0.2        | 1          | worker |
          | **Total**                  | 200         | 0.2         | 1         |        |

          ## Documentation

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak. For a list of IBM Cloud Paks, see [IBM Cloud Paks that use IBM Cloud Pak foundational services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator in stand-alone mode or with an IBM Containerized Software, see the IBM Cloud Pak foundational services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).

          ## Installing

          For installation, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ## Configuration

          For configuration, see the [IBM Cloud Platform Common Services documentation](http://ibm.biz/cpcsdocs).

          ### SecurityContextConstraints Requirements

          The IBM Common Service Operator supports running with the OpenShift Container Platform 4.3 default restricted Security Context Constraints (SCCs) and IBM Cloud Pak Security Context Constraints (SCCs).

          For more information about the OpenShift Container Platform Security Context Constraints, see [Managing Security Context Constraints](https://docs.openshift.com/container-platform/4.3/authentication/managing-security-context-constraints.html).

          For more information about the IBM Cloud Pak Security Context Constraints, see [Managing Security Context Constraints](https://ibm.biz/cpkspec-scc).

          ### Limitations

          No
        displayName: IBM Cloud Pak foundational services
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Common Service Operator
          url: https://github.com/IBM/ibm-common-service-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/common-service-operator@sha256:50e585a9e0a6a968af2285516f69f908c360506968bc3ba68b27c5574444de20
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-operator-bundle@sha256:086eaeaaaf440a1d4a254984afacd8d04583aafb9ec11ac072feb42a87b986f7
        - registry.connect.redhat.com/cpopen/common-service-operator@sha256:13f5b6699ab505d02550b09e100f7e205c7fde3f2bc36317025bc795be216761
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-oadp-plugins@sha256:9a537eae7e581322e2f0764ed38a797ba7cd260724a285f24d5095058b7b47a0
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:0fdf1beeb4888f648f6dbd17a29dcc53a715d2c4301bc714a4172b2daf8d13be
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:4209e78f12120b9c9f320496b780c3993a9317825ea17dc7d554452b28c99014
        - registry.marketplace.redhat.com/cp/cpd/pipelines-python-runtime@sha256:98092e96fc3eb1426498acbc3cc65639ea119f76f605ddb6be9b914015159be8
        - registry.marketplace.redhat.com/cp/cpd/run-bash-script@sha256:18f44ab15289625fa2c6b72ece6e06afb51058dcba30b8f38b4fdeedf8c9e51b
        - registry.marketplace.redhat.com/cpopen/ibm-common-service-catalog@sha256:90046eb6170e12439ccc5aa343f19a4d4fc99a10d98b383cf1dc0e020ebeed55
        version: 4.12.0
      entries:
      - name: ibm-common-service-operator.v4.12.0
        version: 4.12.0
      name: v4.12
    defaultChannel: v4.12
    packageName: ibm-common-service-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Isovalent, part of Cisco
      provider-url: https://isovalent.com/
    name: clife
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clife.v1.17.8-cee.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cilium.io/v1alpha1",
                "kind": "CiliumConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "clife"
                  },
                  "name": "ciliumconfig"
                },
                "spec": {
                  "cluster": {
                    "name": "default"
                  },
                  "clusterHealthPort": 9940,
                  "cni": {
                    "binPath": "/var/lib/cni/bin",
                    "chainingMode": "portmap",
                    "confPath": "/var/run/multus/cni/net.d",
                    "exclusive": false
                  },
                  "enterprise": {
                    "featureGate": {
                      "approved": [
                        "CNIChainingMode"
                      ]
                    }
                  },
                  "hubble": {
                    "enabled": true
                  },
                  "ipam": {
                    "mode": "cluster-pool"
                  },
                  "kubeProxyReplacement": false,
                  "operator": {
                    "prometheus": {
                      "enabled": true,
                      "serviceMonitor": {
                        "enabled": true
                      }
                    }
                  },
                  "prometheus": {
                    "enabled": true,
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "securityContext": {
                    "privileged": true
                  },
                  "sessionAffinity": true,
                  "tunnelPort": 4789
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Networking,Security
          createdAt: "2025-09-23T07:55:19Z"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CiliumConfig defines the configuration of Isovalent Networking
              for Kubernetes and all its components
            displayName: CiliumConfig
            kind: CiliumConfig
            name: ciliumconfigs.cilium.io
            version: v1alpha1
        description: Isovalent Networking for Kubernetes, built on Cilium, provides
          eBPF-based networking, observability, and security with optimal scale and
          performance for platform teams operating Kubernetes environments across
          cloud and on-prem infrastructure.
        displayName: Isovalent Networking for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Cilium
        - networking
        - observability
        - security
        - eBPF
        - CNI
        links:
        - name: Isovalent Networking for Kubernetes
          url: https://isovalent.com/product/
        maintainers:
        - email: support@isovalent.com
          name: Isovalent, part of Cisco
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Isovalent, part of Cisco
          url: https://isovalent.com/
        relatedImages:
        - quay.io/isovalent/cilium-envoy-ubi:v1.17.8-cee.1@sha256:c3de0334f90f9b2540cdec7e2356d813f2df93f0d7d8701005eab7161c7663b5
        - quay.io/isovalent/cilium-ubi:v1.17.8-cee.1@sha256:e9816096e481fad5d95209f902e4a7018aeb65d1e3be1574e6b289e21229f87a
        - quay.io/isovalent/startup-script-ubi:v1.17.8-cee.1@sha256:9a6a413262ec59026345aefc743de847891d80e31ef2f41db954acdcee11c2ff
        - registry.connect.redhat.com/isovalent/clife-bundle@sha256:6d2ba514e40559fe3568dd0107f5e6648d43b543b3dbe4319e4114ca05fe20a6
        - quay.io/isovalent/certgen-ubi:v1.17.8-cee.1@sha256:fb63bed641a279cc09114e93821c7b7555c61393ab594ea2b823ce34fa73335e
        - quay.io/isovalent/clife@sha256:ba9bd4a5c35e1d26e395afbc2a4de89f42df1e5e61c9b23959c23194368b7b0e
        - quay.io/isovalent/operator-generic-ubi:v1.17.8-cee.1@sha256:37e4b9f65b2c67d24afeeebebe5d0cb59c09280179de57cd77da2b68e19c3c1b
        - quay.io/isovalent/clustermesh-apiserver-ubi:v1.17.8-cee.1@sha256:b097c823638b4fdbe5bdd2b6040a17fef0c70a020e67a18f8231d60eeac02dd5
        - quay.io/isovalent/hubble-relay-ubi:v1.17.8-cee.1@sha256:88964d4836d59420dc7e0abb758654ceaf2abcfb53e19bbe2c7962aa68874e37
        version: 1.17.8-cee.1
      entries:
      - name: clife.v1.17.8-cee.1
        version: 1.17.8-cee.1
      - name: clife.v1.17.7-cee.1
        version: 1.17.7-cee.1
      - name: clife.v1.17.5-cee.1
        version: 1.17.5-cee.1
      - name: clife.v1.17.4-cee.1
        version: 1.17.4-cee.1
      - name: clife.v0.0.1
        version: 0.0.1
      name: "1.17"
    defaultChannel: "1.17"
    packageName: clife
    provider:
      name: Isovalent, part of Cisco
      url: https://isovalent.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-manager.v0.13.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "mode": "Default"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationConfiguration": {
                    "featureGates": [
                      {
                        "feature": "DefaultClusterSet",
                        "mode": "Enable"
                      }
                    ]
                  },
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              },
              {
                "apiVersion": "operator.open-cluster-management.io/v1",
                "kind": "ClusterManager",
                "metadata": {
                  "name": "cluster-manager"
                },
                "spec": {
                  "addOnManagerImagePullSpec": "quay.io/open-cluster-management/addon-manager:v0.13.1",
                  "deployOption": {
                    "hosted": {
                      "registrationWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 30443
                      },
                      "workWebhookConfiguration": {
                        "address": "management-control-plane",
                        "port": 31443
                      }
                    },
                    "mode": "Hosted"
                  },
                  "placementImagePullSpec": "quay.io/open-cluster-management/placement:v0.13.1",
                  "registrationImagePullSpec": "quay.io/open-cluster-management/registration:v0.13.1",
                  "workImagePullSpec": "quay.io/open-cluster-management/work:v0.13.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/registration-operator:v0.13.1
          createdAt: "2024-04-10T15:27:49Z"
          description: Manages the installation and upgrade of the ClusterManager.
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-cluster-management-io/ocm
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterManager defines the configurations of controllers
              running hub that govern registration and work distribution for attached
              Klusterlets
            displayName: ClusterManager
            kind: ClusterManager
            name: clustermanagers.operator.open-cluster-management.io
            version: v1
        description: "Note: This operator will stop updating, please use the latest
          operator in OperatorHub.io https://operatorhub.io/operator/cluster-manager.
          \n\nThe Cluster Manager provides the multicluster hub, which can manage
          Kubernetes-based clusters across data centers, public clouds, and private
          clouds. This operator supports the installation and upgrade of ClusterManager.\n\n##
          How to Install\nInstall ClusterManager operator by following instructions
          in top right button `Install`. After installing the operator, create an
          instance of the ClusterManager resource to install the ClusterManager.\n\nThe
          pods of ClusterManager are created in the `open-cluster-management-hub`
          namespace:\n\n```\n$ kubectl get pods -n open-cluster-management-hub\nNAME
          \                                                      READY   STATUS    RESTARTS
          \  AGE\ncluster-manager-addon-manager-controller-7fcc67bfdf-x7hbz  1/1     Running
          \  0          61m\ncluster-manager-placement-controller-dbcfd4975-26jpw
          \      1/1     Running   0          61m\ncluster-manager-registration-controller-7949d97899-wwtg4
          \  1/1     Running   0          61m\ncluster-manager-registration-webhook-7f675bc4bb-v4rhj
          \     1/1     Running   0          61m\ncluster-manager-work-webhook-6c9d65fb95-jh97b
          \             1/1     Running   0          61m\n```\n\nThere is a ManagedCluster
          created in the Hub cluster when the managed cluster is registered to the
          Hub cluster.\n\n```\n$ kubectl get managedclusters.cluster.open-cluster-management.io\nNAME
          \                  HUB ACCEPTED   MANAGED CLUSTER URLS     JOINED   AVAILABLE
          \  AGE\ncluster-dev            true                                    True
          \    True        7h20m\ncluster-test           false          https://127.0.0.1:8443
          \                       7h25m\n```\n"
        displayName: Cluster Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open-cluster-management
        - ClusterManager
        links:
        - name: Registration Operator
          url: https://github.com/open-cluster-management-io/ocm
        maintainers:
        - email: zyin@redhat.com
          name: Zhiwei Yin
        - email: deads@redhat.com
          name: David Eads
        - email: jqiu@redhat.com
          name: Jian Qiu
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/community-operator-pipeline-prod/cluster-manager@sha256:a4c8a54148538b371e5cea3b2f2d4109f7b6bc4a7f2e5b52a840c9340fbf5e47
        - quay.io/open-cluster-management/registration-operator:v0.13.1
        version: 0.13.1
      entries:
      - name: cluster-manager.v0.13.1
        version: 0.13.1
      - name: cluster-manager.v0.13.0
        version: 0.13.0
      - name: cluster-manager.v0.12.0
        version: 0.12.0
      - name: cluster-manager.v0.11.0
        version: 0.11.0
      - name: cluster-manager.v0.10.0
        version: 0.10.0
      - name: cluster-manager.v0.9.1
        version: 0.9.1
      - name: cluster-manager.v0.8.0
        version: 0.8.0
      - name: cluster-manager.v0.7.0
        version: 0.7.0
      - name: cluster-manager.v0.6.0
        version: 0.6.0
      - name: cluster-manager.v0.5.0
        version: 0.5.0
      - name: cluster-manager.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: cluster-manager
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: external-dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: external-dns-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "provider": {
                    "type": "AWS",
                    "aws": {
                      "credentials": {
                        "name": "aws-access-key"
                      }
                    }
                  },
                  "source": {
                    "type": "Service"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
          createdAt: 2022-03-01T21:00:46
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## Prerequisites and Requirements
           The ExternalDNS Operator has minimal cluster level permissions. Unfortunately OLM doesn't allow the local role creation in the operand namespace yet.
           Therefore make sure the following commands are run against your cluster before the installation of the operator:
           * Create the operand namespace: `oc create ns external-dns`
           * Grant the operator access to manage operand resources: `oc apply -f https://raw.githubusercontent.com/openshift/external-dns-operator/release-0.1/config/rbac/extra-roles.yaml`
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:c1134bb46172997ef7278b6cefbb0da44e72a9f808a7cd67b3c65d464754cab9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b6fde43491c2e0733c8fe6ee8660f7f6435fe20c0b6d435cfc7f86eeab8b4748
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a6399dfd6fe0e8dc9f59fac4382320d23f9e1709e48ed6da7db69e76354b2646
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:6be1403e418f9b63ec8ebcc8858e602409bc41fdee4586acc627623a78b50eb5
        version: 0.1.2
      entries:
      - name: external-dns-operator.v0.1.2
        version: 0.1.2
      name: alpha
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1
    - currentCSV: external-dns-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
          createdAt: 2024-01-22T09:26:37
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:0a86a1808d9906eb57886d3d3cd5c04cf9578a80860342132d36ecfc411bf807
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:0877eb0a111efe0bff63173a2dfeb7a06a61afd0efbcdba053322a539e0dc667
        - registry.redhat.io/edo/external-dns-rhel8@sha256:17f4127cbb9a9304769e9655c19cbea2260107c7e9371a99a9aa3e85d759c00b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.0.1
      entries:
      - name: external-dns-operator.v1.0.1
        version: 1.0.1
      - name: external-dns-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    - currentCSV: external-dns-operator.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
          createdAt: 2024-01-22T09:56:14
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.1.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:b02bab3af52a2469ced96618f1d122464568ba21ebc0a3a945b78ec4e6bdc80d
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:9ca751e84b8843cc62dc60550e5c4af741bf09a381ef62220e7eb3760ffe8524
        - registry.redhat.io/edo/external-dns-rhel8@sha256:ffcc3f2e9b9feb889e7882465f7f034aba6b9ee0699a1b6c24e83f501a63ecca
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:f11f71448986aa17abec9caadb568a6cc34ef1a7898e6dc20bc6a512830ba476
        version: 1.1.1
      entries:
      - name: external-dns-operator.v1.1.1
        version: 1.1.1
      - name: external-dns-operator.v1.1.0
        version: 1.1.0
      name: stable-v1.1
    - currentCSV: external-dns-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
          createdAt: 2023-11-15T09:24:04
          olm.skipRange: <1.2.0
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-rhel8@sha256:6263a78b2a170e56dfca2173f14dbf14f0a5bddb6ea600bfdef248c7d5f20e4b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:e33874a51971a90917cf437fc6cbeea1da569ba07217f66f1bba96d594f58aed
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:a614ec94c2468e34e129e9ecd4bc0d1895eb67b96e77eb5ad94cd8bd507860f5
        - registry.redhat.io/edo/external-dns-rhel8-operator@sha256:15700d4acb4705ec900c03b77fa50eaa7db9e59f33bbc819e4fdfc0b9f362f82
        version: 1.2.0
      entries:
      - name: external-dns-operator.v1.2.0
        version: 1.2.0
      name: stable-v1.2
    - currentCSV: external-dns-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1alpha1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-aws"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "testextdnsoperator.apacshift.support"
                    }
                  ],
                  "provider": {
                    "type": "AWS"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "Z04015592QJX3EK1YYYYY"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-azure"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-azure1.qe.azure.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "Azure"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "/subscriptions/53b4f551-f0fc-4bea-8cba-11111111111/resourceGroups/test-azure1-nxkxm-rg/providers/Microsoft.Network/dnszones/test-azure1.qe.azure.devcluster.openshift.com"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-bluecat"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "blueCat": {
                      "configFile": {
                        "name": "bluecat-config"
                      }
                    },
                    "type": "BlueCat"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-gcp"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "test-gcp1.qe.gcp.devcluster.openshift.com"
                    }
                  ],
                  "provider": {
                    "type": "GCP"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  },
                  "zones": [
                    "test-gcp1-q6m5v-private-zone"
                  ]
                }
              },
              {
                "apiVersion": "externaldns.olm.openshift.io/v1beta1",
                "kind": "ExternalDNS",
                "metadata": {
                  "name": "sample-infoblox"
                },
                "spec": {
                  "domains": [
                    {
                      "filterType": "Include",
                      "matchType": "Exact",
                      "name": "myzonedomain.com"
                    }
                  ],
                  "provider": {
                    "infoblox": {
                      "credentials": {
                        "name": "infoblox-credentials"
                      },
                      "gridHost": "100.100.100.100",
                      "wapiPort": 443,
                      "wapiVersion": "2.3.1"
                    },
                    "type": "Infoblox"
                  },
                  "source": {
                    "openshiftRouteOptions": {
                      "routerName": "default"
                    },
                    "type": "OpenShiftRoute"
                  }
                }
              }
            ]
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
          createdAt: 2025-09-08T22:55:39
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.3.1
          operatorframework.io/suggested-namespace: external-dns-operator
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/external-dns-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1alpha1
          - description: ExternalDNS describes a managed ExternalDNS controller instance
              for a cluster. The controller is responsible for creating external DNS
              records in supported DNS providers based off of instances of select
              Kubernetes resources.
            displayName: External DNS
            kind: ExternalDNS
            name: externaldnses.externaldns.olm.openshift.io
            version: v1beta1
        description: |-
          The ExternalDNS Operator deploys and manages ExternalDNS, which dynamically manages DNS records in external DNS Providers for specific Kubernetes resources.

          ## How it works
           Follow this link to get an idea of how ExternalDNS Operator works: [flow diagram](https://raw.githubusercontent.com/openshift/external-dns-operator/main/docs/images/external-dns-flow-openshift.png).
        displayName: ExternalDNS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - openshift
        - dns
        - externaldns
        - operator
        links:
        - name: Source code
          url: https://github.com/openshift/external-dns-operator
        maturity: beta
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/edo/external-dns-operator-bundle@sha256:7299850b4a8257303836c8a28d6e4f3cdc29493148e01ad57d7775e4bb32d56c
        - registry.redhat.io/edo/external-dns-rhel9-operator@sha256:1cb36d6418297d9b5fc32ef2157f80697cf0a84008088f1f8860e2a63d5cb723
        - registry.redhat.io/edo/external-dns-rhel9@sha256:591de53fdaae236c8d384deb6c927f4f397bf44ce4580bbc70b57c112cf22315
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:471549d9cc07fcec0f285a05745b3d39a48e9a503b01b4047156a86ec1a30c00
        version: 1.3.1
      entries:
      - name: external-dns-operator.v1.3.1
        version: 1.3.1
      - name: external-dns-operator.v1.3.0
        version: 1.3.0
      name: stable-v1.3
    defaultChannel: stable-v1
    packageName: external-dns-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Net AI
      provider-url: https://netai.tech
    name: netai-demo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netai-demo-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deployments.netai.tech/v1alpha1",
                "kind": "AntennaDeployment",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netai-demo-operator"
                  },
                  "name": "antennadeployment-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, Networking
          com.redhat.openshift.versions: v4.18
          containerImage: ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
          createdAt: "2025-07-09T11:33:08Z"
          description: Net AI Operator that uses EnergAIze engine forecasts to scale
            Red Hat OpenShift resources.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Net AI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AntennaDeployment is the Schema for the antennadeployments
              API
            displayName: Antenna Deployment
            kind: AntennaDeployment
            name: antennadeployments.deployments.netai.tech
            version: v1alpha1
          - description: Demo is the Schema for the Demo API
            displayName: Demo
            kind: Demo
            name: demos.deployments.netai.tech
            version: v1alpha1
        description: POC Operator that uses saved EnergAize forecasts to manage OpenShift
          Custom Resources
        displayName: netai-demo-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - energaize
        - scaling
        - scale
        - manage
        - forecast
        - cells
        - 5g
        - networking
        - net
        - ai
        - network
        - telecommunication
        - xapp
        links:
        - name: Netai Demo Operator
          url: https://netai-demo-operator.domain
        maturity: alpha
        minKubeVersion: 1.27.0
        provider:
          name: Net AI
          url: https://netai.tech
        relatedImages:
        - registry.access.redhat.com/ubi8/ubi@sha256:68fecea0d255ee253acbf0c860eaebb7017ef5ef007c25bee9eeffd29ce85b29
        - registry.connect.redhat.com/net-ai/demo-operator-bundle@sha256:b1a30de1e984da0487a7d042bd940876bedb5a2f43dd73c7895e684a3a35d81e
        - ghcr.io/net-ai-tech/netai-demo-operator-bundle@sha256:0783aba7b039a569b49de2a61ef6dc6ff1d1f301d93c26d623b7f652321871c5
        - ghcr.io/net-ai-tech/netai-demo-operator@sha256:fae0da9ec2f7a9fd6cfe0ec3e9dfb3b10bf6feaa964b81f081088297649c8bcb
        version: 0.0.1
      entries:
      - name: netai-demo-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: netai-demo-operator
    provider:
      name: Net AI
      url: https://netai.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetScaler
      provider-url: https://netscaler.com
    name: netscaler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-operator.v3.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercloudcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCpxWithIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalercpxwithingresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "admSettings": {
                    "admIP": "",
                    "bandWidth": 1000,
                    "bandWidthLicense": false,
                    "cpxCores": 1,
                    "licenseEdition": "PLATINUM",
                    "licenseServerIP": "",
                    "licenseServerPort": 27000,
                    "loginSecret": "",
                    "platform": false,
                    "vCPULicense": false
                  },
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 30,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 5563
                    },
                    "transactions": {
                      "enable": false,
                      "port": 5557
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {}
                  },
                  "extraVolumes": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "k8s",
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "namespaceLabels": "",
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 50,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "128m",
                        "memory": "500Mi"
                      }
                    },
                    "routeLabels": "",
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscalerCpx": {
                    "bgpSettings": {
                      "bgpConfig": [
                        {
                          "bgpRouter": {
                            "localAS": 100,
                            "neighbor": [
                              {
                                "ASOriginationInterval": 10,
                                "address": "",
                                "advertisementInterval": 10,
                                "remoteAS": 100
                              }
                            ]
                          }
                        }
                      ],
                      "bgpPort": 179
                    },
                    "cpxBgpRouter": false,
                    "cpxCommands": [],
                    "cpxLicenseAggregator": "",
                    "cpxShellCommands": [],
                    "daemonSet": false,
                    "enableLivenessProbe": true,
                    "enableStartupProbe": true,
                    "extraVolumeMounts": [],
                    "hostname": "",
                    "image": "quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8",
                    "ingressIP": "",
                    "livenessProbe": {
                      "failureThreshold": 3,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    },
                    "mgmtHttpPort": 9080,
                    "mgmtHttpsPort": 9443,
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsEnableLabels": true,
                    "nsGateway": "192.168.1.1",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "192.168.1.2",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 500,
                      "required": false
                    },
                    "nsProtocol": "HTTP",
                    "nsSvcLbDnsRec": false,
                    "prometheusCredentialSecret": "",
                    "pullPolicy": "IfNotPresent",
                    "service": {
                      "annotations": {},
                      "spec": {}
                    },
                    "startupProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 5,
                      "successThreshold": 1
                    }
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "replicaCount": 1,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalergslbcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "gslbController": {
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteMask": "",
                        "siteName": "",
                        "sitePublicIp": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "imagePullSecrets": [],
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIngressController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleringresscontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "analyticsConfig": {
                    "distributedTracing": {
                      "enable": false,
                      "samplingRate": 100
                    },
                    "endpoint": {
                      "metrics": {
                        "service": ""
                      },
                      "transactions": {
                        "service": ""
                      }
                    },
                    "required": false,
                    "timeseries": {
                      "auditLogs": {
                        "enable": false
                      },
                      "events": {
                        "enable": false
                      },
                      "metrics": {
                        "enable": false,
                        "enableNativeScrape": false,
                        "exportFrequency": 50,
                        "mode": "avro",
                        "schemaFile": "schema.json"
                      },
                      "port": 50002
                    },
                    "transactions": {
                      "enable": false,
                      "port": 50005
                    }
                  },
                  "exporter": {
                    "extraVolumeMounts": [],
                    "image": "quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4",
                    "ports": {
                      "containerPort": 8888
                    },
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {},
                    "serviceMonitorExtraLabels": {}
                  },
                  "extraVolumes": [],
                  "hostAliases": [],
                  "imagePullSecrets": [],
                  "ingressController": {
                    "clusterName": "",
                    "defaultSSLCertSecret": "",
                    "defaultSSLSNICertSecret": "",
                    "disableAPIServerCertVerify": false,
                    "disableOpenshiftRoutes": true,
                    "enableLivenessProbe": true,
                    "enableReadinessProbe": true,
                    "entityPrefix": "",
                    "extraVolumeMounts": [],
                    "ignoreNodeExternalIP": false,
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f",
                    "ingressClass": [],
                    "ipam": false,
                    "jsonLog": false,
                    "kubernetesURL": "",
                    "livenessProbe": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "logLevel": "INFO",
                    "logProxy": "",
                    "multiClusterPrefix": "",
                    "namespaceLabels": "",
                    "nodeLabels": "",
                    "nodeWatch": false,
                    "nsValidateCert": {},
                    "nsncPbr": false,
                    "openshift": true,
                    "optimizeEndpointBinding": false,
                    "podIPsforServiceGroupMembers": false,
                    "profileHttpFrontend": {},
                    "profileSslFrontend": {},
                    "profileTcpFrontend": {},
                    "pullPolicy": "IfNotPresent",
                    "rbacRole": false,
                    "readinessProbe": {
                      "failureThreshold": 3,
                      "initialDelaySeconds": 10,
                      "periodSeconds": 60,
                      "successThreshold": 1
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "32m",
                        "memory": "128Mi"
                      }
                    },
                    "routeLabels": "",
                    "serviceClass": [],
                    "setAsDefaultIngressClass": false,
                    "updateIngressStatus": true
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsConfigDnsRec": false,
                    "nsCookieVersion": 0,
                    "nsDnsNameserver": "",
                    "nsHttp2ServerSide": "OFF",
                    "nsIP": "",
                    "nsLbHashAlgo": {
                      "hashAlgorithm": "default",
                      "hashFingers": 256,
                      "required": false
                    },
                    "nsPort": 443,
                    "nsProtocol": "HTTPS",
                    "nsSNIPs": [],
                    "nsSvcLbDnsRec": false,
                    "nsVIP": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerIpamController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscaleripamcontroller-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "clusterName": "",
                  "image": "quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8",
                  "localIPAM": {
                    "vipRange": ""
                  },
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "reuseIngressVIP": true,
                  "tokenExpirationSeconds": 31536000,
                  "tolerations": []
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerObservabilityExporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscaler-operator"
                  },
                  "name": "netscalerobservabilityexporter-sample"
                },
                "spec": {
                  "acceptLicense": "No",
                  "affinity": {},
                  "elasticsearch": {
                    "enabled": false,
                    "server": "http://elasticsearch:9200"
                  },
                  "headless": false,
                  "image": "quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4",
                  "jsonTransRateLimiting": {
                    "enabled": false,
                    "limit": 100,
                    "queueLimit": 1000,
                    "window": 5
                  },
                  "kafka": {
                    "auditLogs": "no",
                    "broker": "",
                    "dataFormat": "AVRO",
                    "enabled": false,
                    "events": "no",
                    "topic": "HTTP"
                  },
                  "nodePortRequired": false,
                  "nodeSelector": {},
                  "nsoeLogLevel": "INFO",
                  "nstracing": {
                    "enabled": false,
                    "server": "zipkin:9411/api/v1/spans"
                  },
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "replicas": 1,
                  "resources": {},
                  "splunk": {
                    "authToken": "",
                    "enabled": false,
                    "indexPrefix": "adc_nsoe",
                    "server": "http://splunk:8088"
                  },
                  "timeseries": {
                    "enabled": false
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
          createdAt: "2025-04-15T08:51:39Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler CPX with Ingress Controller.
            displayName: Netscaler CPX With Ingress Controller
            kind: NetscalerCpxWithIngressController
            name: netscalercpxwithingresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Ingress Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler Ingress Controller
            kind: NetscalerIngressController
            name: netscaleringresscontrollers.netscaler.com
            version: v1
          - description: To install NetScaler IPAM Controller for NetScaler CPX/VPX/MPX/SDX.
            displayName: Netscaler IPAM Controller
            kind: NetscalerIpamController
            name: netscaleripamcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler Observability Exporter for NetScaler.
            displayName: Netscaler Observability Exporter
            kind: NetscalerObservabilityExporter
            name: netscalerobservabilityexporters.netscaler.com
            version: v1
        description: |
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/unified-ingress-deployment)
          3. [NetScaler CPX with Ingress Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/two-tier-deployment)
          4. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/gslb/gslb)
          5. [NetScaler Observability Exporter](https://docs.netscaler.com/en-us/netscaler-observability-exporter/deploy-coe)
          6. [NetScaler IPAM Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/ipam-installation.html)

          ### Installation
          Refer [installation instructions](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift).

          This operator version contains:
          1. NetScaler Ingress Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          2. NetScaler CPX version [14.1-25.111](https://docs.netscaler.com/en-us/updates?product=NetScaler-ADC-%28includes-NetScaler-Gateway%29&version=14.1&build=25.56)
          3. NetScaler GSLB Controller version [2.2.10](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.2.10)
          4. NetScaler Observability Exporter version [1.10.001](https://github.com/citrix/citrix-helm-charts/releases/tag/nsoe-v1.10.001)
          5. NetScaler Metrics Exporter [1.4.9](https://github.com/citrix/citrix-adc-metrics-exporter/releases/tag/1.4.9)
          6. NetScaler IPAM Controller [2.1.2](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v2.3.15)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-operator-openshift
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cpx@sha256:232596eaacda68c0e2e210f26b64cedc8d70ae12c5e8ae2e1a66896a479ae4f8
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:42102a2e462e27dd1de28ca3f7db02d227c44736f2d57c05037a83ecbff6dd9f
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:e367c8c8ba6171e9ef5dd496592a5c62f4aa764f93d0cb78835aa29024c063d0
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:eca5a7e33260b1e340b044cf2cfaab57bffd77650278d50a5dbb31b64fab8a6f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:89f44c45373da73a51b0b9a02de7efe519dc2d5343b65fe382526cccc92f8f5c
        - quay.io/netscaler/netscaler-ipam-controller@sha256:7c25b59a2e1df0e4f5615a779a9bea939c60a3eca065cb6c2295daf0d03831d8
        - quay.io/netscaler/netscaler-observability-exporter@sha256:626e3a954c67bbb910477352dc740aaab21ff1c1cfb03354becf50beab17d4a4
        - quay.io/netscaler/netscaler-adc-metrics-exporter@sha256:9ffaa5d961dc2083723c6846e250bbb8beea7a1ebb75dce4f5810e28256930f4
        version: 3.2.0
      entries:
      - name: netscaler-operator.v3.2.0
        version: 3.2.0
      - name: netscaler-operator.v3.1.2
        version: 3.1.2
      - name: netscaler-operator.v3.1.1
        version: 3.1.1
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: stable
    - currentCSV: netscaler-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerCloudController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nscc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalercloudcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nscc"
                },
                "spec": {
                  "azure": {
                    "clientsecret": "",
                    "image": "quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb",
                    "location": "",
                    "pullPolicy": "IfNotPresent",
                    "resourcegroupname": "",
                    "resources": {},
                    "subnetname": "",
                    "subscriptionid": "",
                    "tenantid": "",
                    "vnetname": ""
                  }
                }
              },
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "NetscalerGslbController",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netscaler-operator",
                    "app.kubernetes.io/instance": "nsgc",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "netscalergslbcontroller",
                    "app.kubernetes.io/part-of": "netscaler-operator"
                  },
                  "name": "nsgc"
                },
                "spec": {
                  "gslbController": {
                    "acceptLicense": "Yes",
                    "cleanupGSLBSiteConfig": false,
                    "disableAPIServerCertVerify": false,
                    "entityPrefix": "",
                    "image": "quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f",
                    "imagePullSecrets": [],
                    "kubernetesURL": "",
                    "localCluster": "",
                    "localRegion": "",
                    "logLevel": "Debug",
                    "openshift": true,
                    "pullPolicy": "IfNotPresent",
                    "siteData": [
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      },
                      {
                        "secretName": "",
                        "siteIp": "",
                        "siteName": "",
                        "siteRegion": ""
                      }
                    ]
                  },
                  "netscaler": {
                    "adcCredentialSecret": "",
                    "nitroReadTimeout": 20,
                    "nsIP": "",
                    "nsPort": 443,
                    "nsProtocol": "HTTPS"
                  },
                  "resources": {
                    "limits": {},
                    "requests": {
                      "cpu": "32m",
                      "memory": "128Mi"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
          createdAt: "2024-07-31T06:33:49Z"
          description: This is an operator to install Cloud Native portfolio provided
            by NetScaler
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler Cloud Controller for NetScaler VPX/MPX/SDX
              with Azure Redhat OpenShift in Azure.
            displayName: Netscaler Cloud Controller
            kind: NetscalerCloudController
            name: netscalercloudcontrollers.netscaler.com
            version: v1
          - description: To install NetScaler GSLB Controller for NetScaler VPX/MPX/SDX.
            displayName: Netscaler GSLB Controller
            kind: NetscalerGslbController
            name: netscalergslbcontrollers.netscaler.com
            version: v1
        description: |-
          NetScaler provides various products to empower Developer/DevOps managing your applications in OpenShift Cluster. Using this operator you can deploy below controllers for NetScaler appliance:
          1. [NetScaler Cloud Controller](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. [NetScaler GSLB Controller ](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator)

          ### Installation
          Refer [installation instructions](https://github.com/netscaler/netscaler-k8s-ingress-controller/blob/master/docs/deploy/deploy-ns-operator.md).

          This operator version contains:
          1. NetScaler Cloud Controller version [1.2.0](https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/netscaler-cloud-controller)
          2. NetScaler GSLB Controller version [1.43.7](https://github.com/netscaler/netscaler-helm-charts/releases/tag/nsic-v1.43.7)

          ### Support

          For debugging issues, please use [NetScaler Kubernetes Plugin](https://github.com/netscaler/modern-apps-toolkit/tree/main/netscaler-plugin).

          For questions and support, the following channels are available:
          - [NetScaler community portal](https://community.netscaler.com/s/)
          - [Contact us](https://www.netscaler.com/contact)
        displayName: NetScaler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        links:
        - name: Netscaler Operator
          url: https://docs.netscaler.com/en-us/netscaler-k8s-ingress-controller/deploy/gslb-openshift-operator
        maintainers:
        - email: priyanka.sharma@cloud.com
          name: Priyanka Sharma
        - email: arijit.ray@cloud.com
          name: Arijit Ray
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: NetScaler
          url: https://netscaler.com
        relatedImages:
        - quay.io/netscaler/netscaler-cloud-controller@sha256:6897df79411bd1b4ddb5195d570f10e01bd3c084772b918112797a346980f4cb
        - quay.io/netscaler/netscaler-k8s-ingress-controller@sha256:88114c518ecceb54b000f38bfc330efe184bfeb5b5df0ee308950cf29e7a777f
        - registry.connect.redhat.com/citrix/netscaler-bundle@sha256:b405d6148aa680bf31bdc6185c1702916c3219983d8482f27d89ed2f93659aa7
        - registry.connect.redhat.com/citrix/netscaler-operator@sha256:3f71f9a46d7b29e3da98a301965f635c03a022b5535b586d7d5c699341f68193
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:23405618050a4ceea8ea0c748cc7d9d16dbb80f10e4c39f7e82014a3ac80cec9
        version: 3.0.1
      entries:
      - name: netscaler-operator.v3.0.1
        version: 3.0.1
      name: tech-preview
    defaultChannel: stable
    packageName: netscaler-operator
    provider:
      name: NetScaler
      url: https://netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-recyclebin-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-recyclebin-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "recyclebin.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
          createdAt: "2025-09-20T01:20:03Z"
          description: AWS Recycle Bin controller is a service controller for managing
            Recycle Bin resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rule represents the state of an AWS recyclebin Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.recyclebin.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Recycle Bin resources in AWS from within your Kubernetes cluster.

          **About Amazon Recycle Bin**

          Recycle Bin is a data recovery feature that enables you to restore accidentally deleted Amazon EBS snapshots and EBS-backed AMIs. When using Recycle Bin, if your resources are deleted, they are retained in the Recycle Bin for a time period that you specify before being permanently deleted.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Recycle Bin
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recyclebin
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Recycle Bin Developer Resources
          url: https://docs.aws.amazon.com/ebs/latest/userguide/recycle-bin.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: recyclebin maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/recyclebin-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-recyclebin-controller@sha256:9d916ed0af37d60d90f9b4711b1544b9373bd247476278c9c4041c9dbce6acc9
        version: 1.1.1
      entries:
      - name: ack-recyclebin-controller.v1.1.1
        version: 1.1.1
      - name: ack-recyclebin-controller.v1.1.0
        version: 1.1.0
      - name: ack-recyclebin-controller.v1.0.13
        version: 1.0.13
      - name: ack-recyclebin-controller.v1.0.12
        version: 1.0.12
      - name: ack-recyclebin-controller.v1.0.11
        version: 1.0.11
      - name: ack-recyclebin-controller.v1.0.10
        version: 1.0.10
      - name: ack-recyclebin-controller.v1.0.9
        version: 1.0.9
      - name: ack-recyclebin-controller.v1.0.8
        version: 1.0.8
      - name: ack-recyclebin-controller.v1.0.7
        version: 1.0.7
      - name: ack-recyclebin-controller.v1.0.6
        version: 1.0.6
      - name: ack-recyclebin-controller.v1.0.5
        version: 1.0.5
      - name: ack-recyclebin-controller.v1.0.2
        version: 1.0.2
      - name: ack-recyclebin-controller.v1.0.1
        version: 1.0.1
      - name: ack-recyclebin-controller.v1.0.0
        version: 1.0.0
      - name: ack-recyclebin-controller.v0.0.3
        version: 0.0.3
      - name: ack-recyclebin-controller.v0.0.2
        version: 0.0.2
      - name: ack-recyclebin-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-recyclebin-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "Velero",
                "metadata": {
                  "name": "example-velero",
                  "namespace":"openshift-adp"
                },
                "spec": {
                  "backupStorageLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-east-1"
                      },
                      "name": "default",
                      "objectStorage": {
                        "bucket": "myBucket",
                        "prefix": "velero"
                      },
                      "provider": "aws"
                    }
                  ],
                  "defaultVeleroPlugins": [
                    "aws",
                    "csi",
                    "openshift"
                  ],
                  "veleroFeatureFlags": [
                    "EnableCSI"
                  ],
                  "enableRestic": true,
                  "volumeSnapshotLocations": [
                    {
                      "config": {
                        "profile": "default",
                        "region": "us-west-1"
                      },
                      "name": "default",
                      "provider": "aws"
                    }
                  ]
                }
               },
                {
                  "apiVersion": "velero.io/v1",
                  "kind": "Backup",
                  "metadata": {
                    "name": "backup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "BackupStorageLocation",
                  "metadata": {
                    "name": "backupstoragelocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DeleteBackupRequest",
                  "metadata": {
                    "name": "deletebackuprequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "DownloadRequest",
                  "metadata": {
                    "name": "downloadrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeBackup",
                  "metadata": {
                    "name": "podvolumebackup",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "PodVolumeRestore",
                  "metadata": {
                    "name": "podvolumerestore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ResticRepository",
                  "metadata": {
                    "name": "resticrepository",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Restore",
                  "metadata": {
                    "name": "restore",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "Schedule",
                  "metadata": {
                    "name": "schedule",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "ServerStatusRequest",
                  "metadata": {
                    "name": "serverstatusrequest",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              },
              {
                  "apiVersion": "velero.io/v1",
                  "kind": "VolumeSnapshotLocation",
                  "metadata": {
                    "name": "volumesnapshotlocation",
                    "namespace": "openshift-adp"
                  },
                  "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.4.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Velero is a tool to safely backup and restore, perform disaster
              recovery and migrate Kubernetes cluster resources and persistent volumes.
            displayName: Velero
            kind: Velero
            name: veleroes.oadp.openshift.io
            version: v1alpha1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Velero on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-csi@sha256:5c2261effc902d8baa16e38265f9ad1fbe5284a3e6abd2bbc97761c161fa92cf
        - quay.io/konveyor/velero@sha256:e7865cc707664eb6ee767e77ccd933420604c2562736b53e16c8043a51df9d86
        - quay.io/konveyor/openshift-velero-plugin@sha256:aec8b4fc66cfcebd5a35ce99d8c985453ebf25c11453d30256089a8b425990b5
        - quay.io/konveyor/registry@sha256:b74ab3278b79eb89f99b6970c6aa0f5306221a8495b5496faad6cab2f74ff2cd
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c4d12829e42cabe70b662a659875900677dfb415f912f931c3bf4c09a26aed5b
        - quay.io/openshift-community-operators/oadp-operator@sha256:f8c5c6e59f13db8847080ea6d964da6e50e56f7d373bcc4e4278b1a8a80395e4
        - quay.io/konveyor/velero-plugin-for-aws@sha256:2197c6fef32910747526d33c1462b75cb7940aeace65791d0a5ddaa038480e52
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:3300458b1d6c1514dca7d24353ad03bd92db396b2a29f7d3522563160a5f5219
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:62a344e5cf196e55dae92e537648987d758ccc75985bc45e9be6d0646a9e64d0
        - quay.io/konveyor/oadp-operator@sha256:ecb8534fbc245e890b9c7766946d01e1a8b4ce9be359c4ada4c95fd908b679b3
        version: 0.4.2
      entries:
      - name: oadp-operator.v0.4.2
        version: 0.4.2
      - name: oadp-operator.v0.4.1
        version: 0.4.1
      - name: oadp-operator.v0.4.0
        version: 0.4.0
      - name: oadp-operator.v0.3.0
        version: 0.3.0
      name: beta
    - currentCSV: oadp-operator.v0.5.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
          createdAt: "2020-09-08T12:21:00Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          olm.skipRange: '>=0.0.0 <0.5.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that respresents the capture
              of Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://oadp.konveyor.io/
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/konveyor/velero-plugin-for-microsoft-azure@sha256:2d0089cbe715b482fb42e1acd9add70b975a8575636d8b143e6f90cef317ea13
        - quay.io/konveyor/registry@sha256:1be970879206258335357e3cf630138c24b8a5ccab35aae05eccb260e5e8ffb4
        - quay.io/konveyor/velero-plugin-for-csi@sha256:8c36ccfd294dafc415a7f113770cbbd66dbf6ecf8fef3bc4e73adc9878c606b0
        - quay.io/konveyor/velero@sha256:0ea5336dcbf0a1fab561697b24da01a0e5b533ee23553fbfd5f8e002f5a31c3e
        - quay.io/konveyor/oadp-operator@sha256:90609044d52173619ea1e63ddef0ef3e7c1c939ac737b64bd427cbed5374dd27
        - quay.io/konveyor/velero-plugin-for-gcp@sha256:ee463ed2a8019c560961b118f8a87b587ed302f24fd25480c5d4ddad2557e92f
        - quay.io/konveyor/velero-restic-restore-helper@sha256:c16248d027c09cd5b7360ae38484df88ca1cc01f9e470e3eade95076ee1a6b84
        - quay.io/konveyor/openshift-velero-plugin@sha256:221ab0bcd1d95fc166ef7319d187e196b59d2fbf9eefd29d91bd6435d45b5c3c
        - quay.io/konveyor/velero-plugin-for-aws@sha256:1e8bc27d6b60000e1406085d582910cb530af3fb901ef9bc780628df6c766288
        - quay.io/openshift-community-operators/oadp-operator@sha256:b7b4d340157513da2ebef6bc7bc262967d0e825e2baa1f2bc2ab2e170457e015
        version: 0.5.6
      entries:
      - name: oadp-operator.v0.5.6
        version: 0.5.6
      - name: oadp-operator.v0.5.5
        version: 0.5.5
      - name: oadp-operator.v0.5.4
        version: 0.5.4
      - name: oadp-operator.v0.5.3
        version: 0.5.3
      - name: oadp-operator.v0.5.2
        version: 0.5.2
      - name: oadp-operator.v0.5.1
        version: 0.5.1
      - name: oadp-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio Project
      provider-url: ""
    name: apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: 1.0.x
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/apicurio/apicurito-operator:1.0.3
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: candidate
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: latest
    - currentCSV: apicurito.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apicur.io/v1alpha1",
                "kind": "Apicurito",
                "metadata": {
                  "labels": {
                    "app": "apicurito"
                  },
                  "name": "apicurito"
                },
                "spec": {
                  "size": 3
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurito-operator:1.0.3
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Apicurio Project
        relatedImages:
        - quay.io/openshift-community-operators/apicurito@sha256:fa43a6f0259111999fa8d7d13f863929c5bd92e8c2b1bd3871c03aa8eba7884d
        - quay.io/apicurio/apicurito-operator:1.0.3
        version: 1.0.3
      entries:
      - name: apicurito.v1.0.3
        version: 1.0.3
      - name: apicurito.v1.0.2
        version: 1.0.2
      - name: apicurito.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: latest
    packageName: apicurito
    provider:
      name: Apicurio Project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-integration
    name: red-hat-camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.10.x
    - currentCSV: red-hat-camel-k-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1beta1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
          createdAt: "2020-10-06T02:45:00Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:c60664e5da8eb23bbd918b1c13b83565b1f33a1aa8961b425232fa1ef4d1a6f1
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:9e4fb5369e0479d4a68f60725e7d12abdc66c552babe59f1358f6e6b0f26357d
        version: 1.4.2
      entries:
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.4.x
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: 1.6.x
    - currentCSV: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
          createdAt: "2022-11-28T10:43:18Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <1.8.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:82cafd48db8107dd141064ce7530468b11804c8c54235a317eb450a7606bac3e
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:e916f9040d3f4feb26cc1534d3ab2722890b34c7fe146cdce8b6085b47206e22
        version: 1.8.2+0.1675913507.p
      entries:
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: 1.8.x
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
          createdAt: "2024-10-07T13:45:51Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.8.0 <1.10.0'
          olm.substitutesFor: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat Application Foundations", "Red Hat Camel K"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:72ed79d0b4bf82ab58d22a174190f69401da83d80a09785648148760e1803d9d
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:431db7af80d2d6cbaef1d1cb8ebf96d58cd71cba10bc24aeba97b9410da53a8e
        version: 1.10.8+0.1733993735.p
      entries:
      - name: red-hat-camel-k-operator.v1.10.8-0.1731578390.p
        version: 1.10.8+0.1731578390.p
      - name: red-hat-camel-k-operator.v1.10.8
        version: 1.10.8
      - name: red-hat-camel-k-operator.v1.10.8-0.1731485816.p
        version: 1.10.8+0.1731485816.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733135130.p
        version: 1.10.8+0.1733135130.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124835.p
        version: 1.10.8+0.1733124835.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733124325.p
        version: 1.10.8+0.1733124325.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731894055.p
        version: 1.10.8+0.1731894055.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1731893478.p
        version: 1.10.8+0.1731893478.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993120.p
        version: 1.10.8+0.1733993120.p
      - name: red-hat-camel-k-operator.v1.10.8-0.1733993735.p
        version: 1.10.8+0.1733993735.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1723632902.p
        version: 1.10.7+0.1723632902.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724202359.p
        version: 1.10.7+0.1724202359.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724145144.p
        version: 1.10.7+0.1724145144.p
      - name: red-hat-camel-k-operator.v1.10.7-0.1724144664.p
        version: 1.10.7+0.1724144664.p
      - name: red-hat-camel-k-operator.v1.10.7
        version: 1.10.7
      - name: red-hat-camel-k-operator.v1.10.7-0.1723633518.p
        version: 1.10.7+0.1723633518.p
      - name: red-hat-camel-k-operator.v1.10.6
        version: 1.10.6
      - name: red-hat-camel-k-operator.v1.10.5
        version: 1.10.5
      - name: red-hat-camel-k-operator.v1.10.5-0.1705608149.p
        version: 1.10.5+0.1705608149.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1699536334.p
        version: 1.10.4+0.1699536334.p
      - name: red-hat-camel-k-operator.v1.10.4-0.1700072719.p
        version: 1.10.4+0.1700072719.p
      - name: red-hat-camel-k-operator.v1.10.4
        version: 1.10.4
      - name: red-hat-camel-k-operator.v1.10.3
        version: 1.10.3
      - name: red-hat-camel-k-operator.v1.10.2
        version: 1.10.2
      - name: red-hat-camel-k-operator.v1.10.1
        version: 1.10.1
      - name: red-hat-camel-k-operator.v1.10.1-0.1695295065.p
        version: 1.10.1+0.1695295065.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693789561.p
        version: 1.10.1+0.1693789561.p
      - name: red-hat-camel-k-operator.v1.10.1-0.1693378374.p
        version: 1.10.1+0.1693378374.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1682325781.p
        version: 1.10.0+0.1682325781.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1687356936.p
        version: 1.10.0+0.1687356936.p
      - name: red-hat-camel-k-operator.v1.10.0
        version: 1.10.0
      - name: red-hat-camel-k-operator.v1.10.0-0.1679561624.p
        version: 1.10.0+0.1679561624.p
      - name: red-hat-camel-k-operator.v1.10.0-0.1678893647.p
        version: 1.10.0+0.1678893647.p
      - name: red-hat-camel-k-operator.v1.8.2
        version: 1.8.2
      - name: red-hat-camel-k-operator.v1.8.2-0.1673533600.p
        version: 1.8.2+0.1673533600.p
      - name: red-hat-camel-k-operator.v1.8.2-0.1675913507.p
        version: 1.8.2+0.1675913507.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1669858726.p
        version: 1.8.1+0.1669858726.p
      - name: red-hat-camel-k-operator.v1.8.1-0.1667495781.p
        version: 1.8.1+0.1667495781.p
      - name: red-hat-camel-k-operator.v1.8.0
        version: 1.8.0
      - name: red-hat-camel-k-operator.v1.6.7
        version: 1.6.7
      - name: red-hat-camel-k-operator.v1.6.6
        version: 1.6.6
      - name: red-hat-camel-k-operator.v1.6.5
        version: 1.6.5
      - name: red-hat-camel-k-operator.v1.6.4-0.1648537022.p
        version: 1.6.4+0.1648537022.p
      - name: red-hat-camel-k-operator.v1.6.4
        version: 1.6.4
      - name: red-hat-camel-k-operator.v1.6.3
        version: 1.6.3
      - name: red-hat-camel-k-operator.v1.6.2
        version: 1.6.2
      - name: red-hat-camel-k-operator.v1.6.1
        version: 1.6.1
      - name: red-hat-camel-k-operator.v1.6.0
        version: 1.6.0
      - name: red-hat-camel-k-operator.v1.4.2
        version: 1.4.2
      - name: red-hat-camel-k-operator.v1.4.1
        version: 1.4.1
      - name: red-hat-camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: red-hat-camel-k-operator.v1.6.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
          createdAt: "2022-08-15T20:17:25Z"
          description: Red Hat Integration - Camel K is a lightweight integration
            platform, born on Kubernetes, with serverless superpowers.
          olm.skipRange: '>=1.6.7 <=1.6.8'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Red Hat Integration - Camel K
          ==============

          **Red Hat Integration - Camel K** is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Red Hat Integration - Camel K
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Apache Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        maturity: alpha
        minKubeVersion: 1.14.0
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-integration
        relatedImages:
        - registry.redhat.io/integration/camel-k-rhel8-operator@sha256:244e22723124ac6e13a417a94396383f872f15d6f772c80bbd3693a148657526
        - registry.redhat.io/integration/camel-k-rhel8-operator-bundle@sha256:bcc0ef48ec42c88d212173167615feb2cf276c133b26b866ae3453af7ab1ee80
        version: 1.6.10
      entries:
      - name: red-hat-camel-k-operator.v1.6.10
        version: 1.6.10
      - name: red-hat-camel-k-operator.v1.6.9
        version: 1.6.9
      name: stable
    defaultChannel: 1.10.x
    packageName: red-hat-camel-k
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.3.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
          createdAt: "2019-05-01T00:00:00Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/kubeturbo-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — KubeTurbo — that runs in your Kubernetes or Red Hat OpenShift cluster to discover and monitor your environment.
          KubeTurbo runs together with the default scheduler and sends this data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Kubeturbo Operator
          url: https://github.com/turbonomic/kubeturbo/tree/master/deploy/kubeturbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:5d9b56300a3b17178467954b533bee19622f8e96fe9d4ac9745ead2c59b9ec5f
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo-operator@sha256:32178a16dd2115e0a11a2e3dfbfc130ae1d0afd8e293131e76b6fc1a79b36c12
        - registry.marketplace.redhat.com/rhm/turbonomic/kubeturbo@sha256:671be694176d9cc31b8a22c11bbfefc2504ae72d27fd010b315b6839f9ab59d5
        version: 8.3.3
      entries:
      - name: kubeturbo-operator.v8.3.3
        version: 8.3.3
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-efs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-efs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "AccessPoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "FileSystem",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "efs.services.k8s.aws/v1alpha1",
                "kind": "MountTarget",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
          createdAt: "2025-09-19T18:38:13Z"
          description: AWS EFS controller is a service controller for managing EFS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS efs AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.efs.services.k8s.aws
            version: v1alpha1
          - description: FileSystem represents the state of an AWS efs FileSystem
              resource.
            displayName: FileSystem
            kind: FileSystem
            name: filesystems.efs.services.k8s.aws
            version: v1alpha1
          - description: MountTarget represents the state of an AWS efs MountTarget
              resource.
            displayName: MountTarget
            kind: MountTarget
            name: mounttargets.efs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EFS resources in AWS from within your Kubernetes cluster.

          **About Amazon Elastic File System (Amazon EFS)**

          Amazon Elastic File System (Amazon EFS) provides serverless, fully elastic file storage so that you can share file data without provisioning or managing storage capacity and performance. Amazon EFS is built to scale on demand to petabytes without disrupting applications, growing and shrinking automatically as you add and remove files. Because Amazon EFS has a simple web services interface, you can create and configure file systems quickly and easily. The service manages all the file storage infrastructure for you, meaning that you can avoid the complexity of deploying, patching, and maintaining complex file system configurations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Elastic File System (Amazon
          EFS)
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - efs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EFS Developer Resources
          url: https://aws.amazon.com/efs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: efs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/efs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-efs-controller@sha256:338a28600b08d1ead95ed409e3326e03f3c5e02ed840d3be56a950db6b1de061
        version: 1.1.1
      entries:
      - name: ack-efs-controller.v1.1.1
        version: 1.1.1
      - name: ack-efs-controller.v1.1.0
        version: 1.1.0
      - name: ack-efs-controller.v1.0.19
        version: 1.0.19
      - name: ack-efs-controller.v1.0.18
        version: 1.0.18
      - name: ack-efs-controller.v1.0.17
        version: 1.0.17
      - name: ack-efs-controller.v1.0.16
        version: 1.0.16
      - name: ack-efs-controller.v1.0.15
        version: 1.0.15
      - name: ack-efs-controller.v1.0.14
        version: 1.0.14
      - name: ack-efs-controller.v1.0.13
        version: 1.0.13
      - name: ack-efs-controller.v1.0.12
        version: 1.0.12
      - name: ack-efs-controller.v1.0.11
        version: 1.0.11
      - name: ack-efs-controller.v1.0.10
        version: 1.0.10
      - name: ack-efs-controller.v1.0.8
        version: 1.0.8
      - name: ack-efs-controller.v1.0.6
        version: 1.0.6
      - name: ack-efs-controller.v1.0.5
        version: 1.0.5
      - name: ack-efs-controller.v1.0.3
        version: 1.0.3
      - name: ack-efs-controller.v1.0.2
        version: 1.0.2
      - name: ack-efs-controller.v1.0.0
        version: 1.0.0
      - name: ack-efs-controller.v0.0.9
        version: 0.0.9
      - name: ack-efs-controller.v0.0.8
        version: 0.0.8
      - name: ack-efs-controller.v0.0.7
        version: 0.0.7
      - name: ack-efs-controller.v0.0.6
        version: 0.0.6
      - name: ack-efs-controller.v0.0.5
        version: 0.0.5
      - name: ack-efs-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-efs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Protocol Labs
      provider-url: ""
    name: ipfs-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ipfs-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "CircuitRelay",
                "metadata": {
                  "name": "circuitrelay-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "cluster.ipfs.io/v1alpha1",
                "kind": "IpfsCluster",
                "metadata": {
                  "name": "ipfscluster-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage, Networking, Big Data
          containerImage: quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
          createdAt: 2022-04-05T15:50:37-0400
          operatorframework.io/suggested-namespace: ipfs-operator-system
          operators.operatorframework.io/builder: operator-sdk-v1.22.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ipfs-cluster/ipfs-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CircuitRelay is the Schema for the circuitrelays API.
            displayName: Circuit Relay
            kind: CircuitRelay
            name: circuitrelays.cluster.ipfs.io
            version: v1alpha1
          - description: IpfsCluster is the Schema for the ipfs API.
            displayName: Ipfs Cluster
            kind: IpfsCluster
            name: ipfsclusters.cluster.ipfs.io
            version: v1alpha1
        description: "The IPFS Operator enables users to easily create IPFS Cluster
          instances\nwithin their Kubernetes or OpenShift clusters. \n\n### Supported
          Features\n* Horizontal scaling\n* Joining collaborative clusters\n* Controlling
          IPFS node resource usage\n* Specifying reprovider settings  \n\n### Documentation
          \nDocumentation can be found on our [read-the-docs page](https://ipfs-operator.readthedocs.io/en/latest/),
          as well as our [GitHub repository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Getting Help\nIn case you run into issues while using the operator, please
          create an issue in our [GitHub\nrepository](https://github.com/ipfs-cluster/ipfs-operator).\n\n###
          Contributing\nWe welcome contributions of all types and sizes! You can find
          open issues\nin our [GitHub Repository](https://github.com/ipfs-cluster/ipfs-operator/issues),\nas
          well as make your own issue if you feel that something is missing.\n\n###
          License\nIPFS Operator is licensed under the [Apache 2.0 License](https://github.com/ipfs-cluster/ipfs-operator/blob/main/LICENSE)"
        displayName: IPFS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ipfs
        - ipfs cluster
        - decentralized storage
        - peer-to-peer
        - distributed hash table
        - content-addressed
        links:
        - name: IPFS Operator GitHub page
          url: https://github.com/ipfs-cluster/ipfs-operator
        - name: IPFS Cluster
          url: https://ipfscluster.io/
        - name: IPFS
          url: https://ipfs.io/
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Protocol Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/ipfs-operator@sha256:03a61ef328feb626408d42904cdfb9d45f66e022d5a1e1088a15b74455487743
        - quay.io/redhat-et-ipfs/ipfs-operator:v0.0.1
        version: 0.0.2
      entries:
      - name: ipfs-operator.v0.0.2
        version: 0.0.2
      - name: ipfs-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ipfs-operator
    provider:
      name: Protocol Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Yugabyte Inc
      provider-url: ""
    name: yugabyte-platform-operator-bundle-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: yugabyte-platform-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "yugabyte.com/v1alpha1",
                "kind": "YBPlatform",
                "metadata": {
                  "name": "ybplatform-sample"
                },
                "spec": {
                  "domainName": "cluster.local",
                  "helm": {
                    "package": "/opt/yugabyte/helm/yugabyte-latest.tgz",
                    "timeout": 900
                  },
                  "helm2Legacy": false,
                  "image": {
                    "commonRegistry": "",
                    "nginx": {
                      "name": "nginxinc/nginx-unprivileged",
                      "registry": "",
                      "tag": "1.17.4-amd64"
                    },
                    "postgres": {
                      "name": "postgres",
                      "registry": "",
                      "tag": 11.5
                    },
                    "prometheus": {
                      "name": "prom/prometheus",
                      "registry": "",
                      "tag": "v2.2.1"
                    },
                    "pullPolicy": "IfNotPresent",
                    "pullSecret": "yugabyte-k8s-pull-secret",
                    "repository": "registry.connect.redhat.com/yugabytedb/yugabyte-platform",
                    "tag": "latest",
                    "thirdparty-deps": {
                      "name": "yugabyte/thirdparty-deps",
                      "registry": "quay.io",
                      "tag": "latest"
                    }
                  },
                  "ip_version_support": "v4_only",
                  "ocpCompatibility": {
                    "enabled": true
                  },
                  "rbac": {
                    "create": true
                  },
                  "securityContext": {
                    "enabled": false,
                    "fsGroup": 1001,
                    "runAsUser": 1001
                  },
                  "tls": {
                    "certificate": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tDQpNSUlHTHpDQ0JSZWdBd0lCQWdJUUQ2R1ZmL0ViUDhSMFQvcWxjRGZyM3pBTkJna3Foa2lHOXcwQkFRc0ZBREJlDQpNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzDQpkM2N1WkdsbmFXTmxjblF1WTI5dE1SMHdHd1lEVlFRREV4UlNZWEJwWkZOVFRDQlNVMEVnUTBFZ01qQXhPREFlDQpGdzB4T1RBNE1ERXdNREF3TURCYUZ3MHlNVEEzTXpFeE1qQXdNREJhTUI4eEhUQWJCZ05WQkFNTUZDb3VZMnh2DQpkV1F1ZVhWbllXSjVkR1V1WTI5dE1JSUJJakFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBDQp3SmswU25ub1pRYW1hWm0rMVVJdjdRYjBLalhZeVEwckpJSGpsbE03elhFYUFoL0JqVjZ1M0h1TWxrTlZSekVFDQpGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0dTcDJIMGpKbVhuRWNvMFpzDQpwYlVyYWhlcWlxSUUyZ1gwQlpvYldvWGRHc1EydXJiUGxyMGlVSnUxVnp4SGRoTS9nTnFQWEN5bE5YdENrbVI2DQo0eXpIa0Rob21pRTAySTMwTXBlOERxczl4d3lKd3F6T3J2ZnA0eUE0Ukhna2g1clovK1Erc0VQNkIyVzBHQXc2DQpzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVBBTFJzSFh1RW1kTkN5N2pvDQpibzJuUWFObzZBQ1c0U2s5RWJVZXB3SURBUUFCbzRJREpqQ0NBeUl3SHdZRFZSMGpCQmd3Rm9BVVU4b1hXZnhyDQp3QU1oTHhxdTVLcW9ISUpXMm5Vd0hRWURWUjBPQkJZRUZHemw3bjdOMy9GbFdDSDFvdWFCWldKMjI1UStNQjhHDQpBMVVkRVFRWU1CYUNGQ291WTJ4dmRXUXVlWFZuWVdKNWRHVXVZMjl0TUE0R0ExVWREd0VCL3dRRUF3SUZvREFkDQpCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3UGdZRFZSMGZCRGN3TlRBem9ER2dMNFl0DQphSFIwY0RvdkwyTmtjQzV5WVhCcFpITnpiQzVqYjIwdlVtRndhV1JUVTB4U1UwRkRRVEl3TVRndVkzSnNNRXdHDQpBMVVkSUFSRk1FTXdOd1lKWUlaSUFZYjliQUVDTUNvd0tBWUlLd1lCQlFVSEFnRVdIR2gwZEhCek9pOHZkM2QzDQpMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDQVlHWjRFTUFRSUJNSFVHQ0NzR0FRVUZCd0VCQkdrd1p6QW1CZ2dyDQpCZ0VGQlFjd0FZWWFhSFIwY0RvdkwzTjBZWFIxY3k1eVlYQnBaSE56YkM1amIyMHdQUVlJS3dZQkJRVUhNQUtHDQpNV2gwZEhBNkx5OWpZV05sY25SekxuSmhjR2xrYzNOc0xtTnZiUzlTWVhCcFpGTlRURkpUUVVOQk1qQXhPQzVqDQpjblF3Q1FZRFZSMFRCQUl3QURDQ0FYNEdDaXNHQVFRQjFua0NCQUlFZ2dGdUJJSUJhZ0ZvQUhVQTdrdTl0M1hPDQpZTHJoUW1rZnErR2VacU1QZmwrd2N0aURBTVI3aVhxby9jc0FBQUZzVHEwL013QUFCQU1BUmpCRUFpQlg3NXJ0DQpwNkdQbUQxbVl1S1hvcXVRVytIcUtRNE8rV1RXbTNCbVBteWF1d0lnQUoya01WdmduaUx1dEh5Y3AyYzBwdDRsDQptL3M5SWN2Yi9aN09jMU82b2xJQWRnQ0hkYi9uV1h6NGpFT1pYNzN6YnY5V2pVZFdOdjlLdFdEQnRPci9YcUNEDQpEd0FBQVd4T3JUOXNBQUFFQXdCSE1FVUNJSHphY3NCcmpEN1FER1ZRSHA5YUZPMnJ4VEx4NU9GOVE2MTFYaUU2DQpZUEk3QWlFQW9xd3pTd1BVaGt4bVhYNzN0SDVEcVNwaUxuOHdrMXFWeWhZOTVmSmxZa0VBZHdCRWxHVXVzTzdPDQpyOFJBQjlpby9pakEydWFDdnRqTE1iVS8wek9XdGJhQnFBQUFBV3hPclQ1M0FBQUVBd0JJTUVZQ0lRQ1ZLaHQ5DQpWTWhlNTUvZEwrSGN6Qzk5a2xheG9DcW5zOStRNGVOTDZ2Smd4QUloQUsyZmp4a1RadGw4WFZ4WHEwc2NSTWNJDQpiSktHcGJYdEV0N3dFVzRkZDNGS01BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQm1GcDNCK1g2Y21Cb0FLWkhrDQpIWWxzanhIZk01bDQyZWJtSFdGdlQyTi9aVkZ3MEZ3YndBNExyNzlvd01QdU5FbVNNa2pmVVhxR011Q2VlVTdODQo0MlRwUFdPaGNzVHF6ZkxVa0hhMU9rK011L3RUZHBNcmhROEVqbkhMY3FKVE9MbmRSVy8zcE5IM2xrK2FJejNoDQpDVGFueElRVEE2cUxRNThRakszWmFMVzdqWVVNcVJFcnBGVDlBVXkxMDAxMjlOam5HVkhtYmU1bW1ZbEwvQlhMDQpmdHBJUUJJZnhCNXpHbHpTcjRHT3p5RDBES05YU040RU0yM2pIM1ljeUg4eFQrODU4ZkxSaXQ4SFlOQ3RQbkloDQoyRzR6UjV3emZFVDlnZUh6aWtuY0ZMWUpMMEpKMVIrZGVNRE91bmJWOUJqcmZJb3NNSk53TUV2ejFZczlBWXhHDQpxSFNYDQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tDQotLS0tLUJFR0lOIENFUlRJRklDQVRFLS0tLS0KTUlJRXNUQ0NBNW1nQXdJQkFnSVFDS1dpUnMxTFhJeUQxd0swdTZ0VFNUQU5CZ2txaGtpRzl3MEJBUXNGQURCaApNUXN3Q1FZRFZRUUdFd0pWVXpFVk1CTUdBMVVFQ2hNTVJHbG5hVU5sY25RZ1NXNWpNUmt3RndZRFZRUUxFeEIzCmQzY3VaR2xuYVdObGNuUXVZMjl0TVNBd0hnWURWUVFERXhkRWFXZHBRMlZ5ZENCSGJHOWlZV3dnVW05dmRDQkQKUVRBZUZ3MHhOekV4TURZeE1qSXpNek5hRncweU56RXhNRFl4TWpJek16TmFNRjR4Q3pBSkJnTlZCQVlUQWxWVApNUlV3RXdZRFZRUUtFd3hFYVdkcFEyVnlkQ0JKYm1NeEdUQVhCZ05WQkFzVEVIZDNkeTVrYVdkcFkyVnlkQzVqCmIyMHhIVEFiQmdOVkJBTVRGRkpoY0dsa1UxTk1JRkpUUVNCRFFTQXlNREU0TUlJQklqQU5CZ2txaGtpRzl3MEIKQVFFRkFBT0NBUThBTUlJQkNnS0NBUUVBNVMyb2loRW85bm5wZXpvemlEdHg0V1dMTENsbC9lMHQxRVllbUU1bgorTWdQNXZpYUhMeStWcEhQK25kWDVEMThJTkl1dUFWOHdGcTI2S0Y1VTBXTklaaVFwNm1MdElXalVlV0RQQTI4Ck9leWhUbGo5VExrMmJleXRidEZVNnlwYnBXVWx0bXZZNVY4bmdzcEM3bkZSTkNqcGZuREVEMmtSeUp6Tzh5b0sKTUZ6NEo0SkU4TjdOQTF1SndVRUZNVXZITHMwc2NMb1Baa0tjZXdJUm0xUlYyQXhtRlF4SmtkZjdZTjlQY2traQpmMlhnbTNiNDhCWm4wemYwcVhzU2VHdTg0dWE5Z3d6anpJN3RiVEJqYXlUcFQrL1hwV3VCVnY2ZnZhckk2YmlrCktCODU5T1NHUXV3NzNYWGdldUZ3RVBIVElSb1V0a3p1My9FUStMdHd6bmtrZFFJREFRQUJvNElCWmpDQ0FXSXcKSFFZRFZSME9CQllFRkZQS0YxbjhhOEFESVM4YXJ1U3FxQnlDVnRwMU1COEdBMVVkSXdRWU1CYUFGQVBlVURWVwowVXk3WnZDajRoc2J3NWV5UGRGVk1BNEdBMVVkRHdFQi93UUVBd0lCaGpBZEJnTlZIU1VFRmpBVUJnZ3JCZ0VGCkJRY0RBUVlJS3dZQkJRVUhBd0l3RWdZRFZSMFRBUUgvQkFnd0JnRUIvd0lCQURBMEJnZ3JCZ0VGQlFjQkFRUW8KTUNZd0pBWUlLd1lCQlFVSE1BR0dHR2gwZEhBNkx5OXZZM053TG1ScFoybGpaWEowTG1OdmJUQkNCZ05WSFI4RQpPekE1TURlZ05hQXpoakZvZEhSd09pOHZZM0pzTXk1a2FXZHBZMlZ5ZEM1amIyMHZSR2xuYVVObGNuUkhiRzlpCllXeFNiMjkwUTBFdVkzSnNNR01HQTFVZElBUmNNRm93TndZSllJWklBWWI5YkFFQ01Db3dLQVlJS3dZQkJRVUgKQWdFV0hHaDBkSEJ6T2k4dmQzZDNMbVJwWjJsalpYSjBMbU52YlM5RFVGTXdDd1lKWUlaSUFZYjliQUVCTUFnRwpCbWVCREFFQ0FUQUlCZ1puZ1F3QkFnSXdEUVlKS29aSWh2Y05BUUVMQlFBRGdnRUJBSDRqeC9MS05XNVprbEZjCllXczhFamJtMG55ektlWkMyS09WWVI3UDhnZXZLeXNsV200WG80QlN6S3IyMzVGc0o0YUZ0NnlBaXYxZVkwdFoKL1pOMThiT0dTR1N0b0VjL0pFNG9jSXpyOFA1TWcxMWtSWUhibWdZbnIxUnhla2k1bVNlYjM5REd4VHBKRDRrRwpoczVsWE5vbzRjb25VaWlKd0thcUg3dmgyYmFyeWQ4cE1JU2FnODNKVXF5VkdjMnRXUHBPMDMyOS9DV3Eya3J5CnF2NjZPU01qd3VsVXowZFhmNE9IUWFzUjdDTmZJcis0S1NjYzZBQmxRNVJERjg2UEdlRTZrZHdTUWtGaUIvY1EKeXNOeXEwakVEUVRrZmEycGptdVd0TUNOYkJuaEZYQlllamZ1YkloYVViRXYyRk9RQjNkQ2F2K0ZQZzVlRXZlWApUVnlNbkdvPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg==",
                    "enabled": false,
                    "hostname": "localhost",
                    "key": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBd0prMFNubm9aUWFtYVptKzFVSXY3UWIwS2pYWXlRMHJKSUhqbGxNN3pYRWFBaC9CCmpWNnUzSHVNbGtOVlJ6RUVGeEtGUm5nSlpXTVlPcy9sWDU1QmtBS0RocXNmaDl0ZVA4ZFVQU1RTVTN6bWRVc0cKU3AySDBqSm1YbkVjbzBac3BiVXJhaGVxaXFJRTJnWDBCWm9iV29YZEdzUTJ1cmJQbHIwaVVKdTFWenhIZGhNLwpnTnFQWEN5bE5YdENrbVI2NHl6SGtEaG9taUUwMkkzME1wZThEcXM5eHd5Sndxek9ydmZwNHlBNFJIZ2toNXJaCi8rUStzRVA2QjJXMEdBdzZzZDNTYytOTytYVUZoVnlFRDhqZkNsTndPVHZodi9keE5IWHZ4WHd0RWNjZTd1TVAKQUxSc0hYdUVtZE5DeTdqb2JvMm5RYU5vNkFDVzRTazlFYlVlcHdJREFRQUJBb0lCQUJDbDdaZzJlM2tvSHdCbgowdEFhWFl3dW1ySi9YeExaK1kyNVVKUHc4KzFwUm1kQ21YZC81THVZVG1tNnZrZGVIRTFtUTNENDlVNzNSdVBtCmpwdzZVTTZDK1IyTjh6YktKcFo2dzNSWmwzaXlCalhEVzNMZHRSemRVRzMvZHd0Q0RpZ3d4ZVR3ME1RVkJ6aFYKWldlS2o2eStmNE5OekxGTkZocWdsTFBGRWVMTlk3WUgvYUdHdTZwaVpERkFMT0liMlFVUUVvVVRkUEJabDMxSwpKL0Y0cHE1RkV6ekQ0aHMwYXJzSlIxeUd4L0NCRWxqQWRSNWE4Y3FHdGhxYjFMdXJTMVcxT0dNUGkzNFdDblJKCnYwTC9zSU9ORCt5cjdjd3ZQaFFzcmI3Tk9NbGxRSDYwUGdyQ0JyOUVwalFiRm1ZN3NCTS9jYUZTWHVDTmZnTzMKT3hTQjF0RUNnWUVBNHpSMTh2aXl3Ykt5aE5xeEhkNzJ1UkxHS2k4SWZhVmhsNGxFeGdsR0pLNHQ0KzYreXhObAp2OFBkdGg3elJjL1pyQUlBUEtBRnZxa1RnVzkrc2llL3JPcE9PUzA2UnVIWVpJMjdpS3ZiL054N29YTm9TM0Y2CitXRlNJSVJqRFo0eHc2SU1iNUJFS0hOVGtKU1pBMXBkaXhYeFgzeDVTM2hCWTlmekl2NGdVVDhDZ1lFQTJRSDAKbnU5REZ0MDNmWmpvcjRMUkZCQS84NEV1dmdBL2RxMkZ4TzVvMjIrdk43Qk9LRW9IWTRuTzhYdFN2L2MzaVFZdApmTUdGZUR2eXdqRnFPU05SS2JxenMza3ZmY2gxNEhHdFltT29DVTF6Smd2VkF4SytWOFF4YWxMbmhlbjZsaTgwCnJKYmtWWUJXYnFHZHU1d2FPM1RyeGsvbHBMejFvSnliK0FQbmNKa0NnWUVBeHd3anRTeEtOc1hkL2VRMXZReWoKcTdvVjM4ajZzRkRkOGNzeGRDbVZ5UGJMZjErZXlQeUNiSUVrMExQR3ZzZDhWR1FnWHZ4MmEvVm1BRVIwZzVHdApLOGpOcE92Sm5XV29sVmRTRDBjNDRBc3RTR0Y1NG5aVm5QcTAzNFg3M1dkVkE2WUpMUWYvbmt5Q3ppeFE3UTRTCjI1SHF0NU91RS84VGttd1ROdm5aeURNQ2dZQlUxV0xmendFWEt4dktTY0NNL0VMbjZRSlhHeVUrdWJJVm52Z3cKYUZNdGxaZVYyYUowTlp4MTJTWEJXVWljS3ZUVzZFZFlBZjgvWlpBZFFsd1lRaW8wZWZ4T0tRM2J5a1hxbkFYLwo5SjE0SllMaVdMWEpJY2F2WU05QWxYQlU5WWMwdHZlTWtPQnFYR1Z0a2xpclF2Y3VUeUxzdjhjeWczd3ljWWEzCmlpOHI0UUtCZ0NaODZtcEZuUlN2aTdpTWw1MUdybDFQZER2YzZVMURRVkI5empTQlVzak1lamNTQXk4WmVJZlQKUDVoSGc3REFaUEhmUGpYLzZxUnpNQW8zWlFlZVBMdTNEYVgwbFBoekNhalF1bkpIbWZvYlA2QjFOUG45T056bQo0dE0wdEU1YTJVVmJuZW5haVZyaDRhSXlJTTFxVTdHT1ArcDNMNi9hNzVnSVQ2SVJRVlloCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg==",
                    "sslProtocols": ""
                  },
                  "yugaware": {
                    "health": {
                      "email": "",
                      "password": "",
                      "username": ""
                    },
                    "multiTenant": false,
                    "replicas": 1,
                    "service": {
                      "enabled": true,
                      "ip": "",
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": "yugaware",
                    "storage": "100Gi",
                    "storageClass": "standard"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          certified: "true"
          containerImage: registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
          description: Operator for Yugabyte Platform makes it easy to deploy and
            upgrade Yugabyte Platform on OpenShift environments.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/yugabyte-platform-operator-bundle-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/yugabyte/yugabyte-platform-operator
          support: YugabyteDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy and manage Yugabyte Platform as a Kubernetes native
              custom resource.
            displayName: Yugabyte Platform
            kind: YBPlatform
            name: ybplatforms.yugabyte.com
            version: v1alpha1
        description: |2

          Operator for Yugabyte Platform makes it easy to deploy and upgrade Yugabyte Platform on OpenShift environments.
          Use Yugabyte Platform's orchestration and monitoring to manage YugabyteDB universes. Yugabyte Platform can create a YugabyteDB universe with many instances (VMs, pods, machines, etc., provided by IaaS), logically grouped together to form one logical distributed database. Each universe includes one or more clusters. A universe is comprised of one primary cluster and, optionally, one or more read replica clusters. All instances belonging to a cluster run on the same type of cloud provider instance type.

          ### Yugabyte Platform Features
          * **YugabyteDB-as-a-Service on Kubernetes** - Deploy your own YugabyteDB-as-a-Service running on any Kubernetes infrastructure, Yugabyte Platform provides central management for multiple database clusters and Includes 24/7 production support.
          * **Monitoring & Alerting** - Comprehensive monitoring & alerting without any configuration.
          * **Scheduled Backups** - Use Yugabyte Platform to perform a regular scheduled backup of a full YugabyteDB universe or selected tables.

          ### Getting Started
          Here are a few resources for getting started with Yugabyte platform
          * [Quick start guide for Yugabyte Platform](https://docs.yugabyte.com/latest/yugabyte-platform/deploy/) - Install Yugabyte platform on OpenShift.
          * [Quick Start for YugabyteDB](http://docs.yugabyte.com/latest/quick-start/) - install, create a local cluster and read/write from YugabyteDB.
          * [Explore core features](https://docs.yugabyte.com/latest/explore/) - automatic sharding & re-balancing, linear scalability, fault tolerance, tunable reads etc.
          * [Ecosystem integrations](https://docs.yugabyte.com/latest/develop/ecosystem-integrations/) - integrations with Apache Kafka/KSQL, Apache Spark, JanusGraph, KairosDB, Presto and more.
          * [Real world apps](https://docs.yugabyte.com/latest/develop/realworld-apps/) - sample real-world, end-to-end applications built using YugabyteDB.

          ### Help
          Cannot find what you are looking for? Have a question? We love to hear from you - [please file a GitHub issue](https://github.com/YugaByte/yugabyte-db/issues) or talk to us live at [Slack](https://www.yugabyte.com/slack).
        displayName: Yugabyte Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - distributed sql
        links:
        - name: Yugabyte Platform Operator
          url: https://www.yugabyte.com/
        maintainers:
        - email: oss-maintainers@yugabyte.com
          name: Yugabyte
        maturity: alpha
        provider:
          name: Yugabyte Inc
        relatedImages:
        - registry.connect.redhat.com/yugabytedb/yugabyte-platform-operator-bundle@sha256:4cee86e973206b5300e88af088aca700cde3b2e2e5e1c70b9a20289e9f592349
        - registry.marketplace.redhat.com/rhm/yugabytedb/yugabyte-platform-operator@sha256:f227c9e2c625b04160d952349ddba552ac66dae8e7da35f56746f3447c270ce3
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        version: 0.0.6
      entries:
      - name: yugabyte-platform-operator.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: yugabyte-platform-operator-bundle-rhmp
    provider:
      name: Yugabyte Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: rebellions
      provider-url: https://rebellions.ai/
    name: rbln-npu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: candidate
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: fast
    - currentCSV: rbln-npu-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rebellions.ai/v1beta1",
                "kind": "RBLNClusterPolicy",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "rbln-npu-operator"
                  },
                  "name": "rbln-cluster-policy"
                },
                "spec": {
                  "daemonsets": {},
                  "devicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "metricsExporter": {
                    "enabled": true,
                    "image": "rebellions/rbln-metrics-exporter",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "name": "rebellions",
                  "npuFeatureDiscovery": {
                    "enabled": true,
                    "image": "rebellions/rbln-npu-feature-discovery",
                    "imagePullPolicy": "IfNotPresent",
                    "version": "latest"
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true,
                    "image": "rebellions/k8s-device-plugin",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest",
                    "vfioChecker": {
                      "image": "rebellions/rbln-vfio-manager",
                      "version": "latest"
                    }
                  },
                  "vfioManager": {
                    "enabled": true,
                    "image": "rebellions/rbln-vfio-manager",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "system-node-critical",
                    "version": "latest"
                  }
                },
                "status": {
                  "namespace": "rbln-operator-system",
                  "state": "ready"
                }
              }
            ]
          capabilities: Basic Install
          containerImage: docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
          createdAt: "2025-07-14T01:00:30Z"
          description: Deploy and manage Rebellions NPU in Kubernetes
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Rebellions
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RBLNClusterPolicy is the Schema for the RBLNClusterPolicys
              API
            displayName: RBLNCluster Policy
            kind: RBLNClusterPolicy
            name: rblnclusterpolicies.rebellions.ai
            version: v1beta1
        description: RBLN Operator automates the deployment and management of RBLN
          NPU drivers, device plugins, CRDs, firmware updates, and health monitoring
          for seamless integration into Kubernetes.
        displayName: RBLN operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rebellions
        - ai
        - AI
        - NPU
        - AI acceleration
        - ML inference
        links:
        - name: Rbln K8s Operator
          url: https://rbln-npu-operator.domain
        maintainers:
        - email: heesu.sung@rebellions.ai
          name: heesu.sung
        - email: mskim@rebellions.ai
          name: minsik.kim
        maturity: stable
        minKubeVersion: 1.30.0
        provider:
          name: rebellions
          url: https://rebellions.ai/
        relatedImages:
        - registry.connect.redhat.com/rebellions/rbln-npu-operator-bundle@sha256:67bdef4c88b5ac186ee3ec6825db91b61591beb42aabfb371067493623d32379
        - docker.io/rebellions/k8s-device-plugin@sha256:d6a1b7653fa4978dc34e71a62d48a99537218f248c933871bbe5f8d5ff5a6e5d
        - docker.io/rebellions/rbln-metrics-exporter@sha256:f696d75b40e948cad25a1bc14c43e256463acb55477a7787b1be4e7086908240
        - docker.io/rebellions/rbln-npu-feature-discovery@sha256:02ea11356b04e05d2dbbdd7b60286aec239e846e4beb26fdf515f61535b3e885
        - docker.io/rebellions/rbln-npu-operator@sha256:f3dfd53e159c0ab05d65ca93f8d783659b33acd0a015b38975c26106b14a1600
        - docker.io/rebellions/rbln-vfio-manager@sha256:c9abcf882ddceb17a0f6b65463de1d4232d5d773994ff75e0aa2beef4c8ce87d
        version: 0.0.1
      entries:
      - name: rbln-npu-operator.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: rbln-npu-operator
    provider:
      name: rebellions
      url: https://rebellions.ai/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3e0fde49a8e8c5f780a76087bbc42507a42fea0e1fea572ae05ba66d27338365
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:3b36dcea36ebf4b2e95fac6691f6fc8f1242f7e5c38195a9af28bfad514c28cb
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:36dc3710ac11d86e1cf0a3c6994ae0e6feb6c8ff1e8110997f27f4e573c259bc
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:64cadb49077eff7863dd678c4aec70f4101f17ee8e0bf7ca64ce7010c9bf4b9b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v6.2.10-41
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/outbound",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager",
                      "versionTag": "6.2.10-41"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise",
                    "versionTag": "6.2.14-9.rhel8-openshift"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/outbound@sha256:3c251754ae8650157465a56ef3a7b41d5a0ad4e9db09ef042e0014076f25e8a2
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:4981d96081cc3d10cd26f20aab792e9c195b05bdcd609367011b7e9c7e6759b1
        version: 6.2.10-41
      entries:
      - name: redis-enterprise-operator.v6.2.10-41
        version: 6.2.10-41
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: preview
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle@sha256:83b7ecec8cfb96d4d33c86f95c70d5090e6b50d8962af3806785d627a8978ce0
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-apigateway-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-apigateway-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Resource",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "RestAPI",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "Stage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "apigateway.services.k8s.aws/v1alpha1",
                "kind": "VPCLink",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
          createdAt: "2025-09-19T21:13:12Z"
          description: AWS API Gateway controller is a service controller for managing
            API Gateway resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: APIIntegrationResponse represents the state of an AWS apigateway
              APIIntegrationResponse resource.
            displayName: APIIntegrationResponse
            kind: APIIntegrationResponse
            name: apiintegrationresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIKey represents the state of an AWS apigateway APIKey resource.
            displayName: APIKey
            kind: APIKey
            name: apikeys.apigateway.services.k8s.aws
            version: v1alpha1
          - description: APIMethodResponse represents the state of an AWS apigateway
              APIMethodResponse resource.
            displayName: APIMethodResponse
            kind: APIMethodResponse
            name: apimethodresponses.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Authorizer represents the state of an AWS apigateway Authorizer
              resource.
            displayName: Authorizer
            kind: Authorizer
            name: authorizers.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Deployment represents the state of an AWS apigateway Deployment
              resource.
            displayName: Deployment
            kind: Deployment
            name: deployments.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Integration represents the state of an AWS apigateway Integration
              resource.
            displayName: Integration
            kind: Integration
            name: integrations.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Method represents the state of an AWS apigateway Method resource.
            displayName: Method
            kind: Method
            name: methods.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Resource represents the state of an AWS apigateway Resource
              resource.
            displayName: Resource
            kind: Resource
            name: resources.apigateway.services.k8s.aws
            version: v1alpha1
          - description: RestAPI represents the state of an AWS apigateway RestAPI
              resource.
            displayName: RestAPI
            kind: RestAPI
            name: restapis.apigateway.services.k8s.aws
            version: v1alpha1
          - description: Stage represents the state of an AWS apigateway Stage resource.
            displayName: Stage
            kind: Stage
            name: stages.apigateway.services.k8s.aws
            version: v1alpha1
          - description: VPCLink represents the state of an AWS apigateway VPCLink
              resource.
            displayName: VPCLink
            kind: VPCLink
            name: vpclinks.apigateway.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon API Gateway resources in AWS from within your Kubernetes cluster.

          **About Amazon API Gateway**

          Amazon API Gateway is an AWS service for creating, publishing, maintaining, monitoring, and securing REST, HTTP, and WebSocket APIs at any scale. API developers can create APIs that access AWS or other web services, as well as data stored in the AWS Cloud. As an API Gateway API developer, you can create APIs for use in your own client applications. Or you can make your APIs available to third-party app developers.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon API Gateway
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apigateway
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon API Gateway Developer Resources
          url: https://aws.amazon.com/api-gateway/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: apigateway maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-apigateway-controller@sha256:44d34788b43d103530115fc53b0252bfdbe6e07a692847a3d32aa21e45892ff9
        - public.ecr.aws/aws-controllers-k8s/apigateway-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-apigateway-controller.v1.3.1
        version: 1.3.1
      - name: ack-apigateway-controller.v1.3.0
        version: 1.3.0
      - name: ack-apigateway-controller.v1.2.5
        version: 1.2.5
      - name: ack-apigateway-controller.v1.2.4
        version: 1.2.4
      - name: ack-apigateway-controller.v1.2.3
        version: 1.2.3
      - name: ack-apigateway-controller.v1.2.2
        version: 1.2.2
      - name: ack-apigateway-controller.v1.2.1
        version: 1.2.1
      - name: ack-apigateway-controller.v1.1.0
        version: 1.1.0
      - name: ack-apigateway-controller.v1.0.6
        version: 1.0.6
      - name: ack-apigateway-controller.v1.0.5
        version: 1.0.5
      - name: ack-apigateway-controller.v1.0.4
        version: 1.0.4
      - name: ack-apigateway-controller.v1.0.2
        version: 1.0.2
      - name: ack-apigateway-controller.v1.0.1
        version: 1.0.1
      - name: ack-apigateway-controller.v1.0.0
        version: 1.0.0
      - name: ack-apigateway-controller.v0.0.3
        version: 0.0.3
      - name: ack-apigateway-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-apigateway-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Syndesis Team
      provider-url: https://github.com/syndesisio/syndesis
    name: syndesis
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: syndesis.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.12
          createdAt: "2021-08-11T12:00:00Z"
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/openshift-community-operators/syndesis@sha256:9e8c08793aebc8857dcd451fa2fde2b3f05b9ff4f3d3cc604f9f752193f58d42
        - quay.io/syndesis/syndesis-operator:1.12
        - centos/postgresql-10-centos7
        version: 1.12.0
      entries:
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.12.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: 1.13.x
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: candidate
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - quay.io/syndesis/syndesis-operator:1.13
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: latest
    - currentCSV: syndesis.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/syndesis/syndesis-operator:1.13
          createdAt: 2022-01-18 19:05:58 UTC
          description: Manages the installation of Syndesis, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Syndesis
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Syndesis operator
          Syndesis is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following syndesis components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
        displayName: Syndesis Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Syndesis Project
          url: https://github.com/syndesisio/syndesis
        - name: Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: syndesis@googlegroups.com
          name: Syndesis Team
        maturity: beta
        provider:
          name: Syndesis Team
          url: https://github.com/syndesisio/syndesis
        relatedImages:
        - centos/postgresql-10-centos7
        - quay.io/openshift-community-operators/syndesis@sha256:fa62677d0e12681f92504d8a15c2b440b2bc0b0e5ebaf6edc0048621806b8cda
        - quay.io/syndesis/syndesis-operator:1.13
        version: 1.13.0
      entries:
      - name: syndesis.v1.13.0
        version: 1.13.0
      - name: syndesis.v1.12.0
        version: 1.12.0
      name: stable
    defaultChannel: latest
    packageName: syndesis
    provider:
      name: Syndesis Team
      url: https://github.com/syndesisio/syndesis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Vertica
      provider-url: http://www.vertica.com
    name: verticadb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: verticadb-operator.v1.11.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "verticadb-operator",
                    "app.kubernetes.io/instance": "eventtrigger-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "eventtrigger",
                    "app.kubernetes.io/part-of": "verticadb-operator"
                  },
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "vertica/vertica-k8s:latest",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "serviceName": "defaultsubcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "vertica/vertica-k8s:23.3.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/vertica/verticadb-operator:1.11.2
          createdAt: "2023-08-02T11:29:39Z"
          description: Operator that manages a Vertica Eon Mode database.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: VerticaAutoscaler is a CR that allows you to autoscale one
              or more subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: matt.spilchen@vertica.com
          name: Matt Spilchen
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/verticadb-operator@sha256:b7691da906474a8ab829966bb3da4a846eaf12909c97bd0d9cb6a1e5e7ff6c03
        - docker.io/vertica/verticadb-operator:1.11.2
        version: 1.11.2
      entries:
      - name: verticadb-operator.v1.11.2
        version: 1.11.2
      - name: verticadb-operator.v1.11.1
        version: 1.11.1
      - name: verticadb-operator.v1.11.0
        version: 1.11.0
      - name: verticadb-operator.v1.10.2
        version: 1.10.2
      - name: verticadb-operator.v1.10.1
        version: 1.10.1
      - name: verticadb-operator.v1.10.0
        version: 1.10.0
      - name: verticadb-operator.v1.9.0
        version: 1.9.0
      - name: verticadb-operator.v1.8.0
        version: 1.8.0
      - name: verticadb-operator.v1.7.0
        version: 1.7.0
      - name: verticadb-operator.v1.6.0
        version: 1.6.0
      - name: verticadb-operator.v1.5.0
        version: 1.5.0
      - name: verticadb-operator.v1.4.0
        version: 1.4.0
      - name: verticadb-operator.v1.3.1
        version: 1.3.1
      - name: verticadb-operator.v1.3.0
        version: 1.3.0
      - name: verticadb-operator.v1.2.0
        version: 1.2.0
      name: stable
    - currentCSV: verticadb-operator.v25.3.1-0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1",
                "kind": "VerticaDB",
                "metadata": {
                  "name": "verticadb-sample"
                },
                "spec": {
                  "communal": {
                    "endpoint": "https://s3.amazonaws.com"
                  },
                  "image": "opentext/vertica-k8s:24.1.0-0-minimal",
                  "subclusters": [
                    {
                      "name": "default_subcluster",
                      "size": 3
                    }
                  ]
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "EventTrigger",
                "metadata": {
                  "name": "eventtrigger-sample"
                },
                "spec": {
                  "matches": [
                    {
                      "condition": {
                        "status": "True",
                        "type": "DBInitialized"
                      }
                    }
                  ],
                  "references": [
                    {
                      "object": {
                        "apiVersion": "vertica.com/v1beta1",
                        "kind": "VerticaDB",
                        "name": "verticadb-sample"
                      }
                    }
                  ],
                  "template": {
                    "metadata": {
                      "generateName": "create-user-table-"
                    },
                    "spec": {
                      "template": {
                        "spec": {
                          "containers": [
                            {
                              "command": [
                                "/opt/vertica/bin/vsql",
                                "-h",
                                "verticadb-sample-defaultsubcluster",
                                "-c",
                                "CREATE TABLE T1 (C1 INT);"
                              ],
                              "image": "opentext/vertica-k8s:23.4.0-0-minimal",
                              "name": "main"
                            }
                          ],
                          "restartPolicy": "OnFailure"
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaAutoscaler",
                "metadata": {
                  "name": "verticaautoscaler-sample"
                },
                "spec": {
                  "customAutoscaler": {},
                  "scalingGranularity": "Subcluster",
                  "serviceName": "default-subcluster",
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaReplicator",
                "metadata": {
                  "name": "verticareplicator-sample"
                },
                "spec": {
                  "source": {
                    "verticaDB": "sourceDBName"
                  },
                  "target": {
                    "verticaDB": "targetDBName"
                  }
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaRestorePointsQuery",
                "metadata": {
                  "name": "verticarestorepointsquery-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              },
              {
                "apiVersion": "vertica.com/v1beta1",
                "kind": "VerticaScrutinize",
                "metadata": {
                  "name": "verticascrutinize-sample"
                },
                "spec": {
                  "verticaDBName": "verticadb-sample"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: docker.io/opentext/verticadb-operator:25.3.1-0
          createdAt: "2025-09-24T00:58:06Z"
          description: Operator that manages a Vertica Eon Mode database.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/vertica/vertica-kubernetes
          support: Vertica
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EventTrigger is the Schema for the eventtriggers API
            displayName: Event Trigger
            kind: EventTrigger
            name: eventtriggers.vertica.com
            version: v1beta1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1
          - description: |-
              VerticaAutoscaler is a CR that allows you to autoscale one or more
              subclusters in a VerticaDB.
            displayName: Vertica Autoscaler
            kind: VerticaAutoscaler
            name: verticaautoscalers.vertica.com
            version: v1beta1
          - description: VerticaDB is the CR that defines a Vertica Eon mode cluster
              that is managed by the verticadb-operator.
            displayName: Vertica DB
            kind: VerticaDB
            name: verticadbs.vertica.com
            version: v1
          - description: VerticaReplicator is the Schema for the verticareplicators
              API
            displayName: Vertica Replicator
            kind: VerticaReplicator
            name: verticareplicators.vertica.com
            version: v1beta1
          - description: VerticaRestorePointsQuery is the Schema for the verticarestorepointsqueries
              API
            displayName: Vertica Restore Points Query
            kind: VerticaRestorePointsQuery
            name: verticarestorepointsqueries.vertica.com
            version: v1beta1
          - description: VerticaScrutinize is the schema for verticascrutinize API
            displayName: Vertica Scrutinize
            kind: VerticaScrutinize
            name: verticascrutinizers.vertica.com
            version: v1beta1
        description: |-
          ### What is Vertica?
          Vertica is a unified analytics platform, based on a massively scalable architecture with the broadest set of analytical functions spanning event and time series, pattern matching, geospatial and end-to-end in-database machine learning. Vertica enables you to easily apply these powerful functions to the largest and most demanding analytical workloads, arming you and your customers with predictive business insights faster than any analytics data warehouse in the market. Vertica provides a unified analytics platform across major public clouds and on-premises data centers and integrates data in cloud object storage and HDFS without forcing you to move any of your data.

          ### What does this operator do?
          This operator manages a Vertica Eon Mode database, and automates the following administrative tasks:

          * Installing Vertica
          * Creating and reviving a Vertica database
          * Restarting and rescheduling DOWN pods to maintain quorum
          * Upgrading Vertica to a new version while keeping the database online
          * Subcluster scaling, both manually and automatically with the Horizontal Pod Autoscaler
          * Service management and health monitoring for pods
          * Load balancing for internal and external traffic
          * Expose a metrics endpoint to integrate with Prometheus to monitor the operator's health

          For a brief overview on how to create a cluster, see the [Vertica GitHub repository](https://github.com/vertica/vertica-kubernetes). For an in-depth look at Vertica on Kubernetes, see the [Vertica documentation](https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm).
        displayName: VerticaDB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - database
        - vertica
        links:
        - name: Documentation
          url: https://www.vertica.com/docs/latest/HTML/Content/Authoring/Containers/ContainerizedVertica.htm
        - name: Vertica Container Images
          url: https://hub.docker.com/u/vertica
        maintainers:
        - email: cchen8@opentext.com
          name: Cai Chen
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Vertica
          url: http://www.vertica.com
        relatedImages:
        - docker.io/opentext/verticadb-operator:25.3.1-0
        - quay.io/community-operator-pipeline-prod/verticadb-operator@sha256:be706a582a4b8cae487835e920288b41229834cd3477b7389adb363db4b4f2ca
        version: 25.3.1-0
      entries:
      - name: verticadb-operator.v25.3.1-0
        version: 25.3.1-0
      - name: verticadb-operator.v25.3.0-0
        version: 25.3.0-0
      - name: verticadb-operator.v25.2.1-0
        version: 25.2.1-0
      - name: verticadb-operator.v25.2.0-0
        version: 25.2.0-0
      - name: verticadb-operator.v25.1.0-0
        version: 25.1.0-0
      - name: verticadb-operator.v24.4.0-0
        version: 24.4.0-0
      - name: verticadb-operator.v2.2.0
        version: 2.2.0
      - name: verticadb-operator.v2.1.3
        version: 2.1.3
      - name: verticadb-operator.v2.1.2
        version: 2.1.2
      - name: verticadb-operator.v2.1.1
        version: 2.1.1
      - name: verticadb-operator.v2.1.0
        version: 2.1.0
      - name: verticadb-operator.v2.0.2
        version: 2.0.2
      - name: verticadb-operator.v2.0.1
        version: 2.0.1
      - name: verticadb-operator.v2.0.0
        version: 2.0.0
      name: v2-stable
    defaultChannel: v2-stable
    packageName: verticadb-operator
    provider:
      name: Vertica
      url: http://www.vertica.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: www.redhat.com
    name: dpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dpu-operator.v4.18.0-202509101149
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "DpuOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "dpuoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dpuoperatorconfig",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "dpuoperatorconfig-sample"
                },
                "spec": {
                  "mode": "host"
                }
              },
              {
                "apiVersion": "config.openshift.io/v1",
                "kind": "ServiceFunctionChain",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dpu-operator",
                    "app.kubernetes.io/instance": "servicefunctionchain-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "servicefunctionchain",
                    "app.kubernetes.io/part-of": "dpu-operator"
                  },
                  "name": "servicefunctionchain-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-10-15T19:06:53Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-dpu-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DpuOperatorConfig is the Schema for the dpuoperatorconfigs
              API
            displayName: Dpu Operator Config
            kind: DpuOperatorConfig
            name: dpuoperatorconfigs.config.openshift.io
            version: v1
          - description: ServiceFunctionChain is the Schema for the servicefunctionchains
              API
            displayName: Service Function Chain
            kind: ServiceFunctionChain
            name: servicefunctionchains.config.openshift.io
            version: v1
        description: This operator will manage and configure data processing unit
          (DPUs) to be used in accelerating/offloading k8s networking functions
        displayName: DPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dpu
        - dpu-operator
        - ipu
        links:
        - name: Dpu Operator
          url: https://github.com/openshift/dpu-operator
        maintainers:
        - email: wizhao@redhat.com
          name: William Zhao
        - email: bnemeth@redhat.com
          name: Balazs Nemeth
        maturity: alpha
        provider:
          name: Red Hat
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-dpu-operator-bundle@sha256:4e98c0f629e938a0c9d011e6a675d45589b044a02b3f9551fc0e8f7cd172c8a0
        - registry.redhat.io/openshift4/ose-dpu-daemon-rhel9@sha256:acb5ab931f4d7a46bb195c59d156441af5e9092b2739bcb0379b89d47ccf4e81
        - registry.redhat.io/openshift4/ose-dpu-rhel9-operator@sha256:5dedc4733081fc21bcb77dbf4f4c032bf418f00bc4371c29c7ce028f51833c0b
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        version: 4.18.0-202509101149
      entries:
      - name: dpu-operator.v4.18.0-202509101149
        version: 4.18.0-202509101149
      - name: dpu-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: dpu-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: dpu-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: dpu-operator.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: dpu-operator.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: dpu-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: dpu-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: dpu-operator.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: dpu-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: dpu-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: dpu-operator.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: dpu-operator.v4.18.0-202503130333
        version: 4.18.0-202503130333
      - name: dpu-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: dpu-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: dpu-operator
    provider:
      name: Red Hat
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator-with-cg-module
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
          createdAt: "2025-09-11T09:32:52Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.1.0 <7.1.6'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition with Cryptographic Module Annual Subscription License
            per Core 17 Operator for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64.\n\nThe download and Use of the Product is strictly subject to the
          terms of the End User License Agreement with Fujitsu Limited found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator with Cryptographic Module
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:4ab17b92016288f1ebfb2106c826a17e75d041f0dba7eaaf7545193c3305a408
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils-with-cryptographic-module@sha256:8b2b6a981c863f7075ced1235c614317b6526626b42405295a9c664736f0a80f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore-with-cryptographic-module@sha256:98a3b0bfa6ec7a9a4a08fb524c331c3de6a9b580d6386a7b5430955ed84b2fc5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter-with-cryptographic-module@sha256:71e1d1e715e2d89757e80d90b8d5b11baa729110cd5f7898df920a4c85f2bf8a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server-with-cryptographic-module@sha256:a012f21f63729ba667af097375e08874156e2cf1b94ae203985d45f11031fea9
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup-with-cryptographic-module@sha256:03b7f3c76e3eb775e4ec7f66f7506ee1feeab56dca96e868ac3f38b9c1ca5567
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore-with-cryptographic-module@sha256:6c0d6e10cfe1868be2986fdde14c6f55f86b12cea01a82199b1e2e0c3154ef75
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils-with-cryptographic-module@sha256:4caf1dc5211ac03893ce9a62b417f412ffd411f13569f3ac75e12ca5128af9e4
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server-with-cryptographic-module@sha256:cad4e36dc2b1e5ca6bf9308d9f93958d4c8f2bbdc14e000ca0b5700fa335df44
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd-with-cryptographic-module@sha256:f743f35012e6fd3d54a6b54baeeee98bd7c65fe52923acd87546066fa9b54d7e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2-with-cryptographic-module@sha256:e510bc2d30df33e4c9d3babb898ff9cbd13704fa3fa69a284bf672765fb03083
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade-with-cryptographic-module@sha256:b8be0344b6cf99b2c5de2b16f04f6f2b05bde861d9b38afd319bbbc4dd72a620
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server-with-cryptographic-module@sha256:7778f4b03861b56371d18f7726b68ad21d0faa7c1a25cb38fe7cac483d81c095
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore-with-cryptographic-module@sha256:edeb7b72a89e05995661e6a17e75b5b477a3ca351e0db146a939a33e88e16c97
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup-with-cryptographic-module@sha256:dd15cf3b6b4a5c959efef8b602abe5d64b4f5a424fb6a3e55079a2040e67cd7a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade-with-cryptographic-module@sha256:caf6e1d2cc31d01958e4adcb4c3c307d761f350d5b0963ce19214f041c14184a
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob-with-cryptographic-module@sha256:809a64600396378ff8c825595d40f72aedbc2b57f394420548262b085a3f2ea5
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit-with-cryptographic-module@sha256:bb661b3aad0b825d56a9696d6d2d9f5c04add2a83e793fc3d4a73980f4a7a0ac
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade-with-cryptographic-module@sha256:1fe690a9c04a55213062ad3bef8c0cff73849f8777c3945b9a8ae2dcff6330fa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2-with-cryptographic-module@sha256:a4c5a59ca18f6fa295b78d570f03991f1a4b4bd6df52f8526a1254ac1a567990
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup-with-cryptographic-module@sha256:5b49aa46386ca954ebcbea28fbabd15929cb64fc2a4d2dd2796c9d2e15b212c6
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2-with-cryptographic-module@sha256:e8aa27cbb57503ec5f4f37beeb2e833ace08d4c583d4361b7ce945c7461a904c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator-with-cryptographic-module@sha256:eaf601e7aef525ebcd4b769c859ed74c7a8077f843b3a75d76713eabf0295745
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils-with-cryptographic-module@sha256:7d169df0b8fe47000891afde79be8dd7978da3f893d1e92f01f57fd8c0b3b4c6
        version: 7.1.6
      entries:
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.6
        version: 7.1.6
      - name: fujitsu-enterprise-postgres-operator-with-cg-module.v7.1.5
        version: 7.1.5
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator-with-cg-module
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: StreamNative
      provider-url: https://streamnative.io
    name: function-mesh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: function-mesh.v0.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "java-function-sample",
                  "namespace": "default"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "input": {
                    "topics": [
                      "persistent://public/default/java-function-input-topic"
                    ],
                    "typeClassName": "java.lang.String"
                  },
                  "java": {
                    "extraDependenciesDir": "random-dir/",
                    "jar": "pulsar-functions-api-examples.jar",
                    "jarLocation": "public/default/nlu-test-java-function"
                  },
                  "logTopic": "persistent://public/default/logging-function-logs",
                  "maxPendingAsyncRequests": 1000,
                  "maxReplicas": 5,
                  "output": {
                    "topic": "persistent://public/default/java-function-output-topic",
                    "typeClassName": "java.lang.String"
                  },
                  "pod": {
                    "annotations": {
                      "managed-function": "true"
                    },
                    "env": [
                      {
                        "name": "EXAMPLE_VARIABLE",
                        "value": "exampleValue"
                      }
                    ],
                    "imagePullSecrets": [
                      {
                        "name": "regcred"
                      }
                    ],
                    "initContainers": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30"
                        ],
                        "image": "busybox:1.28",
                        "name": "init-function"
                      }
                    ],
                    "labels": {
                      "locaction": "mtv"
                    },
                    "sidecars": [
                      {
                        "command": [
                          "sh",
                          "-c",
                          "echo The app is running! \u0026\u0026 sleep 30000"
                        ],
                        "image": "busybox:1.28",
                        "name": "sidecar-function"
                      }
                    ],
                    "volumes": [
                      {
                        "emptyDir": {},
                        "name": "cache-volume"
                      }
                    ]
                  },
                  "pulsar": {
                    "authSecret": "test-auth",
                    "pulsarConfig": "test-pulsar",
                    "tlsSecret": "test-tls"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "secretsMap": {
                    "name": {
                      "key": "username",
                      "path": "test-secret"
                    },
                    "pwd": {
                      "key": "password",
                      "path": "test-secret"
                    }
                  },
                  "volumeMounts": [
                    {
                      "mountPath": "/cache",
                      "name": "cache-volume"
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "FunctionMesh",
                "metadata": {
                  "name": "functionmesh-sample"
                },
                "spec": {
                  "functions": [
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/functionmesh-input-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex1"
                      },
                      "logTopic": "persistent://public/default/logging-function-log",
                      "maxReplicas": 1,
                      "name": "ex1",
                      "output": {
                        "topic": "persistent://public/default/mid-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    },
                    {
                      "autoAck": true,
                      "className": "org.apache.pulsar.functions.api.examples.ExclamationFunction",
                      "clusterName": "test-pulsar",
                      "forwardSourceMessageProperty": true,
                      "input": {
                        "topics": [
                          "persistent://public/default/mid-topic"
                        ],
                        "typeClassName": "java.lang.String"
                      },
                      "java": {
                        "jar": "pulsar-functions-api-examples.jar",
                        "jarLocation": "public/default/nlu-test-functionmesh-ex2"
                      },
                      "logTopic": "persistent://public/default/logging-function-logs",
                      "maxReplicas": 1,
                      "name": "ex2",
                      "output": {
                        "topic": "persistent://public/default/functionmesh-output-topic",
                        "typeClassName": "java.lang.String"
                      },
                      "pulsar": {
                        "pulsarConfig": "mesh-test-pulsar"
                      },
                      "replicas": 1,
                      "resources": {
                        "limits": {
                          "cpu": "0.2",
                          "memory": "1.1G"
                        },
                        "requests": {
                          "cpu": "0.1",
                          "memory": "1G"
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Sink",
                "metadata": {
                  "name": "sink-sample"
                },
                "spec": {
                  "autoAck": true,
                  "className": "org.apache.pulsar.io.elasticsearch.ElasticSearchSink",
                  "clusterName": "test-pulsar",
                  "image": "streamnative/pulsar-io-elastic-search:2.10.0.0-rc10",
                  "input": {
                    "topics": [
                      "persistent://public/default/input"
                    ],
                    "typeClassName": "[B"
                  },
                  "java": {
                    "jar": "connectors/pulsar-io-elastic-search-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "pulsar": {
                    "pulsarConfig": "test-sink"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sinkConfig": {
                    "elasticSearchUrl": "http://quickstart-es-http.default.svc.cluster.local:9200",
                    "indexName": "my_index",
                    "password": "wJ757TmoXEd941kXm07Z2GW3",
                    "typeName": "doc",
                    "username": "elastic"
                  }
                }
              },
              {
                "apiVersion": "compute.functionmesh.io/v1alpha1",
                "kind": "Source",
                "metadata": {
                  "name": "source-sample"
                },
                "spec": {
                  "className": "org.apache.pulsar.io.debezium.mongodb.DebeziumMongoDbSource",
                  "clusterName": "test-pulsar",
                  "forwardSourceMessageProperty": true,
                  "image": "streamnative/pulsar-io-debezium-mongodb:2.10.0.0-rc10",
                  "java": {
                    "jar": "connectors/pulsar-io-debezium-mongodb-2.10.0.0-rc10.nar",
                    "jarLocation": ""
                  },
                  "maxReplicas": 1,
                  "output": {
                    "producerConf": {
                      "maxPendingMessages": 1000,
                      "maxPendingMessagesAcrossPartitions": 50000,
                      "useThreadLocalProducers": true
                    },
                    "topic": "persistent://public/default/destination",
                    "typeClassName": "org.apache.pulsar.common.schema.KeyValue"
                  },
                  "pulsar": {
                    "pulsarConfig": "test-source"
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "0.2",
                      "memory": "1.1G"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "1G"
                    }
                  },
                  "sourceConfig": {
                    "database.whitelist": "inventory",
                    "mongodb.hosts": "rs0/mongo-dbz-0.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-1.mongo.default.svc.cluster.local:27017,rs0/mongo-dbz-2.mongo.default.svc.cluster.local:27017",
                    "mongodb.name": "dbserver1",
                    "mongodb.password": "dbz",
                    "mongodb.task.id": "1",
                    "mongodb.user": "debezium",
                    "pulsar.service.url": "pulsar://test-pulsar-broker.default.svc.cluster.local:6650"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
          createdAt: "2022-10-27T07:49:28Z"
          description: The Function Mesh Operator manages the Pulsar Functions and
            Connectors deployed on a Kubernetes cluster.
          olm.properties: '[{"type":"olm.maxOpenShiftVersion","value":"4.8"}]'
          olm.skipRange: <0.7.0
          operatorhub.io/ui-metadata-max-k8s-version: "1.21"
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/streamnative/function-mesh
          support: StreamNative
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FunctionMesh is the Schema for the functionmeshes API
            displayName: Mesh
            kind: FunctionMesh
            name: functionmeshes.compute.functionmesh.io
            version: v1alpha1
          - description: Function is the Schema for the functions API
            displayName: Function
            kind: Function
            name: functions.compute.functionmesh.io
            version: v1alpha1
          - description: Sink is the Schema for the sinks API
            displayName: Sink
            kind: Sink
            name: sinks.compute.functionmesh.io
            version: v1alpha1
          - description: Source is the Schema for the sources API
            displayName: Source
            kind: Source
            name: sources.compute.functionmesh.io
            version: v1alpha1
        description: |
          [Function Mesh](https://functionmesh.io/) is a serverless framework purpose-built for stream processing applications. It brings powerful event-streaming capabilities to your applications by orchestrating multiple [Pulsar Functions](http://pulsar.apache.org/docs/en/next/functions-overview/) and [Pulsar IO connectors](http://pulsar.apache.org/docs/en/next/io-overview/) for complex stream processing jobs on Kubernetes.
          Prerequisites:
          - Install cert-manager operator first.
        displayName: Function Mesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pulsar
        - function-mesh
        - serverless
        - pulsar-functions
        - pulsar-io-connectors
        - StreamNative
        links:
        - name: Function Mesh
          url: https://function-mesh.io
        maintainers:
        - email: function-mesh@streamnative.io
          name: Function Mesh
        maturity: alpha
        minKubeVersion: v1.17.0
        provider:
          name: StreamNative
          url: https://streamnative.io
        relatedImages:
        - quay.io/streamnativeio/function-mesh@sha256:3e13e5cc3ac4d7387ee0adc02cb5183aad6cf2301f76ab230482609c4da1c0b7
        - registry.connect.redhat.com/streamnative/function-mesh-bundle@sha256:e4377187c20a70d768a253bc0794f3152aa6fcee22ddbf2fc88f45756e9681fe
        - docker.cloudsmith.io/streamnative/mirrors/gcr.io/kubebuilder/kube-rbac-proxy@sha256:67ecb332573384515406ebd71816781366b70adb0eb66345e5980e92603373e1
        version: 0.7.0
      entries:
      - name: function-mesh.v0.7.0
        version: 0.7.0
      name: alpha
    defaultChannel: alpha
    packageName: function-mesh
    provider:
      name: StreamNative
      url: https://streamnative.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-keycloak-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:43:29Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KCOSClientRoleMapping
            name: kcosclientrolemappings.keycloak.org
            version: v1alpha1
          - kind: KCOSRealmRole
            name: kcosrealmroles.keycloak.org
            version: v1alpha1
          - kind: KCOSUserRoleMapping
            name: kcosuserrolemappings.keycloak.org
            version: v1alpha1
          - kind: KeycloakBackup
            name: keycloakbackups.keycloak.org
            version: v1alpha1
          - kind: KeycloakClient
            name: keycloakclients.keycloak.org
            version: v1alpha1
          - kind: KeycloakRealm
            name: keycloakrealms.keycloak.org
            version: v1alpha1
          - kind: Keycloak
            name: keycloaks.keycloak.org
            version: v1alpha1
          - kind: KeycloakUser
            name: keycloakusers.keycloak.org
            version: v1alpha1
        description: Keysight Keycloak Operator
        displayName: Keysight Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Keycloak Operator
          url: https://keysight-keycloak-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-init-container@sha256:90965beaf41c760cb8c8bf906acf08f8fe4770d313de8dfd92d18d7c25348a38
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak-postgresql@sha256:b636955199065e18f9d4679835ae78214843258e5abb4ae758288d6f00df2f8e
        - prod.harbor.keysight.digital/loadcore/kcos-keycloak@sha256:772f73cce6a77b9c46f8035faac99a28602219657d18b75b19b4d905d91942c9
        - prod.harbor.keysight.digital/loadcore/keysight-keycloak-operator@sha256:d74e7e98a114768a6f3327edabce9336743a078844fffee985053b6a90100da8
        - registry.connect.redhat.com/loadcore/loadcore@sha256:7fca6883b092ac7955bb42f893f163403960c81b83e360cfd1eac802e57ea839
        version: 0.1.21
      entries:
      - name: keysight-keycloak-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-keycloak-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Tigera
      provider-url: https://tigera.io/
    name: tigera-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tigera-operator.v1.20.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
          createdAt: "2021-10-05T20:02:14.564507Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.20.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:bd93e66be5cfbc123fe4e8c3f6e8ddf7654217bf26136f06dcf23f28b53063d2
        - registry.connect.redhat.com/tigera/operator@sha256:4d0d814fcfeb628aae8a05cb7a0a416ae85395197950085c907b6cdcf2fb4317
        version: 1.20.4
      entries:
      - name: tigera-operator.v1.20.4
        version: 1.20.4
      - name: tigera-operator.v1.20.3
        version: 1.20.3
      - name: tigera-operator.v1.20.2
        version: 1.20.2
      - name: tigera-operator.v1.20.1
        version: 1.20.1
      name: release-v1.20
    - currentCSV: tigera-operator.v1.23.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
          createdAt: "2022-01-12T22:55:57.409244Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.23.4
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:f258275fe3d48c2ae16346a57f628523a6f2fb7b854b6d43b169056e158537f1
        - registry.connect.redhat.com/tigera/operator@sha256:b1a8b56ff5e1bb5326e1bde4931303358410da9768f0ebf9c08e2c2d8661d742
        version: 1.23.4
      entries:
      - name: tigera-operator.v1.23.4
        version: 1.23.4
      - name: tigera-operator.v1.23.3
        version: 1.23.3
      - name: tigera-operator.v1.23.2
        version: 1.23.2
      - name: tigera-operator.v1.23.1
        version: 1.23.1
      - name: tigera-operator.v1.23.0
        version: 1.23.0
      name: release-v1.23
    - currentCSV: tigera-operator.v1.25.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
          createdAt: "2022-02-17T05:39:13.720885Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.25.1
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:e862708c700df55f99a9c9bb3c453e31f22aef1993d6079d80433eb8da831c94
        - registry.connect.redhat.com/tigera/operator@sha256:cb7f3367ff064cd01dc636a43fd89d5f6c4e6a7004cb7c94b1758b2398e36d71
        version: 1.25.1
      entries:
      - name: tigera-operator.v1.25.1
        version: 1.25.1
      name: release-v1.25
    - currentCSV: tigera-operator.v1.28.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
          createdAt: "2023-05-17T16:13:52.148441156Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.28.13
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:9e8e47045fc67a2b00f9f2fd575fa526d7ddcac50af0b8b6d583b7518ccce120
        - registry.connect.redhat.com/tigera/operator@sha256:de9724b84ca493d3b14c8f801c1049ac4713166b00d9bf3ea0b1986aac5040fa
        version: 1.28.13
      entries:
      - name: tigera-operator.v1.28.13
        version: 1.28.13
      name: release-v1.28
    - currentCSV: tigera-operator.v1.29.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
          createdAt: "2023-03-31T20:49:52.518230079Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.29.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:2a5acd29d8dca6adb318af04dbb15d93d899ed2b28b2ee9aaac09147395efb58
        - registry.connect.redhat.com/tigera/operator@sha256:8e0c026f369f80e8b36d0c94f1fd60f7f53656e57dc5be0b8b12189546874909
        version: 1.29.3
      entries:
      - name: tigera-operator.v1.29.3
        version: 1.29.3
      name: release-v1.29
    - currentCSV: tigera-operator.v1.30.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
          createdAt: "2023-11-17T00:52:31.263564294Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.30.9
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: APIServer
            name: apiservers.operator.tigera.io
            version: v1
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - quay.io/tigera/operator@sha256:431f037ff18b5c867d01312e42671effc55602421aeed25dd3f6109f70596b4a
        - registry.connect.redhat.com/tigera/operator@sha256:1fb1007f7aad744f0e1b0f6a32022c2571ee8558e9db53c85976bbd2a891a6b0
        version: 1.30.9
      entries:
      - name: tigera-operator.v1.30.9
        version: 1.30.9
      name: release-v1.30
    - currentCSV: tigera-operator.v1.32.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Logging & Tracing, Monitoring, Networking, Security
          certified: "true"
          containerImage: quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
          createdAt: "2023-12-15T23:45:03.964812504Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/action-text: Install-time Instructions
          marketplace.openshift.io/remote-workflow: https://docs.projectcalico.org/getting-started/openshift/installation
          olm.skipRange: <1.32.3
          operators.operatorframework.io/builder: operator-sdk-v1.0.1
          operators.operatorframework.io/internal-objects: '["blockaffinities.crd.projectcalico.org","clusterinformations.crd.projectcalico.org","ipamblocks.crd.projectcalico.org","ipamconfigs.crd.projectcalico.org","ipamhandles.crd.projectcalico.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: BGPConfiguration represents BGP specific configuration options
              for the cluster or a specific node.
            displayName: BGPConfiguration
            kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - description: BGPPeer represents a remote BGP peer with which the node(s)
              in a Calico cluster will peer. Configuring BGP peers allows you to peer
              a Calico network with your datacenter fabric (e.g. ToR).
            displayName: BGPPeer
            kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - description: BlockAffinity maintains a block affinity's state.
            displayName: BlockAffinity
            kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - description: ClusterInformation contains the cluster specific information.
            displayName: ClusterInformation
            kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - description: FelixConfiguration represents Felix configuration options
              for the cluster.
            displayName: FelixConfiguration
            kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkPolicy represents an ordered set of rules which
              are applied to a collection of endpoints that match a label selector.
            displayName: GlobalNetworkPolicy
            kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - description: GlobalNetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: GlobalNetworkSet
            kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - description: HostEndpoint represents one or more real or virtual interfaces
              attached to a host that is running Calico. It enforces Calico policy
              on the traffic that is entering or leaving the host’s default network
              namespace through those interfaces.
            displayName: HostEndpoint
            kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - description: Installation configures an installation of Calico or Calico
              Enterprise. At most one instance of this resource is supported. It must
              be named "default". The Installation API installs core networking and
              network policy components, and provides general install-time configuration.
            displayName: Installation
            kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - description: IPAMBlock contains information about a block for IP address
              assignment.
            displayName: IPAMBlock
            kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - description: IPAMConfig contains information about a block for IP address
              assignment.
            displayName: IPAMConfig
            kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - description: IPAMHandle contains information about an IPAMHandle resource.
            displayName: IPAMHandle
            kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - description: IPPool represents a collection of IP addresses from which
              Calico expects endpoint IPs to be assigned.
            displayName: IPPool
            kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - description: KubeControllersConfiguration represents configuration options
              for the Calico Kubernetes controllers.
            displayName: KubeControllersConfiguration
            kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - description: NetworkPolicy represents an ordered set of rules which are
              applied to a collection of endpoints that match a label selector.
            displayName: NetworkPolicy
            kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - description: NetworkSet represents an arbitrary set of IP subnetworks/CIDRs,
              allowing it to be matched by Calico policy. Network sets are useful
              for applying policy to traffic coming from (or going to) external, non-Calico,
              networks.
            displayName: NetworkSet
            kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - description: TigeraStatus represents the most recently observed status
              for Calico or a Calico Enterprise functional area.
            displayName: TigeraStatus
            kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: 'An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift. Its goal is to make
          installation, upgrades, and ongoing lifecycle management of Calico and Calico
          Enterprise as simple and reliable as possible. **Important**: this operator
          should only be installed if the cluster was already provisioned with Calico.'
        displayName: Tigera Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - security
        - monitoring
        links:
        - name: Calico Introduction
          url: https://docs.projectcalico.org/introduction/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.projectcalico.org/getting-started/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
        relatedImages:
        - registry.connect.redhat.com/tigera/operator@sha256:42a4d0d0e24752dbb51daebcfca4d4850d04e79578bfb8b41cdc2a2a289df4a2
        - quay.io/tigera/operator@sha256:715ac9a30f8a9579e44258af20de354715429e11836b493918e9e1a696e9b028
        version: 1.32.3
      entries:
      - name: tigera-operator.v1.32.3
        version: 1.32.3
      name: release-v1.32
    - currentCSV: tigera-operator.v1.38.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.tigera.io/v1",
                "kind": "Installation",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "calicoNetwork": {
                    "linuxDataplane": "BPF"
                  },
                  "variant": "Calico"
                }
              }
            ]
          capabilities: Basic Install
          categories: Logging & Tracing, Monitoring, Networking, Security
          containerImage: quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
          createdAt: "2025-05-06T13:27:07.802897329Z"
          description: An operator which manages the lifecycle of a Calico or Calico
            Enterprise installation on Kubernetes or OpenShift.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          features.operators.openshift.io/valid-subscription: No subscription required
          olm.skipRange: <1.38.0
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/tigera/operator
          support: Tigera
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: BGPConfiguration
            name: bgpconfigurations.crd.projectcalico.org
            version: v1
          - kind: BGPPeer
            name: bgppeers.crd.projectcalico.org
            version: v1
          - kind: BlockAffinity
            name: blockaffinities.crd.projectcalico.org
            version: v1
          - kind: CalicoNodeStatus
            name: caliconodestatuses.crd.projectcalico.org
            version: v1
          - kind: ClusterInformation
            name: clusterinformations.crd.projectcalico.org
            version: v1
          - kind: FelixConfiguration
            name: felixconfigurations.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkPolicy
            name: globalnetworkpolicies.crd.projectcalico.org
            version: v1
          - kind: GlobalNetworkSet
            name: globalnetworksets.crd.projectcalico.org
            version: v1
          - kind: HostEndpoint
            name: hostendpoints.crd.projectcalico.org
            version: v1
          - kind: IPAMBlock
            name: ipamblocks.crd.projectcalico.org
            version: v1
          - kind: IPAMConfig
            name: ipamconfigs.crd.projectcalico.org
            version: v1
          - kind: IPAMHandle
            name: ipamhandles.crd.projectcalico.org
            version: v1
          - kind: IPPool
            name: ippools.crd.projectcalico.org
            version: v1
          - kind: IPReservation
            name: ipreservations.crd.projectcalico.org
            version: v1
          - kind: KubeControllersConfiguration
            name: kubecontrollersconfigurations.crd.projectcalico.org
            version: v1
          - kind: NetworkPolicy
            name: networkpolicies.crd.projectcalico.org
            version: v1
          - kind: NetworkSet
            name: networksets.crd.projectcalico.org
            version: v1
          - kind: ImageSet
            name: imagesets.operator.tigera.io
            version: v1
          - kind: Installation
            name: installations.operator.tigera.io
            version: v1
          - kind: TigeraStatus
            name: tigerastatuses.operator.tigera.io
            version: v1
        description: An operator which manages the lifecycle of a Calico or Calico
          Enterprise installation on Kubernetes or OpenShift.
        displayName: Tigera Operator v1.38
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tigera-operator
        - networking
        - security
        - monitoring
        links:
        - name: Tigera
          url: https://tigera.io/
        - name: Calico Introduction
          url: https://docs.tigera.io/calico/latest/about/
        - name: Install an OpenShift 4 cluster with Calico
          url: https://docs.tigera.io/calico/latest/getting-started/kubernetes/openshift/installation
        maintainers:
        - email: maintainers@tigera.io
          name: Project Calico Maintainers
        maturity: stable
        provider:
          name: Tigera
          url: https://tigera.io/
        relatedImages:
        - quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775
        - registry.connect.redhat.com/tigera/operator@sha256:7894b770bb02b2271e1414d67f300218cc646534c9f2a6917e51dc8bff8ba731
        version: 1.38.0
      entries:
      - name: tigera-operator.v1.38.0
        version: 1.38.0
      name: release-v1.38
    defaultChannel: release-v1.38
    packageName: tigera-operator
    provider:
      name: Tigera
      url: https://tigera.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: awss3-operator-registry
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: awss3operator.1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/screeley44/aws-s3-provisioner:v1.0.0
          createdAt: "2019-05-02 16:12:00"
          description: Manage the full lifecycle of installing, configuring and managing
            AWS S3 Provisioner.
          repository: https://github.com/yard-turkey/aws-s3-provisioner
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - description: instance of an AWS S3 Bucket
            displayName: Object Bucket
            kind: ObjectBucket
            name: objectbuckets.objectbucket.io
            version: v1alpha1
          - description: Request for an AWS S3 Bucket
            displayName: Object Bucket Claim
            kind: ObjectBucketClaim
            name: objectbucketclaims.objectbucket.io
            version: v1alpha1
        description: |
          AWS S3 Operator will deploy the AWS S3 Provisioner which will dynamically or statically
          provision AWS S3 Bucket storage and access. The operator deploys the ObjectBucket (OB) and ObjectBucketClaim (OBC)
          CustomResourceDefinitions. The OB/OBC model follows the traditional Kubernetes PV/PVC pattern, when an OBC is detected
          the operator will act on the OBC to either provision a brand new S3 Bucket in AWS or gain access to an existing
          S3 Bucket in AWS. The operator produces an OB and ConfigMap and Secret which can then be consumed by application pods.

          **Important Note**: Currently, while in preview, this operator does not
          support automatic upgrades. You must remove the old version of the operator
          manually before installing a new version.
          ## Using AWS S3 Operator

          ### Administrator Creates Secret
          This secret will contain the elevated/admin privileges needed by the provisioner
          to properly access and create S3 Buckets and IAM users and policies. The AWS Access ID
          and AWS Secret Key will be needed for this.
          1. Create the Kubernetes Secret for the Provisioner's Owner Access.
          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: s3-bucket-owner [1]
            namespace: s3-provisioner [2]
          type: Opaque
          data:
            AWS_ACCESS_KEY_ID: *base64 encoded value* [3]
            AWS_SECRET_ACCESS_KEY: *base64 encoded value* [4]
          ```
          1. Name of the secret, this will be referenced in StorageClass.
          1. Namespace where the Secret will exist.
          1. Your AWS_ACCESS_KEY_ID base64 encoded.
          1. Your AWS_SECRET_ACCESS_KEY base64 encoded.
          ```
           # kubectl create -f creds.yaml
          secret/s3-bucket-owner created
          ```
          ### Administrator Creates StorageClass
          The StorageClass defines the name of the provisioner and holds other properties that are needed to provision a new bucket, including
          the Owner Secret and Namespace, and the AWS Region.
          #### Greenfield Example:
          For Greenfield, a new, dynamic bucket will be generated.
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            region: us-west-1 [3]
            secretName: s3-bucket-owner [4]
            secretNamespace: s3-provisioner [5]
          reclaimPolicy: Delete [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace where the Secret will exist
          1. reclaimPolicy (Delete or Retain) indicates if the bucket can be deleted when the OBC is deleted.
          **NOTE:** the absence of the `bucketName` Parameter key in the storage class indicates this is a new bucket and its name is based on the bucket name fields in the OBC.
          ```
           # kubectl create -f storageclass-greenfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          #### Brownfield Example:
          For brownfield, the StorageClass defines the name of the provisioner and the name of the existing bucket. It also includes other properties needed by the target
          provisioner, including: the Owner Secret and Namespace, and the AWS Region
          1. Create the Kubernetes StorageClass for the Provisioner.
          ```yaml
          kind: StorageClass
          apiVersion: storage.k8s.io/v1
          metadata:
            name: s3-existing-buckets [1]
          provisioner: aws-s3.io/bucket [2]
          parameters:
            bucketName: my-existing-bucket [3]
            region: us-west-1 [4]
            secretName: s3-bucket-owner [5]
            secretNamespace: s3-provisioner [6]
          ```
          1. Name of the StorageClass, this will be referenced in the User ObjectBucketClaim.
          1. Provisioner name
          1. Name of the existing bucket
          1. AWS Region that the StorageClass will serve
          1. Name of the bucket owner Secret created above
          1. Namespace for that bucket owner secret
          **NOTE:** the storage class's `reclaimPolicy` is ignored for existing buckets.
          ```
           # kubectl create -f storageclass-brownfield.yaml
          storageclass.storage.k8s.io/s3-buckets created
          ```
          ### User Creates ObjectBucketClaim
          An ObjectBucketClaim follows the same concept as a PVC, in that
          it is a request for Object Storage, the user doesn't need to
          concern him/herself with the underlying storage, just that
          they need access to it. The user will work with the cluster/storage
          administrator to get the proper StorageClass needed and will
          then request access via the OBC.
          #### Greenfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            generateBucketName: mybucket [3]
            bucketName: my-awesome-bucket [4]
            storageClassName: s3-buckets [5]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. Name prepended to a random string used to generate a bucket name. It is ignored if bucketName is defined
          1. Name of new bucket which must be unique across all AWS regions, otherwise an error occurs when creating the bucket. If present, this name overrides `generateName`
          1. StorageClass name
          **NOTE:** if both `generateBucketName` and `bucketName` are omitted, and the storage class does _not_ define a bucket name, then a new, random bucket name is generated with no prefix.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          #### Brownfield Request Example:
          1. Create the ObjectBucketClaim.
          ```yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucketClaim
          metadata:
            name: myobc [1]
            namespace: s3-provisioner [2]
          spec:
            storageClassName: s3-existing-buckets [3]
          ```
          1. Name of the OBC
          1. Namespace of the OBC
          1. StorageClass name
          **NOTE:** in the OBC here there is no reference to the bucket's name. This is defined in the storage class and is not a concern of the user creating the claim to this bucket.  An OBC does have fields for defining a bucket name for greenfield use only.
          ```
           # kubectl create -f obc-brownfield.yaml
          objectbucketclaim.objectbucket.io/myobc created
          ```
          ### Results and Recap
          Let's pause for a moment and digest what just happened.
          After creating the OBC, and assuming the S3 provisioner is running, we now have
          the following Kubernetes resources:
          .  a global ObjectBucket (OB) which contains: bucket endpoint info (including region and bucket name), a reference to the OBC, and a reference to the storage class. Unique to S3, the OB also contains the bucket Amazon Resource Name (ARN).Note: there is always a 1:1 relationship between an OBC and an OB.
          .  a ConfigMap in the same namespace as the OBC, which contains the same endpoint data found in the OB.
          .  a Secret in the same namespace as the OBC, which contains the AWS key-pairs needed to access the bucket.
          And of course, we have a *new* AWS S3 Bucket which you should be able to see via the AWS Console.
          *ObjectBucket*
          ```yaml
           # kubectl get ob obc-s3-provisioner-my-awesome-bucket -o yaml
          apiVersion: objectbucket.io/v1alpha1
          kind: ObjectBucket
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            generation: 1
            name: obc-s3-provisioner-my-awesome-bucket
            resourceVersion: "15057"
            selfLink: /apis/objectbucket.io/v1alpha1/objectbuckets/obc-s3-provisioner-my-awesome-bucket
            uid: 0bfe8e84-576d-4c4e-984b-f73c4460f736
          spec:
            Connection:
              additionalState:
                ARN: arn:aws:iam::<accountid>:policy/my-awesome-bucket-vSgD5 [1]
                UserName: my-awesome-bucket-vSgD5 [2]
              endpoint:
                additionalConfig: null
                bucketHost: s3-us-west-1.amazonaws.com
                bucketName: my-awesome-bucket [3]
                bucketPort: 443
                region: us-west-1
                ssl: true
                subRegion: ""
            claimRef: null [4]
            reclaimPolicy: null
            storageClassName: s3-buckets [5]
          ```
          1. The AWS Policy created for this user and bucket.
          1. The new user generated by the Provisioner to access this existing bucket.
          1. The bucket name.
          1. The reference to the OBC (not filled in yet).
          1. The reference to the StorageClass used.
          *ConfigMap*
          ```yaml
           # kubectl get cm myobc -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            BUCKET_HOST: s3-us-west-1.amazonaws.com [1]
            BUCKET_NAME: my-awesome-bucket [2]
            BUCKET_PORT: "443"
            BUCKET_REGION: us-west-1
            BUCKET_SSL: "true"
            BUCKET_SUBREGION: ""
          kind: ConfigMap
          metadata:
            creationTimestamp: "2019-04-01T19:11:38Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "892"
            selfLink: /api/v1/namespaces/s3-provisioner/configmaps/my-awesome-bucket
            uid: 2edcc58a-aff8-4a29-814a-ffbb6439a9cd
          ```
          1. The AWS S3 host.
          1. The name of the new bucket we are gaining access to.
          *Secret*
          ```yaml
           # kubectl get secret my-awesome-bucket -n s3-provisioner -o yaml
          apiVersion: v1
          data:
            AWS_ACCESS_KEY_ID: *the_new_access_id* [1]
            AWS_SECRET_ACCESS_KEY: *the_new_access_key_value* [2]
          kind: Secret
          metadata:
            creationTimestamp: "2019-04-03T15:42:22Z"
            finalizers:
            - objectbucket.io/finalizer
            name: my-awesome-bucket
            namespace: s3-provisioner
            resourceVersion: "15058"
            selfLink: /api/v1/namespaces/s3-provisioner/secrets/screeley-provb-5
            uid: 225c71a5-9d75-4ccc-b41f-bfe91b272a13
          type: Opaque
          ```
          1. The new generated AWS Access Key ID.
          1. The new generated AWS Secret Access Key.
          What happened in AWS? The first thing we do on any OBC request is
          create a new IAM user and generate Access ID and Secret Keys.
          This allows us to also better control ACLs. We also create a policy
          in IAM which we then attach to the user and bucket. We also created a new bucket, called *my-awesome-bucket*.
          When the OBC is deleted all of its Kubernetes and AWS resources will also be deleted, which includes:
          the generated OB, Secret, ConfigMap, IAM user, and policy.
          If the _retainPolicy_ on the StorageClass for this bucket is *"Delete"*, then, in addition to the above cleanup, the physical bucket is also deleted.
          **NOTE:** The actual bucket is only deleted if the Storage Class's _reclaimPolicy_ is "Delete".
          ### User Creates Pod
          Now that we have our bucket and connection/access information, a pod
          can be used to access the bucket. This can be done in several different
          ways, but the key here is that the provisioner has provided the proper
          endpoints and keys to access the bucket. The user then simply references
          the keys.
          1. Create a Sample Pod to Access the Bucket.
          ```yaml
          apiVersion: v1
          kind: Pod
          metadata:
            name: photo1
            labels:
              name: photo1
          spec:
            containers:
            - name: photo1
              image: docker.io/screeley44/photo-gallery:latest
              imagePullPolicy: Always
              envFrom:
              - configMapRef:
                  name: my-awesome-bucket <1>
              - secretRef:
                  name: my-awesome-bucket <2>
              ports:
              - containerPort: 3000
                protocol: TCP
          ```
          1. Name of the generated configmap from the provisioning process
          1. Name of the generated secret from the provisioning process
          *[Note]* Generated ConfigMap and Secret are same name as the OBC!
          Lastly, expose the pod as a service so you can access the url from a browser. In this example,
          I exposed as a LoadBalancer
          ```
            # kubectl expose pod photo1 --type=LoadBalancer --name=photo1 -n your-namespace
          ```
          To access via a url use the EXTERNAL-IP
          ```
            # kubectl get svc photo1
            NAME                         TYPE           CLUSTER-IP       EXTERNAL-IP                                                               PORT(S)          AGE
            photo1                       LoadBalancer   100.66.124.105   a00c53ccb3c5411e9b6550a7c0e50a2a-2010797808.us-east-1.elb.amazonaws.com   3000:32344/TCP   6d
          ```
          **NOTE:** This is just one example of a Pod that can utilize the bucket information,
          there are several ways that these pod applications can be developed and therefore
          the method of getting the actual values needed from the Secrets and ConfigMaps
          will vary greatly, but the idea remains the same, that the pod consumes the generated
          ConfigMap and Secret created by the provisioner.
        displayName: AWS S3 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - awss3provisioner
        - aws
        - S3
        - provisioner
        - buckets
        links:
        - name: AWS S3 Provisioner
          url: https://github.com/yard-turkey/aws-s3-provisioner
        - name: Provisioner Examples
          url: https://github.com/yard-turkey/aws-s3-provisioner/tree/master/examples
        - name: Operator Documentation
          url: https://github.com/yard-turkey/awss3operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Community
        maturity: alpha
        provider:
          name: Community
        relatedImages:
        - quay.io/openshift-community-operators/awss3-operator-registry@sha256:9ab3a2cd00b8704a8bae8e9b75f2a95d0a3e60fa3cb52589db8a4e7641b73027
        - quay.io/screeley44/aws-s3-provisioner:v1.0.0
        version: 1.0.1
      entries:
      - name: awss3operator.1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: awss3-operator-registry
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-odlm
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operand-deployment-lifecycle-manager.v4.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandBindInfo",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "bindings": {
                    "public": {
                      "configmap": "mongodb-configmap",
                      "secret": "mongodb-secret"
                    }
                  },
                  "description": "Binding information that should be accessible to MongoDB adopters",
                  "operand": "mongodb-atlas-kubernetes",
                  "registry": "example-service"
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "services": [
                    {
                      "name": "jaeger",
                      "spec": {
                        "jaeger": {
                          "strategy": "allinone"
                        }
                      }
                    },
                    {
                      "name": "mongodb-atlas-kubernetes",
                      "spec": {
                        "atlasDeployment": {
                          "deploymentSpec": {
                            "name": "test-deployment"
                          },
                          "projectRef": {
                            "name": "my-project"
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRegistry",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "operators": [
                    {
                      "channel": "stable",
                      "installMode": "cluster",
                      "name": "jaeger",
                      "namespace": "default",
                      "packageName": "jaeger",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    },
                    {
                      "channel": "stable",
                      "name": "mongodb-atlas-kubernetes",
                      "namespace": "default",
                      "packageName": "mongodb-atlas-kubernetes",
                      "sourceName": "community-operators",
                      "sourceNamespace": "openshift-marketplace"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "OperandRequest",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/managed-by": "operand-deployment-lifecycle-manager",
                    "app.kubernetes.io/name": "operand-deployment-lifecycle-manager"
                  },
                  "name": "example-service"
                },
                "spec": {
                  "requests": [
                    {
                      "operands": [
                        {
                          "name": "jaeger"
                        },
                        {
                          "name": "mongodb-atlas-kubernetes"
                        }
                      ],
                      "registry": "example-service"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, Monitoring, Logging & Tracing, Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
          createdAt: "2025-02-14T21:16:49Z"
          description: The Operand Deployment Lifecycle Manager provides a Kubernetes
            CRD-based API to manage the lifecycle of operands.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          nss.operator.ibm.com/managed-operators: ibm-odlm
          olm.skipRange: '>=1.2.0 <4.5.0'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/operand-deployment-lifecycle-manager
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperandBindInfo is the Schema for the operandbindinfoes API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandBindInfo
            kind: OperandBindInfo
            name: operandbindinfos.operator.ibm.com
            version: v1alpha1
          - description: OperandConfig is the Schema for the operandconfigs API. Documentation
              For additional details regarding install parameters check https://ibm.biz/icpfs39install.
              License By installing this product you accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandConfig
            kind: OperandConfig
            name: operandconfigs.operator.ibm.com
            version: v1alpha1
          - description: OperandRegistry is the Schema for the operandregistries API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRegistry
            kind: OperandRegistry
            name: operandregistries.operator.ibm.com
            version: v1alpha1
          - description: OperandRequest is the Schema for the operandrequests API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperandRequest
            kind: OperandRequest
            name: operandrequests.operator.ibm.com
            version: v1alpha1
          - description: OperatorConfig is the Schema for the operatorconfigs API.
              Documentation For additional details regarding install parameters check
              https://ibm.biz/icpfs39install. License By installing this product you
              accept the license terms https://ibm.biz/icpfs39license
            displayName: OperatorConfig
            kind: OperatorConfig
            name: operatorconfigs.operator.ibm.com
            version: v1alpha1
        description: |-
          # Introduction

          The operand is the instance managed by the operator. ODLM is used to manage the lifecycle of for a group of operands, compared with operator lifecycle manager, ODLM focus on the management of operands but not operators.

          # Details

          ### What are operands?

          Operator is a method of packaging, deploying and managing a Kubernetes application.
          Operands are the services and applications that Operator manage.

          For example, cert-manager operator deploys a cert-manager deployment, then the cert-manager-operator is an operator and the cert-manager deployment is an operand.

          The ODLM will have four CRDs:

          | Resource                 | Short Name | Description                                                                                |
          |--------------------------|------------|--------------------------------------------------------------------------------------------|
          | OperandRequest | opreq | It defines which operator/operand want to be installed in the cluster |
          | OperandRegistry | opreg | It defines the OLM information, like channel and catalog source, for each operator |
          | OperandConfig | opcon | It defines the parameters that should be used to install the operator's operand |
          | OperandBindInfo | opbi | It identifies secrets and/or configmaps that should be shared with requests |

          ### Goal

          1. A single entry point to manage a group of operands
          1. User can select a set of operands to install
          1. The install can be invoked through either OCP UI or CLI

          ## Prerequisites
          - [operator-sdk] version v1.3.0.
          - [go] version 1.15.7+
          - [oc] version v3.11+ or [kubectl] v1.11.3+
          - Access to an Openshift v4.3+ cluster

          ### Resources Required

          #### Minimum scheduling capacity

          | Software                   | Memory (MB) | CPU (cores) |
          | -------------------------- | ----------- | ----------- |
          | Operand Deployment Lifecycle Manager | 200          | 0.2        |
          | **Total**                  | 200         | 0.2         |

          ## Installing

          ### Create Operator NS, Group, Subscription

          ```yaml
          apiVersion: v1
          kind: Namespace
          metadata:
            name: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha2
          kind: OperatorGroup
          metadata:
            name: operatorgroup
            namespace: odlm
          spec:
            targetNamespaces:
            - odlm

          ---
          apiVersion: v1
          data:
            namespaces: odlm
          kind: ConfigMap
          metadata:
            name: namespace-scope
            namespace: odlm

          ---
          apiVersion: operators.coreos.com/v1alpha1
          kind: Subscription
          metadata:
            name: operand-deployment-lifecycle-manager
            namespace: odlm
          spec:
            channel: v3.20
            name: ibm-odlm
            source: opencloud-operators
            sourceNamespace: olm
            config:
              env:
              - name: INSTALL_SCOPE
                value: namespaced
          END
          ```

          ## Configuration

          OperandRegistry defines the OLM information used for installation, like package name and catalog source, for each operator.

          Following is an example of the OperandRegistry CR:

          **NOTE:** The "spec.operators[*].name" parameter must be unique for each entry.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRegistry
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            operators:
            - name: jenkins [3]
              namespace: default [4]
              channel: alpha [5]
              packageName: jenkins-operator [6]
              scope: public [7]
              sourceName: community-operators [8]
              sourceNamespace: openshift-marketplace [9]
              installMode: cluster [10]
              installPlanApproval: Manual [11]
          ```

          The OperandRegistry Custom Resource (CR) lists OLM Operator information for operands that may be requested for installation and/or access by an application that runs in a namespace. The registry CR specifies:

          1. `name` of the OperandRegistry
          2. `namespace` of the OperandRegistry
          3. `name` is the name of the operator, which should be the same as the services name in the OperandConfig and OperandRequest.
          4. `namespace` defines the namespace where the operator and its CR will be deployed. (1) When InstallMode is `cluster`, the operator will be deployed into the `openshift-operators` namespace and the operator CRs will be deployed into the namespace this parameter defines. (2) When InstallMode is empty or set to `namespace`, it is the namespace where both operator and operator CR will be deployed.
          5. `channel` is the name of OLM channel that is subscribed for the operator.
          6. `packageName` is the name of the package in CatalogSource that is subscribed for the operator.
          7. (optional) `scope` is an indicator, either public or private, that dictates whether deployment can be requested from other namespaces (public) or only from the namespace of this OperandRegistry (private). The default value is private.
          8. `sourceName` is the name of the CatalogSource.
          9. `sourceNamespace` is the namespace of the CatalogSource.
          10. (optional) `installMode` is the install mode of the operator, can be either `namespace` (OLM one namespace) or `cluster` (OLM all namespaces). The default value is `namespace`. Operator is deployed in `openshift-operators` namespace when InstallMode is set to `cluster`.
          11. (optional) `installPlanApproval` is the approval mode for emitted installplan. The default value is `Automatic`.

          ### OperandConfig Spec

          OperandConfig defines the individual operand configuration. The OperandConfig Custom Resource (CR) defines the parameters for each operator that is listed in the OperandRegistry that should be used to install the operator instance by specifying an installation CR.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          Kind: OperandConfigs
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            services:
            - name: jenkins [3]
              spec: [4]
                jenkins:
                  port: 8081
          ```

          OperandConfig defines the individual operand deployment config:

          1. `name` of the OperandConfig
          2. `namespace` of the OperandConfig
          3. `name` is the name of the operator, which should be the same as the services name in the OperandRegistry and OperandRequest.
          4. `spec` defines a map. Its key is the kind name of the custom resource. Its value is merged to the spec field of custom resource. For more details, you can check the following topic **How does ODLM create the individual operator CR?**

          #### How does Operator create the individual operator CR

          Jenkins Operator has one CRD: Jenkins:

          The OperandConfig CR has

          ```yaml
          - name: jenkins
            spec:
              jenkins:
                service:
                  port: 8081
          ```

          The Jenkins Operator CSV has

          ```yaml
          apiVersion: operators.coreos.com/v1alpha1
          kind: ClusterServiceVersion
          metadata:
            annotations:
            alm-examples: |-
              [
              {
                "apiVersion":"jenkins.io/v1alpha2",
                "kind":"Jenkins",
                "metadata": {
                  "name":"example"
                },
                "spec":{
                  ...
                  "service":{"port":8080,"type":"ClusterIP"},
                  ...
                }
              }
            ]
          ```

          The ODLM will deep merge the OperandConfig CR spec and Jenkins Operator CSV alm-examples to create the Jenkins CR.

          ```yaml
          apiVersion: jenkins.io/v1alpha2
          kind: Jenkins
          metadata:
            name: example
          spec:
            ...
            service:
              port: 8081
              type: ClusterIP
            ...
          ```

          For day2 operations, the ODLM will patch the OperandConfigs CR spec to the existing Jenkins CR.

          ### OperandRequest Spec

          OperandRequest defines which operator/operand you want to install in the cluster.

          **NOTE:** OperandRequest custom resource is used to trigger a deployment for Operators and Operands.

          #### OperandRequest sample to create custom resource via OperandConfig

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service [1]
            namespace: example-service-ns [2]
          spec:
            requests:
            - registry: example-service [3]
              registryNamespace: example-service-ns [4]
              operands: [5]
              - name: jenkins [6]
                bindings: [7]
                  public:
                    secret: jenkins-operator-credential [8]
                    configmap: jenkins-operator-base-configuration [9]
          ```

          1. `name` of the OperandRequest
          2. `namespace` of the OperandRequest
          3. `registry` identifies the name of the OperandRegistry CR from which this operand deployment is being requested.
          4. (optional) `registryNamespace` identifies the namespace in which the OperandRegistry CR is defined. **Note:** If the `registryNamespace` is not specified then it is assumed that the OperandRegistry CR is in the current (OperandRequest's) namespace.
          5. `operands` in the CR is a list of operands.
          6. `name` of operands in the CR must match a name specification in an OperandRegistry's CR.
          7. (optional) The `bindings` of the operands is a map to get and rename the secret and/or configmap from the provider and create them in the requester's namespace. If the requester wants to rename the secret and/or configmap, they need to know the key of the binding in the OperandBindInfo. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace.
          8. (optional) `secret` names a secret that should be created in the requester's namespace with formatted data that can be used to interact with the service.
          9. (optional) `configmap` names a configmap that should be created in the requester's namespace with formatted data that can be used to interact with the service.

          #### OperandRequest sample to create custom resource via OperandRequest

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandRequest
          metadata:
            name: example-service
            namespace: example-service-ns
          spec:
            requests:
            - registry: example-service
              registryNamespace: example-service-ns
              operands:
              - name: jenkins
                kind: Jenkins [1]
                apiVersion: "jenkins.io/v1alpha2" [2]
                instanceName: "example" [3]
                spec: [4]
                  service:
                    port: 8081
          ```

          1. `kind` of the target custom resource. If it is set in the operand item, ODLM will create CR according to OperandRequest only and get rid of the alm-example and OperandConfig.
          2. `apiVersion` of the target custom resource.
          3. `instanceName` is the name of the custom resource. If `instanceName` is not set, the name of the custom resource will be created with the name of the OperandRequest as a prefix.
          4. `spec` is the spec field of the target CR.

          ### OperandBindInfo Spec

          The ODLM will use the OperandBindInfo to copy the generated secret and/or configmap to a requester's namespace when a service is requested with the OperandRequest CR. An example specification for an OperandBindInfo CR is shown below.

          ```yaml
          apiVersion: operator.ibm.com/v1alpha1
          kind: OperandBindInfo
          metadata:
            name: publicjenkinsbinding [1]
            namespace: example-service-ns [2]
          spec:
            operand: jenkins [3]
            registry: example-service [4]
            description: "Binding information that should be accessible to jenkins adopters" [5]
            bindings: [6]
              public:
                secret: jenkins-operator-credentials-example [7]
                configmap: jenkins-operator-base-configuration-example [8]
          ```

          Fields in this CR are described below.

          1. `name` of the OperandBindInfo
          2. `namespace` of the OperandBindInfo
          3. The `operand` should be the the individual operator name.
          4. The `registry` section must match the name in the OperandRegistry CR in the current namespace.
          5. `description` is used to add a detailed description of a service.
          6. The `bindings` section is used to specify information about the access/configuration data that is to be shared. If the key of the bindings map is prefixed with public, it means the secret and/or configmap can be shared with the requester in the other namespace. If the key of the bindings map is prefixed with private, it means the secret and/or configmap can only be shared within its own namespace. If the key of the bindings map is prefixed with protected, it means the secret and/or configmap can only be shared if it is explicitly declared in the OperandRequest.
          7. The `secret` field names an existing secret, if any, that has been created and holds information that is to be shared with the requester.
          8. The `configmap` field identifies a configmap object, if any, that should be shared with the requester

          ODLM will use the OperandBindInfo CR to pass information to an adopter when they create a OperandRequest to access the service, assuming that both have compatible scopes. ODLM will copy the information from the shared service's "OperandBindInfo.bindinfo[].secret" and/or "OperandBindInfo.bindinfo[].configmap" to the requester namespace.

          **NOTE:** If in the OperandRequest, there is no secret and/or configmap name specified in the bindings or no bindings field in the element of operands, ODLM will copy the secret and/or configmap to the requester's namespace and rename them to the name of the OperandBindInfo + secret/configmap name.

          ### E2E Use Case

          1. User installs ODLM from OLM

              The ODLM will automatically generate two default CRD CRs, since OperandRegistry and OperandConfigs CRs don't define the state, so it should be fine.

          2. **Optional:** Users can update the OperandConfigs CR with their own parameter values

              Users can skip this step if the default configurations match the requirement.

          3. User creates the OperandRequest CR from OLM

              This tells ODLM that users want to install some of the individual operator/operand.

          4. The rest works will be done by OLM and ODLM

              Finally, users will get what they want.

          ### Operator/Operand Upgrade

          - For operator/operand upgrade, you only need to publish your operator OLM to your operator channel, and OLM will handle the upgrade automatically.
          - If there are major version, then you may want to update `channel` in `OperandRegistry` to trigger upgrade.

          ### Limitations

          ODLM depends on NamespaceScope Operator

          ### SecurityContextConstraints Requirements

          The Operand Deployment Lifecycle Manager supports running under the OpenShift Container Platform default restricted security context constraints.
        displayName: Operand Deployment Lifecycle Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Operand Deployment Lifecycle Manager Project
          url: https://github.com/IBM/operand-deployment-lifecycle-manager
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-cs-install-catalog@sha256:d4dff1a8968c18d984e16c4c9f1517c3a10e23a11bf166c64d734773324d4c9d
        - registry.marketplace.redhat.com/cpopen/odlm@sha256:64d35613a927d46dba3aaed44a4960245aad75dc40c82a3866e875639e20a615
        - registry.marketplace.redhat.com/cpopen/operand-deployment-lifecycle-manager-bundle@sha256:68f71f2767a40ca9333f0a670e07493e8a8c13a5c7ca877ea5bd8a544b352d9a
        - registry.connect.redhat.com/cpopen/odlm@sha256:064b597625550ca75dbc9f3c8fd6b2de6e0b9cb7f42d604e871bfc668ad05ac1
        - registry.marketplace.redhat.com/cpopen/cpfs/cpfs-utils@sha256:27195e9758afdf8521971866fd60ea3dafed91a146c3c4165876e5ea476a4622
        - registry.marketplace.redhat.com/cpopen/cpfs/must-gather@sha256:bb074bdf89cbc7c703fadb17639608599ab4f66db9be5617e5f79a0550676790
        version: 4.5.0
      entries:
      - name: operand-deployment-lifecycle-manager.v4.5.0
        version: 4.5.0
      name: v4.5
    defaultChannel: v4.5
    packageName: ibm-odlm
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-team-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-team-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
          createdAt: "2025-04-16T06:57:55Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.4.5
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:60fb1f76b4d4aab569eec5f2c20f7a06d813759f5fb6ac952156b41c4d295a52
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:eccacdf9ac2b0b1f6061aea457c705319f268bfa30631ed9919ae393f0ef7f71
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-team-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-team-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
          createdAt: "2025-07-24T14:54:58Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:052055a6280f60a69caccf89f50fd8bf92da61cd59f1962f61d21467c4191970
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:cf330a9e405a9641c213a74baf28307a6e247c4f03cc0523d071dad8f5f9e2e9
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        version: 2025.5.4-12
      entries:
      - name: uma-team-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-team-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
          createdAt: "2025-08-23T09:01:14Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.7.1.23,990023"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:e3ae2240aeb618f36f41073a5f65b03a7ad5dfdccc7f9feb830506fbd2bfd604
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:412403b096566965b2c92a0a55beed8ebd6d8113a7e1fdf5eb3334769929ca41
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-team-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-team-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
          createdAt: "2025-08-15T03:20:57Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-operator@sha256:b8422d0f3a9e576948ced8244a91abc3283b26217e4c7efc7e2669ca5508930e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:c89d8be470c2fd052c5be7a7dbfb133c9e3528bfa5b3681ffb8b3223c4c6bac2
        version: 2025.8.2-9
      entries:
      - name: uma-team-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-team-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true,
                      "hostmonitoring": {
                        "linux": {
                          "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                        }
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
          createdAt: "2025-07-30T08:27:50Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               tenantID: ""

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             role: team
             namespaces: ""
             adminNamespaceName: broadcom-aiops

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               node:
                 enabled: true
                 hostmonitoring:
                   linux:
                   # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                   groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                 # The below properties define the resources for 'broadcom-monitor-host' deployment that does host monitoring.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA microservice apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:8c12315745ad3564035a4db923698bf63fe32dddbc7d246f5a3867945305713e
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:6c9e04df3f5ddc45788333a85f6aeff8953b008a4b165e9cc2d5a079c6c73ece
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        version: 2025.7.2-19
      entries:
      - name: uma-team-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-team-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
          createdAt: "2025-09-11T11:16:37Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.8,990008"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        - registry.connect.redhat.com/ca/uma-operator@sha256:89e0505b5e526beffda9b13a1ca7b49e5ba144fd0166a7e171e9d9b084915a56
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:5611a2fd53859d55d5dfdfdd117e80a0f30be0a4489ad838d571e872d3b5158a
        version: 2025.9.0-8
      entries:
      - name: uma-team-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-team-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com1/v1alpha1",
                "kind": "UniversalMonitoringAgentsTeam",
                "metadata": {
                  "name": "uma-monitor-team"
                },
                "spec": {
                  "adminNamespaceName": "dx-uma",
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "url": "localhost:5001"
                  },
                  "agentNaming": {
                    "deployment": {
                      "apmia": {
                        "agent": "",
                        "host": "",
                        "process": ""
                      }
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true"
                  },
                  "monitor": {
                    "application": {
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "2048Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "800Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "enabled": true
                      }
                    },
                    "events": {
                      "enabled": false
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    }
                  },
                  "namespaces": "",
                  "role": "team",
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
          createdAt: "2025-09-24T12:24:28Z"
          description: DX APM Universal Monitoring Agent Team For Multi-tenancy acts
            as single deployment that automatically discovers and monitors cloud and
            container infrastructures and containerized application processes given
            namespaces.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-team-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Team Universal Monitoring Agent for Kubernetes monitoring.
            displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
            kind: UniversalMonitoringAgentsTeam
            name: universalmonitoringagentsteams.ca.broadcom.com1
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Team For Multi-tenancy** installed by Application Administrator. Each application team that uses a DX APM tenant installs the UMA Team capability, specifying the namespaces the team owns or wants to monitor. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment(DX APM Universal Monitoring Agent Team For Multi-tenancy). The UMA Team capability reports the application, pod, and node performance metrics only for the specified namespaces. This capability limits your application teams to seeing performance metrics only for their applications running in different environments on the tenant. Example environments include development, verification, pre-production, and production.

          **Prerequisite:**
             1. The **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator needs to be installed by the Cluster Administrator before UMA Team is installed.
             2. UMA Team acts as a client, using a REST API to pull the namespace performance data from the UMA Admin deployment. So any network policy that restricts the communication to the Services running in the UMA Admin namespace, needs to be updated to allow the traffic from UMA team.

          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com1/v1alpha1
           kind: UniversalMonitoringAgentsTeam
           metadata:
             name: uma-monitor-team
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"

             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             role: team
             namespaces: ""
             adminNamespaceName: dx-uma

             monitor:
               application:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
               container:
                 dockerstats:
                   enabled: true
               clusterPerformance:
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'broadcom-monitor' deployment that does cluster and container monitoring.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "2048Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "800Mi"
                   nassClientEnabled: false
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
             agentNaming:
               deployment:
                 apmia:
                   host:
                   process:
                   agent:

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               buildTag: "25.9.1.12,990012"
               daemonset:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []

          ```
        displayName: DX APM Universal Monitoring Agent Team For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - uma-team-operator
        - uma-operator
        links:
        - name: Intro to UMA
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS/implementing-agents/Universal-Monitoring-Agent.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-team-operator-bundle@sha256:9f56876a1ae10a8be23beb65475a1c597d808d9ddf19da347e8d7823d56081d5
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        - registry.connect.redhat.com/ca/uma-operator@sha256:5430695435a756144289b5e96903597dd255771e1b37e4fdfbcc0e3c93847e41
        version: 2025.9.1-12
      entries:
      - name: uma-team-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-team-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://horreum.hyperfoil.io
    name: horreum-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: horreum-operator.v0.7.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha1",
                "kind": "Horreum",
                "metadata": {
                  "name": "horreum"
                },
                "spec": {
                  "nodeHost": "127.0.0.1"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/hyperfoil/horreum-operator:0.7.9
          createdAt: "2023-06-13T10:42:39Z"
          description: Performance results repository
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/hyperfoil/horreum-operator
          support: Red Hat Application Runtimes Performance Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Horreum is the object configuring Horreum performance results
              repository
            displayName: Horreum
            kind: Horreum
            name: horreums.hyperfoil.io
            version: v1alpha1
        description: Performance results repository
        displayName: Horreum
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - performance
        - repository
        - database
        - results
        - hyperfoil
        - benchmark
        links:
        - name: Horreum Operator
          url: https://horreum-operator.domain
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        - email: wreicher@redhat.com
          name: Will Reichert
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat, Inc.
          url: https://horreum.hyperfoil.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/hyperfoil/horreum-operator:0.7.9
        - quay.io/openshift-community-operators/horreum-operator@sha256:a9d82ce368b015e45b93f917d4069978a0ba323f3dd684d470cdfd262e0c9f05
        version: 0.7.9
      entries:
      - name: horreum-operator.v0.7.9
        version: 0.7.9
      - name: horreum-operator.v0.2.0
        version: 0.2.0
      name: alpha
    defaultChannel: alpha
    packageName: horreum-operator
    provider:
      name: Red Hat, Inc.
      url: https://horreum.hyperfoil.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Eclipse Foundation
      provider-url: ""
    name: eclipse-che
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-che.v7.109.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "org.eclipse.che/v1",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "auth": {
                    "identityProviderURL": "",
                    "oAuthClientName": "",
                    "oAuthSecret": ""
                  },
                  "database": {
                    "externalDb": false
                  },
                  "k8s": {
                    "ingressDomain": null,
                    "tlsSecretName": null
                  },
                  "metrics": {
                    "enable": true
                  },
                  "server": {
                    "externalDevfileRegistries": [
                      {
                        "url": "https://registry.devfile.io"
                      }
                    ],
                    "externalDevfileRegistry": true,
                    "externalPluginRegistry": true,
                    "workspaceNamespaceDefault": "<username>-che"
                  },
                  "storage": {
                    "pvcStrategy": "common"
                  }
                }
              },
              {
                "apiVersion": "org.eclipse.che/v2",
                "kind": "CheCluster",
                "metadata": {
                  "name": "eclipse-che",
                  "namespace": "eclipse-che"
                },
                "spec": {
                  "components": {
                    "devfileRegistry": {
                      "disableInternalRegistry": true,
                      "externalDevfileRegistries": [
                        {
                          "url": "https://registry.devfile.io"
                        }
                      ]
                    },
                    "pluginRegistry": {
                      "disableInternalRegistry": true
                    }
                  },
                  "containerRegistry": {},
                  "devEnvironments": {},
                  "gitServices": {},
                  "networking": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
          createdAt: "2025-09-18T09:13:25Z"
          description: A Kube-native development solution that delivers portable and
            collaborative developer workspaces.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/eclipse-che/che-operator
          support: Eclipse Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              The `CheCluster` custom resource allows defining and managing Eclipse Che server installation.
              Based on these settings, the  Operator automatically creates and maintains several ConfigMaps:
              `che`, `plugin-registry` that will contain the appropriate environment variables
              of the various components of the installation. These generated ConfigMaps must NOT be updated manually.
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v2
          - description: The `CheCluster` custom resource allows defining and managing
              a Che server installation
            displayName: Eclipse Che instance Specification
            kind: CheCluster
            name: checlusters.org.eclipse.che
            version: v1
        description: |
          A collaborative Kubernetes-native development solution that delivers OpenShift workspaces and in-browser IDE for rapid cloud application development.
          This operator installs the Plugin and Devfile registries, Dashboard, Gateway and the Eclipse Che server, and configures these services.
          OpenShift OAuth is used directly for authentication. TLS mode is on.

          ## How to Install
          Press the **Install** button, choose the channel and the upgrade strategy, and wait for the **Installed** Operator status.
          When the operator is installed, create a new Custom Resource (CR) of Kind CheCluster (click the **Create New** button).
          The CR spec contains all defaults. You can start using Eclipse Che when the CR status is set to **Available**, and you see a URL to Eclipse Che.

          ## Defaults
          By default, the operator deploys Eclipse Che with:
          * 10Gi storage
          * Auto-generated passwords
          * Bundled Plugin and Devfile registries

          Use `oc edit checluster/eclipse-che -n eclipse-che` to update Eclipse Che default installation options.
          See more in the [Installation guide](https://www.eclipse.org/che/docs/stable/administration-guide/configuring-che/).

          ### Certificates
          Operator uses a default router certificate to secure Eclipse Che routes.
          Follow the [guide](https://www.eclipse.org/che/docs/stable/administration-guide/importing-untrusted-tls-certificates/)
          to import certificates into Eclipse Che.
        displayName: Eclipse Che
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eclipse che
        - workspaces
        - devtools
        - developer
        - ide
        - java
        links:
        - name: Product Page
          url: https://www.eclipse.org/che
        - name: Documentation
          url: https://www.eclipse.org/che/docs
        - name: Operator GitHub Repo
          url: https://github.com/eclipse-che/che-operator
        maintainers:
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: Eclipse Foundation
        relatedImages:
        - quay.io/eclipse/che-dashboard@sha256:8e2285991a725e87d2a9131ea23e08ce97cd5bdf1118f53a55a140c2c97ca25d
        - quay.io/oauth2-proxy/oauth2-proxy@sha256:dcb6ff8dd21bf3058f6a22c6fa385fa5b897a9cd3914c88a2cc2bb0a85f8065d
        - quay.io/che-incubator/che-code@sha256:1df1b75f702952b650e21717b04de5f39c84ca8f55f4afa0f32644ef8638d03a
        - quay.io/che-incubator/che-code@sha256:3b5c7e1ae664c82ddb3eaedc8161eb72a3a1a22c6c5e3f342f4aad972a6282a5
        - quay.io/che-incubator/che-code-server@sha256:2093baab1a73e9682f5892f80582481abfbef99d744472ee6098e2d82fb2899f
        - quay.io/community-operator-pipeline-prod/eclipse-che@sha256:545d98c512c99bc7bf4ccd02c1eb554bbc4bca7da8cc66e0e26934edbfd50c59
        - quay.io/eclipse/che-tls-secret-creator@sha256:cbb2fd8ce903d822377f2093af1aa2543672ed8c516d607a7a9dd857379dac5d
        - quay.io/che-incubator/che-idea-dev-server@sha256:0344c9574a6544c4c0ef93a798a9705ca3a4624508a55c8f94c2bae7196f914a
        - quay.io/eclipse/che-plugin-registry@sha256:bd71aa6817eeda0449056e78a6e832b82dd37b65ec6cba2bac3ffb9defa4048e
        - quay.io/eclipse/che-operator@sha256:5be0ad064feabad071f110547374413cdc8cf909f04374007761dfcc3f8cfcc7
        - quay.io/openshift/origin-oauth-proxy@sha256:870bfe92a4663720775c0dfe5728ecbb10a17f0644eef5f57276ec135034c6a1
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:354fc75eb7a21a934381e93d03ef9d42bc2c8ae8989fdcacecfb39b863b96ced
        - quay.io/che-incubator/header-rewrite-proxy@sha256:bd7873b8feef35f218f54c6251ea224bea2c8bf202a328230019a0ba2941245d
        - quay.io/che-incubator/configbump@sha256:77bd0fc269ff3724583b43ec68e158fcd300e825e1b44e24e2b9e67ed28df6da
        - quay.io/devfile/universal-developer-image@sha256:f76ea662cfdb9e80b93bcb58a393aed8b52aa72e7da73762d30a3ad22c3f06d9
        - quay.io/eclipse/che--traefik@sha256:4cf907247939b5d20bf4eff73abd21cb413c339600dde76dbc94a874b2578a27
        - quay.io/brancz/kube-rbac-proxy@sha256:738c854322f56d63ebab75de5210abcdd5e0782ce2d30c0ecd4620f63b24694d
        - quay.io/eclipse/che-server@sha256:e5a542da0484eb2223ac292e6005cf7e116bec15d377557f92949728e1f5c8bf
        - quay.io/che-incubator/che-idea@sha256:8aae69dc4b0c122491a75400639af0fe92b5e214c6e68ac97cda29fb58b44151
        - quay.io/che-incubator/che-code-server@sha256:357c95297ce2182ef0a792038ea122075fdeb6c8e835099be09e8d4f105eac2a
        version: 7.109.0
      entries:
      - name: eclipse-che.v7.109.0
        version: 7.109.0
      - name: eclipse-che.v7.108.0
        version: 7.108.0
      - name: eclipse-che.v7.107.0
        version: 7.107.0
      - name: eclipse-che.v7.106.0
        version: 7.106.0
      - name: eclipse-che.v7.105.0
        version: 7.105.0
      - name: eclipse-che.v7.104.0
        version: 7.104.0
      - name: eclipse-che.v7.103.0
        version: 7.103.0
      - name: eclipse-che.v7.102.0
        version: 7.102.0
      - name: eclipse-che.v7.101.0
        version: 7.101.0
      - name: eclipse-che.v7.100.0
        version: 7.100.0
      - name: eclipse-che.v7.99.0
        version: 7.99.0
      - name: eclipse-che.v7.98.0
        version: 7.98.0
      - name: eclipse-che.v7.97.0
        version: 7.97.0
      - name: eclipse-che.v7.96.0
        version: 7.96.0
      - name: eclipse-che.v7.95.0
        version: 7.95.0
      - name: eclipse-che.v7.94.0
        version: 7.94.0
      - name: eclipse-che.v7.93.0
        version: 7.93.0
      - name: eclipse-che.v7.92.0
        version: 7.92.0
      - name: eclipse-che.v7.91.0
        version: 7.91.0
      - name: eclipse-che.v7.90.0
        version: 7.90.0
      - name: eclipse-che.v7.89.0
        version: 7.89.0
      - name: eclipse-che.v7.88.0
        version: 7.88.0
      - name: eclipse-che.v7.87.0
        version: 7.87.0
      - name: eclipse-che.v7.86.0
        version: 7.86.0
      - name: eclipse-che.v7.85.0
        version: 7.85.0
      - name: eclipse-che.v7.84.0
        version: 7.84.0
      - name: eclipse-che.v7.83.0
        version: 7.83.0
      - name: eclipse-che.v7.82.0
        version: 7.82.0
      - name: eclipse-che.v7.81.0
        version: 7.81.0
      - name: eclipse-che.v7.80.0
        version: 7.80.0
      - name: eclipse-che.v7.79.0
        version: 7.79.0
      - name: eclipse-che.v7.78.0
        version: 7.78.0
      - name: eclipse-che.v7.77.0
        version: 7.77.0
      - name: eclipse-che.v7.76.0
        version: 7.76.0
      - name: eclipse-che.v7.75.0
        version: 7.75.0
      - name: eclipse-che.v7.74.0
        version: 7.74.0
      - name: eclipse-che.v7.73.0
        version: 7.73.0
      - name: eclipse-che.v7.72.0
        version: 7.72.0
      - name: eclipse-che.v7.71.0
        version: 7.71.0
      - name: eclipse-che.v7.70.0
        version: 7.70.0
      - name: eclipse-che.v7.69.0
        version: 7.69.0
      - name: eclipse-che.v7.68.0
        version: 7.68.0
      - name: eclipse-che.v7.67.0
        version: 7.67.0
      - name: eclipse-che.v7.66.0
        version: 7.66.0
      - name: eclipse-che.v7.65.0
        version: 7.65.0
      - name: eclipse-che.v7.64.0
        version: 7.64.0
      - name: eclipse-che.v7.63.0
        version: 7.63.0
      - name: eclipse-che.v7.62.0
        version: 7.62.0
      - name: eclipse-che.v7.61.0
        version: 7.61.0
      - name: eclipse-che.v7.60.1
        version: 7.60.1
      - name: eclipse-che.v7.60.0
        version: 7.60.0
      - name: eclipse-che.v7.59.0
        version: 7.59.0
      - name: eclipse-che.v7.58.0
        version: 7.58.0
      - name: eclipse-che.v7.57.0
        version: 7.57.0
      - name: eclipse-che.v7.56.0
        version: 7.56.0
      - name: eclipse-che.v7.55.0
        version: 7.55.0
      - name: eclipse-che.v7.54.0
        version: 7.54.0
      - name: eclipse-che.v7.53.0
        version: 7.53.0
      - name: eclipse-che.v7.52.0
        version: 7.52.0
      - name: eclipse-che.v7.51.0
        version: 7.51.0
      - name: eclipse-che.v7.50.0
        version: 7.50.0
      - name: eclipse-che.v7.49.0
        version: 7.49.0
      - name: eclipse-che.v7.48.1
        version: 7.48.1
      - name: eclipse-che.v7.47.0
        version: 7.47.0
      - name: eclipse-che.v7.46.0
        version: 7.46.0
      - name: eclipse-che.v7.45.0
        version: 7.45.0
      - name: eclipse-che.v7.44.0
        version: 7.44.0
      - name: eclipse-che.v7.42.0
        version: 7.42.0
      - name: eclipse-che.v7.41.2
        version: 7.41.2
      - name: eclipse-che.v7.41.1
        version: 7.41.1
      - name: eclipse-che.v7.41.0
        version: 7.41.0
      - name: eclipse-che.v7.40.2
        version: 7.40.2
      - name: eclipse-che.v7.40.1
        version: 7.40.1
      name: stable
    defaultChannel: stable
    packageName: eclipse-che
    provider:
      name: Eclipse Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: group-sync-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: group-sync-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "azure-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "azure": {
                        "credentialsSecret": {
                          "name": "azure-group-sync",
                          "namespace": "group-sync-operator"
                        }
                      },
                      "name": "azure"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "github-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "github": {
                        "credentialsSecret": {
                          "name": "github-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "organization": "redhat-cop"
                      },
                      "name": "github"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "gitlab-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "gitlab": {
                        "credentialsSecret": {
                          "name": "gitlab-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "url": "https://gitlab.apps.openshift.com"
                      },
                      "name": "gitlab"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "keycloak-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "keycloak": {
                        "credentialsSecret": {
                          "name": "keycloak-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "realm": "ocp",
                        "url": "https://keycloak.apps.openshift.com"
                      },
                      "name": "keycloak"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "ldap-groupsync"
                },
                "spec": {
                  "providers": [
                    {
                      "ldap": {
                        "credentialsSecret": {
                          "name": "ldap-group-sync",
                          "namespace": "group-sync-operator"
                        },
                        "insecure": true,
                        "rfc2307": {
                          "groupMembershipAttributes": [
                            "member"
                          ],
                          "groupNameAttributes": [
                            "cn"
                          ],
                          "groupUIDAttribute": "dn",
                          "groupsQuery": {
                            "baseDN": "ou=Groups,dc=example,dc=com",
                            "derefAliases": "never",
                            "filter": "(objectClass=groupofnames)",
                            "scope": "sub"
                          },
                          "tolerateMemberNotFoundErrors": true,
                          "tolerateMemberOutOfScopeErrors": true,
                          "userNameAttributes": [
                            "cn"
                          ],
                          "userUIDAttribute": "dn",
                          "usersQuery": {
                            "baseDN": "ou=Users,dc=example,dc=com",
                            "derefAliases": "never",
                            "scope": "sub"
                          }
                        },
                        "url": "ldap://localhost:389"
                      },
                      "name": "ldap"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupSync",
                "metadata": {
                  "name": "okta-sync"
                },
                "spec": {
                  "providers": [
                    {
                      "name": "okta",
                      "okta": {
                        "appId": "okta-sync-app-id",
                        "credentialsSecret": {
                          "name": "okta-api-token",
                          "namespace": "group-sync-operator"
                        },
                        "extractLoginUsername": true,
                        "profileKey": "login",
                        "url": "https://example.okta.com/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
          createdAt: "2025-08-12T19:19:21Z"
          description: Synchronize groups and users from external providers
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: group-sync-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/group-sync-operator
          support: Red Hat Community of Practice
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupSync is the Schema for the groupsyncs API
            displayName: Group Sync
            kind: GroupSync
            name: groupsyncs.redhatcop.redhat.io
            version: v1alpha1
        description: |
          Synchronizes groups from external providers into OpenShift

          ## Overview

          The OpenShift Container Platform contains functionality to synchronize groups found in external identity providers into the platform. Currently, the functionality that is included in OpenShift to limited to synchronizing LDAP only. This operator is designed to integrate with external providers in order to provide new solutions.

          Group Synchronization is facilitated by creating a `GroupSync` resource. The following describes the high level schema for this resource:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: example-groupsync
          spec:
            providers:
              - <One or more providers to synchronize>
          ```

          ## Authentication

          In most cases, authentication details must be provided in order to communicate with providers. Authentication details are provider specific with regards to the required values. In supported providers, the secret can be referenced in the `credentialsSecret` by name and namespace where it has been created as shown below:

          ```
          credentialsSecret:
            name: <secret_name>
            namespace: <secret_namespace>
          ```

          ## Providers

          Integration with external systems is made possible through a set of pluggable external providers. The following providers are currently supported:

          * [Azure](https://azure.microsoft.com/)
          * [GitHub](https://github.com)
          * [GitLab](https://gitlab.com)
          * [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol)
          * [Keycloak](https://www.keycloak.org/)/[Red Hat Single Sign On](https://access.redhat.com/products/red-hat-single-sign-on)
          * [Okta](https://www.okta.com/)
          * [IBM Security Verify](https://docs.verify.ibm.com/verify)

          The following sections describe the configuration options available for each provider


          ### Azure

          Groups contained within Azure Active Directory can be synchronized into OpenShift. The following table describes the set of configuration options for the Azure provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `authorityHost` | Azure Active Directory Endpoint | `https://login.microsoftonline.com` | No |
          | `baseGroups` | List of groups to start searching from instead of listing all groups in the directory | | No |
          | `credentialsSecret` | Name of the secret containing authentication details (See below) | | Yes |
          | `filter` | Graph API filter | | No |
          | `groups` | List of groups to filter against | | No |
          | `userNameAttributes` | Fields on a user record to use as the User Name | `userPrincipalName` | No |
          | `prune` | Prune Whether to prune groups that are no longer in Azure | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Azure provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: azure-groupsync
          spec:
            providers:
            - name: azure
              azure:
                credentialsSecret:
                  name: azure-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to Azure

          Authentication to Azure can be performed using Application Registration with access to query group information in Azure Active Directory.

          The App Registration must be granted access to the following Microsoft Graph API's:

          * Group.Read.All
          * GroupMember.Read.All
          * User.Read.All

          A secret must be created in the same namespace that contains the `GroupSync` resource:

          The following keys must be defined in the secret

          * `AZURE_TENANT_ID` - Tenant ID
          * `AZURE_CLIENT_ID` - Client ID
          * `AZURE_CLIENT_SECRET` - Client Secret

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic azure-group-sync --from-literal=AZURE_SUBSCRIPTION_ID=<AZURE_SUBSCRIPTION_ID> --from-literal=AZURE_TENANT_ID=<AZURE_TENANT_ID> --from-literal=AZURE_CLIENT_ID=<AZURE_CLIENT_ID> --from-literal=AZURE_CLIENT_SECRET=<AZURE_CLIENT_SECRET>
          ```

          ### GitHub

          Teams stored within a GitHub organization can be synchronized into OpenShift. The following table describes the set of configuration options for the GitHub provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `caSecret` | Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `organization` | Organization to synchronize against | | Yes |
          | `teams` | List of teams to filter against | | No |
          | `url` | Base URL for the GitHub or GitHub Enterprise host (Must contain a trailing slash) | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitHub | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: github-groupsync
          spec:
            providers:
            - name: github
              github:
                organization: ocp
                credentialsSecret:
                  name: github-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitHub

          Authentication to GitLab can be performed using a [Token](https://docs.gitlab.com/ee/security/token_overview.html) or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following token types are supported:

          * Personal Access Token
          * OAuth Token
          * Job Token

          the following key is required:

          * `token` - OAuth token

          Optionally, the `tokenType` key can be specified to indicate the type of token being provided from the following values:

          * OAuth - `oauth`
          * Personal Access Token - `personal`
          * Job Token - `job`

          If no `tokenType` is provided, `oauth` is used by default

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          To specify a token type, such as a Personal Access Token, the following command can be executed:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token> --from-literal=tokenType=personal
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitHub
          * `password` - Password for authenticating with GitHub

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic github-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### GitLab

          Groups stored within a GitLab can be synchronized into OpenShift. The following table describes the set of configuration options for the GitLab provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groups` | List of groups to filter against | | No |
          | `prune` | Prune Whether to prune groups that are no longer in GitLab | `false` | No |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the GitLab instance | `https://gitlab.com` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Github provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: gitlab-groupsync
          spec:
            providers:
            - name: gitlab
              gitlab:
                credentialsSecret:
                  name: gitlab-group-sync
                  namespace: group-sync-operator
          ```

          #### Authenticating to GitLab

          Authentication to GitLab can be performed using an OAuth Personal Access Token or a Username and Password (Note: 2FA not supported). A secret must be created in the same namespace that contains the `GroupSync` resource:

          When using an OAuth token, the following key is required:

          * `token` - OAuth token

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=token=<token>
          ```

          The following keys are required for username and password:

          * `username` - Username for authenticating with GitLab
          * `password` - Password for authenticating with GitLab

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic gitlab-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### LDAP

          Groups stored within an [LDAP](https://en.wikipedia.org/wiki/Lightweight_Directory_Access_Protocol) server can be synchronized into OpenShift. The LDAP provider implements the included features of the [Syncing LDAP groups](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html) feature and makes use of the libraries from the [OpenShift Command Line](https://docs.openshift.com/container-platform/latest/cli_reference/openshift_cli/getting-started-cli.html) tool to streamline the migration to this operator based implementation.

          The configurations of the three primary schemas (`rfc2307`, `activeDirectory` and `augmentedActiveDirectory`) can be directly migrated as is without any modification.

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `groupUIDNameMapping` | User defined name mapping | | No |
          | `rfc2307` | Configuration using the [rfc2307](https://docs.openshift.com/container-platform/latest/authentication/ldap-syncing.html#ldap-syncing-rfc2307_ldap-syncing-groups) schema | | No |
          | `activeDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-activedir_ldap-syncing-groups) schema | | No |
          | `augmentedActiveDirectory` | Configuration using the [activeDirectory](https://docs.openshift.com/container-platform/4.5/authentication/ldap-syncing.html#ldap-syncing-augmented-activedir_ldap-syncing-groups) schema | | No |
          | `url` | Connection URL for the LDAP server | `ldap://ldapserver:389` | No |
          | `whitelist` | Explicit list of groups to synchronize |  | No |
          | `blacklist` | Explicit list of groups to not synchronize |  | No |
          | `prune` | Prune Whether to prune groups that are no longer in LDAP | `false` | No |

          The following is an example using the `rfc2307` schema:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
                credentialsSecret:
                  name: ldap-group-sync
                  namespace: group-sync-operator
                insecure: true
                rfc2307:
                  groupMembershipAttributes:
                  - member
                  groupNameAttributes:
                  - cn
                  groupUIDAttribute: dn
                  groupsQuery:
                    baseDN: ou=Groups,dc=example,dc=com
                    derefAliases: never
                    filter: (objectClass=groupofnames)
                    scope: sub
                  tolerateMemberNotFoundErrors: true
                  tolerateMemberOutOfScopeErrors: true
                  userNameAttributes:
                  - cn
                  userUIDAttribute: dn
                  usersQuery:
                    baseDN: ou=Users,dc=example,dc=com
                    derefAliases: never
                    scope: sub
                url: ldap://ldapserver:389
              name: ldap
          ```

          The examples provided in the OpenShift documented referenced previously can be used to construct the schemas for the other LDAP synchronization types.

          #### Authenticating to LDAP

          If authentication is required in order to communicate with the LDAP server, a secret should be created in the same namespace that contains the `GroupSync` resource. The following keys can be defined:

          * `username` - Username (Bind DN) for authenticating with the LDAP server
          * `password` - Password for authenticating with the LDAP server

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic ldap-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          #### Whitelists and Blacklists

          Groups can be explicitly whitelisted or blacklisted in order to control the groups that are eligible to be synchronized into OpenShift. When running LDAP group synchronization using the command line, this configuration is referenced via separate files, but these are instead specified in the `blacklist` and `whitelist` properties as shown below:

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                whitelist:
                - cn=Online Corporate Banking,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ```
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ldap-groupsync
          spec:
            providers:
            - ldap:
          ...
                blacklist:
                - cn=Finance,ou=Groups,dc=example,dc=com
          ...
              name: ldap
          ```

          ### Keycloak

          Groups stored within Keycloak can be synchronized into OpenShift. The following table describes the set of configuration options for the Keycloak provider:

          | Name | Description | Defaults | Required |
          | ----- | ---------- | -------- | ----- |
          | `ca` | Reference to a resource containing a SSL certificate to use for communication (See below) | | No |
          | `caSecret` | **DEPRECATED** Reference to a secret containing a SSL certificate to use for communication (See below) | | No |
          | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
          | `groups` | List of groups to filter against | | No |
          | `insecure` | Ignore SSL verification | `false` | No |
          | `loginRealm` | Realm to authenticate against | `master` | No |
          | `realm` | Realm to synchronize | | Yes |
          | `scope` | Scope for group synchronization. Options are `one` for one level or `sub` to include subgroups | `sub` | No |
          | `url` | Base URL for the Keycloak server. Older versions (<17.0.0) including Red Hat SSO should include the context path `/auth` appended to the hostname  | | Yes |
          | `prune` | Prune Whether to prune groups that are no longer in Keycloak | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with a Keycloak provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                url: https://keycloak-keycloak-operator.apps.openshift.com/auth
          ```

          #### Authenticating to Keycloak

          A user with rights to query for Keycloak groups must be available. The following permissions must be associated to the user:

          * Password must be set (Temporary option unselected) on the _Credentials_ tab
          * On the _Role Mappings_ tab, select _master-realm_ or _realm-management_ next to the _Client Roles_ dropdown and then select **Query Groups** and **Query Users**.

          A secret must be created in the same namespace that contains the `GroupSync` resource. It must contain the following keys for the user previously created:

          * `username` - Username for authenticating with Keycloak
          * `password` - Password for authenticating with Keycloak

          The secret can be created by executing the following command:

          ```shell
          oc create secret generic keycloak-group-sync --from-literal=username=<username> --from-literal=password=<password>
          ```

          ### Okta

          [Okta Groups](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) assigned to [Okta Applications](https://help.okta.com/en/prod/Content/Topics/Apps/Apps_Apps.htm) can be synchronized into OpenShift. The developer docs for the Okta API that the Okta Syncer uses can be found [here](https://developer.okta.com/docs/reference/api/apps/#list-groups-assigned-to-application).
          The following table describes the set of configuration options for the Okta provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (See below) | | Yes |
            | `groups` | List of groups to filter against | | No |
            | `url` | Okta organization URL (Must contain a trailing slash) | | Yes |
            | `appId` | Application ID of App Groups are assigned to | | Yes |
            | `extractLoginUsername` | Bool to determine if you should extract username from okta login | | No |
            | `profileKey` | Attribute field on Okta User Profile you would like to use as identity | `login` | No |
            | `groupLimit` | Integer to set the maximum number of groups to retrieve from OKTA per request. | `1000` | No |
            | `prune` | Prune Whether to prune groups that are no longer in OKTA | `false` | No |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: okta-sync
          spec:
            providers:
              - name: okta
                okta:
                  credentialsSecret:
                    name: okta-api-token
                    namespace: group-sync-operator
                  url: "https://example.okta.com/"
                  appId: okta-sync-app-id
          ```

          #### Authenticating to Okta

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following key:

            * `okta-api-token` - Okta API Token for interacting with Okta

          The secret can be created by executing the following command:

            ```shell
            oc create secret generic okta-api-token --from-literal=okta-api-token=<OKTA_API_TOKEN> -n group-sync-operator
            ```

          ### IBM Security Verify

          Groups defined in [IBM Security Verify](https://help.okta.com/en/prod/Content/Topics/users-groups-profiles/usgp-main.htm) (ISV) can be synchronized into OpenShift. Currently only the `userName` field from ISV will be synchronized. The developer docs for the ISV API can be found [here](https://docs.verify.ibm.com/verify/page/api-documentation).
          The following table describes the set of configuration options for the provider:

            | Name | Description | Defaults | Required |
            | ----- | ---------- | -------- | ----- |
            | `credentialsSecret` | Reference to a secret containing authentication details (see below) | `''`  | Yes |
            | `groups` | List of groups to synchronize (see below) | `nil`  | Yes |
            | `tenantUrl` | The ISV tenant URL, for example `https://my-isv.verify.ibm.com`) | `''`  | Yes |

          The following is an example of a minimal configuration that can be applied to integrate with an Okta provider:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: ibmsecurityverify-sync
          spec:
              providers:
              - name: ibmsecurityverify
                ibmsecurityverify:
                  credentialsSecret:
                    name: isv-group-sync
                    namespace: group-sync-operator
                  tenantUrl: https://my-isv.verify.ibm.com
                  groups:
                    - name: 'application owners'
                      id: 645001V3V9
                    - name: developer
                      id: 645001V3VA
          ```

          #### Group Objects
          Each group object in the `groups` array must contain an `id` field. The group ID can be retrieved by pulling the group information from the ISV API. Optionally, the object may also contain a `name` which corresponds to the group's display name. When defined, the operator will confirm that the name defined in the YAML matches that received from the API when synchronization occurs; as the group IDs are not human-friendly, using the name can confirm the correct groups are configured. If the names do not match an error will be logged.

          #### Group Names
          The name of each groups created in OpenShift will match the group name in ISV. Any whitespace in the ISV group name will be replaced with a hyphen.

          #### Authenticating to IBM Security Verify

          A secret must be created in the same namespace as the group-sync-operator pod. It must contain the following keys:

          * `clientId` - The API client ID.
          * `clientSecret`- The API client secret.

          See the IBM Security Verify [API documentation](https://docs.verify.ibm.com/verify/docs/api-access) for setting up authentication.

          ### Support for Additional Metadata (Beta)

          Additional metadata based on Keycloak group are also added to the OpenShift groups as Annotations including:

          * Parent/child relationship between groups and their subgroups
          * Group attributes

          ## CA Certificates

          Several providers allow for certificates to be provided in either a _ConfigMap_ or _Secret_ to communicate securely to the target host through the use of a property called `ca`.

          The certificate can be added to a Secret called _keycloak-certs_ using the key `ca.crt` representing the certificate using the following command.

          ```
          oc create secret generic keycloak-certs --from-file=ca.crt=<file>
          ```

          An example of how the CA certificate can be added to the Keycloak provider is shown below:


          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: Secret
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          Alteratively, a _ConfigMap_ can be used instead instead of a _Secret_. This is useful when using the [Certificate injection using Operators](https://docs.openshift.com/container-platform/latest/networking/configuring-a-custom-pki.html#certificate-injection-using-operators_configuring-a-custom-pki) feature.

          The following command can be used to create a _ConfigMap_ containing the certificate:

          ```
          oc create configmap keycloak-certs --from-file=ca.crt=<file>
          ````

          An example of how the CA certificate can be added to the Keycloak provider is shown below:

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            providers:
            - name: keycloak
              keycloak:
                realm: ocp
                credentialsSecret:
                  name: keycloak-group-sync
                  namespace: group-sync-operator
                caSecret:
                  kind: ConfigMap
                  name: keycloak-certs
                  namespace: group-sync-operator
                  key: ca.crt
                url: https://keycloak-keycloak-operator.apps.openshift.com
          ```

          ## Scheduled Execution

          A cron style expression can be specified for which a synchronization event will occur. The following specifies that a synchronization should occur nightly at 3AM

          ```shell
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: GroupSync
          metadata:
            name: keycloak-groupsync
          spec:
            schedule: "0 3 * * *"
            providers:
            - ...
          ```

          If a schedule is not provided, synchronization will occur only when the object is reconciled by the platform.

          ## Metrics

          Prometheus compatible metrics are exposed by the Operator and can be integrated into OpenShift's default cluster monitoring. To enable OpenShift cluster monitoring, label the namespace the operator is deployed in with the label `openshift.io/cluster-monitoring="true"`.

          ```
          oc label namespace <namespace> openshift.io/cluster-monitoring="true"
          ```
        displayName: Group Sync Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - group
        - user
        - security
        - authorization
        links:
        - name: Group Sync Operator
          url: https://github.com/redhat-cop/group-sync-operator
        maintainers:
        - email: ablock@redhat.com
          name: Andrew Block
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/community-operator-pipeline-prod/group-sync-operator@sha256:fe6fda2de67fc4996d3f388e11faa0130fee19453111c829ba0aaa5be5410d01
        - quay.io/redhat-cop/group-sync-operator@sha256:43ef79efc2a77ee5edd4e3123f9dab156a30d3e00fd1fba9f89600b96f0f68d6
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:b8a174be91f3df75dff9656d4c72de28af51a731008e9e96db5eef3c7af58aac
        version: 0.0.36
      entries:
      - name: group-sync-operator.v0.0.36
        version: 0.0.36
      - name: group-sync-operator.v0.0.35
        version: 0.0.35
      - name: group-sync-operator.v0.0.34
        version: 0.0.34
      - name: group-sync-operator.v0.0.33
        version: 0.0.33
      - name: group-sync-operator.v0.0.32
        version: 0.0.32
      - name: group-sync-operator.v0.0.31
        version: 0.0.31
      - name: group-sync-operator.v0.0.30
        version: 0.0.30
      - name: group-sync-operator.v0.0.29
        version: 0.0.29
      - name: group-sync-operator.v0.0.28
        version: 0.0.28
      - name: group-sync-operator.v0.0.27
        version: 0.0.27
      - name: group-sync-operator.v0.0.26
        version: 0.0.26
      - name: group-sync-operator.v0.0.25
        version: 0.0.25
      - name: group-sync-operator.v0.0.24
        version: 0.0.24
      - name: group-sync-operator.v0.0.23
        version: 0.0.23
      - name: group-sync-operator.v0.0.22
        version: 0.0.22
      - name: group-sync-operator.v0.0.21
        version: 0.0.21
      - name: group-sync-operator.v0.0.20
        version: 0.0.20
      - name: group-sync-operator.v0.0.19
        version: 0.0.19
      - name: group-sync-operator.v0.0.18
        version: 0.0.18
      - name: group-sync-operator.v0.0.17
        version: 0.0.17
      - name: group-sync-operator.v0.0.16
        version: 0.0.16
      - name: group-sync-operator.v0.0.15
        version: 0.0.15
      - name: group-sync-operator.v0.0.14
        version: 0.0.14
      - name: group-sync-operator.v0.0.13
        version: 0.0.13
      name: alpha
    defaultChannel: alpha
    packageName: group-sync-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Redis
      provider-url: ""
    name: redis-enterprise-operator-cert-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: redis-enterprise-operator.v6.2.10-45
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "username": "demo@redislabs.com",
                  "redisEnterpriseNodeResources": {
                    "limits": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    },
                    "requests": {
                      "cpu": "4000m",
                      "memory": "4Gi"
                    }
                  },
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "redisEnterpriseCluster": {
                    "name": "rec"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:redis-enterprise-operator*<br>

              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:e6d5ef5583ae26a8312dc331bd6003fda6c7c248de15b6790a128d0cd03965dd
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:b771ef87bf211c17c37df028c202aac97170fb6d7d5d49b3ccb3410deb8212f6
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:ffdf10a1b08115fd4de06f1cd751d36644a9bc9a8c11034cd432e9b66a5d2f73
        - registry.connect.redhat.com/redislabs/services-manager@sha256:8b46f2aacfb3e1b243c0b64975accc55fd80b4de9c9ac4769c3c67a5676addc5
        version: 6.2.10-45
      entries:
      - name: redis-enterprise-operator.v6.2.10-45
        version: 6.2.10-45
      - name: redis-enterprise-operator.v6.2.10-34
        version: 6.2.10-34
      name: 6.2.10
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0
    - currentCSV: redis-enterprise-operator.v7.22.0-11.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:3b96cdc65c3e42fd22eb165950be96b907725903ac74fe4f82cfa0073e28e95a
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:0cfb07f774db131b5b74e5aa91291579c916c4d4cd1a693d7aaed554f0177a68
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:35e0ee6b3ec5cb654b709c5e4a76a1f08182d826b47f077031d9139d4c2cc308
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:088e4b57c6e1ca1680b359078e5eb1bba67343cdbd9a539af26a0ea23a830593
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:d56ba97e95e799210e76615095409143a5dddd5c5a1e1be1d551d7e3f9e4d3ef
        version: 7.22.0-11.3
      entries:
      - name: redis-enterprise-operator.v7.22.0-11.3
        version: 7.22.0-11.3
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.22.0-11.3
    - currentCSV: redis-enterprise-operator.v7.8.4-9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "uiServiceType": "ClusterIP",
                  "bootstrapperImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                      "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "imagePullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Active-Active Database on an Operator
              managed  Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Database on an Operator managed  Redis
              Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b876ec88e1e64c17572e956b1c7e4020f5a423c1275f7f57025823ff7f749c61
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:66ec124da16cc9d0abd66687e088be0b358e20cfd48f5774d34bf5874e6a4e65
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:4ba0c414a74268bed9c6757e447b340db7c912afa864cfb096ddf88f11feeaf9
        - registry.connect.redhat.com/redislabs/services-manager@sha256:803be04534e59cdc7b8898b3bb336676aad1cb77cf266051e226be82679fcf6a
        version: 7.8.4-9.0
      entries:
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.4
    - currentCSV: redis-enterprise-operator.v7.8.6-5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:88d483f5e37b5a901d1ae3d965304edc000c172d2e5e307297d5b7bcd77ae73b
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:b3a45189ae8fc5302fb8a9e15d5747bacce044537923621edf21e9d3b267a27e
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:1fff01daece34aeb7befcb0760d5112c8c7d261f2b86388ece416ac94076a573
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:5125e83d3c93b224d1983c599994e43899a0ba0c6f2fc4eed31476f755b5fc0b
        version: 7.8.6-5.1
      entries:
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: 7.8.6
    - currentCSV: redis-enterprise-operator.v7.22.0-17.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.redislabs.com/v1",
                "kind": "RedisEnterpriseCluster",
                "metadata": {
                  "name": "rec"
                },
                "spec": {
                  "nodes": 3,
                  "persistentSpec": {
                    "enabled": true
                  },
                  "bootstrapperImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise-operator"
                  },
                  "redisEnterpriseServicesRiggerImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/services-manager"
                  },
                  "redisEnterpriseImageSpec": {
                    "repository": "registry.connect.redhat.com/redislabs/redis-enterprise"
                  },
                  "usageMeter": {
                    "callHomeClient": {
                      "imageSpec": {
                        "repository": "registry.connect.redhat.com/redislabs/call-home-client"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseDatabase",
                "metadata": {
                  "name": "redb"
                },
                "spec": {
                  "memorySize": "256MB",
                  "shardCount": 1,
                  "replication": true
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseRemoteCluster",
                "metadata": {
                  "name": "rerc1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "recName": "rec",
                  "recNamespace": "ns1",
                  "apiFqdnUrl": "testapi-rerc1-ns1.redislabs.com",
                  "dbFqdnSuffix": "-example-rerc1-ns1.redislabs.com",
                  "secretName": "redis-enterprise-rerc1"
                }
              },
              {
                "apiVersion": "app.redislabs.com/v1alpha1",
                "kind": "RedisEnterpriseActiveActiveDatabase",
                "metadata": {
                  "name": "reaadb-1",
                  "labels": {
                    "app": "redis-enterprise"
                  }
                },
                "spec": {
                  "participatingClusters": [
                    {
                      "name": "rerc1"
                    },
                    {
                      "name": "rerc2"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/redis-enterprise-operator-cert-rhmp/support?utm_source=openshift_console
          operators.openshift.io/infrastructure-features: '["disconnected"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy a Redis Enterprise Cluster
            displayName: RedisEnterpriseCluster
            kind: RedisEnterpriseCluster
            name: redisenterpriseclusters.app.redislabs.com
            version: v1
          - description: Deploy a Redis Enterprise Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseDatabase
            kind: RedisEnterpriseDatabase
            name: redisenterprisedatabases.app.redislabs.com
            version: v1alpha1
          - description: Deploy a Redis Enterprise Active-Active Database on an operator-managed
              Redis Enteprise Cluster
            displayName: RedisEnterpriseActiveActiveDatabase
            kind: RedisEnterpriseActiveActiveDatabase
            name: redisenterpriseactiveactivedatabases.app.redislabs.com
            version: v1alpha1
          - description: Define Redis Enterprise Clusters used to form Active-Active
              databases
            displayName: RedisEnterpriseRemoteCluster
            kind: RedisEnterpriseRemoteCluster
            name: redisenterpriseremoteclusters.app.redislabs.com
            version: v1alpha1
        description: |-
          Redis Enterprise Software is enterprise grade, distributed, in-memory NoSQL database server, fully compatible with open source Redis by Redis.
          Redis Enterprise Software extends open source Redis and delivers stable high performance, zero-downtime linear scaling and high availability, with significant operational savings.
          Redis Enterprise provides a flexible and cost-effective data platform so developers can focus on rapid, high-quality development for sophisticated use cases that meet the needs of a modern digital enterprise. With Redis Enterprise, you can:
          * Enjoy high performance and record-setting low latencies with minimal operational overhead
          * Develop highly scalable microservices-based and Kubernetes-orchestrated applications
          * Use versatile data structures for a variety of use cases, such as high-speed transactions, user session management, real-time analytics, caching and many more
          * Leverage enterprise-grade operational controls for high availability, security and seamless scale
          * Automate operational best practices for deploying and managing the Redis Enterprise data platform with built-in Kubernetes Operator support
          * Deploy, manage and move applications to and from any cloud for seamless portability

          ### Core capabilities
          * **Speed application development and time to market** - Unburden developers from the complexity of infrastructure operations and empower them to manage data with maximum flexibility. This helps them build modern applications quickly, make changes rapidly and support a variety of different data structures, models, relationships and use cases.
          * **Ensure business continuity with always-on data service** - Maintain service uptime and ensure failsafe high availability, instant failover and automatic recovery to protect your applications against unplanned downtime, outages and data loss.
          * **Design for performance and unmatched user experience** - Deliver the highest level of responsiveness and interactive customer experiences with best-in-class exceptional linear scaling high performance through a shared-nothing architecture and in-memory processing.
          * **Simplify operations with automated lifecycle management and layered orchestration** - Automate database provisioning, management and maintenance, and abstract away the complexities of high availability, seamless scale and zero-downtime upgrades with built-in support for Kubernetes and Operators.
          * **Enjoy multi-level robust security** - Provide granular control to meet self-imposed and regulatory compliance regulations and standards with built-in finegrained security controls and data encryption.
          * **Gain platform independence with flexible deployment options** - Ensure portability with multiple deployment options across any cloud of choice for multicloud and hybrid deployments and on-premises.
          * **Ensure success** - Leverage 24x7 enterprise-grade support backed by expertise in managing and scaling hundreds of thousands of Redis databases for thousands of enterprise customers worldwide.
          * **Future proof investments** - Use the most popular database amongst developers, built on open source innovation and entrenched firmly in the application landscape with over 1B downloads on Docker. Redis preserves your investment for years with easy extensibility and versatility through its Modules and support for over 50 different programming languages

          #### Prerequisites
          A minimum of 3 nodes which support the following requirements:
          * RAM - At least 3GB with 4GB recommended
          * Persistent Storage - At least 10GB of free space with 20GB recommended.
          * A kubernetes version of 1.9 or higher

          #### Recommendations/known issues
          *  Changing the name of the Redis Enterprise Cluster in the spec from the
            default requires running commands to grant permissions to the pods:

            Operator service account:
              Redis Enterprise Cluster service account:
              > *oc adm policy add-scc-to-user redis-enterprise-scc-v2 system:serviceaccount:my-project:<name of Redis Enterprise Cluster>*

              Replace my-project with the actual name of the project/namespace. Replace <name of Redis Enterprise Cluster> with the name of the Redis Enterprise Cluster.

          * When creating a Redis Enterprise Cluster without specifying in its spec a `persistentSpec` with explicit `storageClassName`,
          the default Storage Class of the Kubernetes cluster will be used to dynamically provision storage for the Redis Enterprise Cluster's nodes.
          However, if the Kubernetes cluster doesn't have a default Storage Class configured
          and no specific Storage Class was specified in the Redis Enterprise Cluster Spec, the installation of the cluster could fail.
          To avoid such scenario in cases where a specific Storage Class isn't specified, make sure the Kubernetes Cluster has a default Storage Cluster configured.
          This can be done by running a `kubectl get storageclass` command and see if one of the Storage Classes' names contains a `(default)` mark.
        displayName: Redis Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - redislabs
        - redis
        - database
        - nosql
        - cache
        links:
        - name: Getting Started with Kubernetes and OpenShift
          url: https://docs.redislabs.com/latest/platforms/openshift/
        - name: Deployment files on github
          url: https://github.com/RedisLabs/redis-enterprise-k8s-docs
        - name: Redis Kubernetes Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-architecture/
        - name: Redis Enterprise Kubernetes Operator-based Architecture
          url: https://docs.redislabs.com/latest/platforms/kubernetes/kubernetes-with-operator/
        maintainers:
        - email: support@redislabs.com
          name: Redis, Inc
        maturity: mature
        provider:
          name: Redis
        relatedImages:
        - registry.connect.redhat.com/redislabs/services-manager@sha256:2a8f60b35248abc35f7ea5df97ce2af7fa12eec6c12e0dc5a66c7fa6341907f9
        - registry.connect.redhat.com/redislabs/call-home-client@sha256:c8d27668947e45816dab2ce3786385992579924d6e51aec84447caab6b74c07f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator-bundle-marketplace@sha256:f9f647c465267ec8615b3ea891301c0d8a17bc41c9d71f151b374b0c5f384b3f
        - registry.connect.redhat.com/redislabs/redis-enterprise-operator@sha256:e22083901d1d3c97dc4d2de6fe98f95f713e65624cb1d887bff39b168ac88f4f
        - registry.connect.redhat.com/redislabs/redis-enterprise@sha256:8f23b9314a2bb900079cd31a9bd4ac144cb447987b3f0cfb174c08219537349c
        version: 7.22.0-17.0
      entries:
      - name: redis-enterprise-operator.v7.22.0-17.0
        version: 7.22.0-17.0
      - name: redis-enterprise-operator.v7.22.0-16.0
        version: 7.22.0-16.0
      - name: redis-enterprise-operator.v7.22.0-15.0
        version: 7.22.0-15.0
      - name: redis-enterprise-operator.v7.22.0-11.2
        version: 7.22.0-11.2
      - name: redis-enterprise-operator.v7.22.0-7.0
        version: 7.22.0-7.0
      - name: redis-enterprise-operator.v7.8.6-5.1
        version: 7.8.6-5.1
      - name: redis-enterprise-operator.v7.8.6-3.2
        version: 7.8.6-3.2
      - name: redis-enterprise-operator.v7.8.6-2.2
        version: 7.8.6-2.2
      - name: redis-enterprise-operator.v7.8.6-1.2
        version: 7.8.6-1.2
      - name: redis-enterprise-operator.v7.8.6-1.1
        version: 7.8.6-1.1
      - name: redis-enterprise-operator.v7.8.6-1.0
        version: 7.8.6-1.0
      - name: redis-enterprise-operator.v7.8.4-9.0
        version: 7.8.4-9.0
      name: production
    defaultChannel: production
    packageName: redis-enterprise-operator-cert-rhmp
    provider:
      name: Redis
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Fujitsu
      provider-url: https://www.postgresql.fastware.com/
    name: fujitsu-enterprise-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fujitsu-enterprise-postgres-operator.v7.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPAction",
                "metadata": {
                  "name": "new-fep-action"
                },
                "spec": {
                  "fepAction": {
                    "args": [
                      "new-fep-sts-0",
                      "new-fep-sts-1"
                    ],
                    "type": "reload"
                  },
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "targetClusterName": "new-fep"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPExporter",
                "metadata": {
                  "name": "new-fep-exporter"
                },
                "spec": {
                  "fepExporter": {
                    "exporterLogLevel": "error",
                    "fepClusterList": [
                      "new-fep1"
                    ],
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "userCustomQueries": "usr_example:\n  query: \"SELECT EXTRACT(EPOCH FROM (now() - pg_last_xact_replay_timestamp())) as lag\"\n  master: true\n  metrics:\n    - lag:\n        usage: \"GAUGE\"\n        description: \"Replication lag behind master in seconds\""
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPLogging",
                "metadata": {
                  "name": "new-fep-logging"
                },
                "spec": {
                  "fepLogging": {
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "restartRequired": false,
                    "sysExtraEvent": true,
                    "sysExtraLogging": false
                  }
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPPgpool2",
                "metadata": {
                  "name": "new-fep-pgpool2"
                },
                "spec": {
                  "clientAuthMethod": "scram",
                  "count": 2,
                  "customhba": "local   all         all                               trust\nhost    all         all         127.0.0.1/32          trust\nhost    all         all         ::1/128               trust\n",
                  "customlogsize": "128Mi",
                  "customparams": "listen_addresses = '*'\npcp_listen_addresses = '*'\nnum_init_children = 32\nreserved_connections = 0\nenable_pool_hba = off\nallow_clear_text_frontend_auth = off\nauthentication_timeout = 80\nbackend_weight0 = 1\nbackend_weight1 = 1\nbackend_flag0 = 'ALWAYS_PRIMARY'\nbackend_flag1 = 'DISALLOW_TO_FAILOVER'\nconnection_cache = on\nmax_pool = 4\nlisten_backlog_multiplier = 2\nserialize_accept = off\nchild_life_time = 300\nclient_idle_limit = 0\nchild_max_connections = 0\nconnection_life_time = 0\nreset_query_list = 'ABORT; DISCARD ALL'\nclient_min_messages = info\nlog_min_messages = debug1\nlog_statement = on\nlog_per_node_statement = on\nlog_client_messages = on\nlog_hostname = on\nlog_connections = on\nlog_line_prefix = '%t: pid %p: '\nload_balance_mode = on\nignore_leading_white_space = on\nwhite_function_list = ''\nblack_function_list = 'currval,lastval,nextval,setval'\nblack_query_pattern_list = ''\ndatabase_redirect_preference_list = ''\napp_name_redirect_preference_list = ''\nallow_sql_comments = off\ndisable_load_balance_on_write = 'transaction'\nstatement_level_load_balance = on\nsr_check_period = 0\nsr_check_user = 'postgres'\ndelay_threshold = 0\nlog_standby_delay = 'none'\nssl = on\nssl_ciphers = 'HIGH:MEDIUM:+3DES:!aNULL'\nssl_prefer_server_ciphers = off\nssl_ecdh_curve = 'prime256v1'\nssl_dh_params_file = ''\nrelcache_expire = 0\nrelcache_size = 256\ncheck_temp_table = catalog\ncheck_unlogged_table = on\nenable_shared_relcache = off\nrelcache_query_target = primary\nwd_port0 = 9000\nfailover_on_backend_error = off\n",
                  "custompcp": "none",
                  "customsslcacert": "none",
                  "customsslcert": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                  "customsslkey": "none",
                  "fepclustername": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "limits": {
                    "cpu": "400m",
                    "memory": "512Mi"
                  },
                  "requests": {
                    "cpu": "200m",
                    "memory": "256Mi"
                  },
                  "serviceport": 9999,
                  "statusport": 9898
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v1",
                "kind": "FEPRestore",
                "metadata": {
                  "name": "new-fep-restore"
                },
                "spec": {
                  "fromFEPcluster": "new-fep",
                  "imagePullPolicy": "IfNotPresent",
                  "mcSpec": {
                    "limits": {
                      "cpu": "200m",
                      "memory": "300Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "restoretype": "latest",
                  "sysExtraEvent": true,
                  "sysExtraLogging": false,
                  "toFEPcluster": "new-fep-2"
                }
              },
              {
                "apiVersion": "fep.fujitsu.io/v2",
                "kind": "FEPCluster",
                "metadata": {
                  "name": "new-fep"
                },
                "spec": {
                  "fep": {
                    "customAnnotations": {
                      "allDeployments": {}
                    },
                    "forceSsl": true,
                    "image": {
                      "pullPolicy": "IfNotPresent"
                    },
                    "instances": 1,
                    "mcSpec": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "700Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "512Mi"
                      }
                    },
                    "podAntiAffinity": false,
                    "podDisruptionBudget": false,
                    "servicePort": 27500,
                    "syncMode": "off",
                    "sysExtraEvent": true,
                    "sysExtraLogging": false,
                    "usePodName": true
                  },
                  "fepChildCrVal": {
                    "backup": {
                      "image": {
                        "pullPolicy": "IfNotPresent"
                      },
                      "mcSpec": {
                        "limits": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "200Mi"
                        }
                      },
                      "pgbackrestParams": "# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[global]\nrepo1-retention-full=7\nrepo1-retention-full-type=time\nlog-path=/database/log/backup\n",
                      "postScript": " ",
                      "preScript": " ",
                      "schedule": {
                        "num": 2
                      },
                      "schedule1": {
                        "schedule": "15 0 * * 0",
                        "type": "full"
                      },
                      "schedule2": {
                        "schedule": "15 0 * * 1-6",
                        "type": "incr"
                      },
                      "schedule3": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule4": {
                        "schedule": " ",
                        "type": " "
                      },
                      "schedule5": {
                        "schedule": " ",
                        "type": " "
                      }
                    },
                    "customPgAudit": "# define pg audit custom params here to override defaults.\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\n[output]\nlogger = 'auditlog'\nlog_directory = '/database/log/audit'\nlog_truncate_on_rotation = on\nlog_filename = 'pgaudit-%a.log'\nlog_rotation_age = 1d\nlog_rotation_size = 0\n[rule]\n",
                    "customPgHba": "# define pg_hba custom rules here to be merged with default rules.\n# TYPE     DATABASE        USER        ADDRESS        METHOD\n",
                    "customPgParams": "# define custom postgresql.conf parameters below to override defaults.\n# Current values are as per default FEP deployment\n# If you add a library to shared_preload_libraries, add it after the default value.\n# fsep_operator_security is available in FEP15 and later container images.\nshared_preload_libraries='pgx_datamasking,pg_prewarm,pg_stat_statements,fsep_operator_security'\nsession_preload_libraries='pg_prewarm'\nmax_prepared_transactions = 100\nmax_worker_processes = 30\nmax_connections = 100\nwork_mem = 1MB\nmaintenance_work_mem = 12MB\nshared_buffers = 128MB\neffective_cache_size = 384MB\ncheckpoint_completion_target = 0.8\n# tcp parameters\ntcp_keepalives_idle = 30\ntcp_keepalives_interval = 10\ntcp_keepalives_count = 3\n# logging parameters in default fep installation\n# if log volume is not defined, log_directory should be\n# changed to '/database/userdata/data/log'\nlog_directory = '/database/log'\nlog_filename = 'logfile-%a.log'\nlog_file_mode = 0600\nlog_truncate_on_rotation = on\nlog_rotation_age = 1d\nlog_rotation_size = 0\nlog_checkpoints = on\nlog_line_prefix = '%e %t [%p]: [%l-1] user=%u,db=%d,app=%a,client=%h'\nlog_lock_waits = on\nlog_autovacuum_min_duration = 60s\nlogging_collector = on\npgaudit.config_file='/opt/app-root/src/pgaudit-cfg/pgaudit.conf'\nlog_replication_commands = on\nlog_min_messages = WARNING\nlog_destination = csvlog\n# wal_archive parameters in default fep installation\narchive_mode = on\narchive_command = 'pgbackrest --stanza=backupstanza --config=/database/userdata/pgbackrest.conf archive-push %p'\nwal_level = replica\nmax_wal_senders = 12\nwal_keep_segments = 64\ntrack_activities = on\ntrack_counts = on\npassword_encryption = 'scram-sha-256'\n",
                    "storage": {
                      "archivewalVol": {
                        "size": "1Gi"
                      },
                      "backupVol": {
                        "size": "2Gi"
                      },
                      "dataVol": {
                        "size": "2Gi"
                      },
                      "logVol": {
                        "size": "1Gi"
                      },
                      "tablespaceVol": {
                        "size": "512Mi"
                      },
                      "walVol": {
                        "size": "1200Mi"
                      }
                    },
                    "sysUsers": {
                      "pgRewindPassword": "rewind_password",
                      "pgRewindUser": "rewind_user",
                      "pgdb": "mydb",
                      "pgpassword": "mydbpassword",
                      "pgreplpassword": "repluserpwd",
                      "pgrepluser": "repluser",
                      "pguser": "mydbuser",
                      "tdepassphrase": "tde-passphrase"
                    },
                    "systemCertificates": {
                      "cacrt": "-----BEGIN CERTIFICATE-----\nMIIDTzCCAjegAwIBAgIUYssQ8I74US5g+1+Z7CHuaDgkZnEwDQYJKoZIhvcNAQEL\nBQAwNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9y\nIEt1YmVybmV0ZXMwHhcNMjEwMjA2MDM1MjI4WhcNMzEwMjA0MDM1MjI4WjA3MRAw\nDgYDVQQKDAdGdWppdHN1MSMwIQYDVQQDDBpGRVAgUm9vdCBDQSBmb3IgS3ViZXJu\nZXRlczCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAMs97gUF0xkUzCgL\n7MiiDju9ySr/ziwjvcYU7jA9ML+SLmftMs3HtcYbAmSntqI+MDBSR/FAJTOoytuT\npV+mCFcGj2YAjDpliHPeNcUpbryy4YMChF3+MovkIwGCksxo5rhiWhGmoBYpA48P\n4Xe8SPlzqMzhFvNeKzyiUhvjutS2Y1Ss38lsTaurFPx64vQ2PaC54XzdwMptXtpb\ntYmWSzCpJWwxZ6lF3vitdA2w0tnBWNyctAd0+RIM/fvArxiIqseAux9t0uogm5to\nlRIhvekuxOpXBPEqtIYQ4j9XUW2JH8vUDnzPkPvjrq+A3Ug8OyyfGVrW7+VYXozu\nc4aP7P0CAwEAAaNTMFEwHQYDVR0OBBYEFBzCutQ7S74WEhS5V2sNEJBGyLpmMB8G\nA1UdIwQYMBaAFBzCutQ7S74WEhS5V2sNEJBGyLpmMA8GA1UdEwEB/wQFMAMBAf8w\nDQYJKoZIhvcNAQELBQADggEBAMDwD85RAaWEBptFgLzKw+9xEUy1vcZaonAuA1qc\nT342XTueyAugxkC11HwdCGgGS34VyctfMGqj4AW6pA2ez4tLrbOps4DmV4sw8uBL\n8pgRDgfly3ob9FEg2wa0hmrwX9jH5Bt4vySUE2785uPAqaspT2UNtTBxS85BUi1T\nsKId2Rtil6an281Z81wyWVI6Jm2D4MG0mbsiGcTPlCtdg/UljvDYymXlAvd4vNhl\nk9hDa13TgDqJKgKdTIcmZoNQdpEVgFcO0h9AEUy5AuLqxHq60dLfZ6ESGPlMI7Lm\ni4PzYbCnBmOe+7TnHcPSyrnehs66Ik+oifRd82eYS7vKjFw=\n-----END CERTIFICATE-----",
                      "crt": "-----BEGIN CERTIFICATE-----\nMIIDUTCCAjmgAwIBAgIRAMocW3qMoHrD6qRvMPppMkMwDQYJKoZIhvcNAQELBQAw\nNzEQMA4GA1UECgwHRnVqaXRzdTEjMCEGA1UEAwwaRkVQIFJvb3QgQ0EgZm9yIEt1\nYmVybmV0ZXMwHhcNMjEwMjA2MDQzMjM2WhcNMjYwMjA1MDQzMjM2WjA/MRAwDgYD\nVQQKEwdGdWppdHN1MSswKQYDVQQDEyJGVUpJVFNVIEVudGVycHJpc2UgUG9zdGdy\nZXMgU2VydmVyMIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEA4AI33yvH\nZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I2e4SceTKi6O3C/I1XuvWlpng\n5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4ANB5JWWqDOjrRT3o7nRPGXfila\nbP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYpmjdbfxabTz07ig0+6/cwKoRR\nxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTFYvmAH7gcdssSFBt8NPlUATHE\nsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6Wdgmu5H2pDml8CDNLDv98Aj7i\n+I5SRKKcVPlnuQIDAQABo1AwTjAdBgNVHSUEFjAUBggrBgEFBQcDAQYIKwYBBQUH\nAwIwDAYDVR0TAQH/BAIwADAfBgNVHSMEGDAWgBQcwrrUO0u+FhIUuVdrDRCQRsi6\nZjANBgkqhkiG9w0BAQsFAAOCAQEAm5dxBoI9pScOCvRAchg4CprdRDSJb9K6yB3O\nnCAxnM47iHeXnY3WlnI388kHu8DU7O4ba1tJbGs3KY9KzioPk43pU12jWkO1onoF\n+mTDjx/Ef1cYWA9r5q/LtgTa6Q2sxV4O2x67QW82aAnaxO34dV5zWCPIvAoovZBV\nHRT+BgCg3r2vD1RGKK2nl1aYJtWhO1SZubam+VttdZ/vbM9oOJctxmImsEtBXjkY\nKteePdQtLL5o03JhyXWyRshCq+HMmKf2KgyY8gvydGcP4eLQdBWcW40LcnVq6UjT\n0kJycJEKngMVademq1ZWHGaiYB7hyT6GhgIcHUJ2cKrPgbEh1Q==\n-----END CERTIFICATE-----",
                      "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIEowIBAAKCAQEA4AI33yvHZws+jta6qpV6wzJqF8odIfTIpCfbrVcUUtLFKJ1I\n2e4SceTKi6O3C/I1XuvWlpng5IO65+fQQLO06z1/AuQT78YUn/Wlm9x1aHVsv4AN\nB5JWWqDOjrRT3o7nRPGXfilabP0rGE2mJJcVR9nExJ3IeaktgT3sb8YlXvtchyYp\nmjdbfxabTz07ig0+6/cwKoRRxOK8Uf7f5euE0cI/490J6r5Rs4lgD8sIQNCUFlTF\nYvmAH7gcdssSFBt8NPlUATHEsoFmlW0DKCJWNhTLOht+s6L/1zwTHLjPG2pdkG6W\ndgmu5H2pDml8CDNLDv98Aj7i+I5SRKKcVPlnuQIDAQABAoIBAFPQYKlOzw/+BA0b\nyMIUpdctIMb/54CR/xR0mVw1DbSjigNVPjHUQvB8Y1B2FAITQObgJO06bAv0QdWN\nRb0/v/yYiNJDFjaLjaIAHlO/2+oWrXbFaZqgpVDJhB+e1xaZr2x7XGxm+p925k30\nl6pvIRY+I8JRKvZiV1VZHwL/R3JOtPr++xMZtLVjVOI+f+ySqJ+TZHuAjm49EKxj\ncEmmJ28b7QcziXsvKy00f+zbqLIBKXQdZAFU5eEr1BsDRXdRW+Kf0XIvftuy4BJZ\nvoKT+VGhEvF/qysswL4+6IAO6tpuYnnM0Y2d3sOGoWPkTcQK0MekYKzL/WmtCjNs\n9hodJtECgYEA5EWyhEOf4uOKe5TDp697UCUvXLoOR58FDe/S8XNvScn29jjOkqIg\nOMoqo9xAkJTNTzqn5UUdt1x/pgM2NxlPLFijrc0zQlX3SoOO2ryDd9WNi7YKtN16\nKJqa536WeZu2OEbuAZ+S3GALVy1RPeTNPnUOmKnF06DjDUGzLNCZy10CgYEA+zfw\n952DWuz1U0Z4wvAEqqcgUKXPKrkTXV/iUnjkDkrLYVr0ZofDNTXrdHl+UedFmaOC\ncieZn6DNhcdz5tKtyysGMH3g/qs9PfoGUngvcXsy0Egk04l3x1jc8TTCLqXZXYaQ\nHMsx51n+R58oncPtzYSUOr9qQ6PbC2CstTbFJA0CgYEAjGEsUliAB/jknfEzjXjG\nPdhQUxb8VyE864Az2lah9t/kJzFyIAziAeqZ5GE7t247AGFTBRTHHI8e1Qoemi3P\nWbc9GVIbFs1lIYbcIDpUIyrKPEP8O5QEXtoNLxXTFgAjRGKiVY87spjCAJ+W2ZhO\ne/1it5GYXfgQCYQA2yuBmOUCgYANRkR2YR1axaCk+NlSu6oTdmdPu6M5x7PNQE7O\nOtMaKjua9lppvIzFGAdMDUtueoEEAE7ZR1xnwfB6PDLUpJdIYAqgr1YfPt8qkjaZ\nTv56yZ7CwL0pbF8m6nwqRrZoDp1wwraEvvvxFKFKGY/k3kCHlpTakdjEoDjn3gDi\nRnWeVQKBgCEneMSzucei5LRppRtRaJw/Btll8qlPMlX3W7dxQ3cLwpmLOn0m51Fp\nPIZ44zYK8R6fu4+/sSrlfaIg86Ugeufp6YNxyNROKxUGza5vDIu5OftwWtBeg+UK\nZ8lLWNdX6pp7WMujmF3H1DrkBbauYMUKZ4UxUYtelgHERMePIxwb\n-----END RSA PRIVATE KEY-----"
                    }
                  }
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "false"
          containerImage: quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
          createdAt: "2025-09-10T11:50:37Z"
          description: OpenShift Operator for Fujitsu Enterprise Postgres
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.1.0 <7.3.0'
          operators.openshift.io/valid-subscription: '["Fujitsu Enterprise Postgres
            Advanced Edition Annual Subscription License per Core 17 Operator for
            Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0-ocp
          operators.operatorframework.io/internal-objects: '["fepautoscales.fep.fujitsu.io","fepconfigs.fep.fujitsu.io","fepusers.fep.fujitsu.io","fepcerts.fep.fujitsu.io","fepvolumes.fep.fujitsu.io","fepbackups.fep.fujitsu.io"]'
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: fj-dbaas-tec@dl.jp.fujitsu.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: FEPAction
            name: fepactions.fep.fujitsu.io
            version: v1
          - kind: FEPAutoscale
            name: fepautoscales.fep.fujitsu.io
            version: v1
          - kind: FEPBackup
            name: fepbackups.fep.fujitsu.io
            version: v1
          - kind: FEPCert
            name: fepcerts.fep.fujitsu.io
            version: v1
          - kind: FEPCluster
            name: fepclusters.fep.fujitsu.io
            version: v2
          - kind: FEPConfig
            name: fepconfigs.fep.fujitsu.io
            version: v1
          - kind: FEPExporter
            name: fepexporters.fep.fujitsu.io
            version: v1
          - kind: FEPLogging
            name: feploggings.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2Cert
            name: feppgpool2certs.fep.fujitsu.io
            version: v1
          - kind: FEPPgpool2
            name: feppgpool2s.fep.fujitsu.io
            version: v1
          - kind: FEPRestore
            name: feprestores.fep.fujitsu.io
            version: v1
          - kind: FEPUpgrade
            name: fepupgrades.fep.fujitsu.io
            version: v1
          - kind: FEPUser
            name: fepusers.fep.fujitsu.io
            version: v1
          - kind: FEPVolume
            name: fepvolumes.fep.fujitsu.io
            version: v1
        description: "Fujitsu Enterprise Postgres 17 delivers an enterprise-grade
          PostgreSQL on OpenShift Container Platform.\n\nThis solution provides the
          flexibility of a hybrid cloud solution while delivering an enhanced distribution
          of PostgreSQL to support enterprise-level workloads and provide improved
          deployment and management, availability, performance, data governance and
          security.  \n\nAvailable as a multi-architecture container built for both
          amd64, s390x and ppc64le.\n\nThe download and Use of the Product is strictly
          subject to the terms of the End User License Agreement with Fujitsu Limited
          found at https://www.fast.fujitsu.com/fujitsu-enterprise-postgres-license-agreements.
          Where the Product that has been embedded as a whole or part into a third
          party program, only Authorised Customers may download and use the Product.\n"
        displayName: Fujitsu Enterprise Postgres Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - Fujitsu
        - fep
        links:
        - name: Fujitsu Enterprise Postgres
          url: https://www.postgresql.fastware.com/
        maintainers:
        - email: fj-dbaas-tec@dl.jp.fujitsu.com
          name: Fujitsu
        maturity: stable
        provider:
          name: Fujitsu
          url: https://www.postgresql.fastware.com/
        relatedImages:
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-backup@sha256:42252444e2dcfe5435532f318cce041990d2b71efa26007924717924d30a46ca
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-pgpool2@sha256:8eaace2f9c8e5e844992fe5e008cd1061af33cc79b4d442c4a83a65610bcc502
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-backup@sha256:78a3ad639d0c63858b9db0a5de36ce11a9ea121f4535e9ca07ac7defb689bc0d
        - registry.connect.redhat.com/fujitsu-postgres/fujitsu-enterprise-postgres-operator-bundle@sha256:6cd4cba02a9e5c22062108363f6a052734c398019a19f84afc18a7b05a71d578
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-server@sha256:fa4e8849d75daf331ba7d488bcb52c083cd92bb5f288e4b62c980754e299f618
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-restore@sha256:a6e90dd77634033f2009942b635078cceb0f4c6b15b031624c5dce820713b14c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-restore@sha256:17301c2191c0d75be82d12c5a35c0f74b8d65304bcfb6d4ab312a10ffa286eda
        - quay.io/fujitsu/fujitsu-enterprise-postgres-exporter@sha256:1c0e22aeab65ff3848772424fc3ea6baa6b58b0d88db5413bda160a08435bcff
        - quay.io/fujitsu/fujitsu-enterprise-postgres-operator@sha256:b7dd55c0bfda84e398cbfbb87345fcec218cb866e286e18a111330321ab6ad54
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-backup@sha256:ac268e2af652bdabb8ee5d764a2b335dc73c48b9022eee8c9f3cd422227a4d98
        - quay.io/fujitsu/fujitsu-enterprise-postgres-utils@sha256:57a3fb792480a0d3a8f075c48dee2d02f1d79cf459761e3305cf62799575a885
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-pgpool2@sha256:185af2e4fd1701aded1ff5e44d4336562f5a219c30ba535c2a3097b914c24efe
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-utils@sha256:b7ae535f5e5fa5eb51c6f4cdb406c7b7a9dabe8573939b0431dada6f7d27b87e
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-upgrade@sha256:8a0f7ca1f59ba162424a308e8a70c4717135e2847e61210753b71876c9b0a75f
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-utils@sha256:edeec4d00fd8282aaa7ec6fe3ec4cd84070a6402611002a18addcac87fb8ee94
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-restore@sha256:a81a0ce926572b75aa3151e2d2cc45a0ca411cc0158d69b89247ef70418f3740
        - quay.io/fujitsu/fujitsu-enterprise-postgres-cronjob@sha256:7582964eef14f89f2ca9046298357afd0132fb236f30f44f84cee14692227b0c
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-upgrade@sha256:c0e44447349f02eb9f6105001136402f849d876a288d61997c8c05625facff45
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentd@sha256:3c5534b62695867c5973c9106fc46895ed73a4b9136ad83b2586bf7e71c427d0
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-pgpool2@sha256:c878935ffc796c405d811a6e8a9194aaaaa8652ea7d7b879c82919ce09794b35
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-server@sha256:5c65f8e76f70341c512100760de6792a880ad0d6eebc9e68154b5bbd1332a896
        - quay.io/fujitsu/fujitsu-enterprise-postgres-15-backup@sha256:dd8b84e200c9d0fb32cd567a33968e15e15efc115ff120f54fdd8862c7ea3111
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-server@sha256:3e93a368a6198818e1b7c5525324c8130c5ed8d7b4cb499a47d4bec45ef0347b
        - quay.io/fujitsu/fujitsu-enterprise-postgres-17-server@sha256:1c5823afc7e77d46940214693c12c0ad6ffd6c7d9a156e26399ab55cd5ea1dfa
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-pgpool2@sha256:38aac8108dd3590abcfd854a6ec2c025a68d2170a1ba68d359b45fea6c83cec1
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-upgrade@sha256:bf3b069a71202ea78c7c0c7b36315c8bef8e6a1c11cd0668dcd888329e26ecfb
        - quay.io/fujitsu/fujitsu-enterprise-postgres-fluentbit@sha256:9622d2191525da39e5100e1882dd2cfa3eda740d8094f3fbb71ca8a5734d2843
        - quay.io/fujitsu/fujitsu-enterprise-postgres-16-restore@sha256:09be02274427e3022849fb635a4c9abdc34696838cae39f3afafc8009567db9d
        - quay.io/fujitsu/fujitsu-enterprise-postgres-14-upgrade@sha256:04af997b4f548ffec711b63a06c35715b4e7863537c83697b3ee0c1d95a6aec3
        version: 7.3.0
      entries:
      - name: fujitsu-enterprise-postgres-operator.v7.3.0
        version: 7.3.0
      - name: fujitsu-enterprise-postgres-operator.v7.2.7
        version: 7.2.7
      name: stable
    defaultChannel: stable
    packageName: fujitsu-enterprise-postgres-operator
    provider:
      name: Fujitsu
      url: https://www.postgresql.fastware.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-stats-dashboards
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-stats-dashboards.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Statsdashboard",
                "metadata": {
                  "name": "statsdashboard-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "statsDashboardService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:41:08Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Statsdashboard
            name: statsdashboards.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Stats Dashboard Operator
        displayName: Keysight WAP Stats Dashboard Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Stats Dashboards
          url: https://keysight-wap-stats-dashboards.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-stats-dashboards@sha256:bcdc68bf0b9e9c22d775d704320829d258a9ebc77dfe41e2414061740f87f236
        - registry.connect.redhat.com/loadcore/keysight-wap-stats-dashboards-bundle@sha256:f69f61eb818a1f0f4035ae79487402237b739c55eb104fc4643a7d3ef8d7d0c4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-stats-dashboards.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-stats-dashboards
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Arctera US LLC
      provider-url: https://www.arctera.io
    name: infoscale-licensing-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infoscale-licensing-operator.v8.0.220
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:feeabfe4ecd36205306e3f3bf6c11517cfee735d7ae7eb4350abd01e9d76610b
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          olm.skipRange: <8.0.220
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:6d9eabdb4f645aa1b8b6c08ca94f9ec89466cdc35b3cb02aaba22e160d8fd9ea
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:2f3de2bf46920cfe8c251011fc99841b43999561c1d3fe44850b205e94585a68
        version: 8.0.220
      entries:
      - name: infoscale-licensing-operator.v8.0.220
        version: 8.0.220
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: 8.0.200x
    - currentCSV: infoscale-licensing-operator.v8.0.330
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:5484f8ecf079ee126d24c05e1bcb06467146b3b8e751e90e0fd68f3e933125f0
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:b357b2203466516684c8fbc209757a98a45ec2376981f8699faf7dd5e37486a3
        version: 8.0.330
      entries:
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.300x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: 8.0.400x
    - currentCSV: infoscale-licensing-operator.v8.0.410
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Enterprise"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Arctera US LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Arctera
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Arctera
        - InfoScale
        - Licensing
        - Arctera Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Arctera US LLC
          url: https://www.arctera.io
        relatedImages:
        - registry.connect.redhat.com/arctera/infoscale-license-operator-bundle@sha256:f67479555b113a3ff194370dab48a517005001ce412a1d2ef1ecded9dd1c95f5
        - registry.connect.redhat.com/arctera/infoscale-license-operator@sha256:1ebab8aec17e0e5260678616696be43b720f801127c97820b3a07f3a8cb73e6a
        version: 8.0.410
      entries:
      - name: infoscale-licensing-operator.v8.0.410
        version: 8.0.410
      - name: infoscale-licensing-operator.v8.0.409
        version: 8.0.409
      - name: infoscale-licensing-operator.v8.0.400
        version: 8.0.400
      - name: infoscale-licensing-operator.v8.0.330
        version: 8.0.330
      - name: infoscale-licensing-operator.v8.0.320
        version: 8.0.320
      - name: infoscale-licensing-operator.v8.0.310
        version: 8.0.310
      - name: infoscale-licensing-operator.v8.0.301
        version: 8.0.301
      name: fast
    - currentCSV: infoscale-licensing-operator.v8.0.110
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vlic.veritas.com/v1",
                "kind": "License",
                "metadata": {
                  "name": "license-dev"
                },
                "spec": {
                  "licenseEdition": "Developer"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
          createdAt: "2022-06-22T16:36:59Z"
          description: InfoScale™ Licensing Operator manages the lifecycle of the
            Licensing for InfoScale™ cluster
          operatorframework.io/suggested-namespace: infoscale-vtas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: Veritas Technologies LLC
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: License is the Schema for the licenses API
            displayName: License
            kind: License
            name: licenses.vlic.veritas.com
            version: v1
        description: InfoScale™ Licensing Operator provides license to use Veritas
          Products.
        displayName: InfoScale™ Licensing Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Veritas
        - InfoScale
        - Licensing
        - Veritas Licensing
        - VLIC
        - Storage
        - Data Storage
        - SDS
        - Software Defined Storage
        links:
        - name: InfoScale™ Licensing Operator
          url: https://licensing-operator.domain
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: Veritas Technologies LLC
          url: https://www.veritas.com
        relatedImages:
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator-bundle@sha256:11ce1c35f9c3d9b5ef9539b4402eb8358df843b29ffd94af7a453e5d765e4767
        - registry.connect.redhat.com/veritas-technologies/infoscale-licensing-operator@sha256:e7b8d3c44df6151e61ded91336273e35eeaec04bd19124cf6313dc16d1224660
        version: 8.0.110
      entries:
      - name: infoscale-licensing-operator.v8.0.110
        version: 8.0.110
      - name: infoscale-licensing-operator.v8.0.100
        version: 8.0.100
      name: stable
    defaultChannel: fast
    packageName: infoscale-licensing-operator
    provider:
      name: Arctera US LLC
      url: https://www.arctera.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-wafv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-wafv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wafv2.services.k8s.aws/v1alpha1",
                "kind": "IPSet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
          createdAt: "2025-09-20T21:41:21Z"
          description: AWS WAFV2 controller is a service controller for managing WAFV2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IPSet represents the state of an AWS wafv2 IPSet resource.
            displayName: IPSet
            kind: IPSet
            name: ipsets.wafv2.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS wafv2 RuleGroup
              resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.wafv2.services.k8s.aws
            version: v1alpha1
          - description: WebACL represents the state of an AWS wafv2 WebACL resource.
            displayName: WebACL
            kind: WebACL
            name: webacls.wafv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Web Application Firewall (Amazon WAFV2) resources in AWS from within your Kubernetes cluster.

          **About Amazon WAFV2**

          AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to your protected web application resources. You can protect the following resource types:
          - Amazon CloudFront distribution
          - Amazon API Gateway REST API
          - Application Load Balancer
          - AWS AppSync GraphQL API
          - Amazon Cognito user pool
          - AWS App Runner service
          - AWS Verified Access instance

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon WAFV2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - wafv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon WAFV2 Developer Resources
          url: https://aws.amazon.com/waf/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: wafv2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/wafv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-wafv2-controller@sha256:cb9a426baec724225c0132973d5b90a83672f9b0b75451adc6d7fd858c79ab58
        version: 1.1.1
      entries:
      - name: ack-wafv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-wafv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-wafv2-controller.v1.0.14
        version: 1.0.14
      - name: ack-wafv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-wafv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-wafv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-wafv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-wafv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-wafv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-wafv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-wafv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-wafv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-wafv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-wafv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-wafv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-wafv2-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-wafv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-glue-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-glue-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
          createdAt: "2025-09-19T20:55:49Z"
          description: AWS Glue controller is a service controller for managing Glue
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Job represents the state of an AWS glue Job resource.
            displayName: Job
            kind: Job
            name: jobs.glue.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Glue resources in AWS from within your Kubernetes cluster.

          **About Amazon Glue**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Glue
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - glue
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Glue Developer Resources
          url: https://aws.amazon.com/Glue/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: glue maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/glue-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-glue-controller@sha256:2e1e23de186f3c61be8c7a30f3b295fe3b8b101509cf30bd6b7bff055ab74c82
        version: 0.2.1
      entries:
      - name: ack-glue-controller.v0.2.1
        version: 0.2.1
      - name: ack-glue-controller.v0.2.0
        version: 0.2.0
      - name: ack-glue-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-glue-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MinIO
      provider-url: https://min.io
    name: minio-object-store-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: minio-object-store-operator.v2025.8.19175300
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aistor.min.io/v1",
                "kind": "ObjectStore",
                "metadata": {
                  "annotations": {
                    "prometheus.io/path": "/minio/v2/metrics/cluster",
                    "prometheus.io/port": "9000",
                    "prometheus.io/scrape": "true"
                  },
                  "labels": {
                    "app": "minio"
                  },
                  "name": "my-objectstore"
                },
                "spec": {
                  "cache": {
                    "enabled": false
                  },
                  "certificates": {
                    "disableAutoCert": false
                  },
                  "configuration": {
                    "name": "storage-configuration"
                  },
                  "firewall": {
                    "enabled": false
                  },
                  "mountPath": "/export",
                  "podManagementPolicy": "Parallel",
                  "pools": [
                    {
                      "name": "pool-0",
                      "securityContext": {
                        "fsGroupChangePolicy": "OnRootMismatch"
                      },
                      "servers": 4,
                      "volumeClaimTemplate": {
                        "apiVersion": "v1",
                        "kind": "persistentvolumeclaims",
                        "spec": {
                          "accessModes": [
                            "ReadWriteOnce"
                          ],
                          "resources": {
                            "requests": {
                              "storage": "1Ti"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      "volumesPerServer": 4
                    }
                  ]
                }
              },
              {
                "apiVersion": "aistor.min.io/v1alpha1",
                "kind": "AdminJob",
                "metadata": {
                  "name": "minio-job"
                },
                "spec": {
                  "commands": [
                    {
                      "args": {
                        "name": "iac-bucket"
                      },
                      "op": "make-bucket"
                    }
                  ],
                  "containerSecurityContext": {},
                  "objectStore": {
                    "name": "my-objectstore",
                    "namespace": "my-objectstore"
                  },
                  "securityContext": {},
                  "serviceAccountName": "my-objectstore-sa"
                }
              },
              {
                "apiVersion": "sts.min.io/v1beta1",
                "kind": "PolicyBinding",
                "metadata": {
                  "name": "example-binding"
                },
                "spec": {
                  "application": {
                    "namespace": "my-objectstore",
                    "serviceaccount": "my-objectstore-sa"
                  },
                  "policies": [
                    "consoleAdmin"
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-08-27T14:50:56Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AdminJob
            name: adminjobs.aistor.min.io
            version: v1alpha1
          - kind: ObjectStore
            name: objectstores.aistor.min.io
            version: v1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1alpha1
          - kind: PolicyBinding
            name: policybindings.sts.min.io
            version: v1beta1
        description: "MinIO AIStor extends the most sophisticated object store on
          the market. It is designed for the exascale data \ninfrastructure challenges
          presented by modern AI workloads. AIStor provides new features, along with
          performance and\nscalability improvements, to enable enterprises to store
          all AI data in one infrastructure.\n\nMinIO is the standard for building
          <a href=\"https://min.io/solutions/object-storage-for-ai\">large scale AI
          data infrastructure</a>.\n\nThe MinIO AIStor is a software-defined, Amazon
          S3 compatible object store that will run anywhere - from the public cloud
          to the edge.<br><br>\n\nEnterprises use MinIO to deliver against artificial
          intelligence, machine learning, analytics, application, backup and archival
          workloads - all from a single platform.\n\nThe use cases include storage
          for HDFS replacements, AI datalakes, SIEM datalakes, analytics datalakes,
          object storage as a service, streaming and database warm/hot storage.<br><br>\n\nMinIO’s
          AIStor is the fastest object storage server, with production READ/WRITE
          speeds in excess of 2.2 TiB/s on commodity hardware.\n\nBuilt for the cloud
          operating model it is native to the technologies and architectures that
          define the cloud. These include containerization, orchestration with Kubernetes,
          microservices and multi-tenancy - making it the most widely integrated object
          store in the market today.<br><br>\n\nThe AIStor offers a host of enterprise
          features including inline erasure coding, bit-rot detection, state-of-the-art
          encryption, active-active replication, object locking, lifecycle management,
          observability, key-management, caching, firewall/load balancing, global
          console, search and identity + access management.<br><br>\n\nThe AIStor
          is priced on usable capacity and includes support through the MinIO Subscription
          Network.\n\nThe Subscription Network is designed to optimize deployments
          of MinIO and reduce the risk associated with data loss and data breach while
          providing 24/7/365 direct-to-engineering support.<br><br>\n\n<a href=\"https://resources.min.io/request-a-demo-marketplaces/\">Request
          a demo today</a>"
        displayName: Minio AIStor Object Store
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Storage
        - S3
        - MinIO
        - Object Storage
        links:
        - name: Minio AIStor
          url: https://min.io
        - name: Support
          url: https://subnet.min.io
        - name: Github
          url: https://github.com/minio/aistor
        maintainers:
        - email: dev@min.io
          name: MinIO
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: MinIO
          url: https://min.io
        relatedImages:
        - quay.io/minio/aistor/operator@sha256:f80139efa062fdd83f78318b20f371a827b68e854b03dd9706b9eba4c493aeb2
        - registry.connect.redhat.com/minio/minio-aistor-object-store-operator@sha256:c5af6adc5e7f9bb2d5bf6025b1b8c3e33ce5b9349a0246df1a26b7ec91b2ea2f
        version: 2025.8.19175300
      entries:
      - name: minio-object-store-operator.v2025.8.19175300
        version: 2025.8.19175300
      - name: minio-object-store-operator.v2025.8.1184927
        version: 2025.8.1184927
      - name: minio-object-store-operator.v2025.7.1011319
        version: 2025.7.1011319
      - name: minio-object-store-operator.v2025.6.3065135
        version: 2025.6.3065135
      - name: minio-object-store-operator.v2025.5.12190907
        version: 2025.5.12190907
      name: stable
    defaultChannel: stable
    packageName: minio-object-store-operator
    provider:
      name: MinIO
      url: https://min.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Lightbits Labs
      provider-url: https://www.lightbitslabs.com
    name: lightbits-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: lightbits-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "CSIDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "csidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "csidriver",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "csi.lightbitslabs.com",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "controller": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "tag": "v1.21.0-ubi9"
                  },
                  "node": {
                    "registry": "docker.lightbitslabs.com",
                    "repository": "lightbits-operator/lb-csi-plugin",
                    "sidecars": {
                      "lb-nvme-discovery-client": {
                        "registry": "docker.lightbitslabs.com",
                        "repository": "lightbits-operator/lb-nvme-discovery-client",
                        "tag": "v1.21.0-ubi9"
                      }
                    },
                    "tag": "v1.21.0-ubi9"
                  }
                }
              },
              {
                "apiVersion": "storage.lightbitslabs.com/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "lightbits-operator",
                    "app.kubernetes.io/instance": "cluster1",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cluster",
                    "app.kubernetes.io/part-of": "lightbits-operator"
                  },
                  "name": "cluster1",
                  "namespace": "lightbits-operator"
                },
                "spec": {
                  "addresses": "192.168.40.46,192.168.40.47,192.168.40.48",
                  "apiService": "192.168.40.46:443,192.168.40.47:443,192.168.40.48:443",
                  "clusterAdminSecretName": "lb-secret",
                  "clusterName": "cluster1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          com.redhat.openshift.versions: v4.17
          containerImage: docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
          createdAt: "2025-08-20T11:35:15Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/supported-versions: v4.17
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '[]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          operators.operatorframework.io/suggested-namespace: lightbits-operator
          repository: https://github.com/LightBitsLabs/lightbits-operator
          support: Lightbits Labs <support@lightbitslabs.com>
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster is the Schema for the clusters API
            displayName: Lightbits Cluster
            kind: Cluster
            name: clusters.storage.lightbitslabs.com
            version: v1alpha1
          - description: CSIDriver is the Schema for the csidrivers API
            displayName: Lightbits CSI Driver
            kind: CSIDriver
            name: csidrivers.storage.lightbitslabs.com
            version: v1alpha1
        description: |-
          Operator for deploying and managing Lightbits CSI driver.

          By using the operator users can quickly and easily deploy and upgrade Lightbits CSI Driver.

          ## Requirements

          - Kernel support for NVMe-TCP
          - Kernel support for NVMe Asymmetric Namespace Access (ANA)

          ## CSI driver Supported Features

          - Filesystem Volumes (xfs/ext4)
          - Raw Block Volumes
          - Dynamic/Static Provisioning
          - Volume Snapshots
          - Volume Clones
          - Multi-Tenancy & Credentials
          - Volume Expansion (Online/Offline)
          - Storage Capacity Tracking
          - Host-Side Encryption (LUKS)
          - Publish ReadOnly volumes

          ## How to install

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/install_lightbits_operator.html#installation)

          ## Documentation

          See the official documentation [here](https://lightbitslabs.github.io/lightbits-operator/)
        displayName: Lightbits Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - storage
        - lightbits
        links:
        - name: Lightbits Operator
          url: https://github.com/lightbitslabs/lightbits-operator
        maintainers:
        - email: k8s@lightbitslabs.com
          name: Lightbits Labs
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Lightbits Labs
          url: https://www.lightbitslabs.com
        relatedImages:
        - docker.lightbitslabs.com/lightbits-operator/lightbits-operator@sha256:26c4def57feb42c98d336e20c89644d82ebdee0b43da6e73e90f8e2e2530fd06
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - registry.connect.redhat.com/lightbitslabs/lightbits-operator@sha256:53db496f7a254632a25acf8974aca4041cbddb109499871622ffc2182e7aaa83
        version: 1.2.0
      entries:
      - name: lightbits-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: lightbits-operator
    provider:
      name: Lightbits Labs
      url: https://www.lightbitslabs.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-zookeeper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: zookeeper-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
          description: Stackable Operator for Apache ZooKeeper
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/zookeeper-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for ZookeeperClusterSpec via
              `CustomResource`
            displayName: zookeeperclusters.zookeeper.stackable.tech
            kind: ZookeeperCluster
            name: zookeeperclusters.zookeeper.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ZookeeperZnodeSpec via `CustomResource`
            displayName: zookeeperznodes.zookeeper.stackable.tech
            kind: ZookeeperZnode
            name: zookeeperznodes.zookeeper.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache ZooKeeper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - zookeeper
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/zookeeper-operator@sha256:c934d08ecfc2b375c9d72b95801dc5a7a6e892542b12498809b345fda182d887
        - registry.connect.redhat.com/stackable/stackable-apache-zookeeper-operator@sha256:b4183289cc6d66e36aab6b956d932d1e6537a8975da5b5fa534ff3293e02e83f
        version: 25.7.0
      entries:
      - name: zookeeper-operator.v25.7.0
        version: 25.7.0
      - name: zookeeper-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-zookeeper-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ram-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ram-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
          createdAt: "2025-09-19T22:18:09Z"
          description: AWS RAM controller is a service controller for managing RAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Permission represents the state of an AWS ram Permission
              resource.
            displayName: Permission
            kind: Permission
            name: permissions.ram.services.k8s.aws
            version: v1alpha1
          - description: ResourceShare represents the state of an AWS ram ResourceShare
              resource.
            displayName: ResourceShare
            kind: ResourceShare
            name: resourceshares.ram.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon RAM resources in AWS from within your Kubernetes cluster.

          **About Amazon RAM**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon RAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ram
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon RAM Developer Resources
          url: https://aws.amazon.com/RAM/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ram maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ram-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ram-controller@sha256:a90b07270f0cfa8929e01569895ab9a209ee44dcc452a3e1f6e0e9a52c5f13b9
        version: 1.1.1
      entries:
      - name: ack-ram-controller.v1.1.1
        version: 1.1.1
      - name: ack-ram-controller.v1.1.0
        version: 1.1.0
      - name: ack-ram-controller.v1.0.11
        version: 1.0.11
      - name: ack-ram-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-ram-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecs-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Service",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ecs.services.k8s.aws/v1alpha1",
                "kind": "TaskDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
          createdAt: "2025-09-19T21:02:44Z"
          description: AWS ECS controller is a service controller for managing ECS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS ecs Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.ecs.services.k8s.aws
            version: v1alpha1
          - description: Service represents the state of an AWS ecs Service resource.
            displayName: Service
            kind: Service
            name: services.ecs.services.k8s.aws
            version: v1alpha1
          - description: TaskDefinition represents the state of an AWS ecs TaskDefinition
              resource.
            displayName: TaskDefinition
            kind: TaskDefinition
            name: taskdefinitions.ecs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Service (ECS) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECS**

          Amazon Elastic Container Service (Amazon ECS) is a fully managed container orchestration service that helps you easily deploy, manage, and scale containerized applications. As a fully managed service, Amazon ECS comes with AWS configuration and operational best practices built-in. It's integrated with both AWS and third-party tools, such as Amazon Elastic Container Registry and Docker. This integration makes it easier for teams to focus on building the applications, not the environment. You can run and scale your container workloads across AWS Regions in the cloud, and on-premises, without the complexity of managing a control plane.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECS Developer Resources
          url: https://aws.amazon.com/ecs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecs-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecs-controller@sha256:280d5ec57ad4deb45c145e702242468a4cc04d23112c33f4e72722a269c900e8
        version: 1.1.1
      entries:
      - name: ack-ecs-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecs-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecs-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecs-controller.v1.0.13
        version: 1.0.13
      - name: ack-ecs-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecs-controller.v1.0.11
        version: 1.0.11
      - name: ack-ecs-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecs-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecs-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecs-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecs-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecs-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecs-controller.v1.0.1
        version: 1.0.1
      - name: ack-ecs-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecs-controller.v0.0.8
        version: 0.0.8
      - name: ack-ecs-controller.v0.0.7
        version: 0.0.7
      - name: ack-ecs-controller.v0.0.6
        version: 0.0.6
      - name: ack-ecs-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://your.domain
    name: amq-streams-proxy
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: alpha
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.0.x
    - currentCSV: amq-streams-proxy.v3.0.1-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                  "kind": "KafkaProxy",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "simple"
                  },
                  "spec": {}
                },
                {
                  "kind": "KafkaProxyIngress",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "cluster-ip"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "clusterIP": {
                      "protocol": "TCP"
                    }
                  }
                },
                {
                  "kind": "KafkaService",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "bootstrapServers": "mykafka.example.com:9092",
                     "nodeIdRanges": [
                      {
                        "start": 0,
                        "end": 2
                      }
                    ]
                  }
                },
                {
                  "kind": "VirtualKafkaCluster",
                  "apiVersion": "kroxylicious.io/v1alpha1",
                  "metadata": {
                    "name": "my-cluster"
                  },
                  "spec": {
                    "proxyRef": {
                      "name": "simple"
                    },
                    "targetKafkaServiceRef": {
                      "name": "my-cluster"
                    },
                    "ingresses": [
                      {
                        "ingressRef": {
                          "name": "cluster-ip"
                        }
                      }
                    ]
                  }
                }
            ]
          alm-examples-metadata: |-
            {
              "example-proxy": {
                "description": "An example Streams Proxy instance that proxies a Kafka cluster, exposing it with a virtual kafka cluster using cluster-ip ingress. This configuration does not define any protocol filters so will not transform the protocol in anyway. This configuration does not supply the kafka cluster.  It is assumed a hypothetical cluster at address mykafka.example.com:9092."
              }
            }
          capabilities: Basic Install
          certified: "false"
          containerImage: registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
          createdAt: "2025-09-24T11:54:56Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0-0 <3.0.1-2'
          operators.openshift.io/valid-subscription: '["Red Hat AMQ", "Red Hat Application
            Foundations"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A `KafkaProxy` resource represents an instance of the Kroxylicious
              Proxy.
            displayName: KafkaProxy
            kind: KafkaProxy
            name: kafkaproxies.kroxylicious.io
            version: v1alpha1
          - description: A `VirtualKafkaCluster` resource defines a logical Kafka
              cluster that is accessible to clients over the network.
            displayName: VirtualKafkaCluster
            kind: VirtualKafkaCluster
            name: virtualkafkaclusters.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProxyIngress` resource defines the networking configuration
              that allows Kafka clients to connect to a `VirtualKafkaCluster`.
            displayName: KafkaProxyIngress
            kind: KafkaProxyIngress
            name: kafkaproxyingresses.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaService` resource that defined the backend Kafka
              cluster.
            displayName: KafkaService
            kind: KafkaService
            name: kafkaservices.kroxylicious.io
            version: v1alpha1
          - description: A `KafkaProtocolFilter` resource represents a Kroxylicious
              Proxy filter.
            displayName: KafkaProtocolFilter
            kind: KafkaProtocolFilter
            name: kafkaprotocolfilters.kroxylicious.io
            version: v1alpha1
        description: Red Hat Streams for Apache Kafka Proxy is an Apache Kafka® protocol-aware
          proxy. It can be used to layer uniform behaviours onto a Kafka based system
          in areas such as data-governance, security, policy enforcement and audit
          without needing to change either the applications or the Kafka Cluster.
        displayName: Streams for Apache Kafka Proxy
        installModes:
        - supported: true
          type: AllNamespaces
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        keywords:
        - kafka
        - messaging
        - kafka-streams
        - data-streaming
        - data-streams
        - streaming
        - streams
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-amq
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_streams_for_apache_kafka/3.0
        maintainers:
        - email: customerservice@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://your.domain
        relatedImages:
        - registry.redhat.io/amq-streams/proxy-rhel9@sha256:340811893cd006f9a36f0cb63158badc5c92c936f6cd00f8079ef8062bd4db42
        - registry.redhat.io/amq-streams/proxy-operator-bundle@sha256:f654d5100b027d9fd677deef93905b222953fface1e78dd34333295817086dbf
        - registry.redhat.io/amq-streams/proxy-rhel9-operator@sha256:ec03e2581d5850bb652e2bf63af2dd877bf40f9bfc289d710ed03a66baa4facf
        version: 3.0.1-2
      entries:
      - name: amq-streams-proxy.v3.0.1-2
        version: 3.0.1-2
      - name: amq-streams-proxy.v3.0.1-1
        version: 3.0.1-1
      - name: amq-streams-proxy.v3.0.0-8
        version: 3.0.0-8
      - name: amq-streams-proxy.v3.0.0-5
        version: 3.0.0-5
      name: amq-streams-3.x
    defaultChannel: alpha
    packageName: amq-streams-proxy
    provider:
      name: Red Hat
      url: https://your.domain
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sagemaker-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sagemaker-controller.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "App",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "DataQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Domain",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "EndpointConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "FeatureGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "HyperParameterTuningJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "Model",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelBiasJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelExplainabilityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackage",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelPackageGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ModelQualityJobDefinition",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "MonitoringSchedule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "NotebookInstanceLifecycleConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "ProcessingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TrainingJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "TransformJob",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sagemaker.services.k8s.aws/v1alpha1",
                "kind": "UserProfile",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
          createdAt: "2025-09-22T19:07:36Z"
          description: ACK service controller for Amazon SageMaker manages SageMaker
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: App represents the state of an AWS sagemaker App resource.
            displayName: App
            kind: App
            name: apps.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: DataQualityJobDefinition represents the state of an AWS sagemaker
              DataQualityJobDefinition resource.
            displayName: DataQualityJobDefinition
            kind: DataQualityJobDefinition
            name: dataqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Domain represents the state of an AWS sagemaker Domain resource.
            displayName: Domain
            kind: Domain
            name: domains.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: EndpointConfig represents the state of an AWS sagemaker EndpointConfig
              resource.
            displayName: EndpointConfig
            kind: EndpointConfig
            name: endpointconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS sagemaker Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: FeatureGroup represents the state of an AWS sagemaker FeatureGroup
              resource.
            displayName: FeatureGroup
            kind: FeatureGroup
            name: featuregroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: HyperParameterTuningJob represents the state of an AWS sagemaker
              HyperParameterTuningJob resource.
            displayName: HyperParameterTuningJob
            kind: HyperParameterTuningJob
            name: hyperparametertuningjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: InferenceComponent represents the state of an AWS sagemaker
              InferenceComponent resource.
            displayName: InferenceComponent
            kind: InferenceComponent
            name: inferencecomponents.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: LabelingJob represents the state of an AWS sagemaker LabelingJob
              resource.
            displayName: LabelingJob
            kind: LabelingJob
            name: labelingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelBiasJobDefinition represents the state of an AWS sagemaker
              ModelBiasJobDefinition resource.
            displayName: ModelBiasJobDefinition
            kind: ModelBiasJobDefinition
            name: modelbiasjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelExplainabilityJobDefinition represents the state of
              an AWS sagemaker ModelExplainabilityJobDefinition resource.
            displayName: ModelExplainabilityJobDefinition
            kind: ModelExplainabilityJobDefinition
            name: modelexplainabilityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackageGroup represents the state of an AWS sagemaker
              ModelPackageGroup resource.
            displayName: ModelPackageGroup
            kind: ModelPackageGroup
            name: modelpackagegroups.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelPackage represents the state of an AWS sagemaker ModelPackage
              resource.
            displayName: ModelPackage
            kind: ModelPackage
            name: modelpackages.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ModelQualityJobDefinition represents the state of an AWS
              sagemaker ModelQualityJobDefinition resource.
            displayName: ModelQualityJobDefinition
            kind: ModelQualityJobDefinition
            name: modelqualityjobdefinitions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Model represents the state of an AWS sagemaker Model resource.
            displayName: Model
            kind: Model
            name: models.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: MonitoringSchedule represents the state of an AWS sagemaker
              MonitoringSchedule resource.
            displayName: MonitoringSchedule
            kind: MonitoringSchedule
            name: monitoringschedules.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstanceLifecycleConfig represents the state of an
              AWS sagemaker NotebookInstanceLifecycleConfig resource.
            displayName: NotebookInstanceLifecycleConfig
            kind: NotebookInstanceLifecycleConfig
            name: notebookinstancelifecycleconfigs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: NotebookInstance represents the state of an AWS sagemaker
              NotebookInstance resource.
            displayName: NotebookInstance
            kind: NotebookInstance
            name: notebookinstances.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: PipelineExecution represents the state of an AWS sagemaker
              PipelineExecution resource.
            displayName: PipelineExecution
            kind: PipelineExecution
            name: pipelineexecutions.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: Pipeline represents the state of an AWS sagemaker Pipeline
              resource.
            displayName: Pipeline
            kind: Pipeline
            name: pipelines.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: ProcessingJob represents the state of an AWS sagemaker ProcessingJob
              resource.
            displayName: ProcessingJob
            kind: ProcessingJob
            name: processingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TrainingJob represents the state of an AWS sagemaker TrainingJob
              resource.
            displayName: TrainingJob
            kind: TrainingJob
            name: trainingjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: TransformJob represents the state of an AWS sagemaker TransformJob
              resource.
            displayName: TransformJob
            kind: TransformJob
            name: transformjobs.sagemaker.services.k8s.aws
            version: v1alpha1
          - description: UserProfile represents the state of an AWS sagemaker UserProfile
              resource.
            displayName: UserProfile
            kind: UserProfile
            name: userprofiles.sagemaker.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon SageMaker resources in AWS from within your Kubernetes cluster.

          **About Amazon SageMaker**

          Amazon SageMaker helps data scientists and developers to prepare, build, train, and deploy high-quality machine learning (ML) models quickly by bringing together a broad set of capabilities purpose-built for ML.

          For more information on Amazon SageMaker, visit the [product page](https://aws.amazon.com/sagemaker/).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SageMaker
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sagemaker
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SageMaker Developer Resources
          url: https://aws.amazon.com/sagemaker/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sagemaker maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-sagemaker-controller@sha256:c962ee90891a05197edc08b4cf015fce1e2431cced90bd274e89bb87708420f3
        - public.ecr.aws/aws-controllers-k8s/sagemaker-controller:1.4.1
        version: 1.4.1
      entries:
      - name: ack-sagemaker-controller.v1.4.1
        version: 1.4.1
      - name: ack-sagemaker-controller.v1.4.0
        version: 1.4.0
      - name: ack-sagemaker-controller.v1.3.3
        version: 1.3.3
      - name: ack-sagemaker-controller.v1.3.2
        version: 1.3.2
      - name: ack-sagemaker-controller.v1.3.1
        version: 1.3.1
      - name: ack-sagemaker-controller.v1.3.0
        version: 1.3.0
      - name: ack-sagemaker-controller.v1.2.20
        version: 1.2.20
      - name: ack-sagemaker-controller.v1.2.19
        version: 1.2.19
      - name: ack-sagemaker-controller.v1.2.18
        version: 1.2.18
      - name: ack-sagemaker-controller.v1.2.17
        version: 1.2.17
      - name: ack-sagemaker-controller.v1.2.16
        version: 1.2.16
      - name: ack-sagemaker-controller.v1.2.13
        version: 1.2.13
      - name: ack-sagemaker-controller.v1.2.12
        version: 1.2.12
      - name: ack-sagemaker-controller.v1.2.11
        version: 1.2.11
      - name: ack-sagemaker-controller.v1.2.9
        version: 1.2.9
      - name: ack-sagemaker-controller.v1.2.8
        version: 1.2.8
      - name: ack-sagemaker-controller.v1.2.5
        version: 1.2.5
      - name: ack-sagemaker-controller.v1.2.4
        version: 1.2.4
      - name: ack-sagemaker-controller.v1.2.3
        version: 1.2.3
      - name: ack-sagemaker-controller.v1.2.2
        version: 1.2.2
      - name: ack-sagemaker-controller.v1.2.1
        version: 1.2.1
      - name: ack-sagemaker-controller.v1.2.0
        version: 1.2.0
      - name: ack-sagemaker-controller.v1.1.0
        version: 1.1.0
      - name: ack-sagemaker-controller.v1.0.0
        version: 1.0.0
      - name: ack-sagemaker-controller.v0.6.0
        version: 0.6.0
      - name: ack-sagemaker-controller.v0.5.1
        version: 0.5.1
      - name: ack-sagemaker-controller.v0.5.0
        version: 0.5.0
      - name: ack-sagemaker-controller.v0.4.5
        version: 0.4.5
      - name: ack-sagemaker-controller.v0.4.4
        version: 0.4.4
      - name: ack-sagemaker-controller.v0.4.3
        version: 0.4.3
      - name: ack-sagemaker-controller.v0.4.2
        version: 0.4.2
      - name: ack-sagemaker-controller.v0.4.0
        version: 0.4.0
      - name: ack-sagemaker-controller.v0.3.4
        version: 0.3.4
      - name: ack-sagemaker-controller.v0.3.3
        version: 0.3.3
      - name: ack-sagemaker-controller.v0.3.2
        version: 0.3.2
      - name: ack-sagemaker-controller.v0.3.1
        version: 0.3.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sagemaker-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-bedrockagent-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-bedrockagent-controller.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "bedrockagent.services.k8s.aws/v1alpha1",
                "kind": "Agent",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
          createdAt: "2025-09-20T01:18:56Z"
          description: AWS Bedrock Agent controller is a service controller for managing
            Bedrock Agent resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Agent represents the state of an AWS bedrockagent Agent resource.
            displayName: Agent
            kind: Agent
            name: agents.bedrockagent.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Bedrock Agent resources in AWS from within your Kubernetes cluster.

          **About Amazon Bedrock Agent**

          Amazon Bedrock is a fully managed service that makes high-performing foundation models (FMs) from leading AI companies and Amazon available for your use through a unified API. You can choose from a wide range of foundation models to find the model that is best suited for your use case. Amazon Bedrock also offers a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI. Using Amazon Bedrock, you can easily experiment with and evaluate top foundation models for your use cases, privately customize them with your data using techniques such as fine-tuning and Retrieval Augmented Generation (RAG), and build agents that execute tasks using your enterprise systems and data sources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Bedrock Agent
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - bedrockagent
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Bedrock Agent Developer Resources
          url: https://aws.amazon.com/bedrock/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: bedrock-agent maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-bedrockagent-controller@sha256:6688b26f71500af0e2ffbd0ccc53926b0c41adaaf817a88f0fdfd6fb398075e7
        - public.ecr.aws/aws-controllers-k8s/bedrockagent-controller:0.1.1
        version: 0.1.1
      entries:
      - name: ack-bedrockagent-controller.v0.1.1
        version: 0.1.1
      - name: ack-bedrockagent-controller.v0.1.0
        version: 0.1.0
      - name: ack-bedrockagent-controller.v0.0.6
        version: 0.0.6
      - name: ack-bedrockagent-controller.v0.0.5
        version: 0.0.5
      - name: ack-bedrockagent-controller.v0.0.4
        version: 0.0.4
      - name: ack-bedrockagent-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-bedrockagent-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: docker.io/netapp/trident-operator:25.06.2
          createdAt: "2025-08-03"
          description: Trident Operator, to manage NetApp Trident installations
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy Astra Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-astra-product@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - docker.io/netapp/trident-operator:25.06.2
        - quay.io/community-operator-pipeline-prod/trident-operator:25.6.2
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      - name: trident-operator.v24.10.1
        version: 24.10.1
      - name: trident-operator.v24.10.0
        version: 24.10.0
      - name: trident-operator.v24.6.0
        version: 24.6.0
      - name: trident-operator.v24.2.0
        version: 24.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: sailoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sailoperator.v1.28.0-nightly-2025-09-30
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
          createdAt: "2025-09-30T03:27:25Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - master
          - v1.28-alpha.41429998

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:c9dab4079aa46abb0d61953255436dc64bd6e68fff4751080bb6ecd123bb489a
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - quay.io/community-operator-pipeline-prod/sailoperator:1.28.0-nightly-2025-09-30
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-testing/proxyv2@sha256:28f058629e7969e8441b677b308a09bf5d4968dbd5024d54df8472d1f43e1a8e
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-testing/ztunnel@sha256:f4895a66a99ca0c87e54260d6d5784068a48fe7fbfa25144debad420ea7223fa
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-testing/pilot@sha256:42d64978872f6be8540027002a361933e4fe69590458353e39d2643b2459bb3c
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-testing/install-cni@sha256:a8c8f6fb3048461e08b86056000a3034f7945df067d5ebf43cca9bcb0ffa3916
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        version: 1.28.0-nightly-2025-09-30
      entries:
      - name: sailoperator.v1.28.0-nightly-2025-09-30
        version: 1.28.0-nightly-2025-09-30
      - name: sailoperator.v1.28.0-nightly-2025-09-29
        version: 1.28.0-nightly-2025-09-29
      - name: sailoperator.v1.28.0-nightly-2025-09-28
        version: 1.28.0-nightly-2025-09-28
      - name: sailoperator.v1.28.0-nightly-2025-09-27
        version: 1.28.0-nightly-2025-09-27
      - name: sailoperator.v1.28.0-nightly-2025-09-26
        version: 1.28.0-nightly-2025-09-26
      - name: sailoperator.v1.28.0-nightly-2025-09-25
        version: 1.28.0-nightly-2025-09-25
      - name: sailoperator.v1.28.0-nightly-2025-09-23
        version: 1.28.0-nightly-2025-09-23
      - name: sailoperator.v1.28.0-nightly-2025-09-22
        version: 1.28.0-nightly-2025-09-22
      - name: sailoperator.v1.28.0-nightly-2025-09-21
        version: 1.28.0-nightly-2025-09-21
      - name: sailoperator.v1.28.0-nightly-2025-09-19
        version: 1.28.0-nightly-2025-09-19
      - name: sailoperator.v1.28.0-nightly-2025-09-18
        version: 1.28.0-nightly-2025-09-18
      - name: sailoperator.v1.28.0-nightly-2025-09-17
        version: 1.28.0-nightly-2025-09-17
      - name: sailoperator.v1.28.0-nightly-2025-09-16
        version: 1.28.0-nightly-2025-09-16
      name: 1.28-nightly
    - currentCSV: sailoperator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.23.0"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1alpha1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.23.0"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/maistra-dev/sail-operator:0.1-latest
          createdAt: "2024-08-20T11:54:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1alpha1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1alpha1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1alpha1
          - description: |-
              RemoteIstio represents a remote Istio Service Mesh deployment consisting of one or more
              remote control plane instances (represented by one or more IstioRevision objects).
            displayName: Remote Istio
            kind: RemoteIstio
            name: remoteistios.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.23.0
          - v1.22.4

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - quay.io/maistra-dev/sail-operator:0.1-latest
        - docker.io/istio/install-cni:1.23.0
        - docker.io/istio/pilot:1.22.4
        - docker.io/istio/pilot:1.23.0
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:a5698068f57fb27c6e4b8c385878c4651967ae5f17412c6ea6cf8bd34bfe76a1
        - docker.io/istio/proxyv2:1.22.4
        - docker.io/istio/ztunnel:1.23.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.16.0
        - docker.io/istio/install-cni:1.22.4
        - docker.io/istio/proxyv2:1.23.0
        - docker.io/istio/ztunnel:1.22.4
        version: 0.1.0
      entries:
      - name: sailoperator.v0.1.0
        version: 0.1.0
      - name: sailoperator.v0.1.0-rc.1
        version: 0.1.0-rc.1
      name: candidates
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable
    - currentCSV: sailoperator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.24.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.24.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.0-latest
          createdAt: "2025-03-07T10:17:30Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.24-latest
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/ztunnel:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/proxyv2:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.0
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/install-cni:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/pilot:1.23.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/install-cni:1.23.0
        - gcr.io/istio-release/pilot:1.24.1
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:5416c5fc62c7d9498349ed771bf93bfc0f712ad8da0b8ee109dc49d1e684fa56
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/proxyv2:1.23.3
        - quay.io/sail-dev/sail-operator:1.0-latest
        - gcr.io/istio-release/proxyv2:1.24.3
        version: 1.0.0
      entries:
      - name: sailoperator.v1.0.0
        version: 1.0.0
      name: stable-1.0
    - currentCSV: sailoperator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.25.2"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.25.2"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator:1.25.2
          createdAt: "2025-05-07T19:44:51Z"
          description: Experimental operator for installing Istio service mesh
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.25.2'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          This is an experimental operator for installing Istio service mesh.

          This version of the operator supports the following Istio versions:

          - v1.25-latest
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0
          - v1.23-latest
          - v1.23.5
          - v1.23.4
          - v1.23.3
          - v1.23.2
          - v1.22-latest
          - v1.22.8
          - v1.22.7
          - v1.22.6
          - v1.22.5
          - v1.21.6

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.3
        - gcr.io/istio-release/pilot:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.2
        - gcr.io/istio-release/ztunnel:1.23.2
        - gcr.io/istio-release/ztunnel:1.24.2
        - gcr.io/istio-release/pilot:1.22.8
        - gcr.io/istio-release/pilot:1.23.2
        - gcr.io/istio-release/proxyv2:1.21.6
        - gcr.io/istio-release/proxyv2:1.23.5
        - gcr.io/istio-release/ztunnel:1.24.4
        - gcr.io/istio-release/proxyv2:1.25.1
        - gcr.io/istio-release/proxyv2:1.24.3
        - gcr.io/istio-release/proxyv2:1.23.4
        - gcr.io/istio-release/pilot:1.24.1
        - gcr.io/istio-release/proxyv2:1.22.8
        - gcr.io/istio-release/install-cni:1.22.7
        - gcr.io/istio-release/proxyv2:1.22.7
        - gcr.io/istio-release/install-cni:1.24.1
        - gcr.io/istio-release/install-cni:1.24.0
        - gcr.io/istio-release/ztunnel:1.23.4
        - gcr.io/istio-release/ztunnel:1.25.2
        - gcr.io/istio-release/ztunnel:1.22.8
        - gcr.io/istio-release/ztunnel:1.24.0
        - gcr.io/istio-release/install-cni:1.23.4
        - gcr.io/istio-release/install-cni:1.24.4
        - gcr.io/istio-release/install-cni:1.22.8
        - gcr.io/istio-release/ztunnel:1.23.3
        - gcr.io/istio-release/ztunnel:1.25.1
        - gcr.io/istio-release/pilot:1.21.6
        - gcr.io/istio-release/pilot:1.22.6
        - gcr.io/istio-release/install-cni:1.25.1
        - gcr.io/istio-release/install-cni:1.25.2
        - gcr.io/istio-release/ztunnel:1.24.1
        - gcr.io/istio-release/install-cni:1.23.3
        - gcr.io/istio-release/pilot:1.23.4
        - gcr.io/istio-release/install-cni:1.22.5
        - gcr.io/istio-release/proxyv2:1.24.2
        - quay.io/community-operator-pipeline-prod/sailoperator@sha256:9cc4d5a9f46e4d05f8adb5232459002a073c101cc813330325afb6682e2138ae
        - gcr.io/istio-release/pilot:1.24.2
        - gcr.io/istio-release/install-cni:1.23.2
        - gcr.io/istio-release/pilot:1.24.4
        - gcr.io/istio-release/proxyv2:1.22.6
        - gcr.io/istio-release/pilot:1.23.5
        - gcr.io/istio-release/proxyv2:1.22.5
        - gcr.io/istio-release/pilot:1.22.5
        - gcr.io/istio-release/ztunnel:1.24.3
        - gcr.io/istio-release/install-cni:1.22.6
        - gcr.io/istio-release/pilot:1.22.7
        - gcr.io/istio-release/proxyv2:1.24.0
        - gcr.io/istio-release/proxyv2:1.24.1
        - gcr.io/istio-release/pilot:1.25.1
        - gcr.io/istio-release/pilot:1.24.0
        - gcr.io/istio-release/ztunnel:1.22.7
        - gcr.io/istio-release/pilot:1.25.2
        - gcr.io/istio-release/pilot:1.23.3
        - gcr.io/istio-release/install-cni:1.23.5
        - quay.io/sail-dev/sail-operator:1.25.2
        - gcr.io/istio-release/ztunnel:1.21.6
        - gcr.io/istio-release/install-cni:1.24.2
        - gcr.io/istio-release/ztunnel:1.22.5
        - gcr.io/istio-release/ztunnel:1.22.6
        - gcr.io/istio-release/install-cni:1.21.6
        - gcr.io/istio-release/proxyv2:1.25.2
        - gcr.io/istio-release/ztunnel:1.23.5
        - gcr.io/istio-release/proxyv2:1.24.4
        version: 1.25.2
      entries:
      - name: sailoperator.v1.25.2
        version: 1.25.2
      - name: sailoperator.v1.25.1
        version: 1.25.1
      - name: sailoperator.v1.25.0
        version: 1.25.0
      name: stable-1.25
    - currentCSV: sailoperator.v1.26.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.26.3"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.26.3"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
          createdAt: "2025-08-04T06:23:45Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.26-latest
          - v1.26.3
          - v1.26.2
          - v1.26.1
          - v1.26.0
          - v1.25-latest
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - quay.io/sail-dev/sail-operator@sha256:d37258646d76a9f708a69fbaa081a0a95908a2b0c8837a14d7ecbfcf505f9bd4
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - quay.io/community-operator-pipeline-prod/sailoperator:1.26.3
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:2d567893e5b452d5c0b9c43a43efd4f8e28e639a344e28d90498f13858f13d46
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/proxyv2@sha256:fd734e6031566b4fb92be38f0f6bb02fdba6c199c45c2db5dc988bbc4fdee026
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:c4e33921e030bb22d3b8fabc693851d52c07a7af79a9534addc2f50005fa462a
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        - gcr.io/istio-release/install-cni@sha256:8e782f5077f63c0a15fd4b9a4fa75f8543033c28ccf2457e25afbf1aefb6a72a
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        version: 1.26.3
      entries:
      - name: sailoperator.v1.26.3
        version: 1.26.3
      - name: sailoperator.v1.26.2
        version: 1.26.2
      - name: sailoperator.v1.26.1
        version: 1.26.1
      - name: sailoperator.v1.26.0
        version: 1.26.0
      name: stable-1.26
    - currentCSV: sailoperator.v1.27.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "Istio",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-system",
                  "updateStrategy": {
                    "inactiveRevisionDeletionGracePeriodSeconds": 30,
                    "type": "InPlace"
                  },
                  "version": "v1.27.1"
                }
              },
              {
                "apiVersion": "sailoperator.io/v1",
                "kind": "IstioCNI",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "namespace": "istio-cni",
                  "version": "v1.27.1"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery, Networking, Security
          containerImage: quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
          createdAt: "2025-09-05T10:12:19Z"
          description: The Sail Operator manages the lifecycle of your Istio control
            plane. It provides custom resources for you to deploy and manage your
            control plane components.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/internal-objects: '["wasmplugins.extensions.istio.io","destinationrules.networking.istio.io","envoyfilters.networking.istio.io","gateways.networking.istio.io","proxyconfigs.networking.istio.io","serviceentries.networking.istio.io","sidecars.networking.istio.io","virtualservices.networking.istio.io","workloadentries.networking.istio.io","workloadgroups.networking.istio.io","authorizationpolicies.security.istio.io","peerauthentications.security.istio.io","requestauthentications.security.istio.io","telemetries.telemetry.istio.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/istio-ecosystem/sail-operator
          support: Community based
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: WasmPlugin
            name: wasmplugins.extensions.istio.io
            version: v1alpha1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1alpha3
          - kind: DestinationRule
            name: destinationrules.networking.istio.io
            version: v1beta1
          - kind: EnvoyFilter
            name: envoyfilters.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1alpha3
          - kind: Gateway
            name: gateways.networking.istio.io
            version: v1beta1
          - kind: ProxyConfig
            name: proxyconfigs.networking.istio.io
            version: v1beta1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1alpha3
          - kind: ServiceEntry
            name: serviceentries.networking.istio.io
            version: v1beta1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1alpha3
          - kind: Sidecar
            name: sidecars.networking.istio.io
            version: v1beta1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1alpha3
          - kind: VirtualService
            name: virtualservices.networking.istio.io
            version: v1beta1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1alpha3
          - kind: WorkloadEntry
            name: workloadentries.networking.istio.io
            version: v1beta1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1alpha3
          - kind: WorkloadGroup
            name: workloadgroups.networking.istio.io
            version: v1beta1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1
          - kind: AuthorizationPolicy
            name: authorizationpolicies.security.istio.io
            version: v1beta1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1
          - kind: PeerAuthentication
            name: peerauthentications.security.istio.io
            version: v1beta1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1
          - kind: RequestAuthentication
            name: requestauthentications.security.istio.io
            version: v1beta1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1
          - kind: Telemetry
            name: telemetries.telemetry.istio.io
            version: v1alpha1
          - description: IstioCNI represents a deployment of the Istio CNI component.
            displayName: Istio CNI
            kind: IstioCNI
            name: istiocnis.sailoperator.io
            version: v1
          - description: |-
              IstioRevision represents a single revision of an Istio Service Mesh deployment.
              Users shouldn't create IstioRevision objects directly. Instead, they should
              create an Istio object and allow the operator to create the underlying
              IstioRevision object(s).
            displayName: Istio Revision
            kind: IstioRevision
            name: istiorevisions.sailoperator.io
            version: v1
          - description: IstioRevisionTag references an Istio or IstioRevision object
              and serves as an alias for sidecar injection. It can be used to manage
              stable revision tags without having to use istioctl or helm directly.
              See https://istio.io/latest/docs/setup/upgrade/canary/#stable-revision-labels
              for more information on the concept.
            displayName: Istio Revision Tag
            kind: IstioRevisionTag
            name: istiorevisiontags.sailoperator.io
            version: v1
          - description: |-
              Istio represents an Istio Service Mesh deployment consisting of one or more
              control plane instances (represented by one or more IstioRevision objects).
              To deploy an Istio Service Mesh, a user creates an Istio object with the
              desired Istio version and configuration. The operator then creates
              an IstioRevision object, which in turn creates the underlying Deployment
              objects for istiod and other control plane components, similar to how a
              Deployment object in Kubernetes creates ReplicaSets that create the Pods.
            displayName: Istio
            kind: Istio
            name: istios.sailoperator.io
            version: v1
          - description: ZTunnel represents a deployment of the Istio ztunnel component.
            displayName: ZTunnel
            kind: ZTunnel
            name: ztunnels.sailoperator.io
            version: v1alpha1
        description: |-
          The Sail Operator manages the lifecycle of your Istio control plane. It provides custom resources for you to deploy and manage your control plane components.

          This version of the operator supports the following Istio versions:

          - v1.27-latest
          - v1.27.1
          - v1.27.0
          - v1.26-latest
          - v1.26.4
          - v1.26.3
          - v1.26.2
          - v1.26.0
          - v1.25-latest
          - v1.25.5
          - v1.25.4
          - v1.25.3
          - v1.25.2
          - v1.25.1
          - v1.24-latest
          - v1.24.6
          - v1.24.5
          - v1.24.4
          - v1.24.3
          - v1.24.2
          - v1.24.1
          - v1.24.0

          [See this page](https://github.com/istio-ecosystem/sail-operator/blob/main/bundle/README.md) for instructions on how to use it.
        displayName: Sail Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - servicemesh
        - envoy
        links:
        - name: Istio Project
          url: https://istio.io
        maintainers:
        - email: istio-feedback@redhat.com
          name: OpenShift Service Mesh Team
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/istio-release/install-cni@sha256:9a0ee38b9b1c6240bc71e58cc1f5a9c6963860a98c3b3655955fc47be9b7cf42
        - gcr.io/istio-release/ztunnel@sha256:2402a1c71dce11c409b6da171d8ae27dff62c539d58aac41817499e1fcd7beb3
        - gcr.io/istio-release/ztunnel@sha256:c1fd261b86750a453e26b693737cfaf9575e80be80ac3136580954f84e1033c3
        - gcr.io/istio-release/pilot@sha256:a6e8f88e35f886a45ae645bce076305d1f66a703f80059b6449026159ef767a8
        - gcr.io/istio-release/ztunnel@sha256:56afd26e9384ff41a0244a8c637a0e3aa8654015b8d32b494df8b5023f213ec7
        - gcr.io/istio-release/pilot@sha256:93d6b52a61e7e9891fe114e4f7d503b36e8d1566a22e82e5635084a2001eb14a
        - gcr.io/istio-release/install-cni@sha256:42794fa2bc6862440587a78bacf7ebd3d8cfa63fc176a97ac134569ee50ab96d
        - gcr.io/istio-release/proxyv2@sha256:bc19e6e0978569a7cc886f3dececc3cb3613cf6e457ddf09e3cb849a2b315231
        - gcr.io/istio-release/ztunnel@sha256:73b3ad311c9c9a2618e9c2f10259a3bf5414bb4fc6fd407baa6280bdd8d227f0
        - gcr.io/istio-release/proxyv2@sha256:2cd83a25202af6c5aef48b432f7fed50b6de460c795ce7c84f80bf630f51ca4d
        - gcr.io/istio-release/ztunnel@sha256:eaae7b8ed53a3f18b0291f08db0e3330d7c958d39919fce9dbc43f3027bf2089
        - gcr.io/istio-release/ztunnel@sha256:fabda07d39993a740b69a54bcb6218eccaaf8fa6b4bcdb733664baf729ec9b21
        - gcr.io/istio-release/install-cni@sha256:3c5fff9805d1cd6ae684597d885e4c85a28e4555edd7b7b5b6bcc125b459f12e
        - gcr.io/istio-release/pilot@sha256:4ad6b0c3839a11622e66323a1296c7fc19dcbbb2c371b2e3edd58bd4f763f19f
        - gcr.io/istio-release/install-cni@sha256:d241cd11295b7c7af6c1c7e22096db886b07ab1945b4b7b2fe3d7f95ef41b277
        - gcr.io/istio-release/ztunnel@sha256:45e0288894c9beed590bd4caed1dc04a96be7231de3cdd5e2c45f1c08f194503
        - gcr.io/istio-release/install-cni@sha256:ee1f34354d7043e919216ef6addd2b155349a2e26bf23a35f7405b90724577fd
        - gcr.io/istio-release/install-cni@sha256:3107b1947e755abe698fe2d5c3ed7ba5f228f0b4ff5ff30f484fcc984514b97e
        - gcr.io/istio-release/ztunnel@sha256:38caa7fed18c6d3283892de20ce508aed536622645e4b500395c00f6d1bb26de
        - gcr.io/istio-release/install-cni@sha256:d85b955875b2fbfc2ca8166cf2aa0e4ac3d78dd00f4ec4140a2a22dd17e7deb9
        - gcr.io/istio-release/ztunnel@sha256:ca14b3f869aa998da30c6f8f8d14b2e9cd61a40b78b05196b951cac9613d76d8
        - gcr.io/istio-release/install-cni@sha256:4b10d810228e553bf34924222b7d62fdfafbce5ed67dc54973d168bda384a4cd
        - gcr.io/istio-release/proxyv2@sha256:5b83f93307e7cb0cc0bd11882629c54b475b568fce1304c9afb0df459bab1e8a
        - gcr.io/istio-release/proxyv2@sha256:a9268faf5bb9d621e0506e3db2e9d7178b199dbfc609e19c32c1de5a3aee2fc4
        - gcr.io/istio-release/proxyv2@sha256:ac598d6c98415bfb323b4930ba1423b9c3d17a90c205de1ad2e5f28179275c75
        - gcr.io/istio-release/install-cni@sha256:d28af093e5b7f33e8750364e9077c8b2c464145d1904cae8be48227ec0110b04
        - gcr.io/istio-release/pilot@sha256:1d4fbb7859e515b85b6c2c4787224e4e7cba73aad4573d98e2a83cc028cad392
        - gcr.io/istio-release/pilot@sha256:29cccf38a413fe81a88d02e0bf924ca84a9cf1d6a3c9ece024c0636b1e3cb029
        - gcr.io/istio-release/ztunnel@sha256:7e33000f86ea39490764e5b6244282405a342fab8ac2df692acfeb876ab8b03f
        - gcr.io/istio-release/proxyv2@sha256:921c5ce2c5122facd9a25a7f974aaa4d3679cee38cb3f809e10618384fc3ce7f
        - gcr.io/istio-release/pilot@sha256:4dd632cc5a1a820f56525ad7731d6723451b72366fd782cc2a5f79039a4cfe92
        - gcr.io/istio-release/proxyv2@sha256:a5f2c1804ac4397448149b63a67f6cfd1979d304b99a726667a4859d201c6305
        - gcr.io/istio-release/pilot@sha256:e082a6f27001101942d1e2be71764b17ad350ed0d289038020426e7ee603eb56
        - gcr.io/istio-release/install-cni@sha256:f6e234ab501fabdd9643f1469f90bf0ff3696caa0f6456652e28c1a608644700
        - quay.io/sail-dev/sail-operator@sha256:490213deda4334877867d3dc42e62265f7d4c08842e2b4081cf57ac2b79c721f
        - gcr.io/istio-release/ztunnel@sha256:96daeeb4e71a304edbf2ee2d37bc4bd4b88b9da26d7e6cc0c34c94c1cd6e6cca
        - gcr.io/istio-release/ztunnel@sha256:45807d4c512e9a0e4855eb1a2535532319b698302ebbf4d4eaccf462997bb374
        - gcr.io/istio-release/install-cni@sha256:313c8b38e89137cb455ee96891c2dc5b69f2907b13af324f0cac8b140897f107
        - gcr.io/istio-release/proxyv2@sha256:ee6565e57319e01b5e45b929335eb9dc3d4b30d531b4652467e6939ae81b41f7
        - gcr.io/istio-release/pilot@sha256:32d0f6bd6c1259de78dbc79dfd88cf5f4dc8941613560ff2ca7dbabcf04dfd68
        - gcr.io/istio-release/pilot@sha256:0d06b179025862f13150dda22608e0700a74a0127d0a1b66eebcadf18d3d4bd7
        - gcr.io/istio-release/pilot@sha256:502c5a57143e80f9d8731bf08a25f31f09dbe231284233b9e715b4e62f22639b
        - gcr.io/istio-release/ztunnel@sha256:144093bb2d220b7ee402626dc5ea9ed7ddce2675ebe5ab4bf83646b4762484dd
        - gcr.io/istio-release/proxyv2@sha256:2edf332e1c7ddd46466e337d009eda091bcd43f827b6f460175b772addc8055a
        - gcr.io/istio-release/pilot@sha256:6a155682facaae9179431bf46d2d4c630450fdfa9d58a2210be36b63e445b82b
        - gcr.io/istio-release/pilot@sha256:2961bf88463e475f1e53d82a118405841ecbb9d8a6e0b35c73c77bee53bc0c3e
        - gcr.io/istio-release/install-cni@sha256:7b12bcfabd92c5756e4b12bf82e373f4abcdd3ccf81ac0b93fe0aa6f8ccce63c
        - gcr.io/istio-release/proxyv2@sha256:534f7589b085450b5b94e6c955f5106892df88d81676f53ac5e06a0cf5ec45da
        - gcr.io/istio-release/install-cni@sha256:f8043a943106f8638dc67ed9b8cce5ced0a25d0b8c9217281f7c7ff1dbd41c15
        - gcr.io/istio-release/pilot@sha256:21b6cdaf12e7eb8bb0ad1d493c32a76e7d34f574a41ece86ded0c79e7191e875
        - gcr.io/istio-release/proxyv2@sha256:d902f7f80be93a0521af5e5796d1e989f2107afd0725eca8881eb4c9848bd7bf
        - gcr.io/istio-release/ztunnel@sha256:e3bc4444deb33641b53883a54fff0b11ed4f95c86335dc9af02ab89a9dfe2e46
        - gcr.io/istio-release/install-cni@sha256:1d963aca195f1e06459925431c592cf556112d77b824c61e06555fbc31ee415f
        - gcr.io/istio-release/install-cni@sha256:a80a96d9e4e859d1568fe1db1b12525db8c26d153e4bb783ba4efb3c5b16f877
        - gcr.io/istio-release/ztunnel@sha256:17cc6794403b3ab64629da79fbacccb22f3c7fff46bb53627f2120f5331d7a13
        - gcr.io/istio-release/proxyv2@sha256:445156b5f4a780242d079a47b7d88199cbbb5959c92358469b721af490eca1ae
        - gcr.io/istio-release/proxyv2@sha256:640733abb924663df83761e2ec95a2bddca797daf1ca37672ec0b045969cdeb1
        - gcr.io/istio-release/proxyv2@sha256:9c14df68458fde932fefb0122c4df9b7f241598aeb430e91b42af3c632c4ae91
        - gcr.io/istio-release/pilot@sha256:c8ac4894f2e667eb439d1c960a47ac7fbc80aad3f0adb98f5fd9330d9c515696
        - gcr.io/istio-release/pilot@sha256:7ccda8c22307ee796cb7ba3f4ce224ef533c3c683c063dcc92fd5a244d8a5c9d
        - gcr.io/istio-release/pilot@sha256:f0bcb10a51dbb2b9e5552e7f82aa7fcc2ff4dd6d0d7d2224986036b1645b855b
        - gcr.io/istio-release/ztunnel@sha256:74ee23f374ea28b3455ad2eca2bea62da95bc2a23b865b7c8c301139046650d5
        - gcr.io/istio-release/install-cni@sha256:0f27cb9e3423a3ed8a444dd2137ae05f31a03df2bdd549b21ef8f0f3949e3cef
        - gcr.io/istio-release/pilot@sha256:69e78ea08b624b75ed614e0a846d10fac2a10847d465fb803910e1d23afeaef8
        - gcr.io/istio-release/proxyv2@sha256:75e272cbe8dd31fda4c193aedc251ee2e26593f0b7b4f305225f9ad804cc6f0d
        - gcr.io/istio-release/proxyv2@sha256:d62cc7585450bce8573338c3d147961f4e1df045ffcf5956abea29841cb5aaa4
        - gcr.io/istio-release/pilot@sha256:0bb4d26e88b61cd2f81d39378111a38a151ec5bf0cb65be49c20d1ef41956299
        - gcr.io/istio-release/ztunnel@sha256:ae90c4e06a4560898dbf414226e3ea879b8d0b10387af27b0c2907e8488fc5d1
        - quay.io/community-operator-pipeline-prod/sailoperator:1.27.1
        - gcr.io/istio-release/install-cni@sha256:697243b54b541cf07e562a57ca29fe82b8b268f090e8ebb7d4f33cb4660c3f39
        - gcr.io/istio-release/proxyv2@sha256:91382859f73b1a66ac113631f89ef73dabf9c931d6baf0bd7eb618c0fa958b34
        - gcr.io/istio-release/proxyv2@sha256:49ed9dd2c06383c0a847877a707a3563d0968d83779ad8d13a0c022a48c5c407
        - gcr.io/istio-release/ztunnel@sha256:0e680a1b33d458e5dfd6f91808c8bd61d00a1103084545ea6a1956c935e7554e
        - gcr.io/istio-release/install-cni@sha256:313165b70241f65c79b5984797b8e3bc43a264acfeaeda4f33fe3a7f89a2fdb8
        version: 1.27.1
      entries:
      - name: sailoperator.v1.27.1
        version: 1.27.1
      - name: sailoperator.v1.27.0
        version: 1.27.0
      name: stable-1.27
    defaultChannel: stable
    packageName: sailoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-metering: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: aws-efs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aws-efs-csi-driver-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
          createdAt: "2021-07-14T00:00:00Z"
          description: Install and configure AWS EFS CSI driver.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-cluster-csi-drivers
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/aws-efs-csi-driver-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          AWS EFS CSI Driver Operator provides AWS Elastic File Service (EFS) CSI Driver that enables you to create and mount AWS EFS PersistentVolumes.

          Please follow [AWS EFS CSI Driver Operator documentation](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html) to:

          1. [Install the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-olm-operator-install_persistent-storage-csi-aws-efs). This section also covers details about configuring an AWS role in OCP cluster that uses Secure Token Services (STS).

          2. [Install the AWS EFS CSI driver using the operator](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#persistent-storage-csi-efs-driver-install_persistent-storage-csi-aws-efs)

          3. Finally, [create a StorageClass](https://docs.openshift.com/container-platform/4.18/storage/container_storage_interface/persistent-storage-csi-aws-efs.html#storage-create-storage-class_persistent-storage-csi-aws-efs) that enables dynamic provisioning of PersistentVolumes.
        displayName: AWS EFS CSI Driver Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - efs
        links:
        - name: Documentation
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        - name: Source Repository
          url: https://github.com/openshift/aws-efs-csi-driver-operator
        maintainers:
        - email: aos-storage-staff@redhat.com
          name: Red Hat
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-tools-rhel9@sha256:13013220d7376d23df4780fe65c3bea325bdeb7b7a89d5704aa52e948ed7a2d5
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-operator-bundle@sha256:3fb13020c6a34b329825d268dd70c3c9310075780768312dc52cac1e7134a030
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-container-rhel9@sha256:aa2267c1c871699db37763e938125f290ace4aff2aca6f668b2ce01f7df92b8d
        - registry.redhat.io/openshift4/ose-aws-efs-csi-driver-rhel9-operator@sha256:da50f51a7726dc46d6041b71e3e263f4f88e517aac6bd0790b3871f4900d01d6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        version: 4.18.0-202509100344
      entries:
      - name: aws-efs-csi-driver-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: aws-efs-csi-driver-operator.v4.18.0-202508260754
        version: 4.18.0-202508260754
      - name: aws-efs-csi-driver-operator.v4.18.0-202507221934
        version: 4.18.0-202507221934
      - name: aws-efs-csi-driver-operator.v4.18.0-202507091633
        version: 4.18.0-202507091633
      - name: aws-efs-csi-driver-operator.v4.18.0-202506250134
        version: 4.18.0-202506250134
      - name: aws-efs-csi-driver-operator.v4.18.0-202506110106
        version: 4.18.0-202506110106
      - name: aws-efs-csi-driver-operator.v4.18.0-202506020833
        version: 4.18.0-202506020833
      - name: aws-efs-csi-driver-operator.v4.18.0-202505280604
        version: 4.18.0-202505280604
      - name: aws-efs-csi-driver-operator.v4.18.0-202505150334
        version: 4.18.0-202505150334
      - name: aws-efs-csi-driver-operator.v4.18.0-202505062100
        version: 4.18.0-202505062100
      - name: aws-efs-csi-driver-operator.v4.18.0-202504300105
        version: 4.18.0-202504300105
      - name: aws-efs-csi-driver-operator.v4.18.0-202504160308
        version: 4.18.0-202504160308
      - name: aws-efs-csi-driver-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: aws-efs-csi-driver-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: aws-efs-csi-driver-operator.v4.18.0-202503260133
        version: 4.18.0-202503260133
      - name: aws-efs-csi-driver-operator.v4.18.0-202503190402
        version: 4.18.0-202503190402
      - name: aws-efs-csi-driver-operator.v4.18.0-202503102003
        version: 4.18.0-202503102003
      - name: aws-efs-csi-driver-operator.v4.18.0-202503050303
        version: 4.18.0-202503050303
      - name: aws-efs-csi-driver-operator.v4.18.0-202502260933
        version: 4.18.0-202502260933
      - name: aws-efs-csi-driver-operator.v4.18.0-202502110432
        version: 4.18.0-202502110432
      name: stable
    defaultChannel: stable
    packageName: aws-efs-csi-driver-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: authorino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: authorino-operator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta3",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
          createdAt: "2025-03-25T22:39:23Z"
          description: Enables authentication and authorization for Gateways and applications
            in a Gateway API network.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta3
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **Generally Available**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta3
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/authorino-operator-bundle@sha256:7f3d41672fa8c60725304f0d45e468a4f01defcab7352501737ffddfd32edded
        - registry.redhat.io/rhcl-1/authorino-rhel9-operator@sha256:2d81613856df9725ad20928a7a9895d14482c42effce0a3a0de88f40ec35d411
        - registry.redhat.io/rhcl-1/authorino-rhel9@sha256:413d3796ad0101bf14d3341d31d9367e67fea1e73ebb3fb79124ef4c5dde0c09
        version: 1.2.2
      entries:
      - name: authorino-operator.v1.2.2
        version: 1.2.2
      - name: authorino-operator.v1.2.1
        version: 1.2.1
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: stable
    - currentCSV: authorino-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "authorino.kuadrant.io/v1beta1",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "hosts": [
                    "my-api.io"
                  ],
                  "identity": [
                    {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "in": "authorization_header",
                        "keySelector": "APIKEY"
                      },
                      "name": "api-key-users"
                    }
                  ]
                }
              },
              {
                "apiVersion": "authorino.kuadrant.io/v1beta2",
                "kind": "AuthConfig",
                "metadata": {
                  "name": "my-api-protection"
                },
                "spec": {
                  "authentication": {
                    "api-key-users": {
                      "apiKey": {
                        "selector": {
                          "matchLabels": {
                            "group": "friends"
                          }
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "hosts": [
                    "my-api.io"
                  ]
                }
              },
              {
                "apiVersion": "operator.authorino.kuadrant.io/v1beta1",
                "kind": "Authorino",
                "metadata": {
                  "name": "authorino-sample"
                },
                "spec": {
                  "listener": {
                    "tls": {
                      "enabled": false
                    }
                  },
                  "oidcServer": {
                    "tls": {
                      "enabled": false
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
          createdAt: "2024-11-08T11:06:33Z"
          description: Authorino
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/authorino-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta1
          - description: API to describe the desired protection for a service
            displayName: AuthConfig
            kind: AuthConfig
            name: authconfigs.authorino.kuadrant.io
            version: v1beta2
          - description: API to create instances of authorino
            displayName: Authorino
            kind: Authorino
            name: authorinos.operator.authorino.kuadrant.io
            version: v1beta1
        description: |-
          [Authorino](https://docs.kuadrant.io/authorino/) is Red Hat's Kubernetes-native lightweight external authorization service for tailor-made Zero Trust API security.

          Install this Red Hat official distribution of [Authorino Operator](https://docs.kuadrant.io/authorino-operator/) to manage instances of Authorino in this cluster.

          The current state of this distribution of the operator is: **_Technical Preview_**.

          The Community version of Authorino Operator, based on upstream public images, is available in [OperatorHub.io](https://operatorhub.io/operator/authorino-operator).

          ## Getting started

          After installing the Operator, request an instance of the external authorization service by creating an `Authorino` custom resource.

          **Minimal example (namespaced)**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            listener:
              tls:
                enabled: false
            oidcServer:
              tls:
                enabled: false
          ```

          **Extended example**

          ```yaml
          apiVersion: operator.authorino.kuadrant.io/v1beta1
          kind: Authorino
          metadata:
            name: authorino
          spec:
            clusterWide: true
            authConfigLabelSelectors: environment=production
            secretLabelSelectors: authorino.kuadrant.io/component=authorino,environment=production

            replicas: 2

            evaluatorCacheSize: 2 # mb

            logLevel: info
            logMode: production

            listener:
              ports:
                grpc: 50001
                http: 5001
              tls:
                certSecretRef:
                  name: authorino-server-cert # secret must contain `tls.crt` and `tls.key` entries
              timeout: 2

            oidcServer:
              port: 8083
              tls:
                certSecretRef:
                  name: authorino-oidc-server-cert # secret must contain `tls.crt` and `tls.key` entries

            metrics:
              port: 8080
              deep: true

            healthz:
              port: 8081

            tracing:
              endpoint: rpc://otel-collector.observability.svc.cluster.local:4317
              insecure: true

            volumes:
              items:
                - name: keycloak-tls-cert
                  mountPath: /etc/ssl/certs
                  configMaps:
                    - keycloak-tls-cert
                  items: # details to mount the k8s configmap in the authorino pods
                    - key: keycloak.crt
                      path: keycloak.crt
              defaultMode: 420
          ```

          ### Cluster-wide vs Namespaced

          Namespaced instances only watch auth resources (`AuthConfig` and `Secrets`) created in the same namespace as the Authorino service. Use this mode for dedicated instances that do not require elevated privileges.

          Cluster-wide instances watch resources across the entire cluster (all namespaces.) Deploying and running Authorino in this mode requires elevated privileges.

          ### Multi-tenancy

          Use the `authConfigLabelSelectors` field of the `Authorino` custom resource to narrow the scope of the Authorino instance.

          Only `AuthConfig` custom resources whose labels match the value of this field will be handled by the Authorino instance.

          ## Protect a host

          To protect a host, create an `AuthConfig` custom resource for the host. E.g.:

          ```yaml
          apiVersion: authorino.kuadrant.io/v1beta2
          kind: AuthConfig
          metadata:
            name: my-api-protection
          spec:
            hosts:
            - my-api.io

            authentication:
              "keycloak":
                jwt:
                  issuerUrl: https://keycloak.keycloak.svc.cluster.local:8080/realms/my-realm

            authorization:
              "k8s-rbac":
                kubernetesSubjectAccessReview:
                  user:
                    selector: auth.identity.user.username
                  resourceAttributes:
                    resource:
                      value: my-api
                    verb:
                      selector: request.method
                cache:
                  key:
                    selector: auth.identity.user.username
                  ttl: 30
              "after-2am-only":
                rego: |
                  allow {
                    [hour, _, _] := time.clock(time.now_ns())
                    hour >= 2
                  }
          ```

          Make sure all requests to the host are fisrt checked with the Authorino instance, by configuring an Envoy proxy for external authz:

          ```yaml
          clusters:
          - name: my-api
            …
          - name: authorino
            connect_timeout: 0.25s
            type: STRICT_DNS
            lb_policy: ROUND_ROBIN
            http2_protocol_options: {}
            load_assignment:
              cluster_name: authorino
              endpoints:
              - lb_endpoints:
                - endpoint:
                    address:
                      socket_address:
                        address: authorino-authorino-authorization
                        port_value: 50051
          listeners:
          - filter_chains:
            - filters:
              name: envoy.http_connection_manager
              typed_config:
                "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
                use_remote_address: true
                route_config:
                  name: my-api-config
                  virtual_hosts:
                  - name: my-api-vs
                    domains:
                    - my-api.io
                    routes:
                    - match:
                        prefix: /
                      route:
                        cluster: my-api
                http_filters:
                - name: envoy.filters.http.ext_authz
                  typed_config:
                    "@type": type.googleapis.com/envoy.extensions.filters.http.ext_authz.v3.ExtAuthz
                    transport_api_version: V3
                    failure_mode_allow: false
                    include_peer_certificate: true
                    grpc_service:
                      envoy_grpc:
                        cluster_name: authorino
                      timeout: 1s
          ```

          ...or, if using Istio, by creating an [`AuthorizationPolicy`](https://istio.io/latest/docs/reference/config/security/authorization-policy/#AuthorizationPolicy) custom resource. Use `action: CUSTOM` in the resource and the Authorino authorization service configured in the mesh extension provider settings.

          ## Features

          **Authentication**

          * JWT validation (with OpenID Connect Discovery)
          * OAuth 2.0 Token Introspection (opaque tokens)
          * Kubernetes TokenReview (ServiceAccount tokens)
          * API key authentication
          * X.509 client certificate authentication
          * Anonymous access
          * Proxy-handled (authentication performed by the proxy)

          **Authorization**

          * Built-in simple pattern matching (e.g. JWT claims, request attributes checking)
          * OPA policies (inline Rego and fetch from external registry)
          * Kubernetes SubjectAccessReview (resource and non-resource attributes)
          * Authzed SpiceDB

          **External metadata**

          * HTTP request
          * OpenID Connect User Info
          * UMA-protected resource attributes

          **Custom responses**

          * Header injection (Festival Wristbands tokens, JSON, plain text)
          * Envoy Dynamic Metadata
          * Custom HTTP response (status code, headers, messages, body, etc)

          **Callbacks**

          * HTTP webhooks

          **Caching**

          * OpenID Connect and User-Managed Access configs
          * JSON Web Keys (JWKs) and JSON Web Key Sets (JWKS)
          * Access tokens
          * External metadata
          * Precompiled Rego policies
          * Policy evaluation

          Check out the full [Feature Specification](https://docs.kuadrant.io/authorino/docs/features/) and how-to guides in the [Kuadrant Docs](https://docs.kuadrant.io) website.
        displayName: Red Hat - Authorino (Technical Preview)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Authorino
        - Authorino Operator
        - Kuadrant
        - Authorization
        - Authentication
        links:
        - name: GitHub
          url: https://github.com/kuadrant/authorino-operator
        - name: Documentation
          url: https://docs.kuadrant.io/authorino
        maintainers:
        - email: dcesare@redhat.com
          name: Didier Di Cesare
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-tech-preview/authorino-operator-bundle@sha256:75cb78f7a40d3daedbc11cbbb5505e029328b4120f0de11d4f6423de943e92d8
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9-operator@sha256:eb574781cf39917f8f5d9e5c806f520534601f819ee1ebcbb751832da8938008
        - registry.redhat.io/3scale-tech-preview/authorino-rhel9@sha256:a473dae20e71e3e813ac30ba978f2ab3c5e19d7d39b501ae9103dca892107c87
        - registry.redhat.io/openshift4/ose-cli-rhel9@sha256:3a398158d10018bfac5df291d6998b513cf9a30e5477bcaefddcf66d229627c7
        version: 1.1.3
      entries:
      - name: authorino-operator.v1.1.3
        version: 1.1.3
      - name: authorino-operator.v1.1.2
        version: 1.1.2
      - name: authorino-operator.v1.1.1
        version: 1.1.1
      - name: authorino-operator.v1.1.0
        version: 1.1.0
      - name: authorino-operator.v1.0.2
        version: 1.0.2
      name: tech-preview-v1
    defaultChannel: stable
    packageName: authorino-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: http://marketplace.redhat.com
    name: marketplace-games-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: marketplace-games-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "game.marketplace.redhat.com/v1",
                "kind": "Arcade",
                "metadata": {
                  "name": "arcade-sample"
                },
                "spec": {
                  "size": 1
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/marketplace-games-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Arcade is the Schema for the arcades API
            displayName: Arcade Instance
            kind: Arcade
            name: arcades.game.marketplace.redhat.com
            version: v1
        description: RHM Arcade Operator - a simple operator used to showcase how
          to build an operator around an existing web application that can can run
          on Openshift, Vanilla k8s, or on-boarded to Red Hat Marketplace.
        displayName: Arcade Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - operator
        - rhm
        - red hat
        - marketplace
        - games
        - arcade
        links:
        - name: Marketplace Games Operator
          url: https://github.com/redhat-marketplace/marketplace-games-operator
        maturity: alpha
        provider:
          name: Red Hat Marketplace
          url: http://marketplace.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.connect.redhat.com/rhm-labs/arcade-operator-bundle@sha256:279f5d817f9dc90b202924fc2b2ab59e5303e9dddc1d78d71017d65cdd66d565
        - registry.marketplace.redhat.com/rhm/rhm-labs/arcade-operator@sha256:d6959be6693728d2dc4cd38b69251551d23c5002cb52db2684df31a0a8f7b0ab
        version: 1.0.0
      entries:
      - name: marketplace-games-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: marketplace-games-operator-rhmp
    provider:
      name: Red Hat Marketplace
      url: http://marketplace.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: xspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hpe.com/v1",
                "kind": "XSPC",
                "metadata": {
                  "name": "xspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
          createdAt: "2025-08-26T06:53:59Z"
          description: An operator for managing HPE XP Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XSPC is the Schema for the xspcs API
            displayName: XSPC
            kind: XSPC
            name: xspcs.csi.hpe.com
            version: v1
        description: |-
          ## About
          HPE XP Storage Plug-in for Containers (XSPC) is a plugin that integrates HPE XP storage into Kubernetes based clusters.
          XSPC provides persistent volume provisioning capabilities from HPE's block storage arrays.
        displayName: HPE XP Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - XP7
        - XP8
        - HPE
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Knowledge Base
          url: https://www.hpe.com/us/en/storage/enterprise-xp-storage.html
        maintainers:
        - email: hpexpci@hpe.com
          name: HPE
        maturity: stable
        provider:
          name: HPE
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.connect.redhat.com/hpestorage/xspc-csi-operator-bundle@sha256:658dd12fe1b92bf38dd24601043cb84e27a27b0d987a36a0ec7d99b0514574e2
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver-operator@sha256:196c6236ddca5db67d04772897643dac5ab8ed909aa1ed683cd4490293f82c85
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.connect.redhat.com/hpestorage/xspc-csi-driver@sha256:4c289e5023fd3e59b2cf40335eba9eb53caf562472f16dd97b370773a0ba608b
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: xspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: xspc-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: https://intel.com
    name: intel-device-plugins-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: intel-device-plugins-operator.v0.32.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "DsaDevicePlugin",
                "metadata": {
                  "name": "dsadeviceplugin-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c",
                  "initImage": "registry.connect.redhat.com/intel/intel-idxd-config-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/dsa": "true"
                  },
                  "sharedDevNum": 10
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "GpuDevicePlugin",
                "metadata": {
                  "name": "gpudeviceplugin-sample"
                },
                "spec": {
                  "enableMonitoring": true,
                  "image": "registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0",
                  "preferredAllocationPolicy": "none",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/gpu": "true"
                  },
                  "sharedDevNum": 1
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "QatDevicePlugin",
                "metadata": {
                  "name": "qatdeviceplugin-sample"
                },
                "spec": {
                  "dpdkDriver": "vfio-pci",
                  "image": "registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02",
                  "initImage": "registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90",
                  "kernelVfDrivers": [
                    "4xxxvf"
                  ],
                  "logLevel": 4,
                  "maxNumDevices": 128,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/qat": "true"
                  }
                }
              },
              {
                "apiVersion": "deviceplugin.intel.com/v1",
                "kind": "SgxDevicePlugin",
                "metadata": {
                  "name": "sgxdeviceplugin-sample"
                },
                "spec": {
                  "enclaveLimit": 110,
                  "image": "registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4",
                  "logLevel": 4,
                  "nodeSelector": {
                    "intel.feature.node.kubernetes.io/sgx": "true"
                  },
                  "provisionLimit": 110
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Security, AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
          description: This operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Intel device
            plugins for Kubernetes.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/intel/intel-technology-enabling-for-openshift
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DsaDevicePlugin is the Schema for the dsadeviceplugins API.
              It represents the DSA device plugin responsible for advertising Intel
              DSA hardware resources to the kubelet.
            displayName: Intel DSA Device Plugin
            kind: DsaDevicePlugin
            name: dsadeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              GpuDevicePlugin is the Schema for the gpudeviceplugins API. It represents
              the GPU device plugin responsible for advertising Intel GPU hardware resources to
              the kubelet.
            displayName: Intel GPU Device Plugin
            kind: GpuDevicePlugin
            name: gpudeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              QatDevicePlugin is the Schema for the qatdeviceplugins API. It represents the QAT device
              plugin responsible for advertising Intel QuickAssist Technology hardware resources
              to the kubelet.
            displayName: Intel QuickAssist Technology Device Plugin
            kind: QatDevicePlugin
            name: qatdeviceplugins.deviceplugin.intel.com
            version: v1
          - description: |-
              SgxDevicePlugin is the Schema for the sgxdeviceplugins API. It represents
              the SGX device plugin responsible for advertising SGX device nodes to
              the kubelet.
            displayName: Intel Software Guard Extensions Device Plugin
            kind: SgxDevicePlugin
            name: sgxdeviceplugins.deviceplugin.intel.com
            version: v1
        description: |
          [Intel Device Plugins for Kubernetes](https://github.com/intel/intel-technology-enabling-for-openshift) is a collection of
          [device plugins](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) advertising Intel specific hardware resources
          to the kubelet. Currently the operator has basic support for Intel SGX, Intel GPU, Intel QAT and Intel DSA. Intel IAA device plugin will be supported in a future release.
        displayName: Intel Device Plugins Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - fpga
        - qat
        - sgx
        - gpu
        - dsa
        - dlb
        links:
        - name: Intel Device Plugins Operator
          url: https://github.com/intel/intel-technology-enabling-for-openshift
        maintainers:
        - email: martin.xu@intel.com
          name: Martin Xu
        - email: chaitanya.kulkarni@intel.com
          name: Chaitanya kulkarni
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Intel® Corporation
          url: https://intel.com
        relatedImages:
        - registry.connect.redhat.com/intel/intel-dsa-plugin@sha256:2742a13279cc3f301daa09b6389517024530f658d4e1dd13db495cc94d9ba57c
        - registry.connect.redhat.com/intel/intel-gpu-plugin@sha256:e2c2ce658e78c35c425f16a4f8e85c5f32ce31848d9b53a644a05e7f8b7f71b0
        - registry.connect.redhat.com/intel/intel-idxd-initcontainer@sha256:b74dc43fa81ce14ea97f20ff6b2f726039f6309fdd868d5f45d751d0a8662cc1
        - registry.connect.redhat.com/intel/intel-qat-initcontainer@sha256:34f0b993ca654ea0b386217cba1a44d5ef3da841b3befc780508f5323e95fa90
        - registry.connect.redhat.com/intel/intel-qat-plugin@sha256:8d79dba051b83ec770a4b0fdc3da6ac92264cb19cac8d455b707ed92a6a95d02
        - registry.connect.redhat.com/intel/intel-sgx-plugin@sha256:f2c77521c6dae6b4db1896a5784ba8b06a5ebb2a01684184fc90143cfcca7bf4
        - registry.connect.redhat.com/intel/intel-device-plugins-operator-bundle@sha256:4a422ca897caeabb9fc3be3dfef3746819868de1674ea61c657cad257fbddf12
        - registry.connect.redhat.com/intel/intel-deviceplugin-operator@sha256:5ca1fd11f0ca4e7fb941749c67461c9c66fe9417c9d0f0a9247b6b172b0bcdd5
        version: 0.32.1
      entries:
      - name: intel-device-plugins-operator.v0.32.1
        version: 0.32.1
      - name: intel-device-plugins-operator.v0.29.1
        version: 0.29.1
      - name: intel-device-plugins-operator.v0.29.0
        version: 0.29.0
      - name: intel-device-plugins-operator.v0.28.0
        version: 0.28.0
      name: alpha
    defaultChannel: alpha
    packageName: intel-device-plugins-operator
    provider:
      name: Intel® Corporation
      url: https://intel.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
    name: flink-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flink-kubernetes-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkDeployment",
                "metadata":{
                    "name":"basic-example"
                },
                "spec":{
                    "image":"flink:1.16",
                    "flinkVersion":"v1_16",
                    "flinkConfiguration":{
                        "taskmanager.numberOfTaskSlots":"2"
                    },
                    "serviceAccount":"flink",
                    "jobManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "taskManager":{
                        "resource":{
                            "memory":"2048m",
                            "cpu":1
                        }
                    },
                    "job":{
                        "jarURI":"local:///opt/flink/examples/streaming/StateMachineExample.jar",
                        "parallelism":2,
                        "upgradeMode":"stateless"
                    }
                }
              },
              {
                "apiVersion":"flink.apache.org/v1beta1",
                "kind":"FlinkSessionJob",
                "metadata":{
                  "name":"basic-session-job-example2"
                },
                "spec":{
                  "deploymentName":"basic-session-cluster",
                  "job":{
                      "jarURI":"https://repo1.maven.org/maven2/org/apache/flink/flink-examples-streaming_2.12/1.16.1/flink-examples-streaming_2.12-1.16.1.jar",
                      "parallelism":2,
                      "upgradeMode":"stateless",
                      "entryClass":"org.apache.flink.streaming.examples.statemachine.StateMachineExample"
                  }
                }
              },
              {
                "apiVersion": "flink.apache.org/v1beta1",
                "kind": "FlinkStateSnapshot",
                "metadata": {
                  "name": "example-savepoint"
                },
                "spec": {
                  "savepoint": {
                    "alreadyExists": true,
                    "disposeOnDelete": false,
                    "path": "/flink-data/savepoints/savepoint-98682a-21931bf10c39"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "basic-example": {
                "description": "Example flink deployment"
              },
              "basic-session-job-example2": {
                "description": "Example flink session job"
              }
            }
          capabilities: Deep Insights
          categories: Big Data, Streaming & Messaging
          containerImage: ghcr.io/apache/flink-kubernetes-operator:1.12.0
          createdAt: "2025-06-23 18:42:54"
          description: Flink Kubernetes Operator manages the complete deployment lifecycle
            of Apache Flink applications.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/apache/flink-kubernetes-operator.git
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The FlinkDeployment CR defines Flink Application and Session
              cluster deployments
            displayName: Flink Deployment
            kind: FlinkDeployment
            name: flinkdeployments.flink.apache.org
            version: v1beta1
          - description: The FlinkSessionJob CR defines the session job on the Session
              cluster and each Session cluster can run multiple FlinkSessionJob.
            displayName: Flink Session Job
            kind: FlinkSessionJob
            name: flinksessionjobs.flink.apache.org
            version: v1beta1
          - description: The FlinkStateSnapshots CR is used to describe savepoint
              or checkpoint for a Flink job.
            displayName: Flink State Snapshot
            kind: FlinkStateSnapshot
            name: flinkstatesnapshots.flink.apache.org
            version: v1beta1
        description: |-
          Flink Kubernetes Operator acts as a control plane to manage the complete deployment lifecycle of Apache Flink applications.
          ## Installation
          This operator can be installed from [OperatorHub.io](https://operatorhub.io/operator/flink-kubernetes-operator).

          By default, the Flink operator monitors and run Flink applications in the same namespace as the operator.
          To run Flink jobs in another namespace, users are responsible for setting up the RBAC in that namespace.

          For example, users need to do the following for each additional namespace that runs the Flink applications:

          1. Switch to the namespace by running:

              ```sh
              kubectl config set-context --current --namespace=CHANGEIT
              ```

          2. Create the service account, role, and role binding in the namespace using the commands below:

              ```sh
              kubectl apply -f - <<EOF
              apiVersion: v1
              kind: ServiceAccount
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: Role
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink
              rules:
              - apiGroups:
                - ""
                resources:
                - pods
                - configmaps
                verbs:
                - '*'
              - apiGroups:
                - apps
                resources:
                - deployments
                - deployments/finalizers
                verbs:
                - '*'
              ---
              apiVersion: rbac.authorization.k8s.io/v1
              kind: RoleBinding
              metadata:
                labels:
                  app.kubernetes.io/name: flink-kubernetes-operator
                  app.kubernetes.io/version: 1.12.0
                name: flink-role-binding
              roleRef:
                apiGroup: rbac.authorization.k8s.io
                kind: Role
                name: flink
              subjects:
              - kind: ServiceAccount
                name: flink
              EOF
              ```

          3. Proceed deploying Flink applications using the the Custom Resource Definition below.

              ```sh
              kubectl create -f https://raw.githubusercontent.com/apache/flink-kubernetes-operator/release-1.12.0/examples/basic.yaml
              ```

          See [Flink custom resources](https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/custom-resource/overview/#overview) for more detail.
        displayName: Flink Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flink
        - streaming
        links:
        - name: Website
          url: https://flink.apache.org/
        - name: Documentation
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/
        - name: Mailing list
          url: https://lists.apache.org/list.html?dev@flink.apache.org
        - name: Slack
          url: https://apache-flink.slack.com/join/shared_invite/zt-1llkzbgyt-K2nNGGg88rfsDGLkT09Qzg#/shared-invite/email
        - name: GitHub
          url: https://github.com/apache/flink-kubernetes-operator
        maintainers:
        - email: mbalassi@apache.org
          name: Marton Balassi
        - email: htchang@us.ibm.com
          name: Ted Chang
        - email: jbusche@us.ibm.com
          name: James Busche
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Community
          url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
        relatedImages:
        - ghcr.io/apache/flink-kubernetes-operator:1.12.0
        - quay.io/community-operator-pipeline-prod/flink-kubernetes-operator@sha256:7c919847b88a5969d0c0830e4154512bbca28015cf6fb7fa70cd20824a726685
        version: 1.12.0
      entries:
      - name: flink-kubernetes-operator.v1.12.0
        version: 1.12.0
      - name: flink-kubernetes-operator.v1.11.0
        version: 1.11.0
      - name: flink-kubernetes-operator.v1.6.0
        version: 1.6.0
      - name: flink-kubernetes-operator.v1.5.0
        version: 1.5.0
      - name: flink-kubernetes-operator.v1.4.0
        version: 1.4.0
      - name: flink-kubernetes-operator.v1.3.1
        version: 1.3.1
      - name: flink-kubernetes-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: flink-kubernetes-operator
    provider:
      name: Community
      url: https://nightlies.apache.org/flink/flink-kubernetes-operator-docs-main/docs/try-flink-kubernetes-operator/quick-start/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-01-30T10:20:30Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install the Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle@sha256:8cba51274389030d632c8b82255399217d485146f4797907134e3c4ee55cb4d2
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: multicluster-operators-subscription
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multicluster-operators-subscription.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
          createdAt: "2021-07-19T18:00:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-application:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.3.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.3.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:3bc76999ece42fcb3acd913b32be3fa0e0da056832922819b81119dcfb5f1582
        version: 0.3.0
      entries:
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.3
    - currentCSV: multicluster-operators-subscription.v0.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
          createdAt: "2021-09-29T10:23:00Z"
          description: An operator to subscribe resources from object Bucket, helm
            and github types of channels
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-deployable:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-placementrule:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.4.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:2aeba259864bbddaa8cdd46ab6d8c3495450a0081469a27b9acde53008d91bed
        - quay.io/open-cluster-management/multicloud-integrations:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.4.0
        - quay.io/open-cluster-management/multicluster-operators-channel:2.4.0
        version: 0.3.6
      entries:
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.4
    - currentCSV: multicluster-operators-subscription.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
          createdAt: "2022-03-23T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/open-cluster-management/multicluster-operators-channel:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-subscription:2.5.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:f58f60283f6cf9ea745f4a25378ef6690d98af8fda0eb63f663066631e71450b
        - quay.io/open-cluster-management/multicloud-integrations:2.5.0
        - quay.io/open-cluster-management/multicluster-operators-application:2.5.0
        version: 0.4.0
      entries:
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.5
    - currentCSV: multicluster-operators-subscription.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Channel",
                "metadata": {
                  "name": "dev-helmrepo"
                },
                "spec": {
                  "type": "HelmRepo",
                  "pathname": "http://kubernetes-charts.storage.googleapis.com/"
                },
                "status": {
                  "state": "N/A"
                }
              },
              {
                "apiVersion": "apps.open-cluster-management.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "sub-nginx"
                },
                "spec": {
                  "channel": "ns-ch/predev-ch",
                  "name": "nginx-ingress",
                  "packageFilter": {
                    "version": "=1.26.1"
                  },
                  "packageOverrides": [
                    {
                      "packageAlias": "nginx-ingress",
                      "packageName": "nginx-ingress",
                      "packageOverrides": [
                        {
                          "path": "spec",
                          "value": {
                            "defaultBackend": {
                              "replicaCount": 3
                            }
                          }
                        }
                      ]
                    }
                  ],
                  "placement": {
                    "local": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/multicluster-operators-subscription:2.6.0
          createdAt: "2022-06-06T18:00:00Z"
          description: An operator for deploying resources from object Bucket, helm
            repo or github repo to multiple managed clusters
          olm.skipRange: <0.1.4
          repository: https://github.com/open-cluster-management-io/multicloud-operators-subscription
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Subscribe resources from a channel according to its package
              filters
            displayName: App Subscription
            kind: Subscription
            name: subscriptions.apps.open-cluster-management.io
            version: v1
          - description: subscription status per cluster per app
            displayName: App Subscription Report
            kind: SubscriptionReport
            name: subscriptionreports.apps.open-cluster-management.io
            version: v1alpha1
          - description: subscription status per package
            displayName: App Subscription Status
            kind: SubscriptionStatus
            name: subscriptionstatuses.apps.open-cluster-management.io
            version: v1alpha1
          - description: Represent a helm chart repository
            displayName: Channel
            kind: Channel
            name: channels.apps.open-cluster-management.io
            version: v1
          - description: Represent a helm chart selected by the subscription, for
              internal use only.
            displayName: Helm Release
            kind: HelmRelease
            name: helmreleases.apps.open-cluster-management.io
            version: v1
          - description: Contain a k8s resource template for deployment, for internal
              use only.
            displayName: Deployable
            kind: Deployable
            name: deployables.apps.open-cluster-management.io
            version: v1
          - description: Placement Rules to determine which mangaged clusters will
              be deployed by multicluster applications.
            displayName: Placement Rule
            kind: PlacementRule
            name: placementrules.apps.open-cluster-management.io
            version: v1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1alpha1
          - description: GitOps Cluster to determine which mangaged cluster secret
              will be imported to which argocd server.
            displayName: GitOps Cluster
            kind: GitOpsCluster
            name: gitopsclusters.apps.open-cluster-management.io
            version: v1beta1
          - description: Application selectors to determine which subscriptions and
              deployables are bundled into the application
            displayName: Application
            kind: Application
            name: applications.app.k8s.io
            version: v1beta1
        description: "A multicluster subscription operator subscribes resources from
          multiple types of Channels and get them deployed to multiple managed clusters\n##
          Prerequisites\n- Operator Lifecycle Manager (OLM) needs to be installed.\n\nOLM
          installation can be checked by running the command:\n  ```\n  $ kubectl
          get pods --all-namespaces | grep olm\n  olm        catalog-operator-7f54797f5f-z9l9n
          \        1/1     Running            0          46s\n  olm        olm-operator-65874bdb76-vs6gf
          \            1/1     Running            0          46s\n  olm        operatorhubio-catalog-8wp2d
          \              1/1     Running            0          36s\n  olm        packageserver-7fb4588767-8mxv4
          \           1/1     Running            0          34s\n  olm        packageserver-7fb4588767-d6h8j
          \           1/1     Running            0          25s\n  ```\n\n## How to
          Install\n- Install `Multicluster Subscription Operator` by following instructions
          in top right button `Install`.\n\nThree new pods are created in `openshift-operators`
          namespace\n\n```\n$ kubectl get pods -n openshift-operators\nNAME                                                            READY
          \  STATUS    RESTARTS   AGE\nmulticluster-operators-application-5ccbbc878d-98fdp
          \              4/4     Running   0          49s\nmulticluster-operators-channel-dffcb4f8c-rx5g8
          \                   1/1     Running   0          49s\nmulticluster-operators-hub-subscription-65f7d689d6-6vpr5
          \         1/1     Running   0          49s\nmulticluster-operators-standalone-subscription-86df85d4bf-m4bd9
          \  1/1     Running   0          49s\n```\n\nThe operator is now providing
          new Custom Resources Definitions: `subscriptions.apps.open-cluster-management.io`,
          `channels.apps.open-cluster-management.io`\n\n## Using the Multicluster
          Subscription Operator\nHere is an example to demonstrate how to create one
          Multicluster Subscription CR to subscribe the helm chart nginx-ingress v1.26.1\n\n-
          \ Create a channel `dev-helmrepo` in namespace `dev` to point to a helm
          repo `kubernetes-charts.storage.googleapis.com`\n\n```\napiVersion: v1\nkind:
          Namespace\nmetadata:\n  name: dev\n---\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Channel\nmetadata:\n  name: dev-helmrepo\n  namespace: dev\nspec:\n    type:
          HelmRepo\n    pathname: https://charts.helm.sh/stable/\n    configRef: \n
          \     name: skip-cert-verify\n      apiVersion: v1\n      kind: ConfigMap\n---\napiVersion:
          v1\nkind: ConfigMap\nmetadata:\n  name: skip-cert-verify\n  namespace: dev\ndata:\n
          \ insecureSkipVerify: \"true\"\n```\n\n-  Create a subscription `simple`
          in the `default` namespace to subscribe the nginx-ingress v1.26.1 package.
          The subscription is disabled by default.\n\n```\napiVersion: apps.open-cluster-management.io/v1\nkind:
          Subscription\nmetadata:\n  name: simple\nspec:\n  channel: dev/dev-helmrepo\n
          \ name: nginx-ingress\n  placement:\n    local: false\n  packageFilter:\n
          \   version: \"1.36.x\"\n  packageOverrides:\n  - packageName: nginx-ingress\n
          \   packageAlias: nginx-ingress\n    packageOverrides:\n    - path: spec\n
          \     value:\n        defaultBackend:\n          replicaCount: 3\n  overrides:\n
          \ - clusterName: \"/\"\n    clusterOverrides:\n    - path: \"metadata.namespace\"\n
          \     value: default\n```\n\n- Deploy the nginx-ingress v1.36 package by
          setting the subscription `spec.placement.local` to `true`\n\n```\nkubectl
          patch subscriptions.apps.open-cluster-management.io simple --type='json'
          -p='[{\"op\": \"replace\", \"path\": \"/spec/placement/local\", \"value\":
          true}]'\n```\n\n- Make sure the subscribed nginx-ingress package being deployed
          to the default namespace\n\n```\n$ kubectl get pods -n default |grep nginx\nnginx-ingress-controller-775b4967cb-8nx9m
          \       1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-bfkn8
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-fpcnp
          \  1/1     Running   0          118s\nnginx-ingress-default-backend-659bd647bd-v4nbf
          \  1/1     Running   0          118s\n```\n"
        displayName: Multicluster Subscription Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Red Hat
        - multicluster
        - subscription
        - channel
        links:
        - name: Multicluster Subscription Operator Project
          url: https://github.com/open-cluster-management/multicloud-operators-subscription
        maintainers:
        - email: ocm-dev@open-cluster-management.io
          name: Open Cluster Management Project team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/multicloud-integrations:2.6.0
        - quay.io/stolostron/multicluster-operators-application:2.6.0
        - quay.io/stolostron/multicluster-operators-channel:2.6.0
        - quay.io/stolostron/multicluster-operators-subscription:2.6.0
        - quay.io/openshift-community-operators/multicluster-operators-subscription@sha256:48b37c9f9417ef442d896e79851b176505045b487bddf77bb4e6d91fd5ad3dd1
        version: 0.4.1
      entries:
      - name: multicluster-operators-subscription.v0.4.1
        version: 0.4.1
      - name: multicluster-operators-subscription.v0.4.0
        version: 0.4.0
      - name: multicluster-operators-subscription.v0.3.9
        version: 0.3.9
      - name: multicluster-operators-subscription.v0.3.8
        version: 0.3.8
      - name: multicluster-operators-subscription.v0.3.7
        version: 0.3.7
      - name: multicluster-operators-subscription.v0.3.6
        version: 0.3.6
      - name: multicluster-operators-subscription.v0.3.5
        version: 0.3.5
      - name: multicluster-operators-subscription.v0.3.4
        version: 0.3.4
      - name: multicluster-operators-subscription.v0.3.3
        version: 0.3.3
      - name: multicluster-operators-subscription.v0.3.2
        version: 0.3.2
      - name: multicluster-operators-subscription.v0.3.1
        version: 0.3.1
      - name: multicluster-operators-subscription.v0.3.0
        version: 0.3.0
      - name: multicluster-operators-subscription.v0.2.9
        version: 0.2.9
      - name: multicluster-operators-subscription.v0.2.8
        version: 0.2.8
      - name: multicluster-operators-subscription.v0.2.7
        version: 0.2.7
      name: release-2.6
    defaultChannel: release-2.5
    packageName: multicluster-operators-subscription
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: minhokim
      provider-url: ""
    name: node-discovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-discovery-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.wunicorns.com/v1alpha1",
                "kind": "NodeDiscovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "node-discovery-operator",
                    "app.kubernetes.io/instance": "nodediscovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodediscovery",
                    "app.kubernetes.io/part-of": "node-discovery-operator"
                  },
                  "name": "nodediscovery-sample"
                },
                "spec": {
                  "containers": [
                    {
                      "args": [
                        "tail",
                        "-f",
                        "/dev/null"
                      ],
                      "image": "alicek106/stress",
                      "imagePullPolicy": "IfNotPresent",
                      "name": "stress",
                      "resources": {
                        "limits": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "1000Mi"
                        }
                      },
                      "terminationGracePeriodSeconds": 10
                    }
                  ],
                  "cpus": "0",
                  "enabled": "true",
                  "mems": "0",
                  "nodeName": "worker3"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-06-18T01:32:51Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeDiscovery is the Schema for the nodediscoveries API
            displayName: Node Discovery
            kind: NodeDiscovery
            name: nodediscoveries.cache.wunicorns.com
            version: v1alpha1
        description: find cgroup infor and set value.
        displayName: node-discovery-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cgroup
        - cpulist
        - mems
        - cpus
        - numa
        links:
        - name: Node Discovery Operator
          url: https://node-discovery-operator.domain
        maintainers:
        - email: wunicorns1436@gmail.com
          name: minhokim
        maturity: alpha
        provider:
          name: minhokim
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/caspell/cgroup-discovery@sha256:b437dcde0107124e43a4e2fb8026ec191f44b5e761bedc2c295f47ff489cf343
        - quay.io/community-operator-pipeline-prod/node-discovery-operator@sha256:899baace103c58e7337ae5cd75e3764e6de26966aab7340037c612c26ca7cf87
        version: 0.0.1
      entries:
      - name: node-discovery-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: node-discovery-operator
    provider:
      name: minhokim
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://access.redhat.com/products/red-hat-fuse
    name: fuse-online
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-online.v7.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
          createdAt: 2022-01-31 15:55:46 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:44215c5b244b190c82d9d371a8ffd39b0c83576bd85c980cc72782a5cf9e6e1b
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:5048cd4fab2f98e6000bc60835c539cc40b70dbe265e927cfe185595347d8054
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:5dd1bc8f6b7e724736f4aeac86409bda3e0f64f62b736688e9572d14411232f1
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:df2d55dd8092e9e94d034166890117cf5d1cfd5cf52ef0f90cfa2374cc9e2d06
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:96810c770f75057b8bba6b78a5ba414c8289315d8af1a387f681ae5e4fb7d6dd
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:3f0277e2c278740307c9b2d966f539ba12317e306198cb6c4c1c52a2ad9f4f40
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:ed4faedcb97bb6a45f568586ff19da19da31ecbc304f12e0ae7978381247744a
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:ef229a4c8851ec495f28d0eeab83b137cd02698e283f76ea1b32271b7222621b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:4f8849362e9ed6ff95d318a04e287b7ba33254a2cf7abdaf9b90cab7cbba6646
        - registry.redhat.io/openshift4/ose-prometheus@sha256:7f679d6c03535e300dac366f4bc6dd8d781d96dc47592836713f8bb76be36fa9
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:fcdd0292ebfc7e9edc36241f63b54a9d924a7605329c934ff2b10373af26dbbf
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:53777788253eaa1199e1854d14782619aeec0df9844d86b83742fe0005671cde
        version: 7.10.4
      entries:
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.10.x
    - currentCSV: fuse-online.v7.11.1-0.1687357787.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.11.1-0.1684718286.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:4de91e4527022e1b25df8d3835ddf778f773c20334314be9ba01a885c82287f5
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:658af3fc7819efb7cc39cd3a0504dcddc2ce7788c9b1e5178df869f7c9a742a8
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:ac514973ec428d815d7a248d09d27f2a84f762aedfea63b3b48de9e633267ed5
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:a2b1cd0256f30aff8fa37c332419b4483c0bfe34fb4f574ad8f18b0858969342
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:e3a13ea8b84b9e1abd158d394e11113e5f6c04d879547fa1d3e8638f1ac979d8
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:3d98d775bec85704c9a10c2f0d8e4f8116b7a41494f73c0b763e010262e5cd4d
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:8b9bd777bcbd20b1a1bbbc18274a5076a67bea1f928d772cb243c57fb1da7408
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:251c29e81f8e338f980b475b841647a977807908c388f4c86e18da7bff1b198a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:400a6ee03c3200c19b9b15c5d037a849633156a62e13922fb2e62838024ae69c
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/rhel8/postgresql-12@sha256:9248c4eaa8aeedacc1c06d7e3141ca1457147eef59e329273eb78e32fcd27e79
        version: 7.11.1+0.1687357787.p
      entries:
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.11.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.12.x
    - currentCSV: fuse-online.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta2",
                "kind": "Syndesis",
                "metadata": {
                  "labels": {
                    "app": "syndesis"
                  },
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
          createdAt: 2021-07-21 12:25:15 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          operators.operatorframework.io/builder: operator-sdk-v1.5.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta2
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to the product documentation at https://access.redhat.com/documentation/en-us/red_hat_fuse
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:c668fe1f5d25baddf627acd79df4c8d2ab8b05309d3d41784ddc0a10dd5f34ad
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:d6df09b15599e776b49e1d46aa01728fd02d5629cee65e77dc277c0166554e0a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:f65600506db985689e5b7cd8b838e646d4268a0c4e272e19d7db3577a8dd7ac4
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:c70c5e76dc3db559c722bea2a767eb7dcba1eda786e1ef3a45c4cf32e1aca9ee
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:c3e95e808eecba8ac57632452d3553509131ca783f62c68754a582cf5c55efab
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:75d5772c792a7c795281acec3d59e90021c109e9d7d1110868be0b882768aefb
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:d6f0c804c179fa559d3baed2e2d106c916a0439dc2215c4d34b878bc2e9bc281
        - registry.redhat.io/openshift4/ose-prometheus@sha256:13e9ea359cbcd3e0e94a6589b86b708dc84197f152701fd4092fffa234d11b58
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:31258400bb37f4b8837ef294820f8a728ae83f8031d4e01ff0392690a315a30f
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:eff9f8edf28ccbadcacab2022fff9e6bc93926af97521b49e3968eaf7b6fda65
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:60d5cf00411b03af3f1149a2bb3aeeb196c063d19625f876a094ae6baf46b54b
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:bafd85b0fd1256b8a60a2bb368a7e68f5dbad72718e5059a65660e01a425d8e2
        version: 7.9.3
      entries:
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: 7.9.x
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: candidate
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: latest
    - currentCSV: fuse-online.v7.12.1-0.1706187675.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "syndesis.io/v1beta3",
                "kind": "Syndesis",
                "metadata": {
                  "name": "app"
                },
                "spec": {
                  "addons": {
                    "jaeger": {
                      "enabled": true
                    },
                    "ops": {
                      "enabled": false
                    },
                    "publicApi": {
                      "enabled": false
                    },
                    "todo": {
                      "enabled": false
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
          createdAt: 2022-04-04 12:44:36 UTC
          description: Manages the installation of Fuse Online, a flexible and customizable
            open source platform that provides core integration capabilities as a
            service.
          olm.substitutesFor: fuse-online.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/syndesisio/syndesis/
          support: Fuse Online
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Syndesis is the Schema for the Syndeses API
            displayName: Syndesis
            kind: Syndesis
            name: syndesises.syndesis.io
            version: v1beta3
        description: |
          ### Fuse-Online operator
          Fuse Online is a flexible and customizable, open source platform that provides core integration capabilities as a service.

          This operator installs as well as configures the following Fuse Online components:
          - syndesis-server
          - syndesis-meta
          - syndesis-ui
          - syndesis-db
          - syndesis-prometheus
          - syndesis-proxy
          - syndesis-oauthproxy

          ### Before you begin
          You must configure authentication to Red Hat container registry before you can import and use the Red Hat Fuse OpenShift Image Streams. Follow instruction given below to configure the registration to container registry.

          1. Log in to the OpenShift Server as an administrator, as follow:
              ```
              oc login -u system:admin
              ```
          2. Log in to the OpenShift project where you will be installing the operator.
              ```
              oc project fuse-online
              ```
          3. Create a docker-registry secret using either Red Hat Customer Portal account or Red Hat Developer Program account credentials.
              ```
              oc create secret docker-registry syndesis-pull-secret \
                --docker-server=registry.redhat.io \
                --docker-username=CUSTOMER_PORTAL_USERNAME \
                --docker-password=CUSTOMER_PORTAL_PASSWORD \
                --docker-email=EMAIL_ADDRESS
              ```
              NOTE: You need to create a docker-registry secret in every new namespace where the image streams reside and which use registry.redhat.io.

              If you do not wish to use your Red Hat account username and password to create the secret, it is recommended to create an authentication token using a [registry service account](https://access.redhat.com/terms-based-registry/).

          ### How to install
          - When the operator is installed (you have created a subscription and the operator is running in the selected namespace) and before you create a new CR of Kind Syndesis, you have to link the secret created in the previous section to the operator service account.
          ```
          oc secrets link syndesis-operator syndesis-pull-secret --for=pull
          ```

          - Create a new CR of Kind Syndesis (click the Create New button). The CR spec contains all defaults (see below).

          ### CR Defaults
          The CR definition is pretty simple and an empy CR will trigger a base installation.

          Within the addons section, users are able to enable specific addons. The available addons at the moment are:
          - jaeger: enable jaeger
          - ops: enables monitoring, requires extra CRDs
          - todo: a simple todo application

          To enable addons, set "addon_name": {"enabled": true} in the CR.
          If you enable the todo addon, you should link the previously created syndesis-pull-secret to the builder serviceAccount.
          ```
          oc secrets link builder syndesis-pull-secret --for=pull,mount
          ```

          For a more detailed set of instructions and a complete reference of config options,
          please refer to [Fuse Online documentation](https://access.redhat.com/documentation/en-us/red_hat_fuse)
        displayName: Red Hat Integration - Fuse Online
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - camel
        - integration
        - syndesis
        - fuse
        - online
        links:
        - name: Red Hat Fuse Online Documentation
          url: https://access.redhat.com/documentation/en-us/red-hat-fuse
        - name: Upstream project Syndesis
          url: https://github.com/syndesisio/syndesis
        - name: Upstream Syndesis Operator
          url: https://github.com/syndesisio/syndesis/tree/master/install/operator
        maintainers:
        - email: janstey@redhat.com
          name: Jon Anstey
        - email: parichar@redhat.com
          name: Paul Richardson
        maturity: beta
        provider:
          name: Red Hat
          url: https://access.redhat.com/products/red-hat-fuse
        relatedImages:
        - registry.redhat.io/fuse7/fuse-online-builder-rhel8@sha256:660e87daec112a226322738ebcaa9cf60490bd866e16e48767d20ee3f2d7dd5d
        - registry.redhat.io/fuse7/fuse-online-server-rhel8@sha256:19c580d25c9c52b42f9dad4988788dec2f5701b446103c59679fabf19d98f71a
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator-bundle@sha256:063051dac9fae01ca92e74eb044d7a126b3d493c78846fde5b4793922ef1dc87
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e97e19a7acd23d2021bd0667d9f4f69b2cbbb48ce5a20934ff0883adcd4d6eb0
        - registry.redhat.io/rhel8/postgresql-12@sha256:d8c5112a34ad9a5b93223389fbb64e6429f299c1703876b6652c8810fcacd59d
        - registry.redhat.io/fuse7/fuse-online-ui-rhel8@sha256:d28471a9946d0574f9b135b41f19cb4b56629eae2c462a89f6deac67e7d00131
        - registry.redhat.io/fuse7/fuse-online-upgrade-rhel8@sha256:08c6df76f764abd9e9197ed7252f190cba0a54e0670f3a82a977e47e7e8b4c7b
        - registry.redhat.io/fuse7/fuse-postgres-exporter-rhel8@sha256:7d441bb58b9a469579b96ac1bbf0e3171e8d1e0177dfd4a39240ee462638b5dc
        - registry.redhat.io/fuse7/fuse-online-meta-rhel8@sha256:2946ba7ee37b80c1c3d04928f6cffcb9e2d3265226e7aba0bd12a4e99c771c38
        - registry.redhat.io/openshift4/ose-prometheus@sha256:8d0a70d15c8ca20bae2387b35f39125028c4d6b3c2d27173f46f0a0ce54179c7
        - registry.redhat.io/rhel8/php-72@sha256:48a5e695139a2ee45ee562960324b2e6875ad2689c7f102da7128936153adfd4
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/fuse7/fuse-online-rhel8-operator@sha256:fe47dfb31a4853d2a37fef749996e19717ea5e1fa406f3a1e0ecdc4b78605072
        version: 7.12.1+0.1706187675.p
      entries:
      - name: fuse-online.v7.12.1-0.1706187675.p
        version: 7.12.1+0.1706187675.p
      - name: fuse-online.v7.12.1
        version: 7.12.1
      - name: fuse-online.v7.12.0
        version: 7.12.0
      - name: fuse-online.v7.12.0-0.1697748412.p
        version: 7.12.0+0.1697748412.p
      - name: fuse-online.v7.12.0-0.1697419534.p
        version: 7.12.0+0.1697419534.p
      - name: fuse-online.v7.12.0-0.1696814252.p
        version: 7.12.0+0.1696814252.p
      - name: fuse-online.v7.12.0-0.1694087053.p
        version: 7.12.0+0.1694087053.p
      - name: fuse-online.v7.12.0-0.1693227897.p
        version: 7.12.0+0.1693227897.p
      - name: fuse-online.v7.12.0-0.1689618244.p
        version: 7.12.0+0.1689618244.p
      - name: fuse-online.v7.11.1-0.1679272339.p
        version: 7.11.1+0.1679272339.p
      - name: fuse-online.v7.11.1-0.1682419260.p
        version: 7.11.1+0.1682419260.p
      - name: fuse-online.v7.11.1
        version: 7.11.1
      - name: fuse-online.v7.11.1-0.1669927569.p
        version: 7.11.1+0.1669927569.p
      - name: fuse-online.v7.11.1-0.1673344689.p
        version: 7.11.1+0.1673344689.p
      - name: fuse-online.v7.11.1-0.1675972855.p
        version: 7.11.1+0.1675972855.p
      - name: fuse-online.v7.11.1-0.1677678164.p
        version: 7.11.1+0.1677678164.p
      - name: fuse-online.v7.11.1-0.1687357787.p
        version: 7.11.1+0.1687357787.p
      - name: fuse-online.v7.11.1-0.1681224700.p
        version: 7.11.1+0.1681224700.p
      - name: fuse-online.v7.11.1-0.1684718286.p
        version: 7.11.1+0.1684718286.p
      - name: fuse-online.v7.11.1-0.1682509048.p
        version: 7.11.1+0.1682509048.p
      - name: fuse-online.v7.11.0-0.1665994963.p
        version: 7.11.0+0.1665994963.p
      - name: fuse-online.v7.11.0-0.1665054365.p
        version: 7.11.0+0.1665054365.p
      - name: fuse-online.v7.11.0-0.1662457147.p
        version: 7.11.0+0.1662457147.p
      - name: fuse-online.v7.11.0-0.1659517286.p
        version: 7.11.0+0.1659517286.p
      - name: fuse-online.v7.11.0
        version: 7.11.0
      - name: fuse-online.v7.10.4
        version: 7.10.4
      - name: fuse-online.v7.10.3
        version: 7.10.3
      - name: fuse-online.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-online.v7.10.2-0.1648491127.p
        version: 7.10.2+0.1648491127.p
      - name: fuse-online.v7.10.1
        version: 7.10.1
      - name: fuse-online.v7.10.0
        version: 7.10.0
      - name: fuse-online.v7.9.3
        version: 7.9.3
      - name: fuse-online.v7.9.2
        version: 7.9.2
      - name: fuse-online.v7.9.1
        version: 7.9.1
      - name: fuse-online.v7.9.0
        version: 7.9.0
      name: stable
    defaultChannel: latest
    packageName: fuse-online
    provider:
      name: Red Hat
      url: https://access.redhat.com/products/red-hat-fuse
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://konveyor.io
    name: move2kube-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: move2kube-operator.v0.3.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {
                  "deployment": {
                    "api": {
                      "cpu": "300m",
                      "imageTag": "",
                      "initContainer": false,
                      "memory": "1Gi",
                      "privilegedPods": false,
                      "replicas": 1
                    },
                    "authServer": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1
                    },
                    "database": {
                      "cpu": "300m",
                      "enable": false,
                      "memory": "1Gi",
                      "replicas": 1,
                      "startEmpty": true
                    }
                  },
                  "ingress": {
                    "enable": true,
                    "host": "mydomain.com",
                    "preferRoute": false,
                    "tlsSecretName": ""
                  },
                  "persistentVolumeClaim": {
                    "createNew": true,
                    "enable": false,
                    "name": "",
                    "readWriteMany": true,
                    "storageClassName": "",
                    "storageSize": "1Gi"
                  },
                  "secret": {
                    "api": {
                      "configYAML": "",
                      "createNew": true,
                      "enable": false,
                      "name": ""
                    },
                    "authServer": {
                      "adminPassword": "password",
                      "adminUsername": "admin",
                      "createNew": true,
                      "databasePassword": "password",
                      "databaseUsername": "auth-server",
                      "name": "",
                      "realmJSON": "",
                      "standaloneHAXML": ""
                    }
                  },
                  "securityContextContraints": {
                    "enable": false
                  },
                  "serviceAccount": {
                    "createNew": true,
                    "enable": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.10
          createdAt: "2023-09-25T02:56:15Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/konveyor/move2kube-operator:v0.3.10
        - quay.io/openshift-community-operators/move2kube-operator@sha256:cdf288be5ea779e03900af9b57c4d922b8760b88da4fcbf88437f8031d679aef
        version: 0.3.10
      entries:
      - name: move2kube-operator.v0.3.10
        version: 0.3.10
      - name: move2kube-operator.v0.3.10-rc.0
        version: 0.3.10-rc.0
      - name: move2kube-operator.v0.3.9
        version: 0.3.9
      - name: move2kube-operator.v0.3.8
        version: 0.3.8
      name: alpha
    - currentCSV: move2kube-operator.v0.3.15-rc.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
          createdAt: "2024-10-18T16:10:56Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:236a53d6fad949285b6abafcdd1a4541a7748f952c0c96fa2e71500079579ba4
        - quay.io/konveyor/move2kube-operator:v0.3.15-rc.0
        version: 0.3.15-rc.0
      entries:
      - name: move2kube-operator.v0.3.15-rc.0
        version: 0.3.15-rc.0
      - name: move2kube-operator.v0.3.14-rc.0
        version: 0.3.14-rc.0
      - name: move2kube-operator.v0.3.13-rc.0
        version: 0.3.13-rc.0
      - name: move2kube-operator.v0.3.11-rc.0
        version: 0.3.11-rc.0
      name: prerelease
    - currentCSV: move2kube-operator.v0.3.15
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "move2kube.konveyor.io/v1alpha1",
                "kind": "Move2Kube",
                "metadata": {
                  "name": "move2kube-sample"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          containerImage: quay.io/konveyor/move2kube-operator:v0.3.15
          createdAt: "2025-03-07T06:09:48Z"
          description: Konveyor Move2Kube is an open source tool that helps migrate
            your app to run on Kubernetes/Openshift.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/move2kube-operator
          support: https://github.com/konveyor/move2kube-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Move2Kube
            name: move2kubes.move2kube.konveyor.io
            version: v1alpha1
        description: |
          Konveyor Move2Kube is a tool that uses source files such as Docker Compose files or Cloud Foundry manifest files, and even source code, to generate Kubernetes deployment files including object yaml, Helm charts, and operators.

          This operator can be used to start an instance of the Move2Kube UI.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Move2Kube CR.

          By default the operator installs a StatefulSet that brings up a Pod that serves the UI.
          See https://move2kube.konveyor.io/tutorials/ui for a tutorial on how to use the UI.

          Data persistence can be enabled by setting the `persistentVolumeClaim.enable` field to `true` in the CR `spec`. https://github.com/konveyor/move2kube-operator/blob/0e79476bcd873d3272164a47af31b00590dacfb1/helm-charts/move2kube/values.yaml#L26

          ### Documentation

          See https://move2kube.konveyor.io/ for more details.

          ### Getting help

          If you encounter any issues while using this operator, you can create an issue on our [Github repo](https://github.com/konveyor/move2kube-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing

          You can contribute by making PRs to our Github repo https://github.com/konveyor/move2kube-operator
        displayName: Konveyor Move2Kube
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubernetes
        - modernization
        - replatform
        - move2kube
        - konveyor
        - migration
        links:
        - name: Documentation
          url: https://move2kube.konveyor.io/
        - name: Move2Kube Operator
          url: https://github.com/konveyor/move2kube-operator
        maintainers:
        - email: harikrishmenon@gmail.com
          name: Harikrishnan Balagopal
        - email: seshapad@in.ibm.com
          name: Padmanabha V Seshadri
        maturity: alpha
        minKubeVersion: 1.13.0
        provider:
          name: Konveyor
          url: https://konveyor.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/move2kube-operator@sha256:38d03dacca046dda3ba08c39943ec2e7db33f1c7bb897e84fe1308a0982c63f1
        - quay.io/konveyor/move2kube-operator:v0.3.15
        version: 0.3.15
      entries:
      - name: move2kube-operator.v0.3.15
        version: 0.3.15
      - name: move2kube-operator.v0.3.14
        version: 0.3.14
      - name: move2kube-operator.v0.3.13
        version: 0.3.13
      - name: move2kube-operator.v0.3.11
        version: 0.3.11
      name: stable
    defaultChannel: stable
    packageName: move2kube-operator
    provider:
      name: Konveyor
      url: https://konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/limitador-operator
    name: limitador-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: limitador-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "limitador.kuadrant.io/v1alpha1",
                "kind": "Limitador",
                "metadata": {
                  "name": "limitador-sample"
                },
                "spec": {
                  "limits": [
                    {
                      "conditions": [
                        "get_toy == 'yes'"
                      ],
                      "max_value": 2,
                      "name": "toy_get_route",
                      "namespace": "toystore-app",
                      "seconds": 30,
                      "variables": []
                    }
                  ],
                  "listener": {
                    "grpc": {
                      "port": 8081
                    },
                    "http": {
                      "port": 8080
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
          createdAt: "2025-04-07T15:52:04Z"
          description: Enables rate limiting for Gateways and applications in a Gateway
            API network
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/limitador-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Limitador is the Schema for the limitadors API
            displayName: Limitador
            kind: Limitador
            name: limitadors.limitador.kuadrant.io
            version: v1alpha1
        description: Enables rate limiting for Gateways and applications in a Gateway
          API network. - Part of Red Hat Connectivity Link
        displayName: Red Hat - Limitador Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - rate-limit
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/limitador-operator
        - name: Documentation
          url: https://github.com/Kuadrant/limitador/tree/main/docs
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: asnaps@redhat.com
          name: Alex Snaps
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/limitador-operator
        relatedImages:
        - registry.redhat.io/rhcl-1/limitador-rhel9-operator@sha256:edb28f112a0d3536f2f52467f38df685d953f46f1999cbd61f4fd4fa4805cd34
        - registry.redhat.io/rhcl-1/limitador-rhel9@sha256:98f85a05da53fbf7f20845d1969441620407cc13a185fd52714308679f116425
        - registry.redhat.io/rhcl-1/limitador-operator-bundle@sha256:4c56cd8462531cf8ba39316a0c6f7493240bfa4a6ace77352fe7ac1597e2fd90
        version: 1.1.0
      entries:
      - name: limitador-operator.v1.1.0
        version: 1.1.0
      - name: limitador-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: limitador-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/limitador-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NetApp, Inc.
      provider-url: https://www.netapp.com/
    name: trident-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trident-operator.v25.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "trident.netapp.io/v1",
                "kind": "TridentOrchestrator",
                "metadata": {
                  "name": "trident"
                },
                "spec": {
                  "IPv6": false,
                  "debug": true,
                  "nodePrep": [],
                  "imagePullSecrets": [],
                  "imageRegistry": "",
                  "namespace": "trident",
                  "silenceAutosupport": false
                }
              }
            ]
          categories: Storage
          containerImage: docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
          createdAt: "2025-09-03"
          description: Trident Operator, to manage NetApp Trident installations
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.openshift.io/valid-subscription: No subscriptions are required
          repository: https://github.com/netapp/trident
          support: NetApp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Used to deploy NetApp Trident.
            displayName: Trident Orchestrator
            kind: TridentOrchestrator
            name: tridentorchestrators.trident.netapp.io
            version: v1
          - description: Automates AWS FSxN backend configuration
            displayName: Trident Configurator
            kind: TridentConfigurator
            name: tridentconfigurators.trident.netapp.io
            version: v1
        description: |
          NetApp Trident is an open source storage provisioner and orchestrator maintained by NetApp. It enables you to create storage volumes for containerized applications managed by Docker and Kubernetes. For full release information, including patch release changes, see https://docs.netapp.com/us-en/trident/trident-rn.html.
        displayName: NetApp Trident
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - NetApp
        - ONTAP
        - Trident
        - storage
        - performance
        - open source
        - Astra
        - Cloud Native
        - AWS
        - Azure
        - GCP
        - Cloud Volumes
        links:
        - name: GitHub Repository
          url: https://github.com/NetApp/trident
        - name: Trident documentation
          url: https://docs.netapp.com/us-en/trident/index.html
        - name: Support policy
          url: https://mysupport.netapp.com/site/info/version-support
        - name: Release Notes
          url: https://docs.netapp.com/us-en/trident/trident-rn.html
        maintainers:
        - email: ng-trident-openshift-operator@netapp.com
          name: NetApp
        provider:
          name: NetApp, Inc.
          url: https://www.netapp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-external-attacher-rhel9@sha256:145ab76f9606e9d51dcd1e4e159c67021ecf53c24d8d80ed7bbeb2cc7fb5c42b
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:bdee8df520467b30312e7b14d7b6ef3963557ccd38269913de3fa51bb1bceec9
        - docker.io/netapp/trident@sha256:7925717979a84f93b62a2592682cde3443d6350ec4e88f8fd1a7d3ea812e8900
        - registry.connect.redhat.com/netapp/trident-operator@sha256:53a4889e1036af9f858c9fdb1bd58f59b758a301bc773379572cf279f930f5c6
        - registry.redhat.io/openshift4/ose-csi-external-resizer-rhel9@sha256:6781d64bfb01e252595130eec814b9d2fe6426720a003cc61e82e9963767bfd4
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:0ef3bf2a4dcbb5aa829ce9cc5f12efaccaf99b246b4b15abbb66b0135bec4cab
        - docker.io/netapp/trident-autosupport@sha256:21b3afbebfac3cc28c4eb319c2d385187bd8ebb728d8cfd82dbc437cc9b83b33
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:b48bd5bc612c1bbe73d203c70e6caaae2b5360bd31eba20eb06cbd2102cc34fc
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:b88958c38883f5b1e3d307da90aac34a5ceafdb6bd3d989b062cb24ecafd1fbd
        - docker.io/netapp/trident-operator@sha256:c3448b6847550f45849a5cdbb9e0cedb35290d21131bc513d833d95eb42a43e0
        version: 25.6.2
      entries:
      - name: trident-operator.v25.6.2
        version: 25.6.2
      - name: trident-operator.v25.6.1
        version: 25.6.1
      - name: trident-operator.v25.6.0
        version: 25.6.0
      - name: trident-operator.v25.2.1
        version: 25.2.1
      - name: trident-operator.v25.2.0
        version: 25.2.0
      name: stable
    defaultChannel: stable
    packageName: trident-operator
    provider:
      name: NetApp, Inc.
      url: https://www.netapp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed OpenShift Black Belt
      provider-url: https://github.com/rh-mobb/ocm-operator
    name: ocm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ocm-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "GitLabIdentityProvider",
                "metadata": {
                  "name": "gitlab-sample"
                },
                "spec": {
                  "clientID": "gitlab",
                  "clientSecret": {
                    "name": "gitlab"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "gitlab-sample",
                  "mappingMethod": "claim",
                  "url": "https://gitlab.com"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "LDAPIdentityProvider",
                "metadata": {
                  "name": "ldap-sample"
                },
                "spec": {
                  "bindDN": "CN=test,OU=Users,DC=example,DC=com",
                  "bindPassword": {
                    "name": "test"
                  },
                  "clusterName": "my-cluster",
                  "displayName": "ldap-sample",
                  "mappingMethod": "claim",
                  "url": "ldap://test.example.com:389"
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "MachinePool",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "aws": {
                    "spotInstances": {
                      "enabled": false
                    }
                  },
                  "clusterName": "my-cluster",
                  "instanceType": "m5.xlarge",
                  "labels": {
                    "owner": "my-user"
                  },
                  "maximumNodesPerZone": 1,
                  "minimumNodesPerZone": 1,
                  "wait": false
                }
              },
              {
                "apiVersion": "ocm.mobb.redhat.com/v1alpha1",
                "kind": "ROSACluster",
                "metadata": {
                  "name": "rosa-sample"
                },
                "spec": {
                  "accountID": "111111111111",
                  "defaultMachinePool": {
                    "instanceType": "m5.xlarge",
                    "labels": {
                      "owner": "my-owner"
                    },
                    "minimumNodesPerZone": 2
                  },
                  "displayName": "my-cluster",
                  "iam": {
                    "userRole": "arn:aws:iam::111111111111:role/ManagedOpenShift-User-my_user-Role"
                  },
                  "tags": {
                    "owner": "my-owner"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ocm-operator:v0.1.0
          createdAt: "2023-06-02T17:32:25Z"
          operatorframework.io/suggested-namespace: ocm-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ocm-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLabIdentityProvider is the Schema for the gitlabidentityproviders
              API.
            displayName: Git Lab Identity Provider
            kind: GitLabIdentityProvider
            name: gitlabidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: LDAPIdentityProvider is the Schema for the ldapidentityproviders
              API.
            displayName: LDAPIdentity Provider
            kind: LDAPIdentityProvider
            name: ldapidentityproviders.ocm.mobb.redhat.com
            version: v1alpha1
          - description: MachinePool is the Schema for the machinepools API.
            displayName: Machine Pool
            kind: MachinePool
            name: machinepools.ocm.mobb.redhat.com
            version: v1alpha1
          - description: ROSACluster is the Schema for the clusters API.
            displayName: ROSACluster
            kind: ROSACluster
            name: rosaclusters.ocm.mobb.redhat.com
            version: v1alpha1
        description: "# Quick Start\n\n## Prerequisites\n\n- [oc](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html)\n-
          [aws](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html)\n-
          [ROSA Cluster (4.12+)](https://mobb.ninja/docs/rosa/sts/)\n\n> **Warning:**\n>
          This is only tested on ROSA, but may work on other Kubernetes and OpenShift
          variants.  It requires\n> an AWS STS-enabled cluster.  Additionally, OpenShift
          versions 4.12+ are recommended, as they \n> enable [CRD CEL validation](https://kubernetes.io/blog/2022/09/23/crd-validation-rules-beta/).
          \ Versions\n> prior to 4.12 provide a lower version of Kubernetes that does
          not enable this feature.  They may \n> work, but provide no input validation
          when submitting custom resources to the cluster.\n\nBefore installing this
          operator, there are a couple secrets that must be created.\n\n1. OCM Token:
          used to authenticate against the OCM API.  The controller will not start
          without\nthis token as it used for all calls into OCM.\n2. AWS Credentials:
          uses an assume role to a role created in the instructions below in order\nto
          access AWS resources.  This is needed for cluster creation.  See \n[Create
          AWS IAM Policies and Roles](#create-aws-iam-policies-and-roles) for more
          details.\n\n> **Note:**\n> Certain custom resources have their own specific
          prerequisites.  Please use `oc explain` \n> or read the docs [here](https://github.com/rh-mobb/ocm-operator/tree/main/docs)
          for more details.\n\n### Create OCM Token Secret\n\n1. Create a namespace
          where you wish to install the operator:\n\n```bash\noc new-project ocm-operator\n```\n\n2.
          Create a secret containing the OCM_TOKEN.  This token can be obtained form
          \nhttps://console.redhat.com/openshift/token and is used by the operator
          to authenticate \nagainst the OCM API.  This token must exist in the same
          namespace that the operator \nis running and be named `ocm-token`.  It also
          expects the key to be called `OCM_TOKEN` \nas the operator is expecting
          this value as an environment variable.\n\n```bash\noc create secret generic
          ocm-token \\\n  --namespace=ocm-operator \\\n  --from-literal=OCM_TOKEN=${MY_OCM_TOKEN}\n```\n\n###
          Create AWS IAM Policies and Roles\n\nThe operator will need to elevate privileges
          in order to perform things like \ncreating the operator-roles for the clusters.
          \ Because of this, the operator \nmust have a specific role created to allow
          it these permissions.\n\n**NOTE:** please understand what you are doing
          if you deviate from the known good \npolicies.  If errors or more stringent
          security lockdowns are found, please submit a PR \nso that we can get this
          fixed.\n\n1. Set `ACCOUNT_ID` environment variable to define your AWS account
          ID:\n\n```bash\nexport ACCOUNT_ID=111111111111\n```\n\n2. Create the permissions
          boundary.  Because the operator needs to create policies and \nroles, the
          boundary ensures that the operator is not allowed to create additional\npermissions
          outside of the defined boundary.  The sample permission set is located \nat
          `test/aws/boundary.json` in this repository:\n\n```bash\nboundary=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/boundary.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperatorBoundary\" \\\n  --policy-document
          \"$boundary\"\n```\n\n3. Create the policy.  This policy sets what the operator
          is allowed to do.  For any \n`iam` permission, the boundary created above
          is used.  The sample permission set is \nlocated at `test/aws/iam_policy.json`
          in this repository:\n\n```bash\npolicy=$(curl -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/iam_policy.json\")\naws
          iam create-policy \\\n  --policy-name \"OCMOperator\" \\\n  --policy-document
          \"$policy\"\n```\n\n4. Create the role using a trust policy and attach the
          previously created role.  The trust \npolicy is located at `test/aws/trust_policy.json`
          in this repository.  Please note that \nthe trust policy requires an OIDC
          configuration.  The OIDC configuration refers to \n**where the operator
          is running, NOT what it is provisioning**:\n\n```bash\ntrust_policy=$(curl
          -s \"https://raw.githubusercontent.com/rh-mobb/ocm-operator/main/test/aws/trust_policy.json\")\naws
          iam create-role \\\n    --role-name OCMOperator \\\n    --assume-role-policy-document
          \"$trust_policy\"\naws iam attach-role-policy \\\n    --role-name OCMOperator
          \\\n    --policy-arn arn:aws:iam::$ACCOUNT_ID:policy/OCMOperator\n```\n\n5.
          Finally, create the secret containing the assume role credentials.  The
          previous steps allow \nthe operator to assume the role you created in the
          previous step with the permissions created \nvia the previous policies.\n\n```bash\ncat
          <<EOF > /tmp/credentials\n[default]\nrole_arn = arn:aws:iam::$ACCOUNT_ID:role/OCMOperator\nweb_identity_token_file
          = /var/run/secrets/openshift/serviceaccount/token\nEOF\n\noc create secret
          generic aws-credentials \\\n  --namespace=ocm-operator \\\n  --from-file=credentials=/tmp/credentials\n```\n\n##
          Install the Operator\n\n```bash\ncat <<EOF | oc apply -f -\napiVersion:
          operators.coreos.com/v1alpha1\nkind: Subscription\nmetadata:\n  name: ocm-operator\n
          \ namespace: ocm-operator\nspec:\n  channel: v0.1\n  installPlanApproval:
          Automatic\n  name: ocm-operator\n  source: community-operators\n  sourceNamespace:
          ocm-operator\n  startingCSV: ocm-operator.v0.1.0\nEOF\n```\n\n## Provision
          Resources\n\nOnce the operator is available and running, you can provision
          any of the \nresources that it manages.  Also note that documentation is
          always available \nonce the operator is installed as well by using the `oc
          explain` command.  For \nexample, `oc explain rosacluster.spec.clusterName`
          will give you detailed documentation \nabout what the field does.\n\nSee
          the following documentation for details:\n\n* [ROSA Clusters](https://github.com/rh-mobb/ocm-operator/blob/main/docs/clusters.md)\n*
          [Machine Pools](https://github.com/rh-mobb/ocm-operator/blob/main/docs/machinepools.md)\n*
          [Identity Providers](https://github.com/rh-mobb/ocm-operator/blob/main/docs/identityproviders.md)\n"
        displayName: OpenShift Cluster Manager Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ocm
        - rosa
        - redhat
        - machinepool
        - mobb
        links:
        - name: OCM Machine Pool Operator
          url: https://github.com/rh-mobb/ocm-operator
        maintainers:
        - email: dscott@redhat.com
          name: Dustin Scott
        maturity: alpha
        provider:
          name: Managed OpenShift Black Belt
          url: https://github.com/rh-mobb/ocm-operator
        relatedImages:
        - quay.io/openshift-community-operators/ocm-operator@sha256:ca6a3b24ca4ecfcab52ee0402e007d6f14b23fa7f3408d740923806989b170e4
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/mobb/ocm-operator:v0.1.0
        version: 0.1.0
      entries:
      - name: ocm-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ocm-operator
    provider:
      name: Managed OpenShift Black Belt
      url: https://github.com/rh-mobb/ocm-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: community-windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: community-windows-machine-config-operator.v6.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
          createdAt: "2022-08-31T18:57:22Z"
          description: An operator that enables Windows container workloads on OCP
          olm.skipRange: '>=5.0.0 <6.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ## WARNING
          Community distribution of the Windows Machine Config Operator.
          This is a preview build, and is not meant for production workloads.
          Issues with this distribution of WMCO can be opened against the [WMCO repository](https://github.com/openshift/windows-machine-config-operator).
          Please read through the [troubleshooting doc](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/TROUBLESHOOTING.md)
          before opening an issue.
          Please ensure that when installing this operator the starting CSV you subscribe to is supported on the
          version of OKD/OCP you are using. This CSV is meant for OKD/OCP community-4.11.
          ## Documentation
          ### Introduction
          The Windows Machine Config Operator configures Windows instances into nodes, enabling Windows container workloads
          to be ran within OKD/OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/latest/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#configuring-byoh-bring-your-own-host-windows-instances).
          The operator will do all the necessary steps to configure the instance so that it can join the cluster as a worker node.
          ### Pre-requisites
          - [Cluster and OS pre-requisites](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/docs/wmco-prerequisites.md)
          ### Usage
          Please see the usage section of [README.md](https://github.com/openshift/windows-machine-config-operator/blob/community-4.11/README.md#usage).
          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.
        displayName: Community Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: preview
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/community-windows-machine-config-operator@sha256:2b79f0e6b8bc069fac7d6cb5c86783fdc98ccae30d279ca0492f1eca96b6ea3f
        - quay.io/openshift-windows/community-windows-machine-config-operator:community-4.11-6c3601f
        version: 6.0.0
      entries:
      - name: community-windows-machine-config-operator.v6.0.0
        version: 6.0.0
      name: preview
    defaultChannel: preview
    packageName: community-windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-eventbridge-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-eventbridge-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Archive",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Endpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "EventBus",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "eventbridge.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
          createdAt: "2025-09-19T20:51:55Z"
          description: AWS EventBridge controller is a service controller for managing
            EventBridge resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Archive represents the state of an AWS eventbridge Archive
              resource.
            displayName: Archive
            kind: Archive
            name: archives.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Endpoint represents the state of an AWS eventbridge Endpoint
              resource.
            displayName: Endpoint
            kind: Endpoint
            name: endpoints.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: EventBus represents the state of an AWS eventbridge EventBus
              resource.
            displayName: EventBus
            kind: EventBus
            name: eventbuses.eventbridge.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS eventbridge Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.eventbridge.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon EventBridge resources in AWS from within your Kubernetes cluster.

          **About Amazon EventBridge**

          EventBridge is a serverless service that uses events to connect application components together, making it easier for you to build scalable event-driven applications. Use it to route events from sources such as home-grown applications, AWS services, and third- party software to consumer applications across your organization. EventBridge provides a simple and consistent way to ingest, filter, transform, and deliver events so you can build new applications quickly.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EventBridge
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - eventbridge
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EventBridge Developer Resources
          url: https://aws.amazon.com/eventbridge/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: eventbridge maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/eventbridge-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-eventbridge-controller@sha256:3a284fb7884e50ed9976368f312c75fc1f1cb2ec361297b38ee72d0ac6b66fc7
        version: 1.1.1
      entries:
      - name: ack-eventbridge-controller.v1.1.1
        version: 1.1.1
      - name: ack-eventbridge-controller.v1.1.0
        version: 1.1.0
      - name: ack-eventbridge-controller.v1.0.28
        version: 1.0.28
      - name: ack-eventbridge-controller.v1.0.27
        version: 1.0.27
      - name: ack-eventbridge-controller.v1.0.26
        version: 1.0.26
      - name: ack-eventbridge-controller.v1.0.25
        version: 1.0.25
      - name: ack-eventbridge-controller.v1.0.24
        version: 1.0.24
      - name: ack-eventbridge-controller.v1.0.23
        version: 1.0.23
      - name: ack-eventbridge-controller.v1.0.22
        version: 1.0.22
      - name: ack-eventbridge-controller.v1.0.21
        version: 1.0.21
      - name: ack-eventbridge-controller.v1.0.20
        version: 1.0.20
      - name: ack-eventbridge-controller.v1.0.18
        version: 1.0.18
      - name: ack-eventbridge-controller.v1.0.17
        version: 1.0.17
      - name: ack-eventbridge-controller.v1.0.16
        version: 1.0.16
      - name: ack-eventbridge-controller.v1.0.15
        version: 1.0.15
      - name: ack-eventbridge-controller.v1.0.14
        version: 1.0.14
      - name: ack-eventbridge-controller.v1.0.13
        version: 1.0.13
      - name: ack-eventbridge-controller.v1.0.12
        version: 1.0.12
      - name: ack-eventbridge-controller.v1.0.11
        version: 1.0.11
      - name: ack-eventbridge-controller.v1.0.6
        version: 1.0.6
      - name: ack-eventbridge-controller.v1.0.5
        version: 1.0.5
      - name: ack-eventbridge-controller.v1.0.4
        version: 1.0.4
      - name: ack-eventbridge-controller.v1.0.3
        version: 1.0.3
      - name: ack-eventbridge-controller.v1.0.1
        version: 1.0.1
      - name: ack-eventbridge-controller.v1.0.0
        version: 1.0.0
      - name: ack-eventbridge-controller.v0.0.7
        version: 0.0.7
      - name: ack-eventbridge-controller.v0.0.6
        version: 0.0.6
      - name: ack-eventbridge-controller.v0.0.5
        version: 0.0.5
      - name: ack-eventbridge-controller.v0.0.4
        version: 0.0.4
      - name: ack-eventbridge-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-eventbridge-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: ""
    name: nim-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nim-operator-certified.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMCache",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "source": {
                    "ngc": {
                      "modelPuller": "nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6",
                      "pullSecret": "ngc-secret",
                      "authSecret": "ngc-api-secret",
                      "model": {
                        "lora": false,
                        "precision": "fp16",
                        "engine": "tensorrt_llm",
                        "qosProfile": "throughput",
                        "gpus": [
                          {
                            "product": "a100",
                            "ids": []
                          }
                        ]
                      }
                    }
                  },
                  "storage": {
                    "pvc": {
                      "create": true,
                      "storageClass": "local-path",
                      "size": "50Gi",
                      "volumeAccessMode": "ReadWriteOnce"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMService",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "authSecret": "ngc-api-secret",
                  "storage": {
                    "nimCache": {
                      "name": "meta-llama3-1b-instruct",
                      "profile": ""
                    }
                  },
                  "scale": {
                    "enabled": false,
                    "hpa": {
                      "minReplicas": 1,
                      "maxReplicas": 10
                    }
                  },
                  "resources": {
                    "limits": {
                      "nvidia.com/gpu": 1
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMPipeline",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "k8s-nim-operator",
                    "app.kubernetes.io/managed-by": "k8s-nim-operator"
                  },
                  "name": "nimpipeline-sample"
                },
                "spec": {
                  "services": [
                    {
                      "name": "meta-llama3-1b-instruct",
                      "enabled": true,
                      "spec": {
                        "image": {
                          "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                          "tag": "1.12.0",
                          "pullPolicy": "IfNotPresent",
                          "pullSecrets": [
                            "ngc-secret"
                          ]
                        },
                        "authSecret": "ngc-api-secret",
                        "storage": {
                          "nimCache": {
                            "name": "meta-llama3-1b-instruct",
                            "profile": ""
                          }
                        },
                        "replicas": 1,
                        "resources": {
                          "limits": {
                            "nvidia.com/gpu": 1
                          }
                        },
                        "expose": {
                          "service": {
                            "type": "ClusterIP",
                            "port": 8000
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NIMBuild",
                "metadata": {
                  "name": "meta-llama3-1b-instruct",
                  "namespace": "nim-service"
                },
                "spec": {
                  "nimCache": {
                    "name": "meta-llama3-1b-instruct",
                    "profile": "7cc8597690a35aba19a3636f35e7f1c7e7dbc005fe88ce9394cad4a4adeed414"
                  },
                  "image": {
                    "repository": "nvcr.io/nim/meta/llama-3.2-1b-instruct",
                    "tag": "1.12.0",
                    "pullSecrets": [
                     "ngc-secret"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoCustomizer",
                "metadata": {
                  "name": "nemocustomizer-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "scheduler": {
                    "type": "volcano"
                  },
                  "wandb": {
                    "secretName": "wandb-secret",
                    "apiKeyKey": "apiKey",
                    "encryptionKey": "encryptionKey"
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://customizer-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ncsuser",
                      "secretName": "customizer-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "customizer-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ncsdb"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/customizer-api",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  },
                  "mlflow": {
                    "endpoint": "http://mlflow-tracking.nemo.svc.cluster.local:80"
                  },
                  "nemoDatastoreTools": {
                    "image": "nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11"
                  },
                  "modelDownloadJobs": {
                    "image": "nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b",
                    "ngcAPISecret": {
                      "name": "ngc-api-secret",
                      "key": "NGC_API_KEY"
                    },
                    "securityContext": {
                      "fsGroup": 1000,
                      "runAsNonRoot": true,
                      "runAsUser": 1000,
                      "runAsGroup": 1000
                    },
                    "ttlSecondsAfterFinished": 600,
                    "pollIntervalSeconds": 15
                  },
                  "modelConfig": {
                    "name": "nemo-model-config"
                  },
                  "trainingConfig": {
                    "configMap": {
                      "name": "nemo-training-config"
                    },
                    "modelPVC": {
                      "create": true,
                      "name": "finetuning-ms-models-pvc",
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 50
                    },
                    "workspacePVC": {
                      "storageClass": "local-path",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": 10,
                      "mountPath": "/pvc/workspace"
                    },
                    "image": {
                      "repository": "nvcr.io/nvidia/nemo-microservices/customizer",
                      "tag": "25.08"
                    },
                    "env": [
                      { "name": "LOG_LEVEL", "value": "INFO" }
                    ],
                    "networkConfig": [
                      { "name": "NCCL_IB_SL", "value": "0" },
                      { "name": "NCCL_IB_TC", "value": "41" },
                      { "name": "NCCL_IB_QPS_PER_CONNECTION", "value": "4" },
                      { "name": "UCX_TLS", "value": "TCP" },
                      { "name": "UCX_NET_DEVICES", "value": "eth0" },
                      { "name": "HCOLL_ENABLE_MCAST_ALL", "value": "0" },
                      { "name": "NCCL_IB_GID_INDEX", "value": "3" }
                    ],
                    "ttlSecondsAfterFinished": 3600,
                    "timeout": 3600,
                    "tolerations": [
                      {
                        "key": "nvidia.com/gpu",
                        "operator": "Exists",
                        "effect": "NoSchedule"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoGuardrail",
                "metadata": {
                  "name": "nemoguardrails-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "configStore": {
                    "pvc": {
                      "name": "pvc-guardrail-config",
                      "create": true,
                      "storageClass": "",
                      "volumeAccessMode": "ReadWriteOnce",
                      "size": "1Gi"
                    }
                  },
                  "nimEndpoint": {
                    "baseURL": "http://meta-llama3-1b-instruct.nemo.svc.cluster.local:8000/v1"
                  },
                  "databaseConfig": {
                    "host": "guardrail-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "guardraildb",
                    "credentials": {
                      "user": "guardrailuser",
                      "secretName": "guardrail-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/guardrails",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [
                      "ngc-secret"
                    ]
                  },
                  "metrics": {
                    "serviceMonitor": {}
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "1",
                      "ephemeral-storage": "10Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEvaluator",
                "metadata": {
                  "name": "nemoevaluator-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "evaluationImages": {
                    "bigcodeEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc",
                    "lmEvalHarness": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d",
                    "similarityMetrics": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9",
                    "llmAsJudge": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "mtBench": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055",
                    "retriever": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250",
                    "rag": "nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783",
                    "bfcl": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c",
                    "agenticEval": "nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc"
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/evaluator",
                    "tag": "25.06",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "argoWorkflows": {
                    "endpoint": "https://argo-workflows-server.nemo.svc.cluster.local:2746",
                    "serviceAccount": "argo-workflows-executor"
                  },
                  "vectorDB": {
                    "endpoint": "http://milvus.nemo.svc.cluster.local:19530"
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000/v1/hf"
                  },
                  "entitystore": {
                    "endpoint": "http://nemoentitystore-sample.nemo.svc.cluster.local:8000"
                  },
                  "databaseConfig": {
                    "host": "evaluator-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "evaldb",
                    "credentials": {
                      "user": "evaluser",
                      "secretName": "evaluator-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "otel": {
                    "enabled": true,
                    "exporterOtlpEndpoint": "http://evaluator-otel-opentelemetry-collector.nemo.svc.cluster.local:4317"
                  },
                  "replicas": 1
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoEntitystore",
                "metadata": {
                  "name": "nemoentitystore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/entity-store",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "databaseConfig": {
                    "databaseName": "nesdb",
                    "host": "entity-store-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "credentials": {
                      "user": "nesuser",
                      "secretName": "entity-store-pg-existing-secret",
                      "passwordKey": "password"
                    }
                  },
                  "datastore": {
                    "endpoint": "http://nemodatastore-sample.nemo.svc.cluster.local:8000"
                  }
                }
              },
              {
                "apiVersion": "apps.nvidia.com/v1alpha1",
                "kind": "NemoDatastore",
                "metadata": {
                  "name": "nemodatastore-sample",
                  "namespace": "nemo"
                },
                "spec": {
                  "secrets": {
                    "datastoreConfigSecret": "nemo-ms-nemo-datastore",
                    "datastoreInitSecret": "nemo-ms-nemo-datastore-init",
                    "datastoreInlineConfigSecret": "nemo-ms-nemo-datastore-inline-config",
                    "giteaAdminSecret": "gitea-admin-credentials",
                    "lfsJwtSecret": "nemo-ms-nemo-datastore--lfs-jwt"
                  },
                  "databaseConfig": {
                    "credentials": {
                      "user": "ndsuser",
                      "secretName": "datastore-pg-existing-secret",
                      "passwordKey": "password"
                    },
                    "host": "datastore-pg-postgresql.nemo.svc.cluster.local",
                    "port": 5432,
                    "databaseName": "ndsdb"
                  },
                  "pvc": {
                    "name": "pvc-shared-data",
                    "create": true,
                    "storageClass": "",
                    "volumeAccessMode": "ReadWriteOnce",
                    "size": "10Gi"
                  },
                  "expose": {
                    "service": {
                      "type": "ClusterIP",
                      "port": 8000
                    }
                  },
                  "image": {
                    "repository": "nvcr.io/nvidia/nemo-microservices/datastore",
                    "tag": "25.08",
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": ["ngc-secret"]
                  },
                  "replicas": 1,
                  "resources": {
                    "requests": {
                      "memory": "256Mi",
                      "cpu": "500m"
                    },
                    "limits": {
                      "memory": "512Mi",
                      "cpu": "1"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
          createdAt: Wed Sep 10 05:49:48 PDT 2025
          description: An Operator for deployment and maintenance of NVIDIA NIMs and
            NeMo microservices in a Kubernetes environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          k8sMinVersion: 1.26.0
          repository: github.com/NVIDIA/k8s-nim-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NIM Cache
            displayName: NIMCache
            kind: NIMCache
            name: nimcaches.apps.nvidia.com
            version: v1alpha1
          - description: NIM Service
            displayName: NIMService
            kind: NIMService
            name: nimservices.apps.nvidia.com
            version: v1alpha1
          - description: NIM Pipeline
            displayName: NIMPipeline
            kind: NIMPipeline
            name: nimpipelines.apps.nvidia.com
            version: v1alpha1
          - description: NIM Build
            displayName: NIMBuild
            kind: NIMBuild
            name: nimbuilds.apps.nvidia.com
            version: v1alpha1
          - description: NEMO guardrail Service
            displayName: NemoGuardrail
            kind: NemoGuardrail
            name: nemoguardrails.apps.nvidia.com
            version: v1alpha1
          - description: NEMO datastore Service
            displayName: NemoDatastore
            kind: NemoDatastore
            name: nemodatastores.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Eval Service
            displayName: NemoEvaluator
            kind: NemoEvaluator
            name: nemoevaluators.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Customizer Service
            displayName: NemoCustomizer
            kind: NemoCustomizer
            name: nemocustomizers.apps.nvidia.com
            version: v1alpha1
          - description: NEMO Entitystore Service
            displayName: NemoEntitystore
            kind: NemoEntitystore
            name: nemoentitystores.apps.nvidia.com
            version: v1alpha1
        description: |-
          ## NVIDIA NIM Operator
          The NVIDIA NIM Operator enables Kubernetes cluster administrators to operate the software components and services necessary to run NVIDIA NIMs in various domains such as reasoning, retrieval, speech, and biology. Additionally, it allows the use of NeMo Microservices to fine-tune, evaluate, or apply guardrails to your models.
          The Operator manages the life cycle of the following microservices and the models they use:
          ### NVIDIA NIM models, such as:
          * Reasoning LLMs
          * Retrieval - Embedding, Reranking, etc.
          * Speech
          * Biology
          ### NeMo core microservices:
          * NeMo Customizer
          * NeMo Evaluator
          * NeMo Guardrails
          ### NeMo platform component microservices:
          * NeMo Data Store
          * NeMo Entity Store
        displayName: The NVIDIA NIM Operator for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        links:
        - name: NVIDIA NIM
          url: https://docs.nvidia.com/nim/index.html
        - name: Documentation
          url: https://docs.nvidia.com/nim-operator/latest
        maintainers:
        - email: operator_feedback@nvidia.com
          name: shivamerla
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: NVIDIA
        relatedImages:
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-retriever@sha256:cb99fb3ea5b4e9989832679d4c6028573c28b99e43744e008ac5c9748b7dd250
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-custom-eval@sha256:eb51edf668481d1f5d2b100c850958dadf184fd8779433e45da40fa2a35ec4b9
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-rag@sha256:b5e35cf277185e0901a4bbba35780b3673be71ae2c5e8efcdbab7710a8a6f783
        - nvcr.io/nvidia/cloud-native/k8s-nim-operator@sha256:f42dd5c3dc269553ee46725ed4b1ab2d2f253ef06117f1d9d52faaa22d8fe60d
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-llm-as-a-judge@sha256:188916fce714453d48ff20d95e1ca09fc5520ccb6e9e8da7493948697e919055
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-lm-eval-harness@sha256:58e7c3da7f1edfed2ff508fa8abab7660366c2c17bb55d17aa68159d6e16be6d
        - nvcr.io/nvidia/nemo-microservices/nds-v2-huggingface-cli@sha256:7958e5244217de8af0ed723dfe7ea8e291b3afbb2a0aa73f5c27b218a1149b11
        - nvcr.io/nim/meta/llama-3.2-1b-instruct@sha256:1ff21ad32124d2382f8c733879394e7849cc3df5172651a24fcb671ac4fe77c6
        - nvcr.io/nvidia/nemo-microservices/customizer-api@sha256:c590fa19121af2aac0379933deace3991f27c3ef88a8043560d6e3a172273b0b
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-agentic-eval@sha256:b38bbc5e61df0930204efa096cf4f40db4655b92d03741190513b019d67fc7dc
        - nvcr.io/nvidia/nemo-microservices/eval-factory-benchmark-bfcl@sha256:95f45f66140d481d544b36b16bc3de7ad5e090d76225fd73ed70461b9bd9907c
        - nvcr.io/nvidia/nemo-microservices/eval-tool-benchmark-bigcode@sha256:62f74c32d1ed07925ddb906de8975025bf63d73e82e7faefb840339aaf21a9dc
        - registry.connect.redhat.com/nvidia/nim-operator-bundle@sha256:3dbd307d32a39e37a02ac8fa9c9fb15a0c2fd6dc5aea12d55abb061e86365124
        version: 3.0.0
      entries:
      - name: nim-operator-certified.v3.0.0
        version: 3.0.0
      - name: nim-operator-certified.v2.0.2
        version: 2.0.2
      - name: nim-operator-certified.v2.0.1
        version: 2.0.1
      - name: nim-operator-certified.v2.0.0
        version: 2.0.0
      - name: nim-operator-certified.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: stable
    packageName: nim-operator-certified
    provider:
      name: NVIDIA
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-directory-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-directory-operator.v23.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyDirectory",
                "metadata": {
                  "name": "ibmsecurityverifydirectory-sample"
                },
                "spec": {
                  "pods": {
                    "configMap": {
                      "proxy": {
                        "key": "config.yaml",
                        "name": "isvd-proxy-config"
                      },
                      "server": {
                        "key": "config.yaml",
                        "name": "isvd-server-config"
                      }
                    },
                    "image": {
                      "label": "latest",
                      "repo": "icr.io/isvd"
                    },
                    "proxy": {
                      "replicas": 1
                    },
                    "serviceAccountName": "isvd"
                  },
                  "replicas": {
                    "pvcs": [
                      "replica-1",
                      "replica-2"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
          createdAt: "2023-04-18T22:10:32Z"
          description: The IBM Security Verify Directory operator is used to scale
            an IBM Security Verify Directory Server environment.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ibm-security/verify-directory-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyDirectory is the Schema for the ibmsecurityverifydirectories
              API
            displayName: IBMSecurity Verify Directory
            kind: IBMSecurityVerifyDirectory
            name: ibmsecurityverifydirectories.ibm.com
            version: v1
        description: "IBM Security Verify Directory is a scalable, standards-based
          identity directory that helps simplify identity and directory management.
          Verify Directory helps consolidate identity silos into a single identity
          source. Verify Directory is purpose-built to provide a directory foundation
          that can help provide a trusted identity data infrastructure that assists
          in enabling mission-critical security and authentication. It is designed
          to deliver a reliable, scalable, standards-based identity data platform
          that interoperates with a broad range of operating systems and applications.
          Verify Directory supports Lightweight Directory Access Protocol (LDAP) V3,
          offering a flexible and highly scalable LDAP infrastructure. \n\nThe operator
          can be used to manage deployed instances of the IBM Security Verify Directory
          server, providing an easy mechanism to scale the number of servers which
          are used to handle the workload.\n\nSee the project [Readme](https://github.com/ibm-security/verify-directory-operator/blob/master/README.md)
          for further information and details.\n\n"
        displayName: IBM Security Verify Directory
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Security
        - Identity and Access
        links:
        - name: IBM Security Verify Directory Documentation
          url: https://www.ibm.com/docs/en/svd
        maintainers:
        - email: iamdev@us.ibm.com
          name: IBM Security Verify Directory Development Team
        maturity: stable
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-directory-operator@sha256:9d93e3aeef3cfbba0b655de7783ba9fd1f4cb4ee5de4389005b7c31d971d4947
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d99a8d144816b951a67648c12c0b988936ccd25cf3754f3cd85ab8c01592248f
        - icr.io/isvd/verify-directory-operator@sha256:d1dd7037584c6818dcf8d5aa69472278833bb8e37fba2aec5a69185ba88dcc34
        version: 23.4.1
      entries:
      - name: ibm-security-verify-directory-operator.v23.4.1
        version: 23.4.1
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-directory-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-listener-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: listener-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
          description: Stackable Listener Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/listener-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Listener Class
            displayName: Listener Class
            kind: ListenerClass
            name: listenerclasses.listeners.stackable.tech
            version: v1alpha1
          - description: Represents a Listener
            displayName: Listener
            kind: Listener
            name: listeners.listeners.stackable.tech
            version: v1alpha1
          - description: Represents PodListeners of a Pod
            displayName: PodListeners
            kind: PodListeners
            name: podlisteners.listeners.stackable.tech
            version: v1alpha1
        description: |-
          This is an operator for Kubernetes that provisions network listeners according to the cluster policy, and injects connection parameters into Pods.
          The Stackable Listener Operator is part of the Stackable Data Platform, a curated selection of the best open source data apps like Kafka, Druid, Trino or
          Spark, all working together seamlessly. Based on Kubernetes, it runs everywhere – on prem or in the cloud.

          See it in action in one of our [demos](https://stackable.tech/en/demos/) or follow this
          [usage overview](https://docs.stackable.tech/home/stable/listener-operator/usage.html).

          NOTE: Make sure you install this operator in a namespace called "stackable-operators". Failing to do so will result in a broken installation.
        displayName: Stackable Listener Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - listener
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/listener-operator@sha256:2d184d2d61aca2b5c17e471f02b732069dd58518d3a3f675da5ecb695cefde0b
        - quay.io/stackable/listener-operator@sha256:2d65b37982d6a116ff3518b1858151e08284c50e2a2e75fb268874fd8b1912bf
        - quay.io/stackable/sig-storage/csi-node-driver-registrar@sha256:89a09e870909deb6125393e30f3dbcd1e0769d3c112b58622419f5c6490b7a1b
        - quay.io/stackable/sig-storage/csi-provisioner@sha256:5b30fb9f8c85860b7021d6139aaf04c2104b1acc99ca6b68a0f6e34a6a7fc886
        version: 25.7.0
      entries:
      - name: listener-operator.v25.7.0
        version: 25.7.0
      - name: listener-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-listener-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: global-load-balancer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: global-load-balancer-operator.v1.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "hello-global-record"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "loadBalancingPolicy": "Multivalue",
                  "name": "hello.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSRecord",
                "metadata": {
                  "name": "route53-geoproximity-global-record-with-healthcheck"
                },
                "spec": {
                  "endpoints": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3",
                      "loadBalancerServiceRef": {
                        "name": "cluster-default",
                        "namespace": "openshift-ingress"
                      }
                    }
                  ],
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "healthCheck": {
                    "failureThreshold": 3,
                    "httpGet": {
                      "host": "geoproximity-hc.${global_base_domain}",
                      "path": "/healthz",
                      "port": 80,
                      "scheme": "HTTP"
                    },
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 2
                  },
                  "loadBalancingPolicy": "Geoproximity",
                  "name": "geoproximity-hc.my.base.domain",
                  "ttl": 60
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "external-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "externalDNS": {
                      "annotations": {
                        "type": "global"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalDNSZone",
                "metadata": {
                  "name": "route53-global-dns-zone"
                },
                "spec": {
                  "domain": "myglobalzone.mybasedomain.io",
                  "provider": {
                    "route53": {
                      "credentialsSecretRef": {
                        "name": "route53-global-zone-credentials",
                        "namespace": "mynamespace"
                      },
                      "zoneID": "Z083892239ROKNEP70BDY"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "external-dns-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "external-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GlobalRouteDiscovery",
                "metadata": {
                  "name": "route53-globalroutediscovery"
                },
                "spec": {
                  "clusters": [
                    {
                      "clusterCredentialRef": {
                        "name": "secret1",
                        "namespace": "cluster1"
                      },
                      "clusterName": "cluster1"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret2",
                        "namespace": "cluster2"
                      },
                      "clusterName": "cluster2"
                    },
                    {
                      "clusterCredentialRef": {
                        "name": "secret3",
                        "namespace": "cluster3"
                      },
                      "clusterName": "cluster3"
                    }
                  ],
                  "defaultLoadBalancingPolicy": "Multivalue",
                  "defaultTTL": 30,
                  "globalZoneRef": {
                    "name": "route53-global-dns-zone"
                  },
                  "routeSelector": {
                    "matchLabels": {
                      "route-type": "global"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking, Cloud Provider
          certified: "false"
          containerImage: quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
          createdAt: "2021-12-02T17:33:23Z"
          description: This operator creates automation around a DNS to operate as
            a global load balancer for a set of OpenShift clusters.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/global-load-balancer-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DNSEndpoint
            name: dnsendpoints.externaldns.k8s.io
            version: v1alpha1
          - description: GlobalDNSRecord is the Schema for the globaldnsrecords API
            displayName: Global DNSRecord
            kind: GlobalDNSRecord
            name: globaldnsrecords.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalDNSZone is the Schema for the globaldnszones API
            displayName: Global DNSZone
            kind: GlobalDNSZone
            name: globaldnszones.redhatcop.redhat.io
            version: v1alpha1
          - description: GlobalRouteDiscovery is the Schema for the globalroutediscoveries
              API
            displayName: Global Route Discovery
            kind: GlobalRouteDiscovery
            name: globalroutediscoveries.redhatcop.redhat.io
            version: v1alpha1
        description: "The global-load-balancer-operator implements automation to program
          a DNS service to act as global load balancer for applications deployed to
          multiple OpenShift clusters.\nThis operator is designed to be deployed to
          a control cluster which will watch the load balanced clusters (controlled
          clusters).\nThere are two main concepts (APIs) provided by this operator:\n\n1.
          GlobalDNSZone\n2. GlobalDNSRecord\n\n## GlobalDNSZone\n\nThe `GlobalDNSZone`
          CR allows you to configure a zone which will contain global load balanced
          records and the provider used to populate it.\nHere is an example of GlobalDNSZone:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\nkind: GlobalDNSZone\nmetadata:\n  name: external-dns-zone\nspec:\n
          \ # Add fields here\n  domain: global.myzone.io\n  provider:\n    externalDNS:\n
          \     annotations:\n        type: global\n```\n\nHere is a table summarizing
          the supported providers and their capabilities:\n\n| Provider  | Supports
          Health Checks  | Supports Multivalue LB | Supports Latency LB  | Supports
          GeoProximity LB  |\n|:--:|:--:|:---:|:---:|:---:|\n| External-dns  | no
          \ | yes | no  | no  |\n| Route53  | yes | yes(*)  | yes(*)  | yes(*)  |\n\n(*)
          only if all controlled clusters run on AWS.\n\n## GlobalDNSRecord\n\nThe
          `GlobalDNSRecord` CR allows you to specify the intention to create a global
          dns record. Here is an example\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalDNSRecord\nmetadata:\n  name: hello-global-record\nspec:\n  name:
          hello.global.myzone.io\n  endpoints:\n  - clusterName: cluster1\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster1\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  - clusterName:
          cluster2\n    clusterCredentialRef:\n      name: kubeconfig\n      namespace:
          cluster2\n    loadBalancerServiceRef:\n      name: router-default\n      namespace:
          openshift-ingress\n  - clusterName: cluster3\n    clusterCredentialRef:\n
          \     name: kubeconfig\n      namespace: cluster3\n    loadBalancerServiceRef:\n
          \     name: router-default\n      namespace: openshift-ingress\n  ttl: 60\n
          \ loadBalancingPolicy: Multivalue\n  globalZoneRef:\n    name: external-dns-zone\n
          \ healthCheck:\n    failureThreshold: 3\n    httpGet:\n      host: hello.global.myzone.io\n
          \     httpHeaders:\n        - name: ciao\n          value: hello\n      path:
          /readyz\n      port: 80\n      scheme: HTTP\n    periodSeconds: 10\n    successThreshold:
          1\n    timeoutSeconds: 1     \n```\n\nThe list of endpoints identifies the
          set of `LoadBalancer` type of services to watch on the remote clusters (at
          the moment only LoadBalancer services are supported). These are the `LoadBalancer`
          services created by the [ingress controller](https://docs.openshift.com/container-platform/4.5/networking/ingress-operator.html)
          on which the routers rely. Each entry of this list requires a reference
          to the loadbalancer service on the remote cluster and a reference to a local
          secret containing the credential to connect to the remote cluster.\n\nThis
          secret must contain one entry called `kubeconfig` with a kubeconfig file
          with a default context pointing to the remote cluster. The account used
          by the kubeconfig file (presumably a service account) will require at a
          minimum cluster-level permissions as described in this [cluster role](./docs/scripts/remote-account-cluster-role.yaml).\n\nThe
          `globalZoneRef` field refers to a local (same namespace) `GlobalZone` CR.
          The DNS record represented by this `GlobalDNSRecord`, will be created in
          the referenced zone.\n\n`ttl` is the TTL of the crated DNS record.\n\n`loadBalancingPolicy`
          is the load balancing policy for this global record. It must match one of
          the policy supported by the provider of the referenced `GlobalZone`.\n\nFinally,
          `healthcheck` represent a [probe](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.18/#probe-v1-core)
          to be used to test the health of a record. This field will be ignored if
          the provider does not support health checks.\n\n## External DNS Provider\n\nThe
          [external-dns](https://github.com/kubernetes-sigs/external-dns) provider
          delegates to external-dns the creation of the actual DNS records by creating
          a `DNSEndpoint` object.\nThe `DNSEndpoint` object will be created in the
          same namespace as the `GlobalDNSRecord` and will be owned by it.\nThe `DNSEdnpoint`
          object will have the same labels as the `GlobalDNSRecord` and the annotations
          specified in the GlobalDNSZone configuration.\nExternal-dns should be configured
          to watch for DNSEnpoints at the cluster level and to point to the desired
          provider.\nDetails on configuration can be found at the external-dns git
          repository.\nThe External-dns should be used as a fall back option when
          other options are not available as it does not support health checks and
          advanced load balancing policies.\n\n## AWS Route53 provider\n\nAWS Route53
          provider uses the Route53 service as a global loadbalancer and offers advanced
          routing capabilities via [route53 traffic policies](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/traffic-flow.html)
          (note that traffic policies will trigger an expense).\nThe following routing
          polices are currently supported:\n\n1. [Multivalue](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-multivalue)\n2.
          [Geoproximity](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-geoproximity)\n3.
          [Latency](https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/routing-policy.html#routing-policy-latency)\n\nAWS
          Route53 provider at the moment requires that all the controlled clusters
          run in AWS.\n\nIf health checks are defined, a route53 health check originating
          from any reason (you have to ensure connectivity) will be created for each
          of the endpoint. Because the endpoint represent s a shared ELB (shared with
          other apps, that is) and the health check is app specific, we cannot use
          the ELB health check, so the route53 endpoint is created with one of the
          two IP exposed by the ELB. This is suboptimal, but it works in most situations.\n\n##
          Global Route Auto Discovery\n\nThe aboev examples showed how to create global
          DNS records. This can be good in some situations, but most of the times
          in an openshift deployment global DNS records will point to routes that
          are intended to be global.\nThe global-load-balancer operator can auto-discover
          these routes and automatically create the corresponding `GloablDNSRecord`.
          The `GlobalRouteDiscovery` CRD is used to configure the discovery process,
          here is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          GlobalRouteDiscovery\nmetadata:\n  name: route53-globalroutediscovery\nspec:\n
          \ clusters:\n  - clusterName: cluster1\n    clusterCredentialRef:\n      name:
          ${cluster1_secret_name}\n      namespace: cluster1\n...\n  routeSelector:\n
          \   matchLabels:\n      route-type: global\n  defaultLoadBalancingPolicy:
          Multivalue\n  defaultTTL: 30 \n  globalZoneRef:\n    name: route53-global-dns-zone\n```\n\nThis
          global discovery route will discover routes in the provided list of cluster.
          Only the routes that match the route selector will be considered global.
          \nThe default load balancing policy and default TTL can be expressed in
          the `GlobalRouteDiscovery` CR. However with the following annotations, it's
          possible to configure route-specific values:\n\n* `global-load-balancer-operator.redhat-cop.io/load-balancing-policy`
          to set the load balancing policy\n* `global-load-balancer-operator.redhat-cop.io/ttl`
          to set the TTL\n\nThe globalZoneRef refers to the global zone to be used
          for the created `GlobalDNSRecords`.\n\nHealth checks will also be automatically
          discovered. If the pods behind the route expose a readiness check of httpGet
          kind, that configuration will be used to create the GlobalDNSRecord health
          check.\nWhen more than one container is present in the pod, by default the
          first container will be examined for health check. This behavior can be
          overridden with the this annotation on the route: `global-load-balancer-operator.redhat-cop.io/container-probe`
          where the value will container the name of the container with teh correct
          readiness probe.\n\nIf routes with the same namespace and name exists in
          multiple cluster, the following conditions must be met:\n\n* all host names
          must be the same\n* all load balancing policy must be the same\n* all TTLs
          must be the same\n* all discovered readiness checks must be the same\n\n##
          Examples\n\nThese examples are intended to help you setting up working configuration
          with each of the providers\n\n### Cluster Set up\n\nTwo approaches for cluster
          setup are provided\n\n1. [One cluster, three ingress-gateways.](https://github.com/redhat-cop/global-load-balancer-operator/docs/one-cluster-three-ingress-gateways.md)
          This approach is intended for development purposes and has the objective
          to keep resource consumption at the minimum.\n2. [Control cluster and three
          controlled clusters in different regions.](https://github.com/redhat-cop/global-load-balancer-operator/docs/three-clusters.md).
          This approach represents a more realistic set-up albeit it consumes more
          resources.\n\nYou can also set up the cluster on your own, at the end the
          following conditions must be met:\n\nThree namespace `cluster1` `cluster2`
          `cluster3` are created.\nthe following environment variables are initialized
          for each cluster:\n\n1. <cluster>_secret_name. Pointing to a secret in each
          of the cluster namespaces containing a valid kubeconfig fot that cluster\n2.
          <cluster>_service_name.  Pointing to the name of the loadbalancer service
          to be used for that cluster.\n3. <cluster>_service_namespace. Pointing to
          the namespace of the loadbalancer service to be used for that cluster.\n\nHere
          are examples for the supported provider:\n\n1. [Setting up external-dns
          as provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/external-dns-provider.md)\n2.
          [Setting up route53 as a provider](https://github.com/redhat-cop/global-load-balancer-operator/docs/aws-route53-provider.md)\n"
        displayName: Global Load Balancer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Networking
        - Cloud Provider
        - Global Load Balancing
        - GLB
        - GTM
        - route53
        - DNS
        links:
        - name: GigHub Repository
          url: https://github.com/redhat-cop/global-load-balancer-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/global-load-balancer-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/global-load-balancer-operator@sha256:50771c320c8b7faea144f765535e89979f9f81e61d34c8dd2e86139fa68e222e
        - quay.io/redhat-cop/global-load-balancer-operator@sha256:c0b6ac9b03571a74a56b28302d720e8aa4442233f7867d05ed365e022f7cb8f5
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.4.2
      entries:
      - name: global-load-balancer-operator.v1.4.2
        version: 1.4.2
      - name: global-load-balancer-operator.v1.4.1
        version: 1.4.1
      - name: global-load-balancer-operator.v1.4.0
        version: 1.4.0
      - name: global-load-balancer-operator.v1.3.0
        version: 1.3.0
      - name: global-load-balancer-operator.v1.2.0
        version: 1.2.0
      - name: global-load-balancer-operator.v1.1.0
        version: 1.1.0
      - name: global-load-balancer-operator.v1.0.7
        version: 1.0.7
      - name: global-load-balancer-operator.v1.0.6
        version: 1.0.6
      - name: global-load-balancer-operator.v1.0.5
        version: 1.0.5
      - name: global-load-balancer-operator.v1.0.4
        version: 1.0.4
      - name: global-load-balancer-operator.v1.0.3
        version: 1.0.3
      - name: global-load-balancer-operator.v1.0.2
        version: 1.0.2
      - name: global-load-balancer-operator.v1.0.1
        version: 1.0.1
      - name: global-load-balancer-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: global-load-balancer-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: unsupported
      operatorframework.io/arch.s390x: unsupported
      operatorframework.io/os.linux: supported
      provider: Konveyor
      provider-url: https://www.konveyor.io
    name: konveyor-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: konveyor-operator.v0.8.0-alpha.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
          createdAt: "2025-08-13T23:43:34Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-alpha.4'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/kantra@sha256:06ae932d9c2f51035bc5a488dd61e8a31c344e72615b013419199168701ea3fb
        - quay.io/konveyor/tackle2-ui@sha256:867bf24dd1687a79f2a911c22c1059ca7bd025a307d290290265653a696b8e98
        - quay.io/konveyor/kai-solution-server@sha256:83987d12d8290e085b6b66de56d347a8bf262815a5f99796177cb9120513ee3a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:ba865a74287ff94375ed316f348ee2c1d962d8d350c8afae5dbb0aee9b64fdea
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:1f4fd5ed8e398b713e733fb552e7294f22b3927d4e017f2ac15ea20024d3de32
        - quay.io/konveyor/java-external-provider@sha256:42a156f1d81352f5e09fa0ffe2b51a4d93a1f476755eb0bac9a1d24bd3a89631
        - quay.io/konveyor/postgresql-15-c9s@sha256:d75c5a1a0fa686ccf701e5e55633aa73d0cbb2cd9b5d8d4d4e80d7c46d78c121
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7a5750fb76935e55b7b59e8363f21f5f9b48b39492d7f3e58e2e6a84d07ab027
        - quay.io/konveyor/tackle2-addon-platform@sha256:12efa5b7fbf6d829879b19e2820dcf782e648059f51544ebd25251dfc67fc7ca
        - quay.io/konveyor/tackle-keycloak-init@sha256:07c2ca07922d273bd2a3d9156ad8af0e5bf973aeb316496ee65d7201c381c1b1
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-hub@sha256:6be8fe8307e6686b178ace42f6db5cab8f04c5e2e29e02d2edcb5dc28ee337e9
        - quay.io/konveyor/tackle2-operator@sha256:c1d1942b3bb3c4e99d9b5484493e27a432455608fee4548f1d352be34a8d6ded
        - quay.io/konveyor/generic-external-provider@sha256:d58e867b51d5e76cd3597cfb7f2478d221b966f84509dbdd44c609ff5d9169be
        version: 0.8.0-alpha.4
      entries:
      - name: konveyor-operator.v0.8.0-alpha.4
        version: 0.8.0-alpha.4
      - name: konveyor-operator.v0.8.0-alpha.2
        version: 0.8.0-alpha.2
      - name: konveyor-operator.v0.8.0-alpha.1
        version: 0.8.0-alpha.1
      - name: konveyor-operator.v0.7.0-alpha.2
        version: 0.7.0-alpha.2
      - name: konveyor-operator.v0.7.0-alpha.1
        version: 0.7.0-alpha.1
      - name: konveyor-operator.v0.6.0-alpha.2
        version: 0.6.0-alpha.2
      - name: konveyor-operator.v0.6.0-alpha.1
        version: 0.6.0-alpha.1
      - name: konveyor-operator.v0.5.0-alpha.1
        version: 0.5.0-alpha.1
      - name: konveyor-operator.v0.4.0-alpha.1
        version: 0.4.0-alpha.1
      - name: konveyor-operator.v0.3.0-alpha.6
        version: 0.3.0-alpha.6
      - name: konveyor-operator.v0.3.0-alpha.5
        version: 0.3.0-alpha.5
      - name: konveyor-operator.v0.3.0-alpha.4
        version: 0.3.0-alpha.4
      - name: konveyor-operator.v0.3.0-alpha.3
        version: 0.3.0-alpha.3
      - name: konveyor-operator.v0.3.0-alpha.2
        version: 0.3.0-alpha.2
      - name: konveyor-operator.v0.3.0-alpha.1
        version: 0.3.0-alpha.1
      - name: konveyor-operator.v0.2.0-alpha.3
        version: 0.2.0-alpha.3
      - name: konveyor-operator.v0.2.0-alpha.2
        version: 0.2.0-alpha.2
      - name: konveyor-operator.v0.2.0-alpha.1
        version: 0.2.0-alpha.1
      name: alpha
    - currentCSV: konveyor-operator.v0.8.0-beta.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
          createdAt: "2025-09-17T19:50:24Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0-beta.5'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:fed49c5ff14bb232e63d2dc180d3b530da63f11bf97055fc9d86cbb9e47104a4
        - quay.io/konveyor/postgresql-15-c9s@sha256:157a2941ae971d8f72b2f8af1d4a29deeb5548012a978fa986c1c6e52395da4d
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/konveyor/tackle2-ui@sha256:78398832e092557639531230cf6d023bb07b589f3587afeaacf8c68640abc454
        - quay.io/konveyor/tackle2-hub@sha256:1611913558c9f1203a844b43969d7b1539dab3578b83460e8383ee6757c3c34c
        - quay.io/konveyor/tackle2-operator@sha256:019a79dec40445b6289948e173e5bf2ebfa9b495a7bb077abd634c0667953a54
        - quay.io/konveyor/tackle-keycloak-init@sha256:0b4bfc1fea756f10a0657861c4c422303a2349d8d9b04a23238a32ceaf106ac5
        - quay.io/konveyor/java-external-provider@sha256:0cbdd91cc5057889b8daa2f7611fa0c944fab951c73bdf88a3e28795e9a5fc33
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/kai-solution-server@sha256:1e69cf8a342090a03a9dddb1b28ae3627052b6d636ffa290a11761b94fd10648
        - quay.io/konveyor/generic-external-provider@sha256:a6c550bad7a09a1c51e226247f11c00d8453987698832cb566582842219f1f3c
        - quay.io/konveyor/tackle2-addon-discovery@sha256:a29a0248e222d9ac3225e9d3df603f1383fa8844856bf68ea618d303f0b79883
        - quay.io/konveyor/kantra@sha256:ad6f5c05e5e950e33a745accd03ad933fb7bb4907f44455f88a2b5c27e6dd15e
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:759dc7d6836f8c8830ec145e3088bddc1b63bd40ff3047079fa8f0e986e74c30
        version: 0.8.0-beta.5
      entries:
      - name: konveyor-operator.v0.8.0-beta.5
        version: 0.8.0-beta.5
      - name: konveyor-operator.v0.8.0-beta.4
        version: 0.8.0-beta.4
      - name: konveyor-operator.v0.8.0-beta.2
        version: 0.8.0-beta.2
      - name: konveyor-operator.v0.8.0-beta.1
        version: 0.8.0-beta.1
      - name: konveyor-operator.v0.6.0-beta.2
        version: 0.6.0-beta.2
      - name: konveyor-operator.v0.6.0-beta.1
        version: 0.6.0-beta.1
      - name: konveyor-operator.v0.5.0-beta.3
        version: 0.5.0-beta.3
      - name: konveyor-operator.v0.5.0-beta.2
        version: 0.5.0-beta.2
      - name: konveyor-operator.v0.5.0-beta.1
        version: 0.5.0-beta.1
      - name: konveyor-operator.v0.3.0-beta.3
        version: 0.3.0-beta.3
      - name: konveyor-operator.v0.3.0-beta.2
        version: 0.3.0-beta.2
      - name: konveyor-operator.v0.3.0-beta.1.2
        version: 0.3.0-beta.1.2
      - name: konveyor-operator.v0.3.0-beta.1.1
        version: 0.3.0-beta.1.1
      - name: konveyor-operator.v0.3.0-beta.1
        version: 0.3.0-beta.1
      name: beta
    - currentCSV: konveyor-operator.v0.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
          createdAt: "2023-05-25T15:23:45Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.1.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/postgresql-12-centos7@sha256:1287d947464fbf5275feb31d106d2feb856c6117f0eaec8d947f2e03bd8618c5
        - quay.io/konveyor/tackle2-addon@sha256:35be344bbd3393f5634f52aafd21e9c3964d1e0ba5011ecad1561600d476bbe9
        - quay.io/konveyor/tackle-keycloak-init@sha256:b0347e5a608352c4b7bb796b65a67d417e98ba4896db71a487fa26243214cfcf
        - quay.io/openshift-community-operators/konveyor-operator@sha256:04e63481c5104bd034283d5d4b0567167f62480d2c61154b1d16dbd41c344dfb
        - quay.io/konveyor/tackle2-operator@sha256:9688d0763d8c7812beee715a061c3c198006448855d7b891d45833f62cf869fb
        - quay.io/konveyor/tackle2-ui@sha256:99650804e65e2b6cb15a37c00189d7f463e4d19a0f866a79e9c09c8420a4d028
        - quay.io/konveyor/tackle2-addon-windup@sha256:81f0625cb82eb24ded51c7258f229f57778820c62e3be6ec05fb2b12d2970825
        - quay.io/konveyor/tackle2-hub@sha256:3d15d158ffb6f1fc4ca5aaf29a12173fe344c289773902c9161056bdf69acc92
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-pathfinder@sha256:f9c5aa49f1d27f477cdd82c21441ae0287acdcdc9b2b3fcf197884007dc5f7dd
        version: 0.1.2
      entries:
      - name: konveyor-operator.v0.1.2
        version: 0.1.2
      - name: konveyor-operator.v0.1.1
        version: 0.1.1
      - name: konveyor-operator.v0.1.0
        version: 0.1.0
      name: konveyor-0.1
    - currentCSV: konveyor-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "move2kube-plan",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "image": "quay.io/konveyor/tackle-addon-move2kube-plan:latest"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
          createdAt: "2023-07-06T06:17:40Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.2.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
        description: |+
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.
          * Pathfinder, a service to manage the assessment questionnaires.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)

        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-hub@sha256:d9c4f21f44098a2526474968b6d88094dca8b5eaeb9042716332c9cb67df47cd
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/postgresql-12-centos7@sha256:6016530ab1fce71d29421ce9a97b3a89fbbb126b87b1b079c6aadebb7ff5664c
        - quay.io/konveyor/tackle-keycloak-init@sha256:ca6c85bdf8417d4be116c2098b8d8ad5ae8b7ab0b636c4cbe4a392aa424e2d47
        - quay.io/konveyor/tackle2-ui@sha256:d10c1e717d160c5eb2581649a8dd7cdf4922a74d92db014f72f42a48a3134341
        - quay.io/openshift-community-operators/konveyor-operator@sha256:d4c590033b1a2bcd67375465298e5cb6d136fe091abb12b8d009aa683c23ef62
        - quay.io/konveyor/tackle2-addon@sha256:11693e16363bcd98249209d63a29f0c6cfffab3a4071649cbeea2fd276bc2440
        - quay.io/konveyor/tackle-pathfinder@sha256:190269b061b2065ca8ba9357bf9b959a2f587a547a44769ae0afd1ac86b4b197
        - quay.io/konveyor/tackle2-addon-windup@sha256:538dfdb134d53a876d126dbd16679638584a606ef2b9e8fbe69ed1f29733c03a
        - quay.io/konveyor/tackle2-operator@sha256:1860bb099ac64d9a2af27692f0671b7240e4d19746fffa9f9ae2356ff2818e10
        version: 0.2.1
      entries:
      - name: konveyor-operator.v0.2.1
        version: 0.2.1
      - name: konveyor-operator.v0.2.0
        version: 0.2.0
      name: konveyor-0.2
    - currentCSV: konveyor-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
          createdAt: "2024-05-02T22:12:42Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.3.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:bbf58d2ea613489f91efae450421afafb9052416b0d7971606e2729e89fd7de6
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:848db711d9e8a329b689b40003293fce19c43faedd31afd80d9ae9d5c11cda49
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-hub@sha256:bf8803721d06d1a39251e22892343a013d5ef0fb4c1883c95e1eb55aa683f486
        - quay.io/konveyor/tackle2-operator@sha256:cd2305373f50ce927393e43f5735915e46c32d6f8a76091ee2808218885bba72
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/tackle-keycloak-init@sha256:a63702cc0cb36830a7830ed70ca344a9aacc201ba6f3c94806ca2e93160df4dd
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:51bda0c6577c15684583d607a4f0cebef760814102423f4b2025016093a19fc1
        version: 0.3.2
      entries:
      - name: konveyor-operator.v0.3.2
        version: 0.3.2
      - name: konveyor-operator.v0.3.1
        version: 0.3.1
      - name: konveyor-operator.v0.3.0
        version: 0.3.0
      name: konveyor-0.3
    - currentCSV: konveyor-operator.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
          createdAt: "2024-06-05T03:08:51Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          olm.skipRange: '>=0.0.0 <0.4.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-ui@sha256:cb1255fc5fa4608fc659c3455f3164664b960cc2d9a3b2208967798e3532df78
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:e29f175a799597c84f0663a66767bba24feb78f802516056b35b5beb14d9eb41
        - quay.io/konveyor/origin-oauth-proxy@sha256:05f6bd1216b921b1f47a3821f0e22b84ed6b141bb2ada474c12bd9fde613108c
        - quay.io/konveyor/postgresql-12-centos7@sha256:c65e48c23994a077f618075f06faa8e24a58f64f78e47902f7cf4174c43b4b1b
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:e84ea36d206a8c143794e3986940377e77c7618225cad594bdf9a4047d7823dc
        - quay.io/konveyor/tackle2-operator@sha256:496db8ca859608d04008ac36cf7bdc9cafd7214294d33a9c5646083c361ffcb6
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle2-hub@sha256:b87acd8e2ee4a7edc4369bc375f64e2b98bc479d09556439e2ee4d74b74970a9
        - quay.io/konveyor/tackle-keycloak-init@sha256:4c7e07b1fce05f7dca47edecc0ba0ce373cc38e48879cc7eed72510d2865a18e
        version: 0.4.0
      entries:
      - name: konveyor-operator.v0.4.0
        version: 0.4.0
      name: konveyor-0.4
    - currentCSV: konveyor-operator.v0.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
          createdAt: "2024-08-27T15:48:02Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.5.3'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:334a631538d31efc8d82adae4f52303336800fa3037b5055c72400918d81955c
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:2afbe8bea09904e084e026cd0a294fbe1c7b05d0e387f6633b09c5eaf4665e02
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle2-operator@sha256:474d739ed470701b628f1f9e971925ac7fbe0b543f3e36305d932b963b1225fd
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:2d58f12044b0951c87c851527240196db148acb9797e6ac49f9474a6e6039d71
        - quay.io/konveyor/postgresql-15-c9s@sha256:039ce764cadd03b4cfea954700139b8ceff415a9a0970baa7f2cf60f7f929174
        - quay.io/konveyor/tackle2-addon-discovery@sha256:461f960bb1d29e719fb472fe9be9cf3deca17d984687957f54546a09047b0d3c
        - quay.io/konveyor/tackle2-hub@sha256:faf91ca68e24979642a44fa013b1fb9edfee415eee835ccef9dce2eb418e2909
        - quay.io/konveyor/tackle2-ui@sha256:36f29fe01d3078f0e100b24d1ed126850c5a773f4b0ae922fd4c4487de1ac29d
        - quay.io/konveyor/java-external-provider@sha256:33fb5d24b998f01c481a0a03e3f1bf3f7a6c93f5b662885102a0c966026e3375
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/tackle-keycloak-init@sha256:5f652be5914ae6e23537b9fe02ccc9cd638c08de5462d9f58135a405c81f5287
        version: 0.5.3
      entries:
      - name: konveyor-operator.v0.5.3
        version: 0.5.3
      - name: konveyor-operator.v0.5.1
        version: 0.5.1
      - name: konveyor-operator.v0.5.0
        version: 0.5.0
      name: konveyor-0.5
    - currentCSV: konveyor-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
          createdAt: "2025-03-05T21:14:52Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:27c1731859a4a639d2750cff8d1246281f4b0466bb5b069169c3e170c3c19b30
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:7b8bdecbe3b3def1ec810ad53f604c7a71daebcfa7433cbb3a0db937580d087c
        - quay.io/konveyor/postgresql-15-c9s@sha256:55f8014c1edafab8f9085228a0d207c4e98820c615be34b935ce98e942d388b1
        - quay.io/konveyor/tackle-keycloak-init@sha256:a3741269aa15edd2f9fb7df853a03a942d16a88eec114f3e54f82ef63b82adc4
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle2-operator@sha256:8ee143c473570efd27b6f93332dc272871e11c66cc20f381c3a1e6e5688abf68
        - quay.io/konveyor/tackle2-ui@sha256:12b0eda13a63508b9c13d4f9309bbfbe19d564627cf0de672307af1a6ec993fb
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:0b5cbb322fbe96335522410376c2e8505a8302088c0e454f3e5e7e28ba7c96a6
        - quay.io/konveyor/tackle2-hub@sha256:52a8b9be038fbfbc4e22172a4a8c7191d2b06fd8670e5df2ec442e3ce00d8c8a
        - quay.io/konveyor/tackle2-addon-discovery@sha256:3eb6f395c451e2c9a867871b3984a25f9f8e6ddcfe388f37fbd6d46a61a70b1e
        - quay.io/konveyor/java-external-provider@sha256:15148cde9a6fedef2207fe36f0ac3bd1ecdc6e3ac76f1cbec9eb97ffcf247231
        version: 0.6.1
      entries:
      - name: konveyor-operator.v0.6.1
        version: 0.6.1
      - name: konveyor-operator.v0.6.0
        version: 0.6.0
      name: konveyor-0.6
    - currentCSV: konveyor-operator.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
          createdAt: "2025-07-10T15:16:17Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/generic-external-provider@sha256:facca8861559a658710ad3aef8b325731595bf944cce1907715d6efd33887072
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:7ed52ea9b95650241b5c2c01914775e29be23f7ed75fabb6e7c71969f3f0728f
        - quay.io/konveyor/tackle2-operator@sha256:f5870bc26b2a07ddacf00ed5b5f7df7d1068489010a80bcc543686c6c9576fbb
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-addon-discovery@sha256:60951801aa539ee59cd43f3c5f42a2ca751ca70e995cbb258b7d55ebb24d8450
        - quay.io/konveyor/tackle2-hub@sha256:52f66607f3134a4165ce691c6e9ce992d93927e9591055daafdc1bfc1dd09dd9
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:16c4f165143b8bab7f3eb08c815c2cae0dbf7fffd9ce7a625357be5806c24c5c
        - quay.io/konveyor/tackle2-ui@sha256:f87b04762fd163f40c49f3bcffd473f0c3f80a643576c6a3e2c2b4a14395c09f
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/origin-oauth-proxy@sha256:2c811da843057119b246784f21dc7eec490669a8569e8295d8e914cdce555b24
        - quay.io/konveyor/postgresql-15-c9s@sha256:41460103a0bab189cd77acb6e741d9c5a6d3506479f7e56df40eb156d9b2cadd
        - quay.io/konveyor/tackle-keycloak-init@sha256:dab4cb1229c8a729d14b4cecb4aa3b5c4d48e94c36e84d785c922fc08d120617
        - quay.io/konveyor/java-external-provider@sha256:a0ef733e9fce3845d02b5360440add85427acad697d2157ef92242dec06bd73b
        version: 0.7.1
      entries:
      - name: konveyor-operator.v0.7.1
        version: 0.7.1
      - name: konveyor-operator.v0.7.0
        version: 0.7.0
      name: konveyor-0.7
    - currentCSV: konveyor-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Schema",
                "metadata": {
                  "name": "myschema",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "domain": "mydomain",
                  "subject": "mysubject",
                  "variant": "myvariant",
                  "versions": [
                    {
                      "definition": {
                        "$schema": "https://json-schema.org/draft/2020-12/schema",
                        "properties": {
                          "names": {
                            "description": "Application names. Each may be a glob expression.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 0,
                            "type": "array"
                          },
                          "spaces": {
                            "description": "Space names.",
                            "items": {
                              "minLength": 1,
                              "type": "string"
                            },
                            "minItems": 1,
                            "type": "array"
                          }
                        },
                        "required": [
                          "spaces"
                        ],
                        "title": "Coordinates",
                        "type": "object"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
          createdAt: "2025-09-29T20:41:09Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.8.0'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - kind: Schema
            name: schemas.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:8c82ad880ca2a2d5cd4b08125feca16d13d014d15f4f77677306086d1067f9aa
        - quay.io/konveyor/java-external-provider@sha256:8eab08a04e98051f89352250d63523a5b591f35a6497f5d69a5a9d556a94abf9
        - quay.io/konveyor/origin-oauth-proxy@sha256:8988f14db7dde3edcaed0541cecb0b4fd58fab9dbaffb79a6c1c4aee13b9bd73
        - quay.io/konveyor/tackle2-hub@sha256:b3e6db1ffc195143b8e5d54b60949256e154b0f8b667736512a084e015cdf2a7
        - quay.io/konveyor/generic-external-provider@sha256:c7b6618bfff81d73f1b455963107ffdaf19334e6552ce076c009476da3f37f21
        - quay.io/konveyor/kantra@sha256:a5b5591a51143574a6ce6678b141b9cd3e14328c6fad773412dde21b442fe364
        - quay.io/konveyor/tackle2-addon-discovery@sha256:96a5b37834fbed9506014bcbd909258081fd833ef9e16daa104effd0e3a3c6ea
        - quay.io/konveyor/postgresql-15-c9s@sha256:af275f35a307367dcee9a339a6903cf79e3a8a41cabf79fb7cea7015895f0b58
        - quay.io/konveyor/tackle-keycloak-init@sha256:c0051b8c3a3c891fe12012623aee775080ecb47af8a4e75b8be8b0b4855d9f2c
        - quay.io/konveyor/tackle2-ui@sha256:f3f7380be65ff1fb3459c37a3d4865354bab685dc836893f78221f16ba579f70
        - quay.io/konveyor/tackle2-addon-platform@sha256:d0280ade44ba5e15210c07342608576b38e6993af2e479740675a4706cd2c875
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:c28522543b039454d2d97b48198d099ac0b8211cbc58f13468d5f70b9f69b037
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle2-operator@sha256:68b6672fccdf54c65cc83f4c82db023053d6b61409bea880f58b53ccfea7b1ed
        - quay.io/konveyor/kai-solution-server@sha256:bab6e224c50de5b44cc37ce0297a3b91255efc08c1ad0fccf14fef9e5182775f
        version: 0.8.0
      entries:
      - name: konveyor-operator.v0.8.0
        version: 0.8.0
      name: konveyor-0.8
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      - name: konveyor-operator.v0.6.0-rc.2
        version: 0.6.0-rc.2
      - name: konveyor-operator.v0.6.0-rc.1
        version: 0.6.0-rc.1
      - name: konveyor-operator.v0.5.0-rc.1
        version: 0.5.0-rc.1
      - name: konveyor-operator.v0.3.0-rc.4
        version: 0.3.0-rc.4
      - name: konveyor-operator.v0.3.0-rc.3
        version: 0.3.0-rc.3
      - name: konveyor-operator.v0.3.0-rc.2
        version: 0.3.0-rc.2
      - name: konveyor-operator.v0.3.0-rc.1
        version: 0.3.0-rc.1
      name: rc
    - currentCSV: konveyor-operator.v0.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
          createdAt: "2025-03-26T20:25:41Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.6.2'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/konveyor/tackle2-operator@sha256:33cf01f19a7043bcc5bdb12a3ed6bde7a9094981aa2383554111ebec9b4ba498
        - quay.io/konveyor/tackle2-ui@sha256:3c568c6047a044095080b6013b643a4041deae7b45936a443d4c4d708a81b945
        - quay.io/konveyor/generic-external-provider@sha256:bc1112352f1a2c3172c5e7d5e637b5c5a69bf7d98b78f494fa593e8872e5cfd1
        - quay.io/konveyor/postgresql-15-c9s@sha256:7957d3b272b9f4e469bdc0c514ec9e70e8fbdc34a7250b4f4c62f9bb73cbf08d
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:18ba5c587dc613f9c1a8bc9e66081e52f3a37eedeea5e75881e628e9c35ef460
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:734752be111a97e4b7e0030d563c32883f419ad031b2f350cb8e7333f2f37de3
        - quay.io/konveyor/tackle2-hub@sha256:0df49087d0c9341c2120a7c3aa2ba7907c056f94e1602ffd0496ca7113c75382
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:9342435929d8ed447864694383a7fb50d91f0f4913b5f85f4a43ba8bc36f9b82
        - quay.io/konveyor/origin-oauth-proxy@sha256:1db28585d10c8ccc65a38b5d4977e6566661411f51d44ec1fb8b8a562087d86c
        - quay.io/konveyor/tackle-keycloak-init@sha256:e32fa48351f410361e8bc100303deeb40d8a66d7eefb0cf7b1bc16e067a9a795
        - quay.io/konveyor/tackle2-addon-discovery@sha256:1b89266d6bdfa497ba8137e0d33ff53785ac476dd96bf4fd9a185ef0ddf513ce
        - quay.io/konveyor/java-external-provider@sha256:fe85689b57125d678df685012b93c8a567fcb9251777cd711e9b82bdb22a734c
        version: 0.6.2
      entries:
      - name: konveyor-operator.v0.6.2
        version: 0.6.2
      name: release-0.6
    - currentCSV: konveyor-operator.v0.7.0-rc.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Addon",
                "metadata": {
                  "name": "myaddon",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "container": {
                    "image": "quay.io/konveyor/tackle2-addon-analyzer:latest",
                    "imagePullPolicy": "Always",
                    "name": "myaddon",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "512Mi"
                      }
                    }
                  },
                  "task": "mytask"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Extension",
                "metadata": {
                  "name": "myextension",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "addon": "myaddon",
                  "container": {
                    "args": [
                      "--port",
                      "$(PORT)"
                    ],
                    "env": [
                      {
                        "name": "PORT",
                        "value": "${seq:8000}"
                      }
                    ],
                    "image": "quay.io/konveyor/generic-external-provider:latest",
                    "imagePullPolicy": "Always",
                    "name": "myextension",
                    "resources": {
                      "limits": {
                        "cpu": 1,
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": 1,
                        "memory": "1Gi"
                      }
                    }
                  },
                  "metadata": {
                    "provider": {
                      "address": "localhost:$(PORT)",
                      "name": "myextension"
                    }
                  },
                  "selector": "tag:Language=MyLang"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Tackle",
                "metadata": {
                  "name": "tackle",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "feature_auth_required": "false"
                }
              },
              {
                "apiVersion": "tackle.konveyor.io/v1alpha1",
                "kind": "Task",
                "metadata": {
                  "name": "mytask",
                  "namespace": "konveyor-tackle"
                },
                "spec": {
                  "dependencies": [
                    "myaddon"
                  ],
                  "priority": 10
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
          createdAt: "2025-05-21T17:12:48Z"
          description: Konveyor is an open-source application modernization platform
            that helps organizations safely and predictably modernize applications
            to Kubernetes at scale.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <0.7.0-rc.1'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "tackle.konveyor.io/v1alpha1",
              "kind": "Tackle",
              "metadata": {
                "name": "tackle",
                "namespace": "konveyor-tackle"
              },
              "spec": {
                "feature_auth_required": "false"
              }
            }
          operatorframework.io/suggested-namespace: konveyor-tackle
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/konveyor/tackle2-operator
          support: https://github.com/konveyor/tackle2-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tackle Addon
            displayName: Addon
            kind: Addon
            name: addons.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Extension
            displayName: Extension
            kind: Extension
            name: extensions.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle
            displayName: Tackle
            kind: Tackle
            name: tackles.tackle.konveyor.io
            version: v1alpha1
          - description: Tackle Task
            displayName: Task
            kind: Task
            name: tasks.tackle.konveyor.io
            version: v1alpha1
        description: |
          Konveyor is an open-source application modernization platform that helps organizations safely and predictably modernize applications to new technologies, with an initial focus on accelerating the adoption of legacy applications to Kubernetes.

          See the [Konveyor Unified Experience](https://github.com/konveyor/enhancements/tree/master/enhancements/unified_experience) to understand the vision of the project and the [Konveyor Charter](https://github.com/konveyor/community/blob/main/Charter.md) for more information on the community.

          ### Install

          Once you have successfully installed the Operator, proceed to deploy components by creating the required Tackle CR.

          By default, the Operator installs the following components on a target cluster:

          * Hub, to manage the application inventory and coordinate the migration process.
          * UI, the web console to manage the application inventory and drive the migration waves.

          Authentication capabilities may be enabled via the `feature_auth_required` parameter in the Tackle CR. When enabled, the Operator installs the following components:
          * Keycloak, to manage authentication, including with 3rd-party providers.

          ### Examples
          If you are just getting started using Konveyor, take a look at an [example application analysis here](https://github.com/konveyor/example-applications/tree/main/example-1)

          ### Documentation
          Documentation can be found on our [website](https://konveyor.github.io/tackle).

          ### Getting help
          If you encounter any issues while using Konveyor operator, you can create an issue on our [Github repo](https://github.com/konveyor/tackle2-operator/issues), for bugs, enhancements or other requests.
          Or reach out to us in [#konveyor](https://kubernetes.slack.com/archives/CR85S82A2) on kubernetes.slack.com

          ### Contributing
          You can contribute by:

          * Raising any issues you find using Konveyor Operator
          * Fixing issues by opening [Pull Requests](https://github.com/konveyor/tackle2-operator/pulls)
          * Improving [documentation](https://github.com/konveyor/tackle-documentation)
        displayName: Konveyor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - modernization
        - migration
        - konveyor
        - tackle
        links:
        - name: Documentation
          url: https://konveyor.github.io/tackle
        - name: Operator
          url: https://github.com/konveyor/tackle2-operator
        maintainers:
        - email: konveyor-dev@googlegroups.com
          name: Konveyor Community
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Konveyor
          url: https://www.konveyor.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/konveyor-operator@sha256:fbf82685700aaea953a99e98e0c28b56e54ca6972e390f44c07957aa8df29442
        - quay.io/konveyor/generic-external-provider@sha256:de55f78939576c110eb6f084a7ca51338fedba699fbb5135f23392783a5d7b95
        - quay.io/konveyor/tackle2-hub@sha256:bed834eb5488eef23b0d1e8af194b0646cf74c39c197e0d57acc5aed2299d9f5
        - quay.io/konveyor/tackle2-operator@sha256:87cdac762506c121f84f8f872d45067020943c0a1c9de70fba71ad39a99c1c43
        - quay.io/konveyor/origin-oauth-proxy@sha256:6c2847fd9667938b1d1294d988038e885e43cd9fb378862dae3419ce4f21631f
        - quay.io/konveyor/tackle2-addon-analyzer@sha256:46150725a2adbde8dfdc2250614d437a50e64bbf0b4562dabc73715a287de67b
        - quay.io/konveyor/tackle2-ui@sha256:afe9141366a23c072286517e9af5436d14fd1df775aaa2fa13c3911764515777
        - quay.io/konveyor/postgresql-15-c9s@sha256:012cd6611aa9d3f4e0a82636d6817f381c19e3b0809199216fbe7fa3396a2ef5
        - quay.io/konveyor/tackle2-addon-discovery@sha256:89d2b0f8636897eb4dc1957a74a5b5041c0f3267a747adcf2bc6ca56b17b9937
        - quay.io/konveyor/kai@sha256:94bc7683545b58b74f1d8f75c82d9fcba301ca77cdbc5e7d9dcb808bf0fff5e0
        - quay.io/konveyor/keycloak@sha256:dc23dd938a5dc5dfaabc71e44d014928284fafff9122e0395659f1cdcf534da8
        - quay.io/konveyor/tackle-keycloak-init@sha256:c3124971dc31be9201de2a6a1f6b2f886adf4c96731188dd5e750798cde3cfe7
        - quay.io/konveyor/java-external-provider@sha256:61ab20e37916abfeeadde685336660f0e8bbf519403d1f7100d43c1e42a72fbd
        version: 0.7.0-rc.1
      entries:
      - name: konveyor-operator.v0.7.0-rc.1
        version: 0.7.0-rc.1
      name: release-0.7
    defaultChannel: konveyor-0.8
    packageName: konveyor-operator
    provider:
      name: Konveyor
      url: https://www.konveyor.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: https://github.com/openshift/kueue-operator
    name: kueue-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kueue-operator.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "policy": "Disabled"
                    },
                    "integrations": {
                      "frameworks": [
                        "AppWrapper"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "FairSharing"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameOptional"
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob",
                        "Pod",
                        "Deployment",
                        "StatefulSet"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              },
              {
                "apiVersion": "operator.openshift.io/v1alpha1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "gangScheduling": {
                      "byWorkload": "Parallel",
                      "policy": "ByWorkload"
                    },
                    "integrations": {
                      "frameworks": [
                        "RayJob",
                        "RayCluster",
                        "AppWrapper",
                        "PyTorchJob"
                      ]
                    },
                    "preemption": {
                      "preemptionPolicy": "Classical"
                    },
                    "workloadManagement": {
                      "labelPolicy": "QueueNameRequired"
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-05-01T20:58:29Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.operator.openshift.io
            version: v1alpha1
        description: Kueue Operator description. TODO.
        displayName: Kueue Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:de644d640f355815f3aa40ced5456bb5b0e958aeba8bfbdc4ab360fcdb90f26d
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:0c4d9cd97f7579adbf2afd238c953ca5f170408cf27e3b2bf4aa16c108a76881
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:dd765c446564d4397570f4bc056b55989f6c0bae0d70514874e67765c9c77889
        version: 0.1.0
      entries:
      - name: kueue-operator.v0.1.0
        version: 0.1.0
      name: stable-v0.1
    - currentCSV: kueue-operator.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-06-24T02:48:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9-operator@sha256:30659b42f5053dd756190211454c01f049bbc7689dddda3ecd8c406b157873a7
        - registry.redhat.io/kueue-tech-preview/kueue-rhel9@sha256:9a0b85b0752135694193befe00d10f7abdc0c99c6b9f295d59b237fa0fe3e6d2
        - registry.redhat.io/kueue-tech-preview/kueue-must-gather-rhel9@sha256:e5306c95bf6c5f55be6e791fcb66fe9b34f4f982536f4f90f8ed74acf3f6c31d
        - registry.redhat.io/kueue-tech-preview/kueue-operator-bundle@sha256:6f0b5f62f6e549c6f6769dd9a358a14668d7244f9c20753d1690d3e09948588c
        version: 0.2.1
      entries:
      - name: kueue-operator.v0.2.1
        version: 0.2.1
      - name: kueue-operator.v0.2.0
        version: 0.2.0
      name: stable-v0.2
    - currentCSV: kueue-operator.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kueue.openshift.io/v1",
                "kind": "Kueue",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "kueue-operator"
                  },
                  "name": "cluster",
                  "namespace": "openshift-kueue-operator"
                },
                "spec": {
                  "config": {
                    "integrations": {
                      "frameworks": [
                        "BatchJob"
                      ]
                    }
                  },
                  "managementState": "Managed"
                }
              }
            ]
          capabilities: Basic Install
          console.openshift.io/operator-monitoring-default: "true"
          createdAt: "2025-07-15T01:52:59Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kueue-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Kueue
            name: kueues.kueue.openshift.io
            version: v1
        description: Red Hat build of Kueue is a collection of APIs, based on the
          [Kueue](https://kueue.sigs.k8s.io/docs/) open source project, that extends
          Kubernetes to manage access to resources for jobs. Red Hat build of Kueue
          does not replace any existing components in a Kubernetes cluster, but instead
          integrates with the existing Kubernetes API server, scheduler, and cluster
          autoscaler components to determine when a job waits, is admitted to start
          by creating pods, or should be preempted.
        displayName: Red Hat build of Kueue
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kueue-operator
        links:
        - name: Kueue Operator
          url: https://github.com/openshift/kueue-operator
        maintainers:
        - email: aos-node@redhat.com
          name: Node team
        maturity: alpha
        minKubeVersion: 1.28.0
        provider:
          name: Red Hat, Inc
          url: https://github.com/openshift/kueue-operator
        relatedImages:
        - registry.redhat.io/kueue/kueue-rhel9-operator@sha256:5db0af8088cce6e9e7a88faf80855b6eaa1dd8679c6a56dd5ee98c50a3a8e2e4
        - registry.redhat.io/kueue/kueue-rhel9@sha256:16007bd38c480b607b63c82b3ee8721441cc374ee31c04b9e329426314590195
        - registry.redhat.io/kueue/kueue-must-gather-rhel9@sha256:f5f624ef8dbb3676ed4e37335e226e76e5f886c1990c6e0e8c0e4762740ed98d
        - registry.redhat.io/kueue/kueue-operator-bundle@sha256:1bafac48c00fb47972f512ccd5d846efbbccccd2259395de17537b39834d89df
        version: 1.0.1
      entries:
      - name: kueue-operator.v1.0.1
        version: 1.0.1
      - name: kueue-operator.v1.0.0
        version: 1.0.0
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: kueue-operator
    provider:
      name: Red Hat, Inc
      url: https://github.com/openshift/kueue-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-bundle-certified@sha256:a9b8102ab0ec81e5bea84332ef20b023d06660b410e7e2deadb00904d2098613
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: seldon
      provider-url: ""
    name: seldon-deploy-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: seldon-deploy-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1alpha1",
                "kind": "SeldonDeploy",
                "metadata": {
                  "name": "seldondeploy-sample"
                },
                "spec": {
                  "affinity": {},
                  "alibidetect": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306"
                  },
                  "batchjobs": {
                    "mc": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48"
                    },
                    "processor": {
                      "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d"
                    },
                    "pvc": {
                      "defaultSize": "1Gi"
                    },
                    "serviceAccount": "workflow"
                  },
                  "defaultUserID": "",
                  "docker": {
                    "user": "unknown"
                  },
                  "elasticsearch": {
                    "basicAuth": true,
                    "secret": {
                      "name": "elastic-credentials",
                      "passwordKey": "password",
                      "userKey": "username"
                    },
                    "url": "https://elasticsearch-seldon-es-http.seldon-logs:9200"
                  },
                  "enableAppAnalytics": false,
                  "enableAppAuth": false,
                  "env": {
                    "USERID_CLAIM_KEY": "name"
                  },
                  "external": {
                    "protocol": "http"
                  },
                  "fullnameOverride": "",
                  "gitops": {
                    "argocd": {
                      "enabled": false,
                      "namespace": "argocd"
                    },
                    "fileFormat": "json",
                    "git": {
                      "email": "",
                      "secret": "git-creds",
                      "skipVerifyGit": true,
                      "token": "",
                      "user": "",
                      "webhook": {
                        "service": {
                          "create": true,
                          "loadBalancerSourceRanges": {}
                        }
                      }
                    }
                  },
                  "image": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e",
                    "pullPolicy": "Always"
                  },
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [
                      {
                        "host": "chart-example.local",
                        "paths": []
                      }
                    ],
                    "tls": []
                  },
                  "ingressGateway": {
                    "ingressNamespace": "istio-system",
                    "kfServingIngressService": "istio-ingressgateway",
                    "seldonIngressService": "istio-ingressgateway"
                  },
                  "kfserving": {
                    "curlForm": "MODEL_NAME={{ .ModelName }}\u003cbr\u003e\nCLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\nSERVICE_HOSTNAME=$(kubectl get inferenceservice {{ .ModelName }} -o jsonpath='{.status.url}' | cut -d \"/\" -f 3)\u003cbr\u003e\ncurl -v -H \"Host: ${SERVICE_HOSTNAME}\" {{ .KfServingProtocol }}://$CLUSTER_IP/v1/models/$MODEL_NAME:predict -d '{{ .Payload }}'\n",
                    "enabled": false,
                    "explainForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:explain",
                    "protocol": "http",
                    "requestForm": "{{ .KfServingProtocol }}://{{ .IngressServiceName }}/v1/models/{{ .ModelName }}:predict"
                  },
                  "loadtest": {
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c"
                  },
                  "metadata": {
                    "pg": {
                      "enabled": false,
                      "secret": "metadata-postgres"
                    }
                  },
                  "nameOverride": "",
                  "nodeSelector": {},
                  "openshiftMarketplace": {
                    "cleanupClusterServiceVersions": false,
                    "kubectlCleanupImage": "registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382",
                    "prometheus": {
                      "monitorSpecs": {
                        "create": true
                      }
                    },
                    "seldonCore": {
                      "istioGateway": {
                        "create": true,
                        "name": "seldon-gateway",
                        "namespace": "istio-system"
                      },
                      "subscription": {
                        "apiVersion": "operators.coreos.com/v1alpha1",
                        "channel": "stable",
                        "create": true,
                        "istioEnabled": true,
                        "metricsPath": "/metrics",
                        "requestLoggerEndpoint": "http://broker-ingress.knative-eventing.svc.cluster.local/seldon-logs/default"
                      }
                    }
                  },
                  "prometheus": {
                    "knative": {
                      "url": "http://prometheus-system-np.knative-monitoring.svc.cluster.local:8080/api/v1/"
                    },
                    "seldon": {
                      "jwtSecretKey": "jwt-seldon.txt",
                      "jwtSecretName": "jwt-seldon",
                      "namespaceMetricName": "namespace",
                      "resourceMetricsUrl": "https://prometheus-k8s.openshift-monitoring:9091/api/v1/",
                      "serviceMetricName": "exported_service",
                      "url": "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/"
                    }
                  },
                  "rbac": {
                    "clusterWide": true,
                    "create": true,
                    "readNamespaces": true
                  },
                  "replicaCount": 1,
                  "requestLogger": {
                    "create": true,
                    "elasticsearch": {
                      "host": "elasticsearch-seldon-es-http.seldon-logs",
                      "port": "9200",
                      "protocol": "https"
                    },
                    "image": "registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398",
                    "imagePullPolicy": "IfNotPresent",
                    "namespace": {
                      "create": false,
                      "name": "seldon-logs"
                    },
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "600m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "200Mi"
                      }
                    },
                    "trigger": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "broker": "default"
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "800m",
                      "memory": "800Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "200Mi"
                    }
                  },
                  "seldon": {
                    "curlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions -d '{{ .Payload }}'\n",
                    "enabled": true,
                    "explainForm": "http://{{ .ModelName }}-{{ .Predictor }}-explainer.{{ .Namespace }}:9000/v1/models/{{ .ModelName }}:explain",
                    "kfservingV2CurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .GraphModelName }}/infer -d '{{ .Payload }}'\n",
                    "protocol": "http",
                    "seldonRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/api/v0.1/predictions",
                    "tensorFlowCurlForm": "CLUSTER_IP=$(oc get route -n {{ .IngressNamespace }} {{ .IngressServiceName }} -o jsonpath='{.spec.host}')\u003cbr\u003e\ncurl -k -H \"Content-Type: application/json\" {{ .SeldonProtocol }}://$CLUSTER_IP/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict -d '{{ .Payload }}'\n",
                    "tensorflowRequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v1/models/:predict",
                    "v2RequestForm": "{{ .SeldonProtocol }}://{{ .IngressServiceName }}/seldon/{{ .Namespace }}/{{ .ModelName }}/v2/models/{{ .ModelName }}/infer"
                  },
                  "service": {
                    "port": 80,
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "create": true
                  },
                  "serviceAccountName": "seldon-deploy",
                  "skipVerifyHttpCalls": true,
                  "tolerations": [],
                  "virtualService": {
                    "create": true,
                    "gateways": [
                      "istio-system/seldon-gateway"
                    ],
                    "prefix": "/seldon-deploy/"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
          createdAt: "2020-12-11 15:00:00"
          description: The Seldon deploy operator allows the creation of Seldon Deploy
            applications to provide management, monitoring and updating of machine
            learning models from a range of technologies. It includes the ability
            to run explanations as well as outlier, drift and adversarial detection
            against running models. There is a full metrics and logging integration
            via Prometheus and ELK stacks as well as full GitOps operations to allow
            full auditing and recovery of updated models.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/seldon-deploy-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.7.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/SeldonIO/seldon-deploy-operator
          support: Seldon Dev
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: SeldonDeploy
            name: seldondeploys.machinelearning.seldon.io
            version: v1alpha1
        description: The Seldon deploy operator allows the creation of Seldon Deploy
          applications to provide management, monitoring and updating of machine learning
          models from a range of technologies. It includes the ability to run explanations
          as well as outlier, drift and adversarial detection against running models.
          There is a full metrics and logging integration via Prometheus and ELK stacks
          as well as full GitOps operations to allow full auditing and recovery of
          updated models.
        displayName: Seldon Deploy Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - machine learning
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://deploy.seldon.io
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        provider:
          name: seldon
        relatedImages:
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-request-logger@sha256:7df6b8f60441f1855fc223c919527dba7c72885206b4d72645e08bf998aab398
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-loadtester@sha256:fd71904fa9f5f7c470fe12bfb4ebed5cdd02d6c984b26db6a228611c43eed87c
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy-operator@sha256:8d36acf08b6f8905749a0a876c152cea43cd3f3e34b445901c095c9452ef0e05
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-deploy@sha256:88d1450309588b6c1692811f195791131aa6cbc54d89987d879fa31d4cb5650e
        - registry.connect.redhat.com/seldonio/seldon-deploy-operator-bundle@sha256:f3098b04f92d7272de172e91315b79a1b32eeff60227a3c70a2f62361be6eee8
        - registry.marketplace.redhat.com/rhm/seldonio/alibi-detect-server@sha256:4b0edf72477f54bdcb850457582f12bcb1338ca64dc94ebca056897402708306
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - registry.marketplace.redhat.com/rhm/seldonio/seldon-batch-processor@sha256:a07c0e2ec60dc91316ed40c3c5549d0efce9f8bc47caf626621ab471a6ee450d
        - registry.marketplace.redhat.com/rhm/seldonio/mc-ubi@sha256:5947afb33ef0b708ea0952032794c417b1fca8e3b80f86f7bf7bde4351459a48
        - registry.marketplace.redhat.com/rhm/seldonio/kubectl@sha256:15075893422caf0a42805f2ff53d0996471f3ececf09632caaf39402da41f382
        version: 1.2.1
      entries:
      - name: seldon-deploy-operator.v1.2.1
        version: 1.2.1
      - name: seldon-deploy-operator.v1.2.0
        version: 1.2.0
      - name: seldon-deploy-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: seldon-deploy-operator-rhmp
    provider:
      name: seldon
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenTelemetry Community
      provider-url: ""
    name: opentelemetry-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.136.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          containerImage: ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator
          createdAt: "2025-09-29T13:00:19Z"
          description: Provides the OpenTelemetry components, including the Collector
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/open-telemetry/opentelemetry-operator
          support: OpenTelemetry Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.
        displayName: Community OpenTelemetry Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: jpkroehling@redhat.com
          name: Juraci Paixão Kröhling
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: OpenTelemetry Community
        relatedImages:
        - quay.io/community-operator-pipeline-prod/opentelemetry-operator@sha256:5dc347a5085c10823a122fab0bd42afd5d9b9aa19ba17e8992e9492000886629
        - ghcr.io/open-telemetry/opentelemetry-operator/opentelemetry-operator:0.136.0
        - quay.io/brancz/kube-rbac-proxy:v0.13.1
        version: 0.136.0
      entries:
      - name: opentelemetry-operator.v0.136.0
        version: 0.136.0
      - name: opentelemetry-operator.v0.135.0
        version: 0.135.0
      - name: opentelemetry-operator.v0.134.0
        version: 0.134.0
      - name: opentelemetry-operator.v0.132.0
        version: 0.132.0
      - name: opentelemetry-operator.v0.131.0
        version: 0.131.0
      - name: opentelemetry-operator.v0.129.1
        version: 0.129.1
      - name: opentelemetry-operator.v0.127.0
        version: 0.127.0
      - name: opentelemetry-operator.v0.126.0
        version: 0.126.0
      - name: opentelemetry-operator.v0.125.0
        version: 0.125.0
      - name: opentelemetry-operator.v0.124.0
        version: 0.124.0
      - name: opentelemetry-operator.v0.123.0
        version: 0.123.0
      - name: opentelemetry-operator.v0.122.0
        version: 0.122.0
      - name: opentelemetry-operator.v0.121.0
        version: 0.121.0
      - name: opentelemetry-operator.v0.120.0
        version: 0.120.0
      - name: opentelemetry-operator.v0.119.0
        version: 0.119.0
      - name: opentelemetry-operator.v0.118.0
        version: 0.118.0
      - name: opentelemetry-operator.v0.117.0
        version: 0.117.0
      - name: opentelemetry-operator.v0.116.0
        version: 0.116.0
      - name: opentelemetry-operator.v0.115.0
        version: 0.115.0
      - name: opentelemetry-operator.v0.114.1
        version: 0.114.1
      - name: opentelemetry-operator.v0.112.0
        version: 0.112.0
      - name: opentelemetry-operator.v0.111.0
        version: 0.111.0
      - name: opentelemetry-operator.v0.110.0
        version: 0.110.0
      - name: opentelemetry-operator.v0.109.0
        version: 0.109.0
      - name: opentelemetry-operator.v0.108.0
        version: 0.108.0
      - name: opentelemetry-operator.v0.107.0
        version: 0.107.0
      - name: opentelemetry-operator.v0.106.0
        version: 0.106.0
      - name: opentelemetry-operator.v0.105.0
        version: 0.105.0
      name: alpha
    defaultChannel: alpha
    packageName: opentelemetry-operator
    provider:
      name: OpenTelemetry Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Aerospike
      provider-url: ""
    name: aerospike-kubernetes-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aerospike-kubernetes-operator.v2.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "memory-size": 3000000000,
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:6.3.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
          createdAt: 2023-05-04T09:32:47UTC
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          operators.operatorframework.io/builder: operator-sdk-v1.10.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:2.5.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/openshift-community-operators/aerospike-kubernetes-operator@sha256:bfa83ff729ef8b3a4ebca27b02e223327d3e1351d8bc78a06c18873d288dfb02
        version: 2.5.0
      entries:
      - name: aerospike-kubernetes-operator.v2.5.0
        version: 2.5.0
      - name: aerospike-kubernetes-operator.v2.4.0
        version: 2.4.0
      - name: aerospike-kubernetes-operator.v2.3.0
        version: 2.3.0
      - name: aerospike-kubernetes-operator.v2.2.1
        version: 2.2.1
      - name: aerospike-kubernetes-operator.v2.2.0
        version: 2.2.0
      - name: aerospike-kubernetes-operator.v2.1.0
        version: 2.1.0
      name: alpha
    - currentCSV: aerospike-kubernetes-operator.v4.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "asdb.aerospike.com/v1",
                "kind": "AerospikeCluster",
                "metadata": {
                  "name": "aerocluster",
                  "namespace": "aerospike"
                },
                "spec": {
                  "aerospikeConfig": {
                    "namespaces": [
                      {
                        "name": "test",
                        "replication-factor": 2,
                        "storage-engine": {
                          "data-size": 1073741824,
                          "type": "memory"
                        }
                      }
                    ],
                    "network": {
                      "fabric": {
                        "port": 3001
                      },
                      "heartbeat": {
                        "port": 3002
                      },
                      "service": {
                        "port": 3000
                      }
                    },
                    "service": {
                      "feature-key-file": "/etc/aerospike/secret/features.conf"
                    }
                  },
                  "image": "aerospike/aerospike-server-enterprise:8.1.0.0",
                  "podSpec": {
                    "multiPodPerHost": true
                  },
                  "size": 2,
                  "storage": {
                    "volumes": [
                      {
                        "aerospike": {
                          "path": "/etc/aerospike/secret"
                        },
                        "name": "aerospike-config-secret",
                        "source": {
                          "secret": {
                            "secretName": "aerospike-secret"
                          }
                        }
                      }
                    ]
                  },
                  "validationPolicy": {
                    "skipWorkDirValidate": true,
                    "skipXdrDlogFileValidate": true
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackup",
                "metadata": {
                  "name": "aerospikebackup",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "aerospike-cluster": {
                      "aerospike-aerospikebackup-test-cluster": {
                        "credentials": {
                          "password": "admin123",
                          "user": "admin"
                        },
                        "seed-nodes": [
                          {
                            "host-name": "aerocluster.aerospike.svc.cluster.local",
                            "port": 3000
                          }
                        ]
                      }
                    },
                    "backup-routines": {
                      "aerospike-aerospikebackup-test-routine": {
                        "backup-policy": "test-policy",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "local"
                      },
                      "aerospike-aerospikebackup-test-routine1": {
                        "backup-policy": "test-policy1",
                        "incr-interval-cron": "@hourly",
                        "interval-cron": "@daily",
                        "namespaces": [
                          "test"
                        ],
                        "source-cluster": "aerospike-aerospikebackup-test-cluster",
                        "storage": "s3Storage"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeBackupService",
                "metadata": {
                  "name": "aerospikebackupservice-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "config": {
                    "backup-policies": {
                      "test-policy": {
                        "parallel": 3
                      },
                      "test-policy1": {
                        "parallel": 3
                      }
                    },
                    "service": {
                      "http": {
                        "port": 8081
                      }
                    },
                    "storage": {
                      "local": {
                        "local-storage": {
                          "path": "/tmp/localStorage"
                        }
                      },
                      "s3Storage": {
                        "s3-storage": {
                          "bucket": "aerospike-kubernetes-operator-test",
                          "s3-profile": "default",
                          "s3-region": "us-east-1"
                        }
                      }
                    }
                  },
                  "image": "aerospike/aerospike-backup-service:3.2.0",
                  "secrets": [
                    {
                      "secretName": "aws-secret",
                      "volumeMount": {
                        "mountPath": "/root/.aws/credentials",
                        "name": "aws-secret",
                        "subPath": "credentials"
                      }
                    }
                  ],
                  "service": {
                    "type": "LoadBalancer"
                  }
                }
              },
              {
                "apiVersion": "asdb.aerospike.com/v1beta1",
                "kind": "AerospikeRestore",
                "metadata": {
                  "name": "aerospikerestore-sample",
                  "namespace": "aerospike"
                },
                "spec": {
                  "backupService": {
                    "name": "aerospikebackupservice-sample",
                    "namespace": "aerospike"
                  },
                  "config": {
                    "backup-data-path": "aerospike-aerospikebackup-test-routine/backup/1733506015827/data/test",
                    "destination": {
                      "credentials": {
                        "password": "admin123",
                        "user": "admin"
                      },
                      "label": "destinationCluster",
                      "seed-nodes": [
                        {
                          "host-name": "aerocluster.test.svc.cluster.local",
                          "port": 3000
                        }
                      ]
                    },
                    "policy": {
                      "no-generation": true,
                      "no-indexes": true,
                      "parallel": 3
                    },
                    "source": {
                      "local-storage": {
                        "path": "/tmp/localStorage"
                      }
                    }
                  },
                  "type": "Full"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
          createdAt: "2025-09-15T10:14:05Z"
          description: The Aerospike Kubernetes Operator automates the deployment
            and management of Aerospike enterprise clusters on Kubernetes.
          olm.skipRange: '>=4.0.0 <4.1.1'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/aerospike/aerospike-kubernetes-operator/
          support: Aerospike
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AerospikeBackup is the Schema for the aerospikebackup API
            displayName: Aerospike Backup
            kind: AerospikeBackup
            name: aerospikebackups.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeBackupService is the Schema for the aerospikebackupservices
              API
            displayName: Aerospike Backup Service
            kind: AerospikeBackupService
            name: aerospikebackupservices.asdb.aerospike.com
            version: v1beta1
          - description: AerospikeCluster is the schema for the AerospikeCluster API
            displayName: Aerospike Cluster
            kind: AerospikeCluster
            name: aerospikeclusters.asdb.aerospike.com
            version: v1
          - description: AerospikeRestore is the Schema for the aerospikerestores
              API
            displayName: Aerospike Restore
            kind: AerospikeRestore
            name: aerospikerestores.asdb.aerospike.com
            version: v1beta1
        description: |
          The Aerospike Kubernetes Operator automates the deployment and management of Aerospike enterprise clusters on Kubernetes. The operator allows you to deploy multi-node Aerospike clusters, recover automatically from node failures, scale up or down automatically as load changes, ensure nodes are evenly split across racks or zones, automatically update to new versions of Aerospike and manage configuration changes in your clusters.

          ### Capabilities:

           -   Deploy Aerospike clusters
           -   Scale up and down existing Aerospike clusters
           -   Version upgrade and downgrade
           -   Configure persistent storage and resource allocation
           -   Standardize and validate configurations
           -   Cluster security management
           -   Aerospike cluster monitoring
           -   Backup and restore Aerospike clusters
        displayName: Aerospike Kubernetes Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Aerospike
        links:
        - name: Documentation
          url: https://docs.aerospike.com/cloud/kubernetes/operator
        - name: Aerospike
          url: https://aerospike.com
        maintainers:
        - email: developers@aerospike.com
          name: Aerospike, Inc.
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Aerospike
        relatedImages:
        - docker.io/aerospike/aerospike-kubernetes-operator:4.1.1
        - quay.io/community-operator-pipeline-prod/aerospike-kubernetes-operator@sha256:63cdb7ea6cd6bda4c6ab2c1f3467489c2cef6c8a95836763b9c88b20c43e8168
        version: 4.1.1
      entries:
      - name: aerospike-kubernetes-operator.v4.1.1
        version: 4.1.1
      - name: aerospike-kubernetes-operator.v4.1.0
        version: 4.1.0
      - name: aerospike-kubernetes-operator.v4.0.2
        version: 4.0.2
      - name: aerospike-kubernetes-operator.v4.0.1
        version: 4.0.1
      - name: aerospike-kubernetes-operator.v4.0.0
        version: 4.0.0
      - name: aerospike-kubernetes-operator.v3.4.1
        version: 3.4.1
      - name: aerospike-kubernetes-operator.v3.4.0
        version: 3.4.0
      - name: aerospike-kubernetes-operator.v3.3.1
        version: 3.3.1
      - name: aerospike-kubernetes-operator.v3.3.0
        version: 3.3.0
      - name: aerospike-kubernetes-operator.v3.2.2
        version: 3.2.2
      - name: aerospike-kubernetes-operator.v3.2.1
        version: 3.2.1
      - name: aerospike-kubernetes-operator.v3.2.0
        version: 3.2.0
      - name: aerospike-kubernetes-operator.v3.1.0
        version: 3.1.0
      - name: aerospike-kubernetes-operator.v3.0.0
        version: 3.0.0
      name: stable
    defaultChannel: stable
    packageName: aerospike-kubernetes-operator
    provider:
      name: Aerospike
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SAP LinuxLab - Red Hat
      provider-url: https://github.com/sap-linuxlab
    name: sap-data-intelligence-observer-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observer-operator.v0.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "observer-operator",
                    "app.kubernetes.io/instance": "sdiobserver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sdiobserver",
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
          description: Configure and enable SAP Data Intelligence to run on OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          The observer shall be run in a dedicated namespace. It must be deployed
          before the SDI installation is started. This is the prerequisite for the
          successful installation of the SDI on OpenShift. Please refer to https://access.redhat.com/articles/5100521
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - quay.io/openshift-community-operators/sap-data-intelligence-observer-operator@sha256:5217360266f62fc8ba087fbc730074c14aacf62dbd124fdc3aa46650d51a243b
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.5
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 0.0.3
      entries:
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: alpha
    - currentCSV: observer-operator.v1.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sdi.sap-redhat.io/v1alpha1",
                "kind": "SDIObserver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/part-of": "observer-operator"
                  },
                  "name": "sdiobserver-sample",
                  "namespace": "sdi-observer"
                },
                "spec": {
                  "manageSDINodeConfig": true,
                  "sdiNamespace": "sdi",
                  "sdiVSystemRoute": {
                    "managementState": "Managed"
                  },
                  "slcbNamespace": "sap-slcbridge",
                  "slcbRoute": {
                    "managementState": "Managed"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
          createdAt: "20240726"
          description: Configure and enable SAP Data Intelligence 3 to run on OpenShift
            4
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mjiao/sap-data-intelligence/tree/master/observer-operator
          support: RedHat-SAPLinuxLab
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SDIObserver is the Schema for the sdiobservers API
            displayName: SDIObserver
            kind: SDIObserver
            name: sdiobservers.sdi.sap-redhat.io
            version: v1alpha1
        description: Operator for monitoring SAP Data Intelligence (SDI) namespace
          and modifying objects in there that enable running of SDI on top of OpenShift.
          It must be deployed before the SDI installation is started. This is the
          prerequisite for the successful installation of the SDI on OpenShift. Please
          refer to https://access.redhat.com/articles/7042265#ocp-install-sdi-observer-operator
          for detailed information.
        displayName: SAP Data Intelligence 3 - Observer Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sdi
        - sap data intelligence
        links:
        - name: SAP Data Intelligence 3 - Observer Operator
          url: https://github.com/redhat-sap/sap-data-intelligence/tree/master/observer-operator
        maintainers:
        - email: mjiao@redhat.com
          name: Manjun Jiao
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: SAP LinuxLab - Red Hat
          url: https://github.com/sap-linuxlab
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sap-data-intelligence-observer-operator@sha256:79aa33cb7e4b570fc9c962fc6bd59708c7e28e81e2b188608780fa085db91442
        - quay.io/redhat-sap-cop/sdi-observer-operator:0.1.32
        version: 1.0.9
      entries:
      - name: observer-operator.v1.0.9
        version: 1.0.9
      - name: observer-operator.v1.0.8
        version: 1.0.8
      - name: observer-operator.v1.0.7
        version: 1.0.7
      - name: observer-operator.v1.0.6
        version: 1.0.6
      - name: observer-operator.v1.0.5
        version: 1.0.5
      - name: observer-operator.v1.0.4
        version: 1.0.4
      - name: observer-operator.v1.0.3
        version: 1.0.3
      - name: observer-operator.v1.0.2
        version: 1.0.2
      - name: observer-operator.v1.0.1
        version: 1.0.1
      - name: observer-operator.v0.0.3
        version: 0.0.3
      name: stable
    defaultChannel: stable
    packageName: sap-data-intelligence-observer-operator
    provider:
      name: SAP LinuxLab - Red Hat
      url: https://github.com/sap-linuxlab
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Apache Software Foundation
      provider-url: ""
    name: camel-k
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: candidate
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: latest
    - currentCSV: camel-k-operator.v1.12.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.0
          createdAt: "2023-02-27T09:01:55Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:60a9e4b05e7eb17ebfe58dd55db96ab2c68580d9282117f74f21869cb63e4a73
        - docker.io/apache/camel-k:1.12.0
        version: 1.12.0
      entries:
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable
    - currentCSV: camel-k-operator.v1.10.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.10.4
          createdAt: "2023-01-28T13:49:05Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:1461576c2e06c2ee8d84fd40836184566a346ed596b99cb81fe6a08d0ad99096
        - docker.io/apache/camel-k:1.10.4
        version: 1.10.4
      entries:
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.10
    - currentCSV: camel-k-operator.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.11.1
          createdAt: "2023-01-30T08:45:04Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.11.1
        - quay.io/openshift-community-operators/camel-k@sha256:2b8369f729c18384f45a458e88c9f551eb00f26792b40dbf286e6ea2837e9400
        version: 1.11.1
      entries:
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.11
    - currentCSV: camel-k-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.12.1
          createdAt: "2023-06-12T08:46:58Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:6154ff0f72c9c49f1b7ea0edbe93e0a3b04f6e63abbd851d0e6706796050d672
        - docker.io/apache/camel-k:1.12.1
        version: 1.12.1
      entries:
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.12
    - currentCSV: camel-k-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.5.1
          createdAt: "2021-08-31T09:03:41Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:21ce00e81ce495c142cbe9611f494cfb26b52b3c09a507bee0572c1e1e3333d7
        - docker.io/apache/camel-k:1.5.1
        version: 1.5.1
      entries:
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.5
    - currentCSV: camel-k-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.6.1
          createdAt: "2021-10-27T21:06:20Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.6.1
        - quay.io/openshift-community-operators/camel-k@sha256:26df9d918e942f7fc637212df87a114b5edd52570da18cc212b3499354cdb993
        version: 1.6.1
      entries:
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.6
    - currentCSV: camel-k-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": null
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.7.0
          createdAt: "2021-11-16T10:54:12Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.8.1+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:11dac0881f5ed66b5d42b2d69b13aba31cd56e6b32157cb3ce977be197189b79
        - docker.io/apache/camel-k:1.7.0
        version: 1.7.0
      entries:
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.7
    - currentCSV: camel-k-operator.v1.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "flow": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.8.2
          createdAt: "2022-03-07T07:13:19Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:1.8.2
        - quay.io/openshift-community-operators/camel-k@sha256:cba8cde86a870b31f7dae9a554f4c496ed81731f91a8e64cd62b0caa88efaa38
        version: 1.8.2
      entries:
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.8
    - currentCSV: camel-k-operator.v1.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "Openshift",
                  "resources": {}
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {},
                  "resources": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:1.9.2
          createdAt: "2022-05-23T09:05:22Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: KameletBinding is the Schema for the kamelets binding API
            displayName: Kamelet Binding
            kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - quay.io/openshift-community-operators/camel-k@sha256:06d31bb777b37f8f7c8981efc24ca26a2f9c56a1c579243d0c7bdddb34c76a98
        - docker.io/apache/camel-k:1.9.2
        version: 1.9.2
      entries:
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-1.9
    - currentCSV: camel-k-operator.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "set-body": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "set-body": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1alpha1",
                "kind": "KameletBinding",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "messaging.knative.dev/v1",
                      "kind": "InMemoryChannel",
                      "name": "example"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world"
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1alpha1",
                      "kind": "Kamelet",
                      "name": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.0.1
          createdAt: "2023-09-04T12:32:38Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - kind: KameletBinding
            name: kameletbindings.camel.apache.org
            version: v1alpha1
          - kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1alpha1
          - kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers.

          ## Installation

          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step.
          Also, You can edit the `IntegrationPlatform`, to configure Camel K.
          The configuration from the `IntegrationPlatform` will apply to the Camel integrations created in the same namespace/project.

          ## Running an Integration

          After the initial setup, you can run a Camel integration on the cluster by creating an example `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            name: example
          spec:
            sources:
            - name: Example.java
              content: |
                import org.apache.camel.builder.RouteBuilder;

                public class Example extends RouteBuilder {
                    @Override
                    public void configure() throws Exception {
                        from("timer:tick")
                            .setBody(constant("Hello World!"))
                        .to("log:info?skipBodyLineSeparator=false");
                    }
                }
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.11.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.0.1
        - quay.io/openshift-community-operators/camel-k@sha256:d96f1b35ef30d6701f2a52b5e660dbd5378ec82a933afd38d2eb603c9ca084fd
        version: 2.0.1
      entries:
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-2.0
    - currentCSV: camel-k-operator.v2.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Build",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "CamelCatalog",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Integration",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {
                  "flows": [
                    {
                      "from": {
                        "parameters": {
                          "period": "1000"
                        },
                        "steps": [
                          {
                            "setBody": {
                              "constant": "Hello from Camel K"
                            }
                          },
                          {
                            "to": "log:info"
                          }
                        ],
                        "uri": "timer:yaml"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationKit",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationPlatform",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "camel-k"
                },
                "spec": {
                  "build": {},
                  "kamelet": {},
                  "profile": "OpenShift"
                },
                "status": {
                  "build": {
                    "maven": {
                      "settings": {}
                    },
                    "registry": {}
                  },
                  "kamelet": {}
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "IntegrationProfile",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "debug"
                },
                "spec": {
                  "traits": {
                    "logging": {
                      "level": "DEBUG"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Kamelet",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "example-source"
                },
                "spec": {
                  "definition": {
                    "description": "Produces periodic events with a custom payload",
                    "properties": {
                      "message": {
                        "description": "The message to generate",
                        "title": "Message",
                        "type": "string"
                      },
                      "period": {
                        "default": 1000,
                        "description": "The time interval between two events",
                        "title": "Period",
                        "type": "integer"
                      }
                    },
                    "required": [
                      "message"
                    ],
                    "title": "Example Timer"
                  },
                  "template": {
                    "from": {
                      "parameters": {
                        "period": "#property:period"
                      },
                      "steps": [
                        {
                          "setBody": {
                            "constant": "#property:message"
                          }
                        },
                        {
                          "to": "kamelet:sink"
                        }
                      ],
                      "uri": "timer:tick"
                    }
                  }
                }
              },
              {
                "apiVersion": "camel.apache.org/v1",
                "kind": "Pipe",
                "metadata": {
                  "labels": {
                    "app": "camel-k"
                  },
                  "name": "timer-to-log-example"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "log-sink"
                    }
                  },
                  "source": {
                    "properties": {
                      "message": "Hello world",
                      "period": 2000
                    },
                    "ref": {
                      "apiVersion": "camel.apache.org/v1",
                      "kind": "Kamelet",
                      "name": "timer-source"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/apache/camel-k:2.8.0
          createdAt: "2025-09-21T09:07:43Z"
          description: Apache Camel K is a lightweight integration platform, born
            on Kubernetes, with serverless superpowers.
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/internal-objects: '["builds.camel.apache.org","integrationkits.camel.apache.org","camelcatalogs.camel.apache.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/apache/camel-k
          support: Camel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Build is the Schema for the builds API
            displayName: Build
            kind: Build
            name: builds.camel.apache.org
            version: v1
          - description: CamelCatalog is the Schema for the camelcatalogs API
            displayName: Camel Catalog
            kind: CamelCatalog
            name: camelcatalogs.camel.apache.org
            version: v1
          - description: IntegrationKit is the Schema for the integrationkits API
            displayName: Integration Kit
            kind: IntegrationKit
            name: integrationkits.camel.apache.org
            version: v1
          - description: IntegrationPlatform is the Schema for the integrationplatforms
              API
            displayName: Integration Platform
            kind: IntegrationPlatform
            name: integrationplatforms.camel.apache.org
            version: v1
          - description: IntegrationProfile is the Schema for the integrationprofiles
              API
            displayName: Integration Profile
            kind: IntegrationProfile
            name: integrationprofiles.camel.apache.org
            version: v1
          - description: Integration is the Schema for the integrations API
            displayName: Integration
            kind: Integration
            name: integrations.camel.apache.org
            version: v1
          - description: Kamelet is the Schema for the kamelets API
            displayName: Kamelet
            kind: Kamelet
            name: kamelets.camel.apache.org
            version: v1
          - description: Pipe is the Schema for the pipes API
            displayName: Pipe
            kind: Pipe
            name: pipes.camel.apache.org
            version: v1
        description: |
          Apache Camel K
          ==============

          Apache Camel K is a lightweight integration platform, born on Kubernetes, with serverless superpowers:
          the easiest way to build and manage your Camel applications on Kubernetes.

          ## Installation

          You may need to specify the container registry where to store the container images created for the Camel applications.
          To start using Camel K, install the operator and then create the following `IntegrationPlatform`:
          ```
          apiVersion: camel.apache.org/v1
          kind: IntegrationPlatform
          metadata:
            name: camel-k
            labels:
              app: "camel-k"
          spec:
            build:
              registry:
                address: docker.io/my-org
          ```

          An `IntegrationPlatform` resource is automatically created by default on OpenShift, so you can skip this step when using Openshift cluster.

          ## Running an Integration

          Run a Camel application on the cluster by creating a sample `Integration`, e.g.:
          ```
          apiVersion: camel.apache.org/v1
          kind: Integration
          metadata:
            annotations:
              camel.apache.org/operator.id: camel-k
            name: test
          spec:
            flows:
            - from:
                parameters:
                  period: "1000"
                steps:
                - setBody:
                    simple: Hello Camel from ${routeId}
                - log: ${body}
                uri: timer:yaml
          ```
          Log the result of the running application:
          ```
          kubectl wait --for=condition=ready integration test --timeout=180s
          kubectl logs deployment/test -f
          ```
        displayName: Camel K Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - apache
        - kamel
        - kubernetes
        - serverless
        - microservices
        links:
        - name: Camel K source code repository
          url: https://github.com/apache/camel-k
        maintainers:
        - email: users@camel.apache.org
          name: The Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: The Apache Software Foundation
        relatedImages:
        - docker.io/apache/camel-k:2.8.0
        - quay.io/community-operator-pipeline-prod/camel-k@sha256:c20f28e8f40b32c6f10f494e35df85210148f962c7a685d1f29b1d5024314f05
        version: 2.8.0
      entries:
      - name: camel-k-operator.v2.8.0
        version: 2.8.0
      - name: camel-k-operator.v2.7.0
        version: 2.7.0
      - name: camel-k-operator.v2.6.0
        version: 2.6.0
      - name: camel-k-operator.v2.5.1
        version: 2.5.1
      - name: camel-k-operator.v2.5.0
        version: 2.5.0
      - name: camel-k-operator.v2.4.0
        version: 2.4.0
      - name: camel-k-operator.v2.3.3
        version: 2.3.3
      - name: camel-k-operator.v2.3.2
        version: 2.3.2
      - name: camel-k-operator.v2.3.1
        version: 2.3.1
      - name: camel-k-operator.v2.3.0
        version: 2.3.0
      - name: camel-k-operator.v2.2.0
        version: 2.2.0
      - name: camel-k-operator.v2.1.0
        version: 2.1.0
      - name: camel-k-operator.v2.0.1
        version: 2.0.1
      - name: camel-k-operator.v2.0.0
        version: 2.0.0
      - name: camel-k-operator.v1.12.1
        version: 1.12.1
      - name: camel-k-operator.v1.12.0
        version: 1.12.0
      - name: camel-k-operator.v1.11.1
        version: 1.11.1
      - name: camel-k-operator.v1.11.0
        version: 1.11.0
      - name: camel-k-operator.v1.10.4
        version: 1.10.4
      - name: camel-k-operator.v1.10.3
        version: 1.10.3
      - name: camel-k-operator.v1.10.2
        version: 1.10.2
      - name: camel-k-operator.v1.10.1
        version: 1.10.1
      - name: camel-k-operator.v1.10.0
        version: 1.10.0
      - name: camel-k-operator.v1.9.2
        version: 1.9.2
      - name: camel-k-operator.v1.9.1
        version: 1.9.1
      - name: camel-k-operator.v1.9.0
        version: 1.9.0
      - name: camel-k-operator.v1.8.2
        version: 1.8.2
      - name: camel-k-operator.v1.8.1
        version: 1.8.1
      - name: camel-k-operator.v1.8.0
        version: 1.8.0
      - name: camel-k-operator.v1.7.0
        version: 1.7.0
      - name: camel-k-operator.v1.6.1
        version: 1.6.1
      - name: camel-k-operator.v1.6.0
        version: 1.6.0
      - name: camel-k-operator.v1.5.1
        version: 1.5.1
      - name: camel-k-operator.v1.5.0
        version: 1.5.0
      - name: camel-k-operator.v1.4.1
        version: 1.4.1
      - name: camel-k-operator.v1.4.0
        version: 1.4.0
      name: stable-v2
    defaultChannel: stable-v2
    packageName: camel-k
    provider:
      name: The Apache Software Foundation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-dependencies
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-dependencies.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          createdAt: "2025-09-11T15:25:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          deploys dependencies.
        displayName: Data Foundation Dependencies
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-dependencies-operator-bundle@sha256:6c2223357b672b1a59d22eeab742d3c241d87d5524935974876c9371ad650afa
        version: 4.18.11-rhodf
      entries:
      - name: odf-dependencies.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-dependencies.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-dependencies.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-dependencies.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-dependencies.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-dependencies.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-dependencies.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-dependencies.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-dependencies.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-dependencies.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-dependencies.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-dependencies.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-dependencies
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc
      provider-url: ""
    name: ingress-node-firewall
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ingress-node-firewall.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewall",
                "metadata": {
                  "name": "ingressnodefirewall-demo2"
                },
                "spec": {
                  "ingress": [
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmp": {
                              "icmpType": 8
                            },
                            "protocol": "ICMP"
                          }
                        },
                        {
                          "action": "Deny",
                          "order": 20,
                          "protocolConfig": {
                            "protocol": "TCP",
                            "tcp": {
                              "ports": "8000-9000"
                            }
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "172.16.0.0/12"
                      ]
                    },
                    {
                      "rules": [
                        {
                          "action": "Deny",
                          "order": 10,
                          "protocolConfig": {
                            "icmpv6": {
                              "icmpType": 128
                            },
                            "protocol": "ICMPv6"
                          }
                        }
                      ],
                      "sourceCIDRs": [
                        "fc00:f853:ccd:e793::0/64"
                      ]
                    }
                  ],
                  "interfaces": [
                    "eth0"
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "do-node-ingress-firewall": "true"
                    }
                  }
                }
              },
              {
                "apiVersion": "ingressnodefirewall.openshift.io/v1alpha1",
                "kind": "IngressNodeFirewallConfig",
                "metadata": {
                  "name": "ingressnodefirewallconfig",
                  "namespace": "ingress-node-firewall-system"
                },
                "spec": {
                  "debug": false,
                  "ebpfProgramManagerMode": false,
                  "nodeSelector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
          createdAt: "2024-09-13T12:46:10Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.11.0 <4.18.0'
          operatorframework.io/suggested-namespace: openshift-ingress-node-firewall
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/ingress-node-firewall
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: IngressNodeFirewallConfig
            name: ingressnodefirewallconfigs.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewallNodeState
            name: ingressnodefirewallnodestates.ingressnodefirewall.openshift.io
            version: v1alpha1
          - kind: IngressNodeFirewall
            name: ingressnodefirewalls.ingressnodefirewall.openshift.io
            version: v1alpha1
        description: The Ingress node firewall operator, manages the configuration
          of cluster scoped Ingress Node firewall rules for Kubernetes.
        displayName: Ingress Node Firewall Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ingressnodefirewall
        links:
        - name: Ingress Node Firewall
          url: https://github.com/openshift/ingress-node-firewall
        maintainers:
        - email: mmahmoud@redhat.com
          name: msherif1234
        maturity: alpha
        provider:
          name: Red Hat, Inc
        relatedImages:
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9-operator@sha256:d71a5f9b01e6580ea78f71b23f1c9107be2bd1ec85e27318928bfff761de021d
        - registry.redhat.io/openshift4/ingress-node-firewall-rhel9@sha256:577ab5d4780ba01e130ca5713fb7b3764244cdcb937b6b00e6bb17638f15fea2
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ingress-node-firewall-operator-bundle@sha256:6c6a76287353a02c05164f9a0a3e07d27b75f0e5d403eb1b47f1e9dc7c9883d8
        version: 4.18.0-202509100344
      entries:
      - name: ingress-node-firewall.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ingress-node-firewall.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ingress-node-firewall.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ingress-node-firewall.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ingress-node-firewall.v4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: ingress-node-firewall.v4.18.0-202506062012
        version: 4.18.0-202506062012
      - name: ingress-node-firewall.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ingress-node-firewall.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ingress-node-firewall.v4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: ingress-node-firewall.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ingress-node-firewall.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ingress-node-firewall.v4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: ingress-node-firewall.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ingress-node-firewall.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ingress-node-firewall.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ingress-node-firewall
    provider:
      name: Red Hat, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/component: kmm
      app.kubernetes.io/name: kmm
      app.kubernetes.io/part-of: kmm
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: kernel-module-management
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kernel-module-management.v2.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.0.2'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel-module-management
        - out-of-tree
        - kernel
        - kmm
        - module
        - modules
        - oot
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:25f39f20f325f64676bf8ef14cd8a60cfbdc77116753a89d4bb667ea2a86172c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62c44057a27abd41cf3daf60b8736da59a82dadb65fa6c32bef65d19bd49ea49
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:192e08b121a04ae69aea60c35199e8f5e3c42305e1ebd024a605a371de4ab472
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:e39a0b57e73d92c5db330ac9b8a184eeb400397f14c05c1ef75c952d91e4f5b2
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:1022205d712a2ed9d6e76968961c180d4e66a2e0fb991221b0b02ed94ed830d2
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:94b4ddc45cb34d4ffe794a51b99c868cb709f90cc4eb8b28284d70c526a0a226
        version: 2.0.2
      entries:
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.0
    - currentCSV: kernel-module-management.v2.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.1.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:14e100e5853444114029116255ebfb7a54e5ed51503ba969740ad48b5ca80f70
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:5024be9a3190288983e3c42e8a2dbb49aa44cd596237ad9ed0e72de5090dc9b9
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a85037778c53c3766f39efe4c4d2523d01ae7743e5c5693b05bd8b2cbb50b1d8
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:1a9a19df17b464d1708b2905d03b8726e4480ca1f11aa97a15ff4c049a24ea4b
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:611556d999ea97f9dbb78fec5963baa02de7a26626d7096ba63b177e3bcf9707
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ea3e5fbf7eaf6bd29031e2876d7e24859779fb4d3d7ef27c26a1d310f137ad17
        version: 2.1.1
      entries:
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.1
    - currentCSV: kernel-module-management.v2.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.2.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:ece58f1e360eccf7d970dca2f9da0c4d99a692648f165f677f11ff080649cf62
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:a224708950f8defb259ef0da49011d361c5c0378404350640600c6e8fbdbd606
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:c3dc9ad4ebc68cd85a074b520903a9310e72a90a73bdf4aa7c42b7ed66b7bcba
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:b1d98c0f14448ae8eb80821cbec236150c338674aae2a2b37f8d0ab0d5f1430c
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:ce1d68fd96f5003285f9fbc0c7abd8aa23830dbdd9b76629d38d4d2238e30032
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:53fddf0b226eb579c71c86112a8d9b6b1d90a1498d6f6985932b16b8be532f35
        version: 2.2.1
      entries:
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.2
    - currentCSV: kernel-module-management.v2.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidationOCP",
                "metadata": {
                  "name": "preflightvalidationocp-sample"
                },
                "spec": {
                  "pushBuiltImage": true,
                  "releaseImage": "ocp-release-image"
                }
              }
            ]
          capabilities: Basic Install
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.3.0'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        links:
        - name: kernel module management
          url: https://github.com/rh-ecosystem-edge/kernel-module-management
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:f9fde334af68220943017a7491130e6a82f15e63522f5bb3931aa7d9b00ccc40
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:6bd423056b555d39337c0e3a9c31e7732e80f34d0c01f0766a6aadb2aff586f8
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:750dcc75f908cdda86932386fa1a28edaf65d7b5379ef6d0447586a064db906f
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:06ac089cc89b303e525fd71d6439b958f34ab944a72209f1cc8cb562ecfd1112
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:c5005b74402f680d11c9ea682a9a0d3f74b6ab0d01d25298d232f910f1a29c34
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:055c2d6e80bd1ec225f0cb2041572ebb28cfd1a4ed82ba47ff0d638d52ded2a8
        version: 2.3.0
      entries:
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.3
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: release-2.4
    - currentCSV: kernel-module-management.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "Module",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "module-sample"
                },
                "spec": {
                  "moduleLoader": {
                    "container": {
                      "kernelMappings": [],
                      "modprobe": {
                        "moduleName": "sample"
                      }
                    }
                  },
                  "selector": {
                    "node-role.kubernetes.io/worker": ""
                  }
                }
              },
              {
                "apiVersion": "kmm.sigs.x-k8s.io/v1beta1",
                "kind": "PreflightValidation",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/component": "kmm",
                    "app.kubernetes.io/name": "kmm",
                    "app.kubernetes.io/part-of": "kmm"
                  },
                  "name": "preflightvalidation-sample"
                },
                "spec": {
                  "kernelVersion": "4.18.0-372.32.1.el8_6.x86_64",
                  "pushBuiltImage": true
                }
              }
            ]
          capabilities: Seamless Upgrades
          certified: "true"
          containerImage: registry.redhat.io/kmm/kernel-module-management-rhel9-operator
          createdAt: "2025-06-19T20:27:41Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <2.4.1'
          operatorframework.io/suggested-namespace: openshift-kmm
          operators.openshift.io/must-gather-image: registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-ecosystem-edge/kernel-module-management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ModuleBuildSignConfig
            name: modulebuildsignconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: ModuleImagesConfig
            name: moduleimagesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: Module describes how to load a module on different kernel
              versions
            displayName: Module
            kind: Module
            name: modules.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: NodeModulesConfig
            name: nodemodulesconfigs.kmm.sigs.x-k8s.io
            version: v1beta1
          - description: PreflightValidation initiates a preflight validations for
              all Modules on the current Kubernetes cluster.
            displayName: Preflight Validation
            kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidation
            name: preflightvalidations.kmm.sigs.x-k8s.io
            version: v1beta2
          - description: PreflightValidationOCP initiates a preflight validations
              for all Modules on the current OCP cluster.
            displayName: Preflight Validation OCP
            kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta1
          - kind: PreflightValidationOCP
            name: preflightvalidationsocp.kmm.sigs.x-k8s.io
            version: v1beta2
        description: Kubernetes operator managing out of tree kernel modules
        displayName: Kernel Module Management
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kernel
        - kmm
        - module
        - modules
        - oot
        - out-of-tree
        maintainers:
        - email: edge-kmm@redhat.com
          name: Red Hat Ecosystem - Partner Accelerators Team
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/kmm/kernel-module-management-rhel9-operator@sha256:8e0a3cfb1cb92d6d3c9ab5a3030593d8df1378c3ce755d598fb77b08ea5793d3
        - registry.redhat.io/kmm/kernel-module-management-signing-rhel9@sha256:38d811a00b88a096d9fa8f650dae132c43db7057aac61c98932494ce94d67fff
        - registry.redhat.io/kmm/kernel-module-management-webhook-server-rhel9@sha256:34bd5c4f17411d5aa71edb1d6482b9a4cffffe4983b0a599db080bb09e1ec4d2
        - registry.redhat.io/kmm/kernel-module-management-worker-rhel9@sha256:fc753a1c47c870199bf3c9a6cb4ef704847d525b3df5cd7628e2cc7dd0bbb2aa
        - registry.redhat.io/kmm/kernel-module-management-must-gather-rhel9@sha256:15d84c2191be70931b832cde5a7045350badaeef16d811b07291ec2d4acec699
        - registry.redhat.io/kmm/kernel-module-management-operator-bundle@sha256:088dde6eb9d239a7a34b6e47a1bc3c3f1f74aabe764744eb9554913181b4e0db
        version: 2.4.1
      entries:
      - name: kernel-module-management.v2.4.1
        version: 2.4.1
      - name: kernel-module-management.v2.4.0
        version: 2.4.0
      - name: kernel-module-management.v2.3.0
        version: 2.3.0
      - name: kernel-module-management.v2.2.1
        version: 2.2.1
      - name: kernel-module-management.v2.2.0
        version: 2.2.0
      - name: kernel-module-management.v2.1.1
        version: 2.1.1
      - name: kernel-module-management.v2.1.0
        version: 2.1.0
      - name: kernel-module-management.v2.0.2
        version: 2.0.2
      - name: kernel-module-management.v2.0.1
        version: 2.0.1
      - name: kernel-module-management.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: kernel-module-management
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-certified-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": "registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551",
                    "imagePullPolicy": "IfNotPresent",
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7"
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43",
                    "imagePullPolicy": "IfNotPresent",
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0",
                    "imagePullPolicy": "IfNotPresent",
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": "registry.access.redhat.com/ubi8",
                      "imagePullPolicy": "IfNotPresent",
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": "registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d",
                      "imagePullPolicy": "Always",
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "true"
          containerImage: registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
          createdAt: "2020-12-02 14:19:21"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: Neuvector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm/tree/master/charts/core) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart.  NeuVector Operator versions are tied to NeuVector product versions. Version 1.5.2 of the NeuVector Certified Operator deploys version 5.4.6 of NeuVector.

          **Complete below steps to Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.6 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.6, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s                    neuvector/default


          #Deploying the NeuVector Operator#


          Please refer to the instructions [here](https://github.com/neuvector/neuvector-operator/blob/master/README.md)


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Monitoring
        - Networking
        - Security
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - registry.connect.redhat.com/neuvector/controller@sha256:65ee3723c7c007db5ce3cd0493a785d4163d972cf957de6dec956b842807a551
        - registry.connect.redhat.com/neuvector/enforcer@sha256:72a628a5c811c95bb267bbd0a5ffc9d74dbaca03a92aae615d43c7601b4d3b43
        - registry.connect.redhat.com/neuvector/manager@sha256:89a42c9ee86f98ef56b5e695e23bdcc6334361ebd29cfb7729d7c86f04da98b0
        - registry.connect.redhat.com/neuvector/neuvector-operator-bundle@sha256:261b77cfb315c608505e5659b2d01dc4e1ea96cd37fae25f417cd7df148f4ce4
        - registry.connect.redhat.com/neuvector/neuvector-operator@sha256:e2f19c50ebc4127a83c943dbcfa1312ec46dec71bd36f6d99178b320d8da970b
        - registry.connect.redhat.com/neuvector/scanner@sha256:b632f1d31998a4c8bad9f97216e8aa56cd7e6446bd70d2089d6c66845b964e8d
        - registry.access.redhat.com/ubi8@sha256:534c2c0efa4150ede18e3f9d7480d3b9ec2a52e62bc91cd54e08ee7336819619
        version: 1.5.2
      entries:
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-certified-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Snyk Ltd.
      provider-url: ""
    name: snyk-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snyk-operator.v1.92.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1alpha1",
                "kind": "SnykMonitor",
                "metadata": {
                  "name": "snyk-monitor"
                },
                "spec": {
                  "clusterName": "",
                  "image": {
                    "pullPolicy": "Always",
                    "repository": "snyk/kubernetes-monitor",
                    "tag": "1.92.0-ubi8"
                  },
                  "integrationApi": "",
                  "monitorSecrets": "snyk-monitor",
                  "scope": "Cluster",
                  "temporaryStorageSize": "50Gi",
                  "pvc": {
                    "enabled": false,
                    "create": false,
                    "name": "snyk-monitor-pvc",
                    "storageClassName": null
                  },
                  "initContainerImage": {
                    "repository": "busybox",
                    "tag": "latest"
                  },
                  "requests": {
                    "memory": "400Mi"
                  },
                  "limits": {
                    "memory": "2Gi"
                  },
                  "nodeAffinity": {
                    "disableBetaArchNodeSelector": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Security
          certified: "false"
          containerImage: docker.io/snyk/kubernetes-operator:1.92.0
          createdAt: "2022-06-09T17:49:36Z"
          description: A Kubernetes Operator for creating and managing Snyk controller
            instances.
          olm.skipRange: <1.68.2
          repository: https://github.com/snyk/kubernetes-monitor
          support: Snyk Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a deployment of the Snyk controller, which scans
              container images in the cluster for vulnerabilities.
            displayName: Snyk Monitor
            kind: SnykMonitor
            name: snykmonitors.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          A Kubernetes Operator for creating and managing Snyk Kubernetes controller instances.

          Snyk integrates with Kubernetes, enabling you to import and test your running workloads and identify vulnerabilities in their associated images and configurations that might make those workloads less secure. Once imported, Snyk continues to monitor those workloads, identifying additional security issues as new images are deployed and the workload configuration changes.

          ## Prerequisites

          To start using the Operator and the Snyk controller that it manages, you will need to create a secret containing your Snyk integrationId and Docker config file:

          ```
          kubectl create namespace snyk-monitor
          kubectl create secret generic snyk-monitor -n snyk-monitor --from-literal=dockercfg.json={} --from-literal=integrationId=${SNYK_INTEGRATION_ID}
          ```

          Refer to the [Snyk documentation](https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview) for more details.
        displayName: Snyk Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - snyk
        - controller
        - monitor
        - operator
        - security
        - cluster
        - container
        - image
        - security
        - vulnerabilities
        - advisories
        - audit
        - scan
        - scanning
        links:
        - name: Website
          url: https://snyk.io
        - name: Documentation
          url: https://support.snyk.io/hc/en-us/articles/360003916138-Kubernetes-integration-overview
        maintainers:
        - email: support@snyk.io
          name: Snyk Ltd.
        maturity: stable
        provider:
          name: Snyk Ltd.
        relatedImages:
        - docker.io/snyk/kubernetes-operator:1.92.0
        - quay.io/openshift-community-operators/snyk-operator@sha256:97908144406700ae9c038316661dc7904f3e0f4e01b1bbdf68259b59dfb86d6c
        version: 1.92.0
      entries:
      - name: snyk-operator.v1.92.0
        version: 1.92.0
      - name: snyk-operator.v1.90.2
        version: 1.90.2
      - name: snyk-operator.v1.90.1
        version: 1.90.1
      - name: snyk-operator.v1.88.4
        version: 1.88.4
      - name: snyk-operator.v1.85.2
        version: 1.85.2
      - name: snyk-operator.v1.82.1
        version: 1.82.1
      - name: snyk-operator.v1.81.1
        version: 1.81.1
      - name: snyk-operator.v1.80.0
        version: 1.80.0
      - name: snyk-operator.v1.73.0
        version: 1.73.0
      - name: snyk-operator.v1.72.0
        version: 1.72.0
      - name: snyk-operator.v1.70.4
        version: 1.70.4
      - name: snyk-operator.v1.70.1
        version: 1.70.1
      name: stable
    defaultChannel: stable
    packageName: snyk-operator
    provider:
      name: Snyk Ltd.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: stolostron
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: stolostron.v0.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
          createdAt: "2025-02-24T01:42:20Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.6.3'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MulticlusterHub defines the configuration for an instance
              of a multicluster hub, a central point for managing multiple Kubernetes-based
              clusters. The deployment of multicluster hub components is determined
              based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/observability-e2e-test@sha256:03403aeb59b364f578761ae4bcb302bd3589baff500d733775e8cb4f2f961236
        - quay.io/stolostron/grc-policy-framework-tests@sha256:9b84dbafef35bd568b2bf79283b6850ba1ff57d70239a27cc0cc815fb1a55515
        - quay.io/stolostron/config-policy-controller@sha256:8cf54eb6f5b49a60ed165fa4b70e4984a1362f2f31e8304c3a22e8a6a953a606
        - quay.io/stolostron/governance-policy-framework-addon@sha256:047124271d0b30176a94454b293c82c2eb67a6f06757fc22c54c455fd0913157
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/prometheus-alertmanager@sha256:b98480f8ba6ffd19fcee3b412757ad34428e6feab7e4196bb13f2f940306e2c9
        - quay.io/stolostron/rbac-query-proxy@sha256:096ad23c9b9c7e90451c09b48ded67a82575e75c8cc19f22c459d63ae8b5a786
        - quay.io/stolostron/kube-rbac-proxy@sha256:5b521e4f0e5e54c53cea714020069aa796da23d1482d14e4b7c908098ac69ae9
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/must-gather@sha256:2bf9d3c15f32a6bfe0a7a8d39b9dc62ac6cbb6fe47b7cf1f6bc5e30b17e5e849
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/governance-policy-addon-controller@sha256:fa0738abc0368f9dadd8b0845d868512d26be3525a1a07b9de8a9cefafaa42b5
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:51a8acc47098175db148ebc19a10cdba2803d4a9f5884dd38b1c839702398789
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cadee58c1c1040fad5a46fc9be5fae4a1944845c5e736c50fa192192d5c17a36
        - quay.io/stolostron/observatorium@sha256:398ecee4d3f12e84e48b12009a4b02fc8b360dfc0c1607ffecc26e6bc04d0ea2
        - quay.io/stolostron/volsync-addon-controller@sha256:d9dad837f2f735674e5711c7fff1cfe2c0c92160fceb7d1548a62348dd43b81c
        - quay.io/stolostron/prometheus-operator@sha256:f80e79ea87cde66955e1e46d264e414c072085da9db6876ee2ae68c87d2452e5
        - quay.io/stolostron/grafana@sha256:51092f6754106a90aff28cdd80c2f0447c85c1bbc72db15878d0b65b93e1ead6
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/multicluster-operators-channel@sha256:501a5053bd3fd5cc9de34e7e020890d731321b090a0e5bf11b8b3350d002f573
        - quay.io/stolostron/observatorium-operator@sha256:5eb4c0da92c62b664aaf3bc187924870275e687440c3b0a3ad4f1c760c8b35e9
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:9c85fb91c4be3287cac2a376465836689df885902099096f0f76ffca50a6d311
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/klusterlet-addon-controller@sha256:39e6940eb956fe9640342e11ecbcbc7d6df0169e1b340a60afc19aff7dd713a5
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/console@sha256:eed52f3ab860a3a224ce178d92765ca50d72d6261f04df040a39577881738252
        - quay.io/stolostron/search-e2e@sha256:43e1bbfb7259612ffb408108c71b825f6e61a268c7e405b8d855f6878b6f96dd
        - quay.io/stolostron/helloworld@sha256:56f0d641d4f07b922f1c28c536eacad747c56a564e0a65f084106c2ffc9bd99d
        - quay.io/stolostron/governance-policy-propagator@sha256:c21574a2421fefe3fcabeef14640d6fe699d950106c9ed056b3ba9be3cbeabe0
        - quay.io/stolostron/insights-metrics@sha256:f2f4cf14090252af32832ff6df8a9f896e5314751199156c4e5ae0755e4515e6
        - quay.io/stolostron/multicluster-operators-application@sha256:59802fcfae5b6fd0bebe58e98412885e71f32464b8dbbf85364ddaf090aa3e00
        - quay.io/stolostron/prometheus@sha256:3d1810567f67ae1a7bb024a1c4e494ee510cfb14c7ed21a8b5e63e41fa7bcd68
        - quay.io/stolostron/siteconfig-operator@sha256:97dc128666780ec9a4c955f392ab98f81f0d6038ec2159e8b1d55aab393cf118
        - quay.io/stolostron/search-v2-operator@sha256:85e53a97240f6fd1c4251101146cc2bcc51eb91d9b171efb5b58edade9e0244e
        - quay.io/stolostron/multicloud-integrations@sha256:0081065f9e59c5ed41087344668ac86012f796cdbbf3575fd8491c7dfd24fbf8
        - quay.io/stolostron/insights-client@sha256:978d8b7c3e692272024b71b8c005b752c08dc643dde8258ad75961ddcdcce67f
        - quay.io/stolostron/kube-state-metrics@sha256:cd24916e55ad905e386fa3df8a648272114353a0d75283224a4dddb1955b2687
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/thanos@sha256:53eee303836070a7c7621bfb2a094e3809146497d1653dc65d16055aacea332c
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/search-collector@sha256:54ece9741641e1bb6e30f15d97a65c92b9c6408ba2572a2f6658f537134adff1
        - quay.io/stolostron/thanos-receive-controller@sha256:d73f98e93a258adffc1a4bd875c79a9629166599e68ef777a790a1f3146a2929
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:92369b6deb21ab3cb0a12ae5c74b455cfefb8698826bca39e26648e608e48a2b
        - quay.io/stolostron/acm-cli@sha256:8c8c059ce2f6d131c410972bf8fd0b7e399cf17cd044c7ddc89a9668c0ceac4c
        - quay.io/stolostron/cert-policy-controller@sha256:98dd74d7375ca2b546075d5c8f06a9e2c4f2ce3b77247dccf36bcaa46bf564c7
        - quay.io/stolostron/multicluster-operators-subscription@sha256:a056ce7d438f68d33ca3dfd42ec3d9a34b347cdf05499066412784109726e7ad
        - quay.io/stolostron/multicluster-observability-operator@sha256:bed0dd30646d7fce894a625e6cf52e7829c272f47dde0baf1f788e9f85f66e42
        - quay.io/stolostron/search-indexer@sha256:4f9be902fe976fd88a77b737cc1f56f00d48740a6a62661468328cf3bb9da31a
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/prometheus-config-reloader@sha256:2b26d2fb5c69e54aecfbcaab46e652d136d726f0604844031c8771874f1fd769
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/multiclusterhub-operator@sha256:09f1f4266938fba88376a50c3e2837c5517b4f3a947cf86638a17e52eb17c1f6
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/cluster-permission@sha256:443d8b30e3bb83c139c64ddfc38b0d5e6d7a25a4fae783303ab2f2c27d015b26
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/submariner-addon@sha256:9c72b5177de8ddc56dd3544997e9bd222895545b7c578e43e4051596fb1fa873
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/cluster-backup-controller@sha256:c85f88d42b42e4366f8ff7ad6c3e6b375e98c5b6a12a9c24a6333929e48f1811
        - quay.io/stolostron/search-v2-api@sha256:694523242aabb8db1f27b5afcd8ab2f63a9cdb6056c33a79badf8e68a337738d
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        version: 0.6.3
      entries:
      - name: stolostron.v0.6.3
        version: 0.6.3
      - name: stolostron.v0.6.2
        version: 0.6.2
      - name: stolostron.v0.6.1
        version: 0.6.1
      - name: stolostron.v0.6.0
        version: 0.6.0
      name: community-0.6
    - currentCSV: stolostron.v0.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion": "operator.open-cluster-management.io/v1",
            "kind": "MultiClusterHub", "metadata": {"name": "multiclusterhub", "namespace":
            "open-cluster-management"}, "spec": {}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
          createdAt: "2025-06-09T15:10:05Z"
          description: Advanced management of OpenShift and Kubernetes clusters
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.7.0 <0.7.1'
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.open-cluster-management.io/v1",
            "kind":"MultiClusterHub","metadata":{"name":"multiclusterhub","namespace":"stolostron"},
            "spec": {}}'
          operatorframework.io/suggested-namespace: stolostron
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware",
            "fips"]'
          repository: https://github.com/stolostron/stolostron
          support: Stolostron Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              MulticlusterHub defines the configuration
              for an instance of a multicluster hub, a central point for managing multiple
              Kubernetes-based clusters. The deployment of multicluster hub components
              is determined based on the configuration that is defined in this resource.
            displayName: MultiClusterHub
            kind: MultiClusterHub
            name: multiclusterhubs.operator.open-cluster-management.io
            version: v1
        description: 'Stolostron provides the multicluster hub, a central management
          console for managing multiple Kubernetes-based clusters across data centers,
          public clouds, and private clouds. You can use the hub to create Red Hat
          OpenShift Container Platform clusters on selected providers, or import existing
          Kubernetes-based clusters. After the clusters are managed, you can set compliance
          requirements to ensure that the clusters maintain the specified security
          requirements. You can also deploy business applications across your clusters. '
        displayName: Stolostron
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - install
        - multiclusterhub
        links:
        - name: Documentation
          url: https://github.com/stolostron/stolostron
        maintainers:
        - email: acm-contact@redhat.com
          name: Red Hat
        maturity: pre-release
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/stolostron/origin-cli@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/multicluster-operators-channel@sha256:318d885ab33d5afe17334e831c955f1c3cdf804281b7cc8523330820ac22eee2
        - quay.io/stolostron/search-e2e@sha256:1c7dcc9f3588620b3570a493c2d744de059f425b572b9cb5725142b89414a904
        - quay.io/stolostron/multiclusterhub-operator-tests@sha256:4b29cc08c9e35054a4a820472ea19cd51e043ab1a90495418a36922df74b9540
        - quay.io/stolostron/insights-metrics@sha256:861dc03fff55c1d9aff96b5ed15b9d38d5932df1c00a4423e35ead19da6eda03
        - quay.io/stolostron/origin-oauth-proxy@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/governance-policy-framework-addon@sha256:b1b24c40d68415525e2bae21810fc439d3e96fcc5fa75c92e7f2eee284422bfd
        - quay.io/stolostron/search-operator-bundle@sha256:d3fbacdcc760259c7abda625c6964d1ad4382b6988a98be0f8f33bbd951dc7be
        - quay.io/stolostron/postgresql-13@sha256:f8c4c4364ab8d27d2c774cbd02a0d02b9d6e3cb6e8d69af728cdd889df312b56
        - quay.io/stolostron/analytics-recommends@sha256:1d4a3cfb193992bed84578661affc31be25523966af3aea9f13d1d52498bac8b
        - quay.io/stolostron/prometheus-alertmanager@sha256:a9da6f01cfa61fef803f09ef3f6b27e33a07b97fc1f1e11b6a6c8e13b8a4b90e
        - quay.io/stolostron/search-indexer@sha256:58320b54847e3bb6baa558bda65e9d0ae2198df480ba27f46f96ba1454614508
        - quay.io/stolostron/siteconfig-operator@sha256:16b00f6c4d81eea26e663ca5596526d85bf7777728d1f0ca7e4dd765c64d2bb5
        - quay.io/stolostron/grafana@sha256:cebaed7e5abe0064c65ea69bd4e14794d153de9172c4c3b6de4730f23fed7d24
        - quay.io/stolostron/cluster-proxy@sha256:decd4ebdad2fcadd15fcc4be977ac85f5bb6331afcd96dc862626ae8e1f58691
        - quay.io/stolostron/kube-state-metrics@sha256:3503df61f3d9ebf3cbf66deef865450b32c7403839fdaca42504febaac00687c
        - quay.io/stolostron/endpoint-monitoring-operator@sha256:54c6a2eec497eafdbf861fb95d78d961fb24906b197fc69b4e96d3e8b53e6496
        - quay.io/stolostron/cluster-backup-controller@sha256:e576c2de1f931b7d33da2dbc03d71ce323766bd9158bf4b128221ddd1ab3fb9f
        - quay.io/stolostron/cluster-templates-operator@sha256:133e5cede92a5dbe9373982c851ca40134fc4de8665e54de60ecc7738f9f50f7
        - quay.io/stolostron/helloprow-go@sha256:0671c58cc5c1f3fb9e00ae58ad6ef2c1f470867e26637b823261e9ab9279f05e
        - quay.io/stolostron/governance-policy-addon-controller@sha256:6ecc7557168177474423cea60eadc07d8382adb3e97dee669a8590859a41b08c
        - quay.io/stolostron/thanos@sha256:9d504e2aa1001135eb6e454c58c620d102f2b0a85016b7587452dd1595794c93
        - quay.io/stolostron/rbac-query-proxy@sha256:1c198b6388a72ccfe2c5b68d57a2b22893fb65d977992ec44ee8cad6ee1ca976
        - quay.io/stolostron/search-v2-api@sha256:786866ca8c71a8263ddef45ca9bd9e48cfd1b3ad6215725fbf6f79967bda20ad
        - quay.io/stolostron/observability-e2e-test@sha256:4e60d3e3e1b5dbab3bcc1c4d37a4940a2a42a065a9df539fb243fb174dde6716
        - quay.io/stolostron/postgresql-12@sha256:da0b8d525b173ef472ff4c71fae60b396f518860d6313c4f3287b844aab6d622
        - quay.io/stolostron/multiclusterhub-operator@sha256:ab6660283a25bdf2ea9afe8c76bfbc32f4610d69af2bb13a7bca37a1f5961bdd
        - quay.io/stolostron/submariner-addon@sha256:51359ccdc0ea29bcaf7ec2b190878e9f3333b1530cb1128e142d893fb87eba8b
        - quay.io/stolostron/console@sha256:ba89196efe30c812a767d58a41170b56e2b04167a72e7f13180bf40474bdc824
        - quay.io/stolostron/volsync-addon-controller@sha256:224a0a03a25e5ce61f4c50f6252370876ac0cbf4f9711e997e98a00d6a124fe9
        - quay.io/stolostron/bailer@sha256:b6d0f5ff086a0a19e98ef4b60ac4e1c3f61e775863ee772d7829395800eb6492
        - quay.io/stolostron/cluster-lifecycle-e2e@sha256:1d997d1238d92c852e35162be21c338ff87216e4f92bce24203b997da827197c
        - quay.io/stolostron/multicluster-global-hub-operator@sha256:1efd8e470b88a992e2f35ce18bdf4970f91af7ba917b8d92627b47ce8c10eb40
        - quay.io/stolostron/multicloud-integrations@sha256:665e9385531674a5849aa661eb207a0c6e39ea563fc891c57678cddbf7878d8e
        - quay.io/stolostron/origin-oauth-proxy-415-and-below@sha256:41762eda03298362d3f97e1fd0be4af429d22658b520f94023941655e062d1a1
        - quay.io/stolostron/helloworld@sha256:4465550664bd210cf0df8edfd706de2fc51cacf01b66a962b5607935f9666ebc
        - quay.io/stolostron/flightctl-ocp-ui@sha256:6cff6425850d2b4a8b885a684afb01244d7cd9de7035d7ad71b1e10e5bab25e9
        - quay.io/stolostron/flightctl-periodic@sha256:a3ed745dc65a5886c4cc05348cf4822bfbf2c6d28b85a39fa0b8f3e1f53fdf8a
        - quay.io/stolostron/acm-cli@sha256:2f0cec3bed4ad77f573e8cd51a7e2c62d68a59e4e261ea801d5ac203fd2724fc
        - quay.io/stolostron/search-v2-operator@sha256:2ae9ab48615966ccf91b2022ccda12913a3db3cfbad5351dd8a20648b01137d8
        - quay.io/stolostron/search-collector@sha256:2b3dfd21c0ba7b7b84fac390002e47fae927cd5fdf13e5d1d960703aa6c5bede
        - quay.io/stolostron/insights-client@sha256:2cbd20252f10a5a64f93af50f117f05a865aa9d578e5f6e55e4249470b761604
        - quay.io/stolostron/postgresql-16@sha256:db2f73275f9cc0f2d262fadb8e8a5da72b089f9d22f7c859bb1364d9f9a8b622
        - quay.io/stolostron/config-policy-controller@sha256:b1718d5facb95c5572ac658c899f9fce0a7e8c08de93804de3327cd50cdb073b
        - quay.io/stolostron/kube-rbac-proxy@sha256:e5555e256a518ca34bea01c959bbb5497582b53dbfaaf326e1852430c175637a
        - quay.io/stolostron/flightctl-worker@sha256:4dddded7d2d4cce63972c8ca81f213c2cedb551cc645ec85495a9692dc36bb4c
        - quay.io/stolostron/applifecycle-backend-e2e@sha256:05da11b343938866c5afcfe844bcd15d9c513fdb10fd7ea0d3473b455b427c2a
        - quay.io/stolostron/thanos-receive-controller@sha256:f445bbfc18bbc4646e0432842022d3be5305809e52ddcb9daf0ec000844d5854
        - quay.io/stolostron/multicluster-operators-subscription@sha256:d7c137379781ce6b697f73a464b89d04fbc20c3da00ce21a195ac79ab57a6dc7
        - quay.io/stolostron/node-exporter@sha256:80f4bba97684bcebd2e8be3460f47db90dec6432f7f1ae025105d9a438345f52
        - quay.io/stolostron/origin-oauth-proxy-416-and-up@sha256:0b1051e9e7a18a81a9156bbc182424ad3eb480968256a4765cfd07c13bf4931a
        - quay.io/stolostron/cert-policy-controller@sha256:6deb41867f8781f4e34f69a54d4aeca190d588ba2991279d39cb5a82c2a5c5be
        - quay.io/stolostron/prometheus-operator@sha256:8c1c12562ec00da8eaedb7d428694ebdd3a749c0b1ce03fe2b5663ae40e1c701
        - quay.io/stolostron/prometheus@sha256:b959035ff754050ca092863bb74c35ba9622d9adc9e8eb7caf3b430cfe211730
        - quay.io/stolostron/redis-7-c9s@sha256:6f06170c1684fdd234ef8edbb65b73b47affbec609800493d427d5169b7dfe22
        - quay.io/stolostron/grafana-dashboard-loader@sha256:6c6c2aa528373536d5db66fffeb87a85b79f3cf723e5d9eeaefcb74f91c1a8a1
        - quay.io/stolostron/cluster-proxy-addon@sha256:5c9df70888b1cbc5209b0217bd0312991e6320cd9442e72e208f0af1498aa2f1
        - quay.io/stolostron/governance-policy-propagator@sha256:2b77422fb03177f647c04716e7242bc08f241a1ef9c8968e492ae57816f4c4fb
        - quay.io/stolostron/metrics-collector@sha256:c7433ee4ccdb56717802acf258530e2d6d237787d760a8fbcfd642790254d5ab
        - quay.io/stolostron/flightctl-api@sha256:64fb1b655bef6aeff03cf64b6450155d3cb570ed4713f084590fe515f4fbad02
        - quay.io/stolostron/observatorium-operator@sha256:2f11444703addec163a00c99529c228594fb52df13de47f08faac2e446bec483
        - quay.io/stolostron/observatorium@sha256:18a06117ef114a1e68474fd9832adacd30b03f60ec1aa55374285a20e5311f13
        - quay.io/community-operator-pipeline-prod/stolostron@sha256:cf782e9e2a62a2fc71c81032e6bf40f9ba15d37c33feff2df551edb111c4e17c
        - quay.io/stolostron/multicluster-operators-application@sha256:9f804b4f2da910e0d7f122176f72c0b0ef4b30785fde0369a66f95f4bac58a82
        - quay.io/stolostron/origin-oauth-proxy-49-and-up@sha256:0e72ed64cdf76828392ef8cfe4ee49d644a013ac00ff1590d6c53774d77615e3
        - quay.io/stolostron/prometheus-config-reloader@sha256:024ee6dda10c6d22ddeab57d7d6503b7aa9ac78981cfcec8ffae94b4914500e3
        - quay.io/stolostron/klusterlet-addon-controller@sha256:4cb2f4f4e14cd67a4b2f9549cba714b635d9e1c2d8b869b0e0573fd7a41bc13e
        - quay.io/stolostron/iam-policy-controller@sha256:27cefc2f923d7ed216f7e650b569551668d1f67f0047b6b98902d37032f6ab92
        - quay.io/stolostron/grc-policy-framework-tests@sha256:327122b13ec79ea488d6f424806549e36ebfbabd055f9b40217ba541f4a118a9
        - quay.io/stolostron/origin-oauth-proxy-48@sha256:c92a1480b2628d50d8fe158df9451598406e2e5cc7177cf5efcb81caba120afd
        - quay.io/stolostron/cluster-permission@sha256:e45139d459951b2e751018fe7ecabe1ff22b90ac87be81fb8f17a12f7b7f6a65
        - quay.io/stolostron/flightctl-ui@sha256:fea717f3b8b4c343ec60ce7d74a691f1124b6847d55e30a73dc0d50ea610d370
        - quay.io/stolostron/must-gather@sha256:ad8db72d559d06f4d467f78213d4f8f2cb00f8e76e942d8e75394eb7dadd29d2
        - quay.io/stolostron/multicluster-observability-operator@sha256:3d95297297a301fca1915268fec71d7a6ecef98c6c2ab8494d9befacb2e169fa
        version: 0.7.1
      entries:
      - name: stolostron.v0.7.1
        version: 0.7.1
      - name: stolostron.v0.7.0
        version: 0.7.0
      name: community-0.7
    defaultChannel: community-0.7
    packageName: stolostron
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-secretsmanager-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-secretsmanager-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secretsmanager.services.k8s.aws/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
          createdAt: "2025-09-19T21:14:43Z"
          description: AWS Secrets Manager controller is a service controller for
            managing Secrets Manager resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret represents the state of an AWS secretsmanager Secret
              resource.
            displayName: Secret
            kind: Secret
            name: secrets.secretsmanager.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Secrets Manager resources in AWS from within your Kubernetes cluster.

          **About Amazon Secrets Manager**

          AWS Secrets Manager helps you manage, retrieve, and rotate database credentials, application credentials, OAuth tokens, API keys, and other secrets throughout their lifecycles. [Many AWS services that use secrets](https://docs.aws.amazon.com/secretsmanager/latest/userguide/integrating.html) store them in Secrets Manager.

          Secrets Manager helps you improve your security posture, because you no longer need hard-coded credentials in application source code. Storing the credentials in Secrets Manager helps avoid possible compromise by anyone who can inspect your application or the components. You replace hard-coded credentials with a runtime call to the Secrets Manager service to retrieve credentials dynamically when you need them.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Secrets Manager
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - secretsmanager
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Secrets Manager Developer Resources
          url: https://aws.amazon.com/secrets-manager/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: secretsmanager maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-secretsmanager-controller@sha256:c92ea9b492b6585a816668bd2a77d65471e339e02e25e44c9558a34b352db593
        - public.ecr.aws/aws-controllers-k8s/secretsmanager-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-secretsmanager-controller.v1.1.1
        version: 1.1.1
      - name: ack-secretsmanager-controller.v1.1.0
        version: 1.1.0
      - name: ack-secretsmanager-controller.v1.0.14
        version: 1.0.14
      - name: ack-secretsmanager-controller.v1.0.13
        version: 1.0.13
      - name: ack-secretsmanager-controller.v1.0.12
        version: 1.0.12
      - name: ack-secretsmanager-controller.v1.0.11
        version: 1.0.11
      - name: ack-secretsmanager-controller.v1.0.10
        version: 1.0.10
      - name: ack-secretsmanager-controller.v1.0.9
        version: 1.0.9
      - name: ack-secretsmanager-controller.v1.0.8
        version: 1.0.8
      - name: ack-secretsmanager-controller.v1.0.7
        version: 1.0.7
      - name: ack-secretsmanager-controller.v1.0.6
        version: 1.0.6
      - name: ack-secretsmanager-controller.v1.0.5
        version: 1.0.5
      - name: ack-secretsmanager-controller.v1.0.4
        version: 1.0.4
      - name: ack-secretsmanager-controller.v1.0.3
        version: 1.0.3
      - name: ack-secretsmanager-controller.v1.0.1
        version: 1.0.1
      - name: ack-secretsmanager-controller.v1.0.0
        version: 1.0.0
      - name: ack-secretsmanager-controller.v0.0.11
        version: 0.0.11
      - name: ack-secretsmanager-controller.v0.0.10
        version: 0.0.10
      - name: ack-secretsmanager-controller.v0.0.9
        version: 0.0.9
      - name: ack-secretsmanager-controller.v0.0.8
        version: 0.0.8
      - name: ack-secretsmanager-controller.v0.0.7
        version: 0.0.7
      - name: ack-secretsmanager-controller.v0.0.2
        version: 0.0.2
      - name: ack-secretsmanager-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-secretsmanager-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-healthcheck-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-healthcheck-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
          createdAt: "2023-11-02 11:44:50"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.3.0 <0.4.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.12/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:151e7d9bc500b7252af46d0856e3e16559ff310a1d0efedcc5d914164bbdef55
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:60d95ea8240ecf701ba974b6df3ee296a7fb3ef5393e0dcf43af15e782f2d153
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:8d32dee5c72ec1c94e4e92f8d1da935a384e40e379f237c432086a2781f4ecb9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:796753816645b35cd08da53d925df5e4fb12df8b2a14db98db361f0ff787a028
        version: 0.4.1
      entries:
      - name: node-healthcheck-operator.v0.4.1
        version: 0.4.1
      name: 4.12-eus
    - currentCSV: node-healthcheck-operator.v0.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
          createdAt: "2023-10-30 14:21:41"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.4.0 <0.6.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:b8d5db068bfc81031c2eef86e8649e4fc008b7480e08c9cfab2ab4497098eecf
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:85054b34e119c16cc95cda9dc3f5d63f2d7bfc70634f70973410b2a1b1fa58d4
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:309ba89dece3ee9901ed66c40bd28d71f20c7d91a7feff552e28ea0c772b9178
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:90cceb176fa903cc8e9aa4c0f7316560a8a354a3138dc26abbfc9884c1c576b5
        version: 0.6.1
      entries:
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.14-eus
    - currentCSV: node-healthcheck-operator.v0.8.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
          createdAt: "2024-08-09T14:54:37Z"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.8.2'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          SNR is installed automatically when installing NHC.

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-health-check-operator
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:4811046e26846b1f12a3b20ddc6209980897c11b171ddd61d26b43630cc1d4b4
        - registry.redhat.io/workload-availability/node-remediation-console-rhel8@sha256:942f69efc7ae05d6c4d9c33335e4be71bfae2e6f9ef062dcd7887eebfe02120e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel8@sha256:062cd9def39a087ae3802cd5208e2f0d5bfca29ba76ebd5139d88348e8c9e923
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:63bd0faf0b64e4672014205f9dcfeec6c436edd7dc2aaf83cd053d6e4569d90a
        version: 0.8.2
      entries:
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.16-eus
    - currentCSV: node-healthcheck-operator.v0.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
          createdAt: "2025-08-04 12:38:05"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <0.9.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.7/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        - registry.redhat.io/workload-availability/node-healthcheck-must-gather-rhel9@sha256:8e9926558c3d8082f298e9e5f58bc5b473020acd44c2ce853c2916cdad71d649
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:1b464c682cd29c8fb7bbc566a1575b9002ad61163c2b6e94138e0e4b0ce51140
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:f174ada653bb5fb7e58219510d0b10d868273a70db5c486e0b3ad05a1be50640
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:37f62407284fe574b973a1a685371d191b710ad70a6bfe8172f19a4568c11cf6
        version: 0.9.1
      entries:
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: 4.18-eus
    - currentCSV: node-healthcheck-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-resource-deletion-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
          createdAt: "2022-08-15 12:34:34"
          description: Detect failed Nodes and trigger remediation with e.g. Self
            Node Remediation.
          olm.skipRange: '>=0.2.0 <0.3.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.18.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: Self Node Remediation
          By default NHC depends on the “Self Node Remediation” (SNR) operator, which
          is installed automatically.
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods and VolumeAttachments on the
          failed Node are no longer active and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-health-check-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b5786bbbef725badf3dfcc2c2c7a86ead5ebb584c978c47aae8b9a62e241b80d
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:8223bda16ac5ffac4b7407c849c4c30ee006ddafd34cad8a03d2af8314d56f58
        - registry.redhat.io/workload-availability/node-healthcheck-rhel8-operator@sha256:8e1dbc35126101e1025bd4c0eac4f4c614eb779ccafbb55050541a27ba7e9706
        version: 0.3.1
      entries:
      - name: node-healthcheck-operator.v0.3.1
        version: 0.3.1
      name: candidate
    - currentCSV: node-healthcheck-operator.v0.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "remediation.medik8s.io/v1alpha1",
                "kind": "NodeHealthCheck",
                "metadata": {
                  "name": "nodehealthcheck-sample"
                },
                "spec": {
                  "minHealthy": "51%",
                  "remediationTemplate": {
                    "apiVersion": "self-node-remediation.medik8s.io/v1alpha1",
                    "kind": "SelfNodeRemediationTemplate",
                    "name": "self-node-remediation-automatic-strategy-template",
                    "namespace": "openshift-operators"
                  },
                  "selector": {
                    "matchExpressions": [
                      {
                        "key": "node-role.kubernetes.io/worker",
                        "operator": "Exists"
                      }
                    ]
                  },
                  "unhealthyConditions": [
                    {
                      "duration": "300s",
                      "status": "False",
                      "type": "Ready"
                    },
                    {
                      "duration": "300s",
                      "status": "Unknown",
                      "type": "Ready"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          console.openshift.io/plugins: '["node-remediation-console-plugin"]'
          containerImage: registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
          createdAt: "2025-09-01 18:07:56"
          description: Detect failed Nodes and trigger remediation with a remediation
            operator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.9.0 <0.10.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-healthcheck-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeHealthCheck is the Schema for the nodehealthchecks API
            displayName: Node Health Check
            kind: NodeHealthCheck
            name: nodehealthchecks.remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Hardware is imperfect, and software contains bugs. When node level failures such as kernel hangs or dead NICs
          occur, the work required from the cluster does not decrease - workloads from affected nodes need to be
          restarted somewhere.

          However some workloads, such as RWO volumes and StatefulSets, may require at-most-one semantics.
          Failures affecting these kind of workloads risk data loss and/or corruption if nodes (and the workloads
          running on them) are assumed to be dead whenever we stop hearing from them. For this reason it is important
          to know that the node has reached a safe state before initiating recovery of the workload.

          Unfortunately it is not always practical to require admin intervention in order to confirm the node’s true status.
          In order to automate the recovery of exclusive workloads, we provide operators for failure detection
          and remediation.

          ### Failure detection: Node Health Check operator
          The “Node Health Check” (NHC) operator checks each Node’s set of
          NodeConditions (eg. NotReady) against the criteria and thresholds defined in
          NodeHealthCheck configuration. If the Node is deemed to be in a failed
          state, NHC will initiate recovery by using the SIG Cluster API's “External
          Remediation” API to instantiate the configured remediation template which
          specifies the mechanism/controller to be used.

          ### Failure handling: External remediators
          There are multiple remediators for handling node failure that we recommend:
          - Self Node Remediation (SNR)
          - Fence Agents Remediation (FAR)
          - Machine Deletion Remediation (MDR)

          #### Self Node Remediation (SNR)
          SNR uses watchdog timers and heuristics to ensure nodes enter a safe state
          (no longer hosting workloads) within a known and finite period of time,
          before signaling to the system that all Pods on the failed Node are no longer active
          and can be relocated elsewhere.
          In the case of transient errors, the watchdog’s actions will also result in
          the node rebooting and rejoining the cluster - restoring capacity.

          #### Fence Agents Remediation (FAR)
          FAR uses well-known agents to fence unhealthy nodes, and eventually FAR remediates the nodes.
          The remediation includes rebooting the unhealthy node using a fence agent,
          and then evicting workloads from the unhealthy node.

          #### Machine Deletion Remediation (MDR)
          MDR is limited to OpenShift, and it uses Machine API for reprovisioning unhealthy nodes by deleting their machines.
        displayName: Node Health Check Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NHC
        - Self Node Remediation
        - SNR
        - Remediation
        - Fencing
        - medik8s
        - k8s
        links:
        - name: Node Healthcheck Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance
        - name: Source Code
          url: https://github.com/medik8s/node-healthcheck-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-healthcheck-rhel9-operator@sha256:2160f2688e5b098c0bae7597cebcd380019b82ae02f4506d07c2ed85f0b3664d
        - registry.redhat.io/workload-availability/node-remediation-console-rhel9@sha256:cc0f671b126cde10476f8cc2061c6903932e7bf5e3dbf4e8bdda9add70d5deea
        - registry.redhat.io/workload-availability/node-healthcheck-operator-bundle@sha256:d7333e377a5aace0d42e42ca936fa7a84d288b181eda4747f9367a9539f2bedc
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:3d3333285fd6736d11ea830fb7fe1f2b8d3e304d682a876458e18eb1173f271d
        version: 0.10.0
      entries:
      - name: node-healthcheck-operator.v0.10.0
        version: 0.10.0
      - name: node-healthcheck-operator.v0.9.1
        version: 0.9.1
      - name: node-healthcheck-operator.v0.9.0
        version: 0.9.0
      - name: node-healthcheck-operator.v0.8.2
        version: 0.8.2
      - name: node-healthcheck-operator.v0.8.1
        version: 0.8.1
      - name: node-healthcheck-operator.v0.7.0
        version: 0.7.0
      - name: node-healthcheck-operator.v0.6.1
        version: 0.6.1
      name: stable
    defaultChannel: stable
    packageName: node-healthcheck-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Instana
      provider-url: ""
    name: instana-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: instana-agent-operator.v2.1.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "instana.io/v1",
                "kind": "InstanaAgent",
                "metadata": {
                  "name": "instana-agent",
                  "namespace": "instana-agent"
                },
                "spec": {
                  "agent": {
                    "configuration_yaml": "# You can leave this empty, or use this to configure your instana agent.\n# See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.\n",
                    "endpointHost": "ingress-red-saas.instana.io",
                    "endpointPort": "443",
                    "env": {},
                    "key": "replace-key"
                  },
                  "cluster": {
                    "name": "my-cluster"
                  },
                  "opentelemetry": {
                    "grpc": {
                      "enabled": true
                    },
                    "http": {
                      "enabled": true
                    }
                  },
                  "zone": {
                    "name": "edited-zone"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Monitoring,Logging & Tracing,OpenShift Optional
          certified: "false"
          containerImage: icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
          description: Fully automated Application Performance Monitoring (APM) for
            microservices.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/instana/instana-agent-operator
          support: Instana
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: InstanaAgent is the Schema for the agents API
            displayName: Instana Agent
            kind: InstanaAgent
            name: agents.instana.io
            version: v1
          - description: InstanaAgentRemote is the Schema for the agents API
            displayName: Instana Agent Remote
            kind: InstanaAgentRemote
            name: agentsremote.instana.io
            version: v1
        description: |
          # Instana

          Instana is an [APM solution](https://www.instana.com/product-overview/) built for microservices that enables IT Ops to build applications faster and deliver higher quality services by automating monitoring, tracing and root cause analysis. The solution is optimized for [Kubernetes](https://www.instana.com/automatic-kubernetes-monitoring/) and [OpenShift](https://www.instana.com/blog/automatic-root-cause-analysis-for-openshift-applications/).

          ## Instana Agent Operator

          This is the Kubernetes Operator for installing the Instana Agent on Kubernetes or OpenShift.

          ## Prerequisites for OpenShift

          Before the agent will be able to run in OpenShift, you need to perform a couple of extra configuration steps.

          You need to set up a project for the Instana Agent and configure it's permissions.

          The project you create here needs to be the namespace where you create the Instana Agent custom resource that the operator will use to deploy the agent.

          For example, create the `instana-agent` project:

              oc new-project instana-agent

          Then, ensure the `instana-agent` service account is in the privileged security context:

              oc adm policy add-scc-to-user privileged -z instana-agent

          The `instana-agent-remote` service account must be allowed to use any userid:

              oc adm policy add-scc-to-user anyuid -z instana-agent-remote -n instana-agent

          This service account will be created by the operator.

          Now you can proceed with installing the operator for the Instana agent.

          ## Installation and Configuration

          First, install this operator from [OperatorHub.io](https://operatorhub.io/), [OpenShift Container Platform](https://www.openshift.com/), or [OKD](https://www.okd.io/).

          Second, create a custom resource with the agent configuration in the target namespace (for now, this MUST always be the `instana-agent` namespace). The operator will pick up the custom resource and install the Instana agent accordingly.

          The following is a minimal template of the custom resource:

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgent
          metadata:
            name: instana-agent
            namespace: instana-agent
          spec:
            zone:
              name: my-zone # (optional) name of the zone of the host
            cluster:
              name: replace-me # replace with the name of your Kubernetes cluster
            agent:
              key: replace-me # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io # the monitoring ingress endpoint
              endpointPort: "443" # the monitoring ingress endpoint port, wrapped in quotes
              env:
                INSTANA_AGENT_TAGS: example
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://ibm.biz/monitoring-k8s
          ```

          Save the template in a file `instana-agent.yaml` and edit the following values:

          * If your target namespace is not `instana-agent`, replace the `namespace:` accordingly.
          * `agent.key` must be set with your Instana agent key.
          * `agent.endpointHost` must be set with the monitoring ingress endpoint, generally either `saas-us-west-2.instana.io` or `saas-eu-west-1.instana.io`.
          * `agent.endpointPort` must be set with the monitoring ingress port, generally "443" (wrapped in quotes).
          * `zone.name` should be set with the name of the Kubernetes cluster that is be displayed in Instana.

          For advanced configuration, you can edit the contents of the `configuration.yaml` file. View documentation [here](https://ibm.biz/monitoring-k8s).

          Apply the custom resource with `kubectl apply -f instana-agent.yaml`. After some time, you should see `instana-agent` Pods being created on each node of your cluster, and your cluster should show on the infrastructure map on your Instana Web interface.

          In addition to the InstanaAgent custom resource, a remote agent can be deployed. This agent will not monitor workload running inside the cluster, but can deploy a single replica of an agent which can get an independent configuration for running remote monitoring for certain sensors.

          ```yaml
          apiVersion: instana.io/v1
          kind: InstanaAgentRemote
          metadata:
            name: remote-agent
            namespace: "instana-agent"
          spec:
            zone:
              name: edited-zone # name of the zone of the host
            hostname:
              name: use-custom-hostname-for-remote-agent # (optional) give your remote agent pod a custom hostname
            agent:
              key: replace-key # replace with your Instana agent key
              endpointHost: ingress-red-saas.instana.io
              endpointPort: "443"
              env: {}
              configuration_yaml: |
                # You can leave this empty, or use this to configure your instana agent.
                # See https://github.com/instana/instana-agent-operator/blob/main/config/samples/instana_v1_extended_instanaagent.yaml for the extended version.
          ```

          ## Uninstalling

          In order to uninstall the Instana agent, simply remove the custom resource with `kubectl delete -f instana-agent.yaml`.

          ## Source Code

          The Instana agent operator is an open source project hosted on [https://github.com/instana/instana-agent-operator](https://github.com/instana/instana-agent-operator/).
        displayName: Instana Agent Operator
        installModes:
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - APM
        - Instana
        links:
        - name: GitHub Repository
          url: https://github.com/instana/instana-agent-operator
        maintainers:
        - email: support@instana.com
          name: Instana
        maturity: stable
        minKubeVersion: 1.11.0
        provider:
          name: Instana
        relatedImages:
        - registry.connect.redhat.com/instana/instana-agent-operator-bundle@sha256:0064fcdc252a04010e605706be377417de4c9be4289358ec0f331d6a37befb0b
        - icr.io/instana/agent@sha256:ed71e6c3e5bdb155853a2db8c916460e6a867fb02627e3a332186188cbbd98dd
        - icr.io/instana/instana-agent-operator@sha256:fe949ae0ca9cc05ffe3569cd7a6f4fde48f5e3ee64f27e20e0b3102733a6e809
        version: 2.1.35
      entries:
      - name: instana-agent-operator.v2.1.35
        version: 2.1.35
      - name: instana-agent-operator.v2.1.34
        version: 2.1.34
      - name: instana-agent-operator.v2.1.33
        version: 2.1.33
      - name: instana-agent-operator.v2.1.32
        version: 2.1.32
      - name: instana-agent-operator.v2.1.31
        version: 2.1.31
      - name: instana-agent-operator.v2.1.30
        version: 2.1.30
      - name: instana-agent-operator.v2.1.29
        version: 2.1.29
      - name: instana-agent-operator.v2.1.28
        version: 2.1.28
      - name: instana-agent-operator.v2.1.27
        version: 2.1.27
      - name: instana-agent-operator.v2.1.25
        version: 2.1.25
      - name: instana-agent-operator.v2.1.24
        version: 2.1.24
      - name: instana-agent-operator.v2.1.23
        version: 2.1.23
      - name: instana-agent-operator.v2.1.22
        version: 2.1.22
      - name: instana-agent-operator.v2.1.21
        version: 2.1.21
      - name: instana-agent-operator.v2.1.20
        version: 2.1.20
      - name: instana-agent-operator.v2.1.19
        version: 2.1.19
      - name: instana-agent-operator.v2.1.18
        version: 2.1.18
      - name: instana-agent-operator.v2.1.17
        version: 2.1.17
      - name: instana-agent-operator.v2.1.16
        version: 2.1.16
      - name: instana-agent-operator.v2.1.15
        version: 2.1.15
      - name: instana-agent-operator.v2.1.14
        version: 2.1.14
      - name: instana-agent-operator.v2.1.13
        version: 2.1.13
      - name: instana-agent-operator.v2.1.12
        version: 2.1.12
      - name: instana-agent-operator.v2.1.11
        version: 2.1.11
      - name: instana-agent-operator.v2.1.10
        version: 2.1.10
      - name: instana-agent-operator.v2.1.9
        version: 2.1.9
      - name: instana-agent-operator.v2.1.8
        version: 2.1.8
      - name: instana-agent-operator.v2.1.7
        version: 2.1.7
      - name: instana-agent-operator.v2.1.6
        version: 2.1.6
      - name: instana-agent-operator.v2.1.5
        version: 2.1.5
      - name: instana-agent-operator.v2.1.4
        version: 2.1.4
      - name: instana-agent-operator.v2.1.3
        version: 2.1.3
      - name: instana-agent-operator.v2.1.2
        version: 2.1.2
      - name: instana-agent-operator.v2.1.1
        version: 2.1.1
      - name: instana-agent-operator.v2.1.0
        version: 2.1.0
      - name: instana-agent-operator.v2.0.32
        version: 2.0.32
      - name: instana-agent-operator.v2.0.31
        version: 2.0.31
      - name: instana-agent-operator.v2.0.30
        version: 2.0.30
      - name: instana-agent-operator.v2.0.27
        version: 2.0.27
      - name: instana-agent-operator.v2.0.19
        version: 2.0.19
      - name: instana-agent-operator.v2.0.18
        version: 2.0.18
      - name: instana-agent-operator.v2.0.17
        version: 2.0.17
      - name: instana-agent-operator.v2.0.13
        version: 2.0.13
      - name: instana-agent-operator.v2.0.12
        version: 2.0.12
      - name: instana-agent-operator.v2.0.11
        version: 2.0.11
      - name: instana-agent-operator.v2.0.10
        version: 2.0.10
      - name: instana-agent-operator.v2.0.9
        version: 2.0.9
      - name: instana-agent-operator.v2.0.8
        version: 2.0.8
      - name: instana-agent-operator.v2.0.7
        version: 2.0.7
      - name: instana-agent-operator.v2.0.5
        version: 2.0.5
      - name: instana-agent-operator.v2.0.4
        version: 2.0.4
      name: stable
    defaultChannel: stable
    packageName: instana-agent-operator
    provider:
      name: Instana
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-nats-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-nats-operator.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          createdAt: "2023-09-19T12:44:10Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NatsCluster
            name: natsclusters.nats.io
            version: v1alpha2
          - kind: NatsServiceRole
            name: natsserviceroles.nats.io
            version: v1alpha2
        description: Keysight Nats Operator
        displayName: Keysight Nats Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Nats Operator
          url: https://keysight-nats-operator.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-nats-operator@sha256:2ee5dc8814274e83a5f82e89eabd0b643e8cde8e3d555a56535a61d05c72fd75
        - registry.connect.redhat.com/loadcore/keysight-nats-operator-bundle@sha256:89bde716c2aa4dba0738d2481a2c4707c59a748f84349f9fb01cb430e69d7beb
        version: 0.1.21
      entries:
      - name: keysight-nats-operator.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-nats-operator
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-trino-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: trino-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
          description: Stackable Operator for Trino
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/trino-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for TrinoClusterSpec via `CustomResource`
            displayName: trinoclusters.trino.stackable.tech
            kind: TrinoCluster
            name: trinoclusters.trino.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for TrinoCatalogSpec via `CustomResource`
            displayName: trinocatalogs.trino.stackable.tech
            kind: TrinoCatalog
            name: trinocatalogs.trino.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Trino
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - trino
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/trino-operator@sha256:b739261a257826a4b14b51a3762338840642c1fac966ff398fa98ab8b24b46d9
        - registry.connect.redhat.com/stackable/stackable-trino-operator@sha256:b4a6e6cb374ca7226cbd1f90563f0d03669ff1168d999ca2ff5fc7be80cac588
        version: 25.7.0
      entries:
      - name: trino-operator.v25.7.0
        version: 25.7.0
      - name: trino-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-trino-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: ThinkParQ
      provider-url: https://www.thinkparq.com
    name: beegfs-csi-driver-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: beegfs-csi-driver-operator.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "beegfs.csi.netapp.com/v1",
                "kind": "BeegfsDriver",
                "metadata": {
                  "name": "csi-beegfs-cr"
                },
                "spec": {
                  "nodeAffinityControllerService": {
                    "preferredDuringSchedulingIgnoredDuringExecution": [
                      {
                        "preference": {
                          "matchExpressions": [
                            {
                              "key": "node-role.kubernetes.io/master",
                              "operator": "Exists"
                            }
                          ]
                        },
                        "weight": 50
                      }
                    ],
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  },
                  "nodeAffinityNodeService": {
                    "requiredDuringSchedulingIgnoredDuringExecution": {
                      "nodeSelectorTerms": [
                        {
                          "matchExpressions": [
                            {
                              "key": "node.openshift.io/os_id",
                              "operator": "NotIn",
                              "values": [
                                "rhcos"
                              ]
                            }
                          ]
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
          createdAt: 2021-06-30:03:11:00
          description: Install and manage the BeeGFS CSI driver
          operatorframework.io/suggested-namespace: beegfs-csi
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ThinkParQ/beegfs-csi-driver
          support: ThinkParQ
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys the BeeGFS CSI driver
            displayName: BeeGFS Driver
            kind: BeegfsDriver
            name: beegfsdrivers.beegfs.csi.netapp.com
            version: v1
        description: |
          The BeeGFS Container Storage Interface (CSI) driver provides high performing and scalable storage for workloads
          running in container orchestrators like Kubernetes. It allows containers to access existing datasets or request
          on-demand ephemeral or persistent high speed storage backed by BeeGFS parallel file systems.

          This operator eases the installation and management of the BeeGFS CSI driver in an OpenShift or other Operator
          Lifecycle Manager enabled cluster.

          ### Documentation

          Compatibility information and a quick start guide for the BeeGFS CSI driver itself are available on our [main GitHub
          page](https://www.github.com/ThinkParQ/beegfs-csi-driver). Additional detailed documentation for the driver is
          available in the [docs directory](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/docs) and for the
          operator is available in the [operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md).

          ### Important Notes

          The BeeGFS client packages and kernel module must be pre-installed on all nodes that will run the driver. See the
          driver ["Getting Started" guide](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#getting-started)
          for additional information.

          The BeeGFS CSI driver is only supported on nodes running a [BeeGFS supported Linux
          distribution](https://doc.beegfs.io/latest/release_notes.html#supported-linux-distributions-and-kernels). Red Hat
          CoreOS (RHCOS), the default distribution in OpenShift environments, is NOT supported. However, Red Hat Enterprise
          Linux (RHEL) nodes can be added to an OpenShift cluster and RHEL is supported for running BeeGFS. By default, the
          operator will install the driver with a node affinity that ensures it does not run on RHCOS nodes. See the driver
          [compatibility
          matrix](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/README.md#interoperability-and-csi-feature-matrix)
          for additional information.

          We have recently published experimental manifests that enable the BeeGFS CSI driver on OpenShift nodes running
          RHCOS. These manifests are not intended for production use. See [this
          README](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/deploy/openshift-beegfs-client/README.md) for more
          details.

          All driver configuration can be specified through the BeegfsDrivers API (in a BeegfsDriver Custom Resource)
          except for BeeGFS connAuth information, which must be placed separately into a Kubernetes Secret. See the [ConnAuth
          Configuration
          section](https://github.com/ThinkParQ/beegfs-csi-driver/blob/master/operator/README.md#connauth-configuration) of the
          operator documentation for additional information.

          Starting with version 1.4.0 of the BeeGFS CSI driver the driver and operator images will be signed using the Cosign
          tool. Refer to the [BeeGFS CSI driver operator
          README](https://www.github.com/ThinkParQ/beegfs-csi-driver/tree/master/operator/README.md)
          for details on how to verify the operator and driver image signatures.
        displayName: BeeGFS CSI Driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - CSI
        - BeeGFS
        - ThinkParQ
        - HPC
        - AI
        - High-Performance Computing
        links:
        - name: BeeGFS introduction
          url: https://www.beegfs.io/c/
        - name: BeeGFS documentation
          url: https://doc.beegfs.io/latest/index.html
        - name: BeeGFS CSI driver source code
          url: https://github.com/ThinkParQ/beegfs-csi-driver
        - name: BeeGFS blog posts
          url: https://www.beegfs.io/c/blog/
        maintainers:
        - email: csi-driver@beegfs.io
          name: ThinkParQ
        maturity: stable
        minKubeVersion: 1.19.0
        provider:
          name: ThinkParQ
          url: https://www.thinkparq.com
        relatedImages:
        - ghcr.io/thinkparq/beegfs-csi-driver-operator:v1.7.0
        - quay.io/community-operator-pipeline-prod/beegfs-csi-driver-operator@sha256:1a9f2aab026c2d1549be5b4624c56d65e813c619fdd5beefb062d5c9fb8981b8
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        version: 1.7.0
      entries:
      - name: beegfs-csi-driver-operator.v1.7.0
        version: 1.7.0
      - name: beegfs-csi-driver-operator.v1.6.0
        version: 1.6.0
      - name: beegfs-csi-driver-operator.v1.5.0
        version: 1.5.0
      - name: beegfs-csi-driver-operator.v1.4.0
        version: 1.4.0
      - name: beegfs-csi-driver-operator.v1.3.0
        version: 1.3.0
      - name: beegfs-csi-driver-operator.v1.2.2
        version: 1.2.2
      - name: beegfs-csi-driver-operator.v1.2.1
        version: 1.2.1
      - name: beegfs-csi-driver-operator.v1.2.0
        version: 1.2.0
      name: stable
    defaultChannel: stable
    packageName: beegfs-csi-driver-operator
    provider:
      name: ThinkParQ
      url: https://www.thinkparq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: m88i Labs
      provider-url: ""
    name: nexus-operator-m88i
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.m88i.io/v1alpha1",
                "kind": "Nexus",
                "metadata": {
                  "name": "nexus3"
                },
                "spec": {
                  "networking": {
                    "expose": false
                  },
                  "persistence": {
                    "persistent": false
                  },
                  "replicas": 1,
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    },
                    "requests": {
                      "cpu": "1",
                      "memory": "2Gi"
                    }
                  },
                  "useRedHatImage": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/m88i/nexus-operator:0.6.0
          createdAt: "2019-11-16T13:12:22Z"
          description: Nexus Operator to deploy and manage Nexus 3.x servers
          operators.operatorframework.io/builder: operator-sdk-v1.4.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/m88i/nexus-operator
          support: m88i Labs
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: Nexus
            kind: Nexus
            name: nexus.apps.m88i.io
            version: v1alpha1
        description: |-
          Creates a new Nexus 3.x deployment in a Kubernetes cluster. Will help DevOps to have a quick Nexus application exposed to the world that can be used in a CI/CD process:

          * Deploys a new Nexus 3.x server based on either Community or Red Hat images
          * Creates an [Ingress](https://kubernetes.io/docs/concepts/services-networking/ingress/) in Kubernetes (1.14+) environments to expose the application to the world
          * On OpenShift, creates a Route to expose the service outside the cluster
          * Automatically creates Apache, Red Hat and JBoss Maven repositories
          * Automatically updates Nexus within the same minor version

          [See our documentation](https://github.com/m88i/nexus-operator/blob/main/README.md) for more installation and usage scenarios.

          If you experience any issues or have any ideas for new features, please [file an issue in our Github repository](https://github.com/m88i/nexus-operator/issues) or send an email to our maillist: [nexus-operator@googlegroups.com](mailto:nexus-operator@googlegroups.com)

          *Please note that the operator is an individual work and it's not provided nor supported by Sonatype.*
        displayName: Nexus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nexus
        - sonatype
        - maven
        - docker
        - ci
        - continuous integration
        - continuous delivery
        - repository
        - repository manager
        - dev tools
        links:
        - name: Documentation
          url: https://github.com/m88i/nexus-operator/blob/main/README.md
        - name: Nexus Operator source code repository
          url: https://github.com/m88i/nexus-operator
        maintainers:
        - email: nexus-operator@googlegroups.com
          name: m88i Labs
        maturity: alpha
        provider:
          name: m88i Labs
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/m88i/nexus-operator:0.6.0
        - quay.io/openshift-community-operators/nexus-operator-m88i@sha256:46cd06d477b4d09dbfca7ca4434ea58f881314138ed744c6390033bd6fb7dc0f
        version: 0.6.0
      entries:
      - name: nexus-operator.v0.6.0
        version: 0.6.0
      name: alpha
    defaultChannel: alpha
    packageName: nexus-operator-m88i
    provider:
      name: m88i Labs
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: ""
    name: ovms-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openvino-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "ModelServer",
                "metadata": {
                  "name": "model-server-sample"
                },
                "spec": {
                  "image_name": "registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0",
                  "deployment_parameters": {
                    "replicas": 1
                  },
                  "service_parameters": {
                    "grpc_port": 8080,
                    "rest_port": 8081
                  },
                  "models_settings": {
                    "single_model_mode": true,
                    "config_configmap_name": "",
                    "config_path": "",
                    "model_name": "resnet",
                    "model_path": "gs://<bucket_name>/<model_dir>",
                    "nireq": 0,
                    "plugin_config": "{\"PERFORMANCE_HINT\":\"LATENCY\"}",
                    "batch_size": "",
                    "shape": "",
                    "model_version_policy": "{\"latest\": { \"num_versions\":1 }}",
                    "layout": "",
                    "target_device": "CPU",
                    "is_stateful": false,
                    "idle_sequence_cleanup": false,
                    "low_latency_transformation": true,
                    "max_sequence_number": 0
                  },
                  "server_settings": {
                    "file_system_poll_wait_seconds": 0,
                    "sequence_cleaner_poll_wait_minutes": 0,
                    "log_level": "INFO",
                    "grpc_workers": 1,
                    "rest_workers": 0
                  },
                  "models_repository": {
                    "https_proxy": "",
                    "http_proxy": "",
                    "storage_type": "google",
                    "models_host_path": "",
                    "models_volume_claim": "",
                    "aws_secret_access_key": "",
                    "aws_access_key_id": "",
                    "aws_region": "",
                    "s3_compat_api_endpoint": "",
                    "gcp_creds_secret_name": "",
                    "azure_storage_connection_string": ""
                  }
                }
              },
              {
                "apiVersion": "intel.com/v1alpha1",
                "kind": "Notebook",
                "metadata": {
                  "name": "v2024.1"
                },
                "spec": {
                  "git_ref": "2024.1",
                  "auto_update_image": false
                }
              }
            ]
          alm-examples-metadata: '{ "model-server-sample":{"description":"Example
            of the model server configuration"}, "v2024":{"description":"Default notebook
            configuration"} }'
          capabilities: Basic Install
          categories: AI/Machine Learning
          containerImage: registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
          createdAt: "2024-05-14T09:29:35Z"
          description: OpenVINO Toolkit Operator manages OpenVINO components in OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/ovms-operator-rhmp/support?utm_source=openshift_console
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.34.1+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Provides inference serving through gRPC & REST
            displayName: Model Server
            kind: ModelServer
            name: modelservers.intel.com
            version: v1alpha1
          - description: Adds a collection of ready-to-run Jupyter notebooks for learning
              and experimenting with the OpenVINO™ Toolkit to Red Hat® OpenShift®
              Data Science
            displayName: Notebook
            kind: Notebook
            name: notebooks.intel.com
            version: v1alpha1
        description: |
          OpenVINO Toolkit Operator manages OpenVINO components in OpenShift.

          Currently there available components are ModelServer and Notebook.

          # Model Server
          [OpenVINO™ Model Server](https://github.com/openvinotoolkit/model_server) (OVMS) is a scalable, high-performance solution for serving machine learning models optimized for Intel® architectures. The server provides an inference service via gRPC or REST API for any models trained in a framework that is supported by [OpenVINO](https://software.intel.com/en-us/openvino-toolki://docs.openvino.ai/latest/index.html).
          Model Server configuration parameters is explained [here](https://github.com/openvinotoolkit/operator/blob/main/docs/modelserver_params.md).
          ## Using the cluster
          OpenVINO Model Server can be consumed as a `Service`. It is called like with `ModelServer` resource with `-ovms` suffix.
          The suffix is omitted when `ovms` phrase is included in the name.
          The service exposes gRPC and REST API interfaces to run inference requests.
          ```
          oc get pods
          NAME                                        READY   STATUS    RESTARTS   AGE
          model-server-sample-ovms-586f6f76df-dpps4   1/1     Running   0          8h

          oc get services
          NAME                       TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
          model-server-sample-ovms   ClusterIP   172.25.199.210   <none>        8080/TCP,8081/TCP   8h
          ```

          # Notebook
          The Notebook resource integrates JupyterHub from OpenShift Data Science or Open Data Hub with a container image that includes [developer
          tools](https://github.com/openvinotoolkit/openvino/blob/master/docs/install_guides/pypi-openvino-dev.md)
          from the OpenVINO toolkit and a set of Jupyter notebook tutorials. It enables selecting a defined image `OpenVINO™ Toolkit` from
          the Jupyter Spawner choice list.

          Create the `Notebook` resource in the same project with JupyterHub and RedHat OpenShift Data Science operator as seen in [documentation](https://github.com/openvinotoolkit/operator/blob/main/docs/notebook_in_rhods.md).
          It builds the image in the cluster based on Dockerfile from [openvino_notebooks](https://github.com/openvinotoolkit/openvino_notebooks).

          ## References
          OpenVINO Model Server on [Github](https://github.com/openvinotoolkit/model_server)

          OpenVINO Model Server Operator on [Github](https://github.com/openvinotoolkit/operator)
        displayName: OpenVINO Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AI Inference OpenVINO
        links:
        - name: OpenVINO Toolkit Operator
          url: https://github.com/openvinotoolkit/operator
        - name: OpenVINO Model Server
          url: https://github.com/openvinotoolkit/model_server
        - name: OpenVINO Notebooks
          url: https://github.com/openvinotoolkit/openvino_notebooks
        maintainers:
        - email: openvino-operator-support@intel.com
          name: Operator Support
        maturity: alpha
        provider:
          name: Intel
        relatedImages:
        - registry.connect.redhat.com/intel/openvino-model-server@sha256:0817c19cfaf2eef85e98b1d1f43f14ad3eacd873b36e7ee3740810ae3b5f60d0
        - registry.connect.redhat.com/intel/ovms-operator-rhmp-bundle@sha256:b66e3cfad24cf28e0a738179cab062667b0103c52cf2e84abb88d5e018e12b2c
        - registry.connect.redhat.com/intel/ovms-operator@sha256:9049f5d4e3c764fdca057b60e41ea62c3ca49207433f593668091850d27977d4
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:da5714b93ac39277affdc10c3ba142d8c032ed0ddb2ba61c04cbddac710464c4
        version: 1.2.0
      entries:
      - name: openvino-operator.v1.2.0
        version: 1.2.0
      - name: openvino-operator.v1.1.0
        version: 1.1.0
      - name: openvino-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: ovms-operator-rhmp
    provider:
      name: Intel
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:27c5d91d62371f3cf0dea4d50044d8a6e0a6015a453c4216a030683ea2474c64
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: koku-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: alpha
    - currentCSV: koku-metrics-operator.v4.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
                "kind": "CostManagementMetricsConfig",
                "metadata": {
                  "name": "costmanagementmetricscfg-sample-v1beta1"
                },
                "spec": {
                  "authentication": {
                    "type": "token"
                  },
                  "packaging": {
                    "max_reports_to_store": 30,
                    "max_size_MB": 100
                  },
                  "prometheus_config": {
                    "collect_previous_data": true,
                    "context_timeout": 120,
                    "disable_metrics_collection_cost_management": false,
                    "disable_metrics_collection_resource_optimization": false
                  },
                  "source": {
                    "check_cycle": 1440,
                    "create_source": false,
                    "name": ""
                  },
                  "upload": {
                    "upload_cycle": 360,
                    "upload_toggle": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
          createdAt: "2025-09-25T19:04:35Z"
          description: A Golang-based OpenShift Operator that generates and uploads
            OpenShift usage metrics to cost management.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "costmanagement-metrics-cfg.openshift.io/v1beta1",
              "kind": "CostManagementMetricsConfig",
              "metadata": {
                "name": "costmanagementmetricscfg-sample-v1beta1"
              },
              "spec": {
                "authentication": {
                  "token_url": "https://sso.redhat.com/auth/realms/redhat-external/protocol/openid-connect/token",
                  "type": "token"
                },
                "api_url": "https://console.redhat.com/",
                "packaging": {
                  "max_reports_to_store": 30,
                  "max_size_MB": 100
                },
                "upload": {
                  "ingress_path": "/api/ingress/v1/upload",
                  "upload_cycle": 360,
                  "upload_toggle": true,
                  "validate_cert": true
                },
                "prometheus_config": {
                  "collect_previous_data": true,
                  "context_timeout": 120,
                  "disable_metrics_collection_cost_management": false,
                  "disable_metrics_collection_resource_optimization": false,
                  "service_address": "https://thanos-querier.openshift-monitoring.svc:9091",
                  "skip_tls_verification": false
                },
                "source": {
                  "check_cycle": 1440,
                  "create_source": false,
                  "name": "",
                  "sources_path": "/api/sources/v1.0/"
                }
              }
            }
          operatorframework.io/suggested-namespace: koku-metrics-operator
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/project-koku/koku-metrics-operator
          support: Cost Management
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CostManagementMetricsConfig is the Schema for the costmanagementmetricsconfig
              API
            displayName: Cost Management Metrics Config
            kind: CostManagementMetricsConfig
            name: costmanagementmetricsconfigs.costmanagement-metrics-cfg.openshift.io
            version: v1beta1
        description: |
          # Koku Metrics Operator
          ## Introduction
          The `koku-metrics-operator` is a component of the [cost managment](https://access.redhat.com/documentation/en-us/cost_management_service) service for Openshift. The operator runs on the latest supported versions of Openshift. This operator obtains OpenShift usage data by querying Prometheus every hour to create metric reports that it uploads to Cost Management at [console.redhat.com](https://console.redhat.com) to be processed. For more information, reach out to <costmanagement@redhat.com>.

          This operator is capable of functioning within a disconnected/restricted network (aka air-gapped mode). In this mode, the operator will store the packaged reports for manual retrieval instead of being uploaded to Cost Management. Documentation for installing an operator within a restricted network can be found [here](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html).

          ## Features and Capabilities
          #### Metrics collection:
          The Koku Metrics Operator (`koku-metrics-operator`) collects the metrics required for Cost Management by:
          * Querying Prometheus to gather the necessary metrics for Cost Management.
          * Writing the results of Prometheus queries to CSV report files.
          * Packaging the CSV report files into tarballs.

          #### Additional Capabilities:
          * Resource Optimization metrics collection.
          * The operator can be configured to gather all previous data within the configured retention period or a maximum of 90 days. The default data collection period is the 14 previous days. This setting is only applicable to newly created KokuMetricsConfigs.
          * The operator can be configured to automatically upload the packaged reports to Cost Management through Red Hat Insights Ingress service.
          * The operator can create an integration in console.redhat.com. An integration is required for Cost Management to process the uploaded packages.
          * PersistentVolumeClaim (PVC) configuration: The KokuMetricsConfig CR can accept a PVC definition and the operator will create and mount the PVC. If one is not provided, a default PVC will be created.
          * Restricted network installation: this operator can function on a restricted network. In this mode, the operator stores the packaged reports for manual retrieval.

          ## New in v4.2.0:
          * Enabled gathering and reporting NVIDIA GPU metrics for resource optimization.
          * (Bugfix) Fixed custom CA certificate validation and configuration issues.
          * (Bugfix) Fixed issue where ROS queries were missing workload metadata for namespaces labeled only with the legacy `insights_cost_management_optimizations` label.

          ## New in v4.1.0:
          * Enabled gathering and reporting namespace metrics to provide more granular data for resource optimization.
          * Updated CSV report filenames to be more descriptive, replacing the previous generic UID-based names with content-reflective names.
          * Introduced a new generic label, `cost_management_optimizations`, for resource optimization while ensuring backward compatibility with the older `insights_cost_management_optimizations` label.

          ## New in v4.0.0:
          * **Virtual Machine Metrics:** Adds capabilities for collecting metrics and generating reports for running virtual machines within your OpenShift environment.
          * **Progress towards FIPS 140 Compliance:** The operator is using the Go Cryptographic Module v1.0.0 to progress toward FIPS 140 compliance. Although this module is not validated at the time of this release, the operator aims to meet stricter security standards when the module does successfully achieve FIPS validation.
          * **API Name Change:** The Custom Resource Definition (CRD) for configuring the upstream `Koku Metrics Operator` was renamed from `KokuMetricsConfig` to `CostManagementMetricsConfig`.

          ### NOTE:
            * The API name change impacts only users of the upstream (community) Koku Metrics Operator.
            * The downstream (Red Hat-supported) Cost Management Metrics Operator is not impacted by this change and users should continue using their existing configurations.

          **Important Upgrade Instructions:**

            If you are upgrading the upstream Koku Metrics Operator to version 4.0.0 or higher, you must manually migrate your configuration. The operator will no longer recognize existing `KokuMetricsConfig` resources.

            To successfully upgrade and retain your operator's configuration, complete the following steps. You can use the provided `oc` commands replacing the angle brackets (`< >`) with your specific values:

            1. Backup existing configuration (recommended):

               Before you upgrade, retrieve the details of your current `KokuMetricsConfig` instance to help you recreate it with the new API name.

               ```
               oc get kokumetricsconfig -n koku-metrics-operator <your-config-name> -o yaml > koku-metrics-config-backup.yaml
               ```

            2. Delete the previous `KokuMetricsConfig` instance (recommended):

                After the operator is upgraded to `version 4.0.0`, delete any existing `KokuMetricsConfig` resources.

                ```
                oc delete kokumetricsconfig -n koku-metrics-operator <your-config-name>
                ```

            3. Create the new `CostManagementMetricsConfig` instance:

                Using the information from your backup (or a new configuration), create a new `CostManagementMetricsConfig` instance. Adjust the spec section based on your specific needs. For more information about configurations, [refer to the configurable parameters](#configurable-parameters).

            4. Apply your new configuration:

                ```
                oc apply -f <your-costmanagementmetricsconfig-file>.yaml -n koku-metrics-operator
                ```

          ## New in v3.3.2:
          * Leader election settings are now configurable via environment variables. These variables can be modified in the operator [Subscription](https://github.com/operator-framework/operator-lifecycle-manager/blob/5a01f50258003e248bd5630df0837fe0bb0f1cb7/doc/design/subscription-config.md). The values should be specified as durations in seconds in the format `<number-of-seconds>s`. The default values for `LEADER_ELECTION_LEASE_DURATION`, `LEADER_ELECTION_RENEW_DEADLINE`, and `LEADER_ELECTION_RETRY_PERIOD` are '60s', '30s', and '5s', respectively.

            ```
            kind: Subscription
            metadata:
            ...
            spec:
              ...
              config:
                env:
                - name: LEADER_ELECTION_LEASE_DURATION
                  value: "60s"
                - name: LEADER_ELECTION_RENEW_DEADLINE
                  value: "30s"
                - name: LEADER_ELECTION_RETRY_PERIOD
                  value: "5s"
            ```

          ## New in v3.3.1:
          * Optimize memory usage when reading CSV files.

          ## New in v3.3.0:
          * Storage reports now contain `node`, `csi_driver`, and `csi_volume_handle` fields.
          * The PVC capacity is now populated using the `kube_persistentvolume_capacity_bytes` metric instead of `kubelet_volume_stats_capacity_bytes`.
          * To receive resource optimization recommendations for your namespaces, you must now first enable each namespace. To enable a namespace, label it with `insights_cost_management_optimizations='true'`. In the CLI, run:
            ```
              oc label namespace NAMESPACE insights_cost_management_optimizations="true" --overwrite=true
            ```
          * __DEPRECATION NOTICE__: Basic authentication is deprecated and will not be supported beyond December 31, 2024. If the default token authentication method cannot be used, you must switch to [service account](https://console.redhat.com/iam/service-accounts) authentication ([more on creating a service account](https://access.redhat.com/articles/7036194)). Once you have created your service account, follow [this documentation](https://access.redhat.com/documentation/en-us/cost_management_service/1-latest/html-single/integrating_openshift_container_platform_data_into_cost_management/index#service-account-authentication_adding-an-ocp-int) on how to configure your operator to use service account authentication. Service-accounts must also be used if manually uploading payloads to console.redhat.com.

          ## New in v3.2.1:
          * The minimum supported configuration for `upload_cycle` is now 60 (minutes).
          * (Bugfix) many-to-many matching not allowed query fix.
          * (Bugfix) Sequentially collect data during initial install.

          ## New in v3.2.0:
          * Support for amd64, arm64, ppc64le, s390x architectures.
          * add liveness and readiness probes to controller Pod.
          * update pod security settings so that the controller Pod can run in Restricted mode [more info](https://sdk.operatorframework.io/docs/best-practices/pod-security-standards/).

          ## New in v3.1.0:
          * Add service-account authentication type.
          * __Deprecation Notice:__ Basic authentication is deprecated and will be removed in a future version of the operator.

          ## New in v3.0.0:
          * Daily report generation: Operator versions prior to v3.0.0 generated sequential reports. Now, reports are generated starting at 0:00 UTC. Any payloads generated throughout a given day will contain all data starting from 0:00 UTC. Once the next day starts, the previous day's reports are packaged, and the new report again starts at 0:00 UTC for the current day.
          * Failed query retry: In an attempt to prevent missing data, the operator will retry queries from the last successful query time, up to 5 times.

          ## New in v2.0.0:
          * Adds metrics and report generation for resource optimization. This feature will collect additional usage metrics and create a new report in the payload. These metrics are enabled by default, but can be disabled by setting `disable_metrics_collection_resource_optimization` to `true`.
          * Collect all available Prometheus data upon CR creation. This feature only applies to newly created KokuMetricsConfigs. The operator will check the monitoring stack configuration in the `openshift-monitoring` namespace. The operator will use the `retention` period set in the `cluster-monitoring-config` ConfigMap if defined, up to a maximum of 90 days. Otherwise it will fall back to collecting 14 days of data, if available. This data collection may be disabled by setting `collect_previous_data` to `false`. Turning this feature off results in the operator collecting metrics from the time the KokuMetricsConfig is created, forward.

          ## Limitations and Pre-Requisites
          #### Limitations (Potential for metrics data loss)
          * An integration **must** exist in console.redhat.com for an uploaded payload to be processed by Cost Management. The operator sends the payload to the Red Hat Insights Ingress service which usually returns successfully, but the operator does not currently confirm with Cost Management that the payload was processed. After Ingress accepts the uploaded payload, it is deleted from the operator. If the data within the payload is not processed, a gap will be introduced in the usage metrics. Data may be recollected by deleting the `KokuMetricsConfig`, creating a new `KokuMetricsConfig`, and setting `collect_previous_data: true`. This re-collection of data will gather all data stored in Prometheus, up to 90 days.

          **Note** The following limitations are specific to operators configured to run in a restricted network:
          * The `koku-metrics-operator` will not be able to generate new reports if the PVC storage is full. If this occurs, the reports must be manually deleted from the PVC so that the operator can function as normal.
          * The default report retention is 30 reports (about one week's worth of data). The reports must be manually downloaded and uploaded to console.redhat.com every week, or they will be deleted and the data will be lost.

          #### Storage configuration prerequisite
          The operator will attempt to create and use the following PVC when installed:

                volume_claim_template:
                  apiVersion: v1
                  kind: PersistentVolumeClaim
                  metadata:
                    name: koku-metrics-operator-data
                  spec:
                    accessModes:
                      - ReadWriteOnce
                    resources:
                      requests:
                        storage: 10Gi

          If a different PVC should be utilized, a valid PVC should be specified in the KokuMetricsConfig CR as described in the appropriate section below. The PVC to be used may exist already, or the operator will attempt to create it.

          To use the default specification, the follow assumptions must be met:
          1. A default StorageClass is defined.
          2. Dynamic provisioning for that default StorageClass is enabled.

          If these assumptions are not met, the operator will not deploy correctly. In these cases, storage must be manually configured. After configuring storage, a valid PVC template should be supplied in the `volume_claim_template` spec of the KokuMetricsConfig CR.

          ## Configurable parameters:
            * `authentication`:
              * `type: token` -> The authentication method for connecting to `console.redhat.com`. The default and preferred method is `token`. `basic` (deprecated) and `service-account` authentication methods are used when the openshift-config pull-secret does not contain a token for `console.redhat.com`.
              * `secret_name` -> The Secret used by the operator when the authentication type is `basic` (deprecated) or `service-account`. This parameter is required **only if** the authentication type is `basic` (deprecated) or `service-account`.
            * `packaging`:
              * `max_reports_to_store: 30` -> The number of reports to store when configured in air-gapped mode. The default is 30, with a minimum of 1 and no maximum. When the operator is not configured in air-gapped mode, this parameter has no effect. Reports are removed as soon as they are uploaded.
              * `max_size: 100` -> The maximum size for packaged files in Megabytes prior to compression. The default is 100, with a minimum of 1 and maximum of 100.
            * `prometheus_config`:
              * `collect_previous_data: true` -> Toggle for collecting all available data in Prometheus **upon KokuMetricsConfig creation** (This parameter will start to appear in KokuMetricsConfigs that were created prior to v2.0.0 but will not have any effect unless the KokuMetricsConfig is deleted and recreated). The default is `true`. The operator will first look for a `retention` period in the `cluster-monitoring-config` ConfigMap in the `openshift-monitoring` namespace and gather data over this time period up to a maximum of 90 days. If this configuration is not set, the default is 14 days. (New in v2.0.0)
              * `disable_metrics_collection_cost_management: false` -> Toggle for disabling the collection of metrics for Cost Management. The default is false. (New in v2.0.0)
              * `disable_metrics_collection_resource_optimization: false` -> Toggle for disabling the collection of metrics for Resource Optimization. The default is false. (New in v2.0.0)
              * `context_timeout: 120` -> The time in seconds before Prometheus queries timeout due to exceeding context timeout. The default is 120, with a minimum of 10 and maximum of 180.
            * `source`:
              * `name: ''` -> The name of the integration the operator will create in `console.redhat.com`. If the name value is empty, the default intergration name is the **cluster id**.
              * `create_source: false` -> Toggle for whether or not the operator will create the integration in `console.redhat.com`. The default is False. This parameter should be switched to True when an integration does not already exist in `console.redhat.com` for this cluster.
              * `check_cycle: 1440` -> The time in minutes to wait between checking if an integration exists for this cluster. The default is 1440 minutes (24 hrs).
            * `upload`:
              * `upload_cycle: 360` -> The time in minutes between payload uploads. The default is 360 (6 hours), minimum is 60 (1 hour).
              * `upload_toggle: true` -> Toggle to turn upload on or off -> true means upload, false means do not upload (false == air-gapped mode). The default is `true`.
              * `upload_wait` -> The amount of time (in seconds) to pause before uploading a payload. The default is a random number between 0 and 35. This is used to decrease service load, but may be set to `0` if desired.
            * `volume_claim_template` -> see the "Storage configuration prerequisite" section above.

          ## Configure the koku-metrics-operator
          **Note** There are separate instructions for configuring the `koku-metrics-operator` to run in a restricted network.
          ##### Configure authentication
          The default authentication for the operator is `token`. No further steps are required to configure token authentication. If `basic` (deprecated) or `service-account` is the preferred authentication method, a Secret which holds the credentials must be created:
          1. On the left navigation pane, select `Workloads` -> `Secrets` -> select Project: `koku-metrics-operator` -> `Create` -> `Key/Value Secret`
          2. Give the Secret a name and add 2 keys (all lowercase) for the respective authentication type. The values for these keys correspond to console.redhat.com credentials:
              * basic auth (deprecated): `username` and `password`
              * service-account auth: `client_id` and `client_secret`

          3. Select `Create`.
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. For `basic` (deprecated) or `service-account` authentication, edit the following values in the spec:
              * Replace `authentication: type:` with `basic` or `service-account`.
              * Add the `secret_name` field under `authentication`, and set it equal to the name of the authentication Secret that was created above. The spec should look similar to the following:

                  * for basic auth type (deprecated)
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: basic
                  ```

                  * for service-account auth type
                  ```
                    authentication:
                      secret_name: SECRET-NAME
                      type: service-account
                  ```

          3. To configure the koku-metrics-operator to create a cost management integration, edit the following values in the `source` field:
              * Replace the `name` field value with the preferred name of the integration to be created.
              * Replace the `create_source` field value with `true`.

              **Note:** if the integration already exists, replace the empty string value of the `name` field with the existing name, and leave `create_source` as false. This will allow the operator to confirm that the integration exists.

          4. If not specified, the operator will create a default PersistentVolumeClaim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec:

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          5. Select `Create`.

          # Restricted Network Usage (disconnected/air-gapped mode)
          ## Installation
          To install the `koku-metrics-operator` in a restricted network, follow the [olm documentation](https://docs.openshift.com/container-platform/latest/operators/admin/olm-restricted-networks.html). The operator is found in the `community-operators` Catalog in the `registry.redhat.io/redhat/community-operator-index:latest` Index. If pruning the index before pushing to the mirrored registry, keep the `koku-metrics-operator` package.

          Within a restricted network, the operator queries prometheus to gather the necessary usage metrics, writes the query results to CSV files, and packages the reports for storage in the PVC. These reports then need to be manually downloaded from the cluster and uploaded to [console.redhat.com](https://console.redhat.com).

          ## Configure the koku-metrics-operator for a restricted network
          ##### Create the KokuMetricsConfig
          Configure the koku-metrics-operator by creating a `KokuMetricsConfig`.
          1. On the left navigation pane, select `Operators` -> `Installed Operators` -> `koku-metrics-operator` -> `KokuMetricsConfig` -> `Create Instance`.
          2. Specify the desired storage. If not specified, the operator will create a default Persistent Volume Claim called `koku-metrics-operator-data` with 10Gi of storage. To configure the koku-metrics-operator to use or create a different PVC, edit the following in the spec:
              * Add the desired configuration to the `volume_claim_template` field in the spec (below is only a template. Any _valid_ `PersistentVolumeClaim` may be defined here):

                ```
                  volume_claim_template:
                    apiVersion: v1
                    kind: PersistentVolumeClaim
                    metadata:
                      name: <insert-name>
                    spec:
                      storageClassName: <insert-class-name>
                      accessModes:
                        - ReadWriteOnce
                      resources:
                        requests:
                          storage: 10Gi
                ```

              **Note:** If using the YAML View, the `volume_claim_template` field must be added to the spec
          3. (Optional) Specify the desired report retention. The operator will retain 30 reports by default. This corresponds to approximately one week's worth of data if using the default packaging cycle. To modify the number of retained reports:
              * Change the `packaging` spec field `max_reports_to_store` to the desired number of reports to retain. Once this max number is reached, the operator will start removing the oldest packages remaining on the PVC:

                ```
                  packaging:
                    max_size_MB: 100
                    max_reports_to_store: 30
                ```

              **Note:** The number of retained reports directly affects the frequency that reports must be manually downloaded from the PVC. Take caution in setting this to a higher number of reports, as the operator cannot write data to the PVC if the storage is full.
          4. To configure the operator to perform in a restricted network, set the `upload_toggle` to `false`:

            ```
              upload:
                upload_cycle: 360
                upload_toggle: false
            ```

          5. Select `Create`.

          ## Download reports from the Operator & clean up the PVC
          If the `koku-metrics-operator` is configured to run in a restricted network, the metric reports will not automatically upload to cost managment. Instead, they need to be manually copied from the PVC for upload to [console.redhat.com](https://console.redhat.com). The default configuration saves one week of reports which means the process of downloading and uploading reports should be repeated weekly to prevent loss of metrics data. To download the reports, complete the following steps:
          1. Create the following Pod, ensuring the `claimName` matches the PVC containing the report data:

            ```
              kind: Pod
              apiVersion: v1
              metadata:
                name: volume-shell
                namespace: koku-metrics-operator
                labels:
                  app: koku-metrics-operator
              spec:
                volumes:
                - name: koku-metrics-operator-reports
                  persistentVolumeClaim:
                    claimName: koku-metrics-operator-data
                containers:
                - name: volume-shell
                  image: busybox
                  command: ['sleep', 'infinity']
                  volumeMounts:
                  - name: koku-metrics-operator-reports
                    mountPath: /tmp/koku-metrics-operator-reports
            ```

          2. Use rsync to copy all of the files ready for upload from the PVC to a local folder:

            ```
            $ oc rsync volume-shell:/tmp/koku-metrics-operator-reports/upload local/path/to/save/folder
            ```

          3. Once confirming that the files have been successfully copied, use rsh to connect to the pod and delete the contents of the upload folder so that they are no longer in storage:

            ```
            $ oc rsh volume-shell
            $ rm /tmp/koku-metrics-operator-reports/upload/*
            ```

          4. (Optional) Delete the pod that was used to connect to the PVC:

            ```
            $ oc delete -f volume-shell.yaml
            ```

          ## Create an Integration
          In a restricted network, the `koku-metrics-operator` cannot automatically create an integration. This process must be done manually. In the console.redhat.com platform, open the [Integrations menu](https://console.redhat.com/settings/integrations/) to begin adding an OpenShift integration to Cost Management:

          Prerequisites:
          * The cluster identifier which can be found in the KokuMetricsConfig CR, the cluster Overview page, or the cluster Help > About.

          Creating an integration:
          1. Navigate to the Integrations menu
          2. Select the `Red Hat` tab
          3. Create a new `Red Hat Openshift Container Platform` integration:
              * give the integration a unique name
              * add the Cost Management application
              * add the cluster identifier
          4. In the Source wizard, review the details and click `Finish` to create the source.

          ## Upload the reports to cost managment
          Uploading reports to cost managment is done through curl:

              $ curl -vvvv -F "file=@FILE_NAME.tar.gz;type=application/vnd.redhat.hccm.tar+tgz" https://console.redhat.com/api/ingress/v1/upload -H "Authorization: Bearer ${ACCESS_TOKEN}"

          where `FILE_NAME` is the name of the report to upload. The `ACCESS_TOKEN` is acquired using a service account. See documentation on [creating and managing a service account](https://docs.redhat.com/en/documentation/red_hat_customer_portal/1/html/creating_and_managing_service_accounts).
        displayName: Koku Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cost
        - management
        - usage
        - monitor
        links:
        - name: Koku Metrics Operator source code repository
          url: https://github.com/project-koku/koku-metrics-operator
        maintainers:
        - email: costmanagement@redhat.com
          name: costmanagement
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/community-operator-pipeline-prod/koku-metrics-operator:4.2.0
        - quay.io/project-koku/koku-metrics-operator@sha256:693db9c90b286c9cc4e12810fda5b0a318a9c53c4d6d4d12f7291d5234229d36
        version: 4.2.0
      entries:
      - name: koku-metrics-operator.v4.2.0
        version: 4.2.0
      - name: koku-metrics-operator.v4.1.0
        version: 4.1.0
      - name: koku-metrics-operator.v4.0.0
        version: 4.0.0
      - name: koku-metrics-operator.v3.3.2
        version: 3.3.2
      - name: koku-metrics-operator.v3.3.1
        version: 3.3.1
      - name: koku-metrics-operator.v3.3.0
        version: 3.3.0
      - name: koku-metrics-operator.v3.2.1
        version: 3.2.1
      - name: koku-metrics-operator.v3.2.0
        version: 3.2.0
      - name: koku-metrics-operator.v3.1.0
        version: 3.1.0
      - name: koku-metrics-operator.v3.0.1
        version: 3.0.1
      - name: koku-metrics-operator.v3.0.0
        version: 3.0.0
      - name: koku-metrics-operator.v2.0.0
        version: 2.0.0
      - name: koku-metrics-operator.v1.2.0
        version: 1.2.0
      - name: koku-metrics-operator.v1.1.9
        version: 1.1.9
      - name: koku-metrics-operator.v1.1.8
        version: 1.1.8
      - name: koku-metrics-operator.v1.1.7
        version: 1.1.7
      - name: koku-metrics-operator.v1.1.6
        version: 1.1.6
      - name: koku-metrics-operator.v1.1.5
        version: 1.1.5
      - name: koku-metrics-operator.v1.1.4
        version: 1.1.4
      - name: koku-metrics-operator.v1.1.3
        version: 1.1.3
      - name: koku-metrics-operator.v1.1.2
        version: 1.1.2
      - name: koku-metrics-operator.v1.1.1
        version: 1.1.1
      - name: koku-metrics-operator.v0.9.8
        version: 0.9.8
      - name: koku-metrics-operator.v0.9.7
        version: 0.9.7
      - name: koku-metrics-operator.v0.9.6
        version: 0.9.6
      - name: koku-metrics-operator.v0.9.5
        version: 0.9.5
      - name: koku-metrics-operator.v0.9.4
        version: 0.9.4
      - name: koku-metrics-operator.v0.9.3
        version: 0.9.3
      - name: koku-metrics-operator.v0.9.2
        version: 0.9.2
      - name: koku-metrics-operator.v0.9.1
        version: 0.9.1
      - name: koku-metrics-operator.v0.9.0
        version: 0.9.0
      name: beta
    defaultChannel: beta
    packageName: koku-metrics-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "docker.io/percona/percona-server-mongodb:7.0.18-11",
                  "imagePullPolicy": "Always",
                  "initImage": "docker.io/percona/percona-server-mongodb-operator:1.20.1",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "docker.io/percona/pmm-client:2.44.1",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "docker.io/percona/percona-backup-mongodb:2.9.1",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: docker.io/percona/percona-server-mongodb-operator:1.20.1
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.20.1
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - docker.io/percona/percona-server-mongodb-operator:1.20.1
        - quay.io/community-operator-pipeline-prod/percona-server-mongodb-operator@sha256:8659c9f25245e541609079970c8193fc6a58a7d323c249a81d743de696dc15d4
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: node-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: node-maintenance-operator.v5.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
          createdAt: "2023-10-20 15:36:51"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=4.10.0 <5.0.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://docs.openshift.com/container-platform/4.11/nodes/nodes/eco-node-maintenance-operator.html
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:9fe87e1b0c49dd00a6ab0c35b68ebf641f347ac5b405d31f3c5ace02a7a30540
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:58846ab62346b1004e11f36bd901c96f8d315958eb39fbb272d8b328faebcf04
        version: 5.0.1
      entries:
      - name: node-maintenance-operator.v5.0.1
        version: 5.0.1
      name: 4.12-eus
    - currentCSV: node-maintenance-operator.v5.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
          createdAt: "2023-10-20 15:50:49"
          description: Node Maintenance Operator for cordoning and draining nodes.
          olm.skipRange: '>=5.0.0 <5.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:35d2c7c2849f89474fd5648a0d2781dde28131303cc8a597ce243a3e314fe0af
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:2efd1976d2fb110391e0e5937b90476d3eeaacb4dd29cd52375f2b8566f17b41
        version: 5.2.1
      entries:
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.14-eus
    - currentCSV: node-maintenance-operator.v5.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
          createdAt: "2024-08-10 19:55:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.2.0 <5.3.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.

          This is a standalone version of NMO. In case OpenShift Virtualization v4.10.z or less is installed in your cluster, then it has an old NMO.
          Please update OpenShift Virtualization version, and use this newer NMO.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:dd0ab94059f782aa31696b15e1b25787cd58bab4905fe64faec7e49bbae3d289
        - registry.redhat.io/workload-availability/node-maintenance-rhel8-operator@sha256:7ca5853885534198544d23067087d80047b0e1421903085a2cb328bd00def147
        version: 5.3.1
      entries:
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.16-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: 4.18-eus
    - currentCSV: node-maintenance-operator.v5.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodemaintenance.medik8s.io/v1beta1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "nodeName": "node02",
                  "reason": "Test node maintenance"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-10 10:59:05"
          description: Node Maintenance Operator for cordoning and draining nodes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.3.0 <5.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/node-maintenance-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.nodemaintenance.medik8s.io
            version: v1beta1
        description: |
          Node Maintenance Operator (NMO)

          This operator will keep nodes cordoned and drained while a matching NodeMaintenance (nm) custom resource exists.
          This is useful when investigating problems with a machine, or performing an operation on the underlying machine that might result in node failure.
        displayName: Node Maintenance Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - Node-maintenance
        - nmo
        links:
        - name: Node Maintenance Operator
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.4/html/remediation_fencing_and_maintenance/node-maintenance-operator
        - name: Source Code
          url: https://github.com/medik8s/node-maintenance-operator
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/node-maintenance-rhel9-operator@sha256:752bff8bebcaf3df1f9d126dc4b07b622a44b02bfcf0f82b9100a7693627079a
        - registry.redhat.io/workload-availability/node-maintenance-operator-bundle@sha256:aee6443d5ab0352b4e4ac5f4bd659d63ad7a9d33e17d53ed4c552e3b69ffc806
        version: 5.4.1
      entries:
      - name: node-maintenance-operator.v5.4.1
        version: 5.4.1
      - name: node-maintenance-operator.v5.4.0
        version: 5.4.0
      - name: node-maintenance-operator.v5.3.1
        version: 5.3.1
      - name: node-maintenance-operator.v5.3.0
        version: 5.3.0
      - name: node-maintenance-operator.v5.2.1
        version: 5.2.1
      name: stable
    defaultChannel: stable
    packageName: node-maintenance-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HashiCorp
      provider-url: https://www.hashicorp.com/
    name: hcp-terraform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hcp-terraform-operator.v2.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "AgentPool",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "agentTokens": [
                    {
                      "name": "token"
                    }
                  ],
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Module",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "module": {
                    "source": "redeux/terraform-cloud-agent/kubernetes",
                    "version": "1.0.1"
                  },
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  },
                  "workspace": {
                    "id": "WORKSPACE_ID"
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Project",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              },
              {
                "apiVersion": "app.terraform.io/v1alpha2",
                "kind": "Workspace",
                "metadata": {
                  "name": "NAME"
                },
                "spec": {
                  "name": "NAME",
                  "organization": "HCP_TF_ORG_NAME",
                  "token": {
                    "secretKeyRef": {
                      "key": "SECRET_KEY",
                      "name": "SECRET_NAME"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
          createdAt: "2025-05-28T13:58:52Z"
          description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
            / Terraform Enterprise resources via Kubernetes Custom Resources.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Valid HCP Terraform / Terraform
            Enterprise subscription is required
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/hashicorp/hcp-terraform-operator
          support: HashiCorp
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              AgentPool manages HCP Terraform Agent Pools, HCP Terraform Agent Tokens and can perform HCP Terraform Agent scaling.
              More infromation:
                - https://developer.hashicorp.com/terraform/cloud-docs/agents/agent-pools
                - https://developer.hashicorp.com/terraform/cloud-docs/users-teams-organizations/api-tokens#agent-api-tokens
                - https://developer.hashicorp.com/terraform/cloud-docs/agents
            displayName: Agent Pool
            kind: AgentPool
            name: agentpools.app.terraform.io
            version: v1alpha2
          - description: |-
              Module implements API-driven Run Workflows.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/run/api
            displayName: Module
            kind: Module
            name: modules.app.terraform.io
            version: v1alpha2
          - description: |-
              Project manages HCP Terraform Projects.
              More information:
              - https://developer.hashicorp.com/terraform/cloud-docs/projects/manage
            displayName: Project
            kind: Project
            name: projects.app.terraform.io
            version: v1alpha2
          - description: |-
              Workspace manages HCP Terraform Workspaces.
              More information:
                - https://developer.hashicorp.com/terraform/cloud-docs/workspaces
            displayName: Workspace
            kind: Workspace
            name: workspaces.app.terraform.io
            version: v1alpha2
        description: HCP Terraform Operator for Kubernetes allows managing HCP Terraform
          / Terraform Enterprise resources via Kubernetes Custom Resources.
        displayName: HCP Terraform Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - terraform
        links:
        - name: HCP Terraform Operator
          url: https://github.com/hashicorp/hcp-terraform-operator
        maintainers:
        - email: hcp-tf-operator@hashicorp.com
          name: HashiCorp
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: HashiCorp
          url: https://www.hashicorp.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:bb84ead437042603254efcb703442be8500a9450b10bbcaaee81b48a6ed869ea
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator-bundle@sha256:0f20db33b1eb069a767089862412846aacb723dc22b44adeb9793495ad20ea50
        - registry.connect.redhat.com/hashicorp/hcp-terraform-operator@sha256:8885a6a5eae092117a670ec5541a3e4a24c6b23807fd17b0446d91b01eb6e421
        version: 2.9.2
      entries:
      - name: hcp-terraform-operator.v2.9.2
        version: 2.9.2
      - name: hcp-terraform-operator.v2.9.1
        version: 2.9.1
      - name: hcp-terraform-operator.v2.9.0
        version: 2.9.0
      - name: hcp-terraform-operator.v2.8.1
        version: 2.8.1
      - name: hcp-terraform-operator.v2.8.0
        version: 2.8.0
      name: alpha
    defaultChannel: alpha
    packageName: hcp-terraform-operator
    provider:
      name: HashiCorp
      url: https://www.hashicorp.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle@sha256:506afdc804dc30e0a720e8e6f151b75a32148844de42d90d8c0648e9108d9ea7
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Elastic
      provider-url: ""
    name: elasticsearch-eck-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: elasticsearch-eck-operator-certified.v3.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
                {
                    "apiVersion": "elasticsearch.k8s.elastic.co/v1",
                    "kind": "Elasticsearch",
                    "metadata": {
                        "name": "elasticsearch-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "nodeSets": [
                            {
                                "name": "default",
                                "config": {
                                    "node.roles": ["master", "data"],
                                    "node.attr.attr_name": "attr_value",
                                    "node.store.allow_mmap": false
                                },
                                "podTemplate": {
                                    "metadata": {
                                        "labels": {
                                            "foo": "bar"
                                        }
                                    },
                                    "spec": {
                                        "containers": [
                                            {
                                                "name": "elasticsearch",
                                                "resources": {
                                                    "requests": {
                                                        "memory": "4Gi",
                                                        "cpu": 1
                                                    },
                                                    "limits": {
                                                        "memory": "4Gi",
                                                        "cpu": 2
                                                    }
                                                }
                                            }
                                        ]
                                    }
                                },
                                "count": 3
                            }
                        ]
                    }
                },
                {
                    "apiVersion": "kibana.k8s.elastic.co/v1",
                    "kind": "Kibana",
                    "metadata": {
                        "name": "kibana-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        },
                        "podTemplate": {
                            "metadata": {
                                "labels": {
                                    "foo": "bar"
                                }
                            },
                            "spec": {
                                "containers": [
                                    {
                                        "name": "kibana",
                                        "resources": {
                                            "requests": {
                                                "memory": "1Gi",
                                                "cpu": 0.5
                                            },
                                            "limits": {
                                                "memory": "2Gi",
                                                "cpu": 2
                                            }
                                        }
                                    }
                                ]
                            }
                        }
                    }
                },
                {
                    "apiVersion": "apm.k8s.elastic.co/v1",
                    "kind": "ApmServer",
                    "metadata": {
                        "name": "apmserver-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "enterprisesearch.k8s.elastic.co/v1",
                    "kind": "EnterpriseSearch",
                    "metadata": {
                        "name": "ent-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "config": {
                            "ent_search.external_url": "https://localhost:3002"
                        },
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                  "apiVersion": "beat.k8s.elastic.co/v1beta1",
                  "kind": "Beat",
                  "metadata": {
                    "name": "heartbeat-sample"
                  },
                  "spec": {
                    "type": "heartbeat",
                    "version": "9.1.0",
                    "elasticsearchRef": {
                      "name": "elasticsearch-sample"
                    },
                    "config": {
                      "heartbeat.monitors": [
                        {
                          "type": "tcp",
                          "schedule": "@every 5s",
                          "hosts": [
                            "elasticsearch-sample-es-http.default.svc:9200"
                          ]
                        }
                      ]
                    },
                    "deployment": {
                      "replicas": 1,
                      "podTemplate": {
                        "spec": {
                          "securityContext": {
                            "runAsUser": 0
                          }
                        }
                      }
                    }
                  }
                },
                {
                  "apiVersion": "agent.k8s.elastic.co/v1alpha1",
                  "kind": "Agent",
                  "metadata": {
                    "name": "agent-sample"
                  },
                  "spec": {
                    "version": "9.1.0",
                    "elasticsearchRefs": [
                      {
                        "name": "elasticsearch-sample"
                      }
                    ],
                    "daemonSet": {},
                    "config": {
                      "inputs": [
                        {
                          "name": "system-1",
                          "revision": 1,
                          "type": "system/metrics",
                          "use_output": "default",
                          "meta": {
                            "package": {
                              "name": "system",
                              "version": "0.9.1"
                            }
                          },
                          "data_stream": {
                            "namespace": "default"
                          },
                          "streams": [
                            {
                              "id": "system/metrics-system.cpu",
                              "data_stream": {
                                "dataset": "system.cpu",
                                "type": "metrics"
                              },
                              "metricsets": [
                                "cpu"
                              ],
                              "cpu.metrics": [
                                "percentages",
                                "normalized_percentages"
                              ],
                              "period": "10s"
                            }
                          ]
                        }
                      ]
                    }
                  }
                },
                {
                    "apiVersion": "maps.k8s.elastic.co/v1alpha1",
                    "kind": "ElasticMapsServer",
                    "metadata": {
                        "name": "ems-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1,
                        "elasticsearchRef": {
                            "name": "elasticsearch-sample"
                        }
                    }
                },
                {
                    "apiVersion": "logstash.k8s.elastic.co/v1alpha1",
                    "kind": "Logstash",
                    "metadata" : {
                        "name": "logstash-sample"
                    },
                    "spec": {
                        "version": "9.1.0",
                        "count": 1
                    }
                }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
          createdAt: "2025-07-30 11:10:34"
          description: Run Elasticsearch, Kibana, APM Server, Beats, Enterprise Search,
            Elastic Agent, Elastic Maps Server and Logstash on Kubernetes and OpenShift
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Elastic Basic license
          repository: https://github.com/elastic/cloud-on-k8s
          support: elastic.co
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Elastic Agent instance
            displayName: Elastic Agent
            kind: Agent
            name: agents.agent.k8s.elastic.co
            version: v1alpha1
          - description: APM Server instance
            displayName: APM Server
            kind: ApmServer
            name: apmservers.apm.k8s.elastic.co
            version: v1
          - description: Beats instance
            displayName: Beats
            kind: Beat
            name: beats.beat.k8s.elastic.co
            version: v1beta1
          - description: Elastic Maps Server instance
            displayName: Elastic Maps Server
            kind: ElasticMapsServer
            name: elasticmapsservers.maps.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch autoscaler
            displayName: Elasticsearch Autoscaler
            kind: ElasticsearchAutoscaler
            name: elasticsearchautoscalers.autoscaling.k8s.elastic.co
            version: v1alpha1
          - description: Instance of an Elasticsearch cluster
            displayName: Elasticsearch Cluster
            kind: Elasticsearch
            name: elasticsearches.elasticsearch.k8s.elastic.co
            version: v1
          - description: Enterprise Search instance
            displayName: Enterprise Search
            kind: EnterpriseSearch
            name: enterprisesearches.enterprisesearch.k8s.elastic.co
            version: v1
          - description: Kibana instance
            displayName: Kibana
            kind: Kibana
            name: kibanas.kibana.k8s.elastic.co
            version: v1
          - description: Logstash instance
            displayName: Logstash
            kind: Logstash
            name: logstashes.logstash.k8s.elastic.co
            version: v1alpha1
          - description: Elastic Stack Config Policy
            displayName: Elastic Stack Config Policy
            kind: StackConfigPolicy
            name: stackconfigpolicies.stackconfigpolicy.k8s.elastic.co
            version: v1alpha1
        description: |-
          Elastic Cloud on Kubernetes (ECK) is the official operator by Elastic for automating the deployment, provisioning, management, and orchestration of Elasticsearch, Kibana, APM Server, Beats, Enterprise Search, Elastic Agent, Elastic Maps Server, and Logstash on Kubernetes.

          Current features:

          *  Elasticsearch, Kibana, APM Server, Enterprise Search, Beats, Elastic Agent, Elastic Maps Server, and Logstash deployments
          *  TLS Certificates management
          *  Safe Elasticsearch cluster configuration and topology changes
          *  Persistent volumes usage
          *  Custom node configuration and attributes
          *  Secure settings keystore updates

          Supported versions:

          * Kubernetes 1.29-1.33
          * OpenShift 4.15-4.19
          * Google Kubernetes Engine (GKE), Azure Kubernetes Service (AKS), and Amazon Elastic Kubernetes Service (EKS)
          * Elasticsearch, Kibana, APM Server: 7.17+, 8+, 9+
          * Enterprise Search: 7.17+, 8+
          * Beats: 7.17+, 8+, 9+
          * Elastic Agent: 7.17+, 8+, 9+
          * Elastic Maps Server: 7.17+, 8+, 9+
          * Logstash 8.12+, 9+

          ECK should work with all conformant installers as listed in these [FAQs](https://github.com/cncf/k8s-conformance/blob/master/faq.md#what-is-a-distribution-hosted-platform-and-an-installer). Distributions include source patches and so may not work as-is with ECK.
          Alpha, beta, and stable API versions follow the same [conventions used by Kubernetes](https://kubernetes.io/docs/concepts/overview/kubernetes-api/#api-versioning).
          See the full [Elastic support matrix](https://www.elastic.co/support/matrix#matrix_kubernetes) for more information.
          See the [Quickstart](https://www.elastic.co/guide/en/cloud-on-k8s/3.1/k8s-quickstart.html) to get started with ECK.
        displayName: Elasticsearch (ECK) Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elasticsearch
        - kibana
        - analytics
        - search
        - database
        - apm
        - logstash
        links:
        - name: Documentation
          url: https://www.elastic.co/guide/en/cloud-on-k8s/3.1/index.html
        maintainers:
        - email: eck@elastic.co
          name: Elastic
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: Elastic
        relatedImages:
        - registry.connect.redhat.com/elastic/eck-operator@sha256:1750178c5dd716c031155f2e7058e19d4b7f1d3fdbb3e8a806b177a71be88500
        - registry.connect.redhat.com/elastic/eck@sha256:063a1f405b87751c5102693451ae5c2e9ad203c6b8a31b062a5b403bb8b2febc
        version: 3.1.0
      entries:
      - name: elasticsearch-eck-operator-certified.v3.1.0
        version: 3.1.0
      - name: elasticsearch-eck-operator-certified.v3.0.0
        version: 3.0.0
      - name: elasticsearch-eck-operator-certified.v2.16.1
        version: 2.16.1
      - name: elasticsearch-eck-operator-certified.v2.16.0
        version: 2.16.0
      - name: elasticsearch-eck-operator-certified.v2.15.0
        version: 2.15.0
      - name: elasticsearch-eck-operator-certified.v2.14.0
        version: 2.14.0
      - name: elasticsearch-eck-operator-certified.v2.13.0
        version: 2.13.0
      - name: elasticsearch-eck-operator-certified.v2.12.1
        version: 2.12.1
      - name: elasticsearch-eck-operator-certified.v2.12.0
        version: 2.12.0
      - name: elasticsearch-eck-operator-certified.v2.11.1
        version: 2.11.1
      - name: elasticsearch-eck-operator-certified.v2.11.0
        version: 2.11.0
      - name: elasticsearch-eck-operator-certified.v2.10.0
        version: 2.10.0
      - name: elasticsearch-eck-operator-certified.v2.9.0
        version: 2.9.0
      - name: elasticsearch-eck-operator-certified.v2.8.0
        version: 2.8.0
      - name: elasticsearch-eck-operator-certified.v2.7.0
        version: 2.7.0
      - name: elasticsearch-eck-operator-certified.v2.6.2
        version: 2.6.2
      - name: elasticsearch-eck-operator-certified.v2.6.1
        version: 2.6.1
      - name: elasticsearch-eck-operator-certified.v2.5.0
        version: 2.5.0
      - name: elasticsearch-eck-operator-certified.v2.4.0
        version: 2.4.0
      - name: elasticsearch-eck-operator-certified.v2.3.0
        version: 2.3.0
      - name: elasticsearch-eck-operator-certified.v2.2.0
        version: 2.2.0
      - name: elasticsearch-eck-operator-certified.v2.1.0
        version: 2.1.0
      - name: elasticsearch-eck-operator-certified.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: elasticsearch-eck-operator-certified
    provider:
      name: Elastic
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
          createdAt: "2024-05-23 19:57:04"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://www.medik8s.io/remediation/fence-agents-remediation/fence-agents-remediation/
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/medik8s/fence-agents-remediation-operator:v0.4.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/fence-agents-remediation@sha256:6ab21048f86979bcc34a84965dc2a97467a79507ab03469fe8c6deed224881f7
        version: 0.4.0
      entries:
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      - name: fence-agents-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-cluster-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:18:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:1a98d55c28a0679134e24b6aafb3c4dc9db0098094c8d2bcfaa41a63c33308e1
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: odr-cluster-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-cluster-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-cluster-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-cluster-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-cluster-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-cluster-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-cluster-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-cluster-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-cluster-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-cluster-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-cluster-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-cluster-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-cluster-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-cluster-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-cluster-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRClusterConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ramen",
                    "app.kubernetes.io/instance": "drclusterconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "drclusterconfig",
                    "app.kubernetes.io/part-of": "ramen"
                  },
                  "name": "drclusterconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "VolumeReplicationGroup",
                "metadata": {
                  "name": "volumereplicationgroup-sample"
                },
                "spec": {
                  "async": {
                    "schedulingInterval": "10m"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ],
                        "name": "cpd"
                      },
                      {
                        "includedResources": [
                          "Deployment"
                        ],
                        "name": "deployments"
                      },
                      {
                        "excludedResources": [
                          ""
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includeClusterResources": true,
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "cpd",
                        "includedResources": [
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      },
                      {
                        "backupName": "deployments",
                        "includedResources": [
                          "Deployment"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "excludedResources": [
                          "ConfigMap",
                          "Secret",
                          "Deployment",
                          "sample1.cpd.ibm.com",
                          "sample2.cpd.ibm.com",
                          "sample3.cpd.ibm.com"
                        ]
                      }
                    ]
                  },
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  },
                  "replicationState": "primary",
                  "s3Profiles": [
                    "s3-profile-of-east",
                    "s3-profile-of-west"
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DRClusterConfig
            name: drclusterconfigs.ramendr.openshift.io
            version: v1alpha1
          - kind: MaintenanceMode
            name: maintenancemodes.ramendr.openshift.io
            version: v1alpha1
          - kind: ProtectedVolumeReplicationGroupList
            name: protectedvolumereplicationgrouplists.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupDestination
            name: replicationgroupdestinations.ramendr.openshift.io
            version: v1alpha1
          - kind: ReplicationGroupSource
            name: replicationgroupsources.ramendr.openshift.io
            version: v1alpha1
          - description: VolumeReplicationGroup is the Schema for the volumereplicationgroups
              API
            displayName: Volume Replication Group
            kind: VolumeReplicationGroup
            name: volumereplicationgroups.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Cluster is a disaster-recovery orchestrator for
          stateful applications, that operates from an Advanced Cluster Management
          (ACM) managed cluster and is controlled by Openshift DR Hub operator to
          orchestrate the life-cycle of an application, and its state on the managed
          cluster.
        displayName: Openshift DR Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-cluster-operator-bundle@sha256:16f820575acdca39c0b7e85edff7b910809f6d184a9a479de2bb2f687c94a6f2
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-cluster-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-cluster-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-cluster-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-cluster-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-cluster-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-cluster-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-cluster-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-cluster-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-cluster-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-cluster-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-cluster-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-cluster-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-cluster-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: rhdh
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast
    - currentCSV: rhdh-operator.v1.1.2-0.1714688890.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
          createdAt: "2023-03-20T16:11:34Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins, configuration settings, and deployment
            mechanisms, which can help streamline the process of setting up a self-managed
            internal developer portal for adopters who are just starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.substitutesFor: rhdh-operator.v1.1.2
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.1.0 <1.2.0'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/janus-idp/operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: jianrzha@redhat.com
          name: Jianrong Zhang
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhel9/postgresql-15@sha256:d5f3d496ecd55688e384cc0627700a97af6924582123bcf00749c375724af3de
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:95ed239eb427c7b1e422ece28c5277b5018efb751c2b472bf502b8216219c047
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:d7ae2456857cab09a9e9acf69c9990ef18deb7f920046b8c2483fbfd0d42bbfc
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a93b382258a34e0b56ef6616a8490e5f44ef1c6aa5d29c0d08d2147ae5b440e6
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:9bea7eabdea44342248dfd6091a8f5c3b6e65884c916b3c2a073a4a64481aa6f
        version: 1.1.2+0.1714688890.p
      entries:
      - name: rhdh-operator.v1.1.2
        version: 1.1.2
      - name: rhdh-operator.v1.1.2-0.1714608881.p
        version: 1.1.2+0.1714608881.p
      - name: rhdh-operator.v1.1.2-0.1714688618.p
        version: 1.1.2+0.1714688618.p
      - name: rhdh-operator.v1.1.2-0.1714688890.p
        version: 1.1.2+0.1714688890.p
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.1
    - currentCSV: rhdh-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
          createdAt: "2024-11-28T17:55:19Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.2.6'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.2/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.2/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: asoro@redhat.com
          name: Armel Soro
        - email: cdaley@redhat.com
          name: Corey Daley
        - email: gazarenk@redhat.com
          name: Gennady Azarenkov
        - email: nboldt@redhat.com
          name: Nick Boldt
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3170da029869d15a246ad31d0debd4b9c03ba13855fb877a754cfdffb2286b80
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:c9051a6c339cd40cbc7474a33b32a6b94913af964e12ca5da0c0b7e1c6f444d5
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:ef29f2b340a6dc52c5014239876d54a54f7c0f0bf4ea917391438a69c94a8532
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:025eb58c69b9cf5992ecf2fccb29a3d36061cbcf62652a1c6878f0c4d767c1fb
        version: 1.2.6
      entries:
      - name: rhdh-operator.v1.2.6
        version: 1.2.6
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.2
    - currentCSV: rhdh-operator.v1.3.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
          createdAt: "2025-02-10T18:12:52Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.3.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.3/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:b4934c45ba595b36e65c73b734c58d1517e7f43b7cfc309ca8bc853733582f76
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:c044fed8485a62a5051ac726525da546936be22f268db05491bf81582e0e6194
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:a7ce92f7f4aadca2781c44839bc357615395587bc8b68a735642967cf8694362
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:5ce579edc10bde5c1c400104f7e0d9f86ca277cd565aaf3dd84b8636f5b949a8
        - registry.redhat.io/rhel9/postgresql-15@sha256:24fb4e7914a6e1464d015be9e5582cc4b9da224137408bd429e7ea4f391aa198
        version: 1.3.5
      entries:
      - name: rhdh-operator.v1.3.5
        version: 1.3.5
      - name: rhdh-operator.v1.3.4
        version: 1.3.4
      - name: rhdh-operator.v1.3.3
        version: 1.3.3
      - name: rhdh-operator.v1.3.1
        version: 1.3.1
      - name: rhdh-operator.v1.1.1
        version: 1.1.1
      - name: rhdh-operator.v1.1.0
        version: 1.1.0
      name: fast-1.3
    - currentCSV: rhdh-operator.v1.4.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "name": "developer-hub"
                },
                "spec": {
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
          createdAt: "2025-04-03T05:42:32Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.4.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.4/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:ae5b52c1b35e82f3c564402e367a2a2028e16806867bba0868643089ba7ed223
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:093c6a66d4faa1d980e4319a048db8b2869eba8f13f40f02615d6696257e6719
        version: 1.4.3
      entries:
      - name: rhdh-operator.v1.4.3
        version: 1.4.3
      - name: rhdh-operator.v1.4.2
        version: 1.4.2
      - name: rhdh-operator.v1.4.1
        version: 1.4.1
      - name: rhdh-operator.v1.4.0
        version: 1.4.0
      name: fast-1.4
    - currentCSV: rhdh-operator.v1.5.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
          createdAt: "2025-06-05T14:31:21Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.5.3'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.5/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Backstage Operator (Upstream)
          url: https://github.com/redhat-developer/rhdh-operator
        - name: RHDH Operator (Downstream)
          url: https://gitlab.cee.redhat.com/rhidp/rhdh/
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:2ad8c87e82df27c2b6ceba94179d1f0a40fe775368663c1b5572611888e0e02b
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:5a19032dda9a0f584b41cbe9e729ed23e6003d90ebbb026eb9e7bb9769d4a4e4
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:8fe167cc649a1c53994c376f3cf3eed8c6e318741a0c08e5aaf07cb0b74e7f8a
        - registry.redhat.io/rhel9/postgresql-15@sha256:74377f5a61dd7e3e0ba5d2fa880b7107025433a026396567c79eac927b4212bc
        version: 1.5.3
      entries:
      - name: rhdh-operator.v1.5.3
        version: 1.5.3
      - name: rhdh-operator.v1.5.2
        version: 1.5.2
      - name: rhdh-operator.v1.5.1
        version: 1.5.1
      name: fast-1.5
    - currentCSV: rhdh-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
          createdAt: "2025-08-27T18:43:12Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.6.5'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.6/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:3b2cf972e1ced0a46d3559da4eb1a1b32c50457293ede4958a13538a84b45316
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:50f20300c81621221260d03d22b1442750e3f1d2e62f3e511d63925d84c4d1f9
        - registry.redhat.io/rhel9/postgresql-15@sha256:5e225c1646251d2509b3a6e45fad23abfb11dd5c45d3224b59178f90deb38a70
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:177a7308f61a6b1ba21bc0f7bf550a93d4478e52479e2c94a2437f2152e3d8ff
        version: 1.6.5
      entries:
      - name: rhdh-operator.v1.6.5
        version: 1.6.5
      - name: rhdh-operator.v1.6.4
        version: 1.6.4
      - name: rhdh-operator.v1.6.3
        version: 1.6.3
      - name: rhdh-operator.v1.6.2
        version: 1.6.2
      - name: rhdh-operator.v1.6.1
        version: 1.6.1
      - name: rhdh-operator.v1.6.0
        version: 1.6.0
      name: fast-1.6
    - currentCSV: rhdh-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha1",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha2",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              },
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Backstage",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "backstage"
                  },
                  "name": "developer-hub"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
          createdAt: "2025-09-12T06:25:11Z"
          description: Red Hat Developer Hub is a Red Hat supported version of Backstage.
            It comes with pre-built plug-ins and configuration settings, supports
            use of an external database, and can help streamline the process of setting
            up a self-managed internal developer portal for adopters who are just
            starting out.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: rhdh-operator
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://gitlab.cee.redhat.com/rhidp/rhdh/
          skipRange: '>=1.0.0 <1.7.1'
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha1
          - description: Backstage is the Schema for the Red Hat Developer Hub backstages
              API. It comes with pre-built plug-ins, configuration settings, and deployment
              mechanisms, which can help streamline the process of setting up a self-managed
              internal developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha2
          - description: |-
              Backstage is the Schema for the Red Hat Developer Hub backstages API.
              It comes with pre-built plug-ins, configuration settings, and deployment mechanisms,
              which can help streamline the process of setting up a self-managed internal
              developer portal for adopters who are just starting out.
            displayName: Red Hat Developer Hub
            kind: Backstage
            name: backstages.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub is an enterprise-grade platform for building developer portals, containing a supported and opinionated framework. It comes with pre-built plug-ins and configuration settings, supports use of an external database, and can help streamline the process of setting up a self-managed internal developer portal for adopters who are just starting out. By implementing a unified and open platform designed to maximize developer skills, ease onboarding, and increase development productivity, focus can be centered on what really matters: writing great code. Red Hat Developer Hub also offers Software Templates to simplify the development process, which can reduce friction and frustration for development teams, boosting their productivity and increasing an organization's competitive advantage.

          System Architects can benefit by implementing a tailored platform with a complementary suite of verified and curated tools and components needed for operations teams to support developers—within a centralized, consistent location. Development teams can experience increased productivity, fewer development team obstacles, and simplified governance of technology choices with self-service and guardrails.

          ## Telemetry data collection

          The telemetry data collection feature is enabled by default. Red Hat Developer Hub sends telemetry data to Red Hat by using the `backstage-plugin-analytics-provider-segment` plugin. To disable this and to learn what data is being collected, see [Red Hat Developer Hub documentation on telemetry data collection](https://access.redhat.com/documentation/en-us/red_hat_developer_hub/1.7/html-single/administration_guide_for_red_hat_developer_hub/index#assembly-rhdh-telemetry_admin-rhdh).

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://access.redhat.com/documentation/en-us/red_hat_developer_hub)
          * [Life Cycle](https://access.redhat.com/node/7025299)
          * [Support Policies](https://access.redhat.com/policy/developerhub-support-policy)
          * [Configuring external PostgreSQL databases](https://docs.redhat.com/en/documentation/red_hat_developer_hub/1.3/html/administration_guide_for_red_hat_developer_hub/assembly-configuring-external-postgresql-databases#assembly-configuring-external-postgresql-databases)
        displayName: Red Hat Developer Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Backstage
        - RHDH
        links:
        - name: Product Page
          url: https://developers.redhat.com/products/developer-hub/overview/
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_developer_hub
        - name: Life Cycle
          url: https://access.redhat.com/node/7025299
        - name: Support Policies
          url: https://access.redhat.com/policy/developerhub-support-policy
        - name: Sources
          url: https://github.com/redhat-developer/rhdh-operator
        maintainers:
        - email: rhdh-notifications@redhat.com
          name: Red Hat Developer Hub Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - registry.redhat.io/rhdh/rhdh-operator-bundle@sha256:f743970668c72ff357ebfc910ddd5110c95a39754862d74d31e108c5993c5ace
        - registry.redhat.io/rhdh/rhdh-rhel9-operator@sha256:d413b6ee53b6271644b9c5d2fc1dc266212b33e419380a6dddbf4613280dfd7a
        - registry.redhat.io/rhel9/postgresql-15@sha256:4d707fc04f13c271b455f7b56c1fda9e232a62214ffc6213c02e41177dd4a13f
        - registry.redhat.io/rhdh/rhdh-hub-rhel9@sha256:62f32e50727c5006766a34daa5b6472b649cd9894f3fe9543b8ecc67e6760e8e
        version: 1.7.1
      entries:
      - name: rhdh-operator.v1.7.1
        version: 1.7.1
      - name: rhdh-operator.v1.7.0
        version: 1.7.0
      name: fast-1.7
    defaultChannel: fast
    packageName: rhdh
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: proactive-node-scaling-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: proactive-node-scaling-operator.v0.4.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "ai-ml"
                },
                "spec": {
                  "nodeSelector": {
                    "machine.openshift.io/cluster-api-machine-type": "ai-ml"
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "workload",
                      "operator": "Equal",
                      "value": "ai-ml"
                    }
                  ],
                  "watermarkPercentage": 20
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NodeScalingWatermark",
                "metadata": {
                  "name": "us-west-2a"
                },
                "spec": {
                  "nodeSelector": {
                    "topology.kubernetes.io/zone": "us-west-2a"
                  },
                  "watermarkPercentage": 20
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "false"
          containerImage: quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
          createdAt: "2022-01-22T14:14:23Z"
          description: This operator provides automation proactively scaling nodes
            before pods get stuck waiting for the cluster autoscaler.
          operatorframework.io/suggested-namespace: proactive-node-scaling-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/proactive-node-scaling-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeScalingWatermark is the Schema for the nodescalingwatermarks
              API
            displayName: Node Scaling Watermark
            kind: NodeScalingWatermark
            name: nodescalingwatermarks.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator makes the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) more proactive. As of now the cluster auto scaler will create new nodes only when a pod is pending because it cannot be allocated due to lack of capacity. This is not a goos user experience as the pending workload has to wait for several minutes as the new node is create and joins the cluster.

          The Proactive Node Scaling Operator improves the user experience by allocating low priority pods that don't do anything. When the cluster is full and a new user pod is created the following happens:

          1. some of the low priority pods are de-scheduled to make room for the user pod, which can then be scheduled. The user workload does not have to wait in this case.

          2. the de-scheduled low priority pods are rescheduled and in doing so the trigger the cluster autoscaler to add new nodes.

          Essentially this operator allows you to trade wasted resources for faster response time.

          In order for this operator to work correctly [pod priorities](https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/) must be defined. The default name for the priority class used by this operator is "proactive-node-autoscaling-pods" and it should have the lowest possible priority, 0. To ensure your regular workloads get a normal priority you should also define a PriorityClass for those with a higher priority than 0 and set globalDefault to true.

          For example:

          ```yaml
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: proactive-node-autoscaling-pods
          value: 0
          globalDefault: false
          description: "This priority class is the priority class used for Proactive Node Scaling Pods."
          ---
          apiVersion: scheduling.k8s.io/v1
          kind: PriorityClass
          metadata:
            name: normal-workload
          value: 1000
          globalDefault: true
          description: "This priority classis the cluster default and should be used for normal workloads."
          ```

          Also for this operator to work the [cluster autoscaler](https://github.com/kubernetes/autoscaler/tree/master/cluster-autoscaler) must be active, see OpenShift instructions [here](https://docs.openshift.com/container-platform/4.6/machine_management/applying-autoscaling.html) on how to turn it on.

          To activate the proactive autoscaling, a CR must be defined, here is an example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: NodeScalingWatermark
          metadata:
            name: us-west-2a
          spec:
            priorityClassName: proactive-node-autoscaling-pods
            watermarkPercentage: 20
            nodeSelector:
              topology.kubernetes.io/zone: us-west-2a
          ```

          The `nodeSelector` selects the nodes observed by this operator, which are also the nodes on which the low priority pods will be scheduled. The nodes observed by the cluster autoscaler should coincide with the nodes selected by this operator CR.

          The `watermarkPercentage` define the percentage of capacity of user workload that will be allocated to low priority pods. So in this example 20% of the user allocated capacity will be allocated via low priority pods. This also means that when the user workload reaches 80% capacity of the nodes selected by this CR (and the autoscaler), the cluster will start to scale.
        displayName: Proactive Node Scaling Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Capacity Management
        - Node Autoscaling
        - Proactive Autoscaling
        - Elastic Capacity
        links:
        - name: Proactive Node Scaling Operator
          url: https://github.com/redhat-cop/proactive-node-scaling-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/proactive-node-scaling-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/proactive-node-scaling-operator@sha256:0ae01e9adc67ff7341fa3faf51a67c02206c41342374cab0bb2e6079d84fef4a
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/proactive-node-scaling-operator@sha256:892303babb45090a6458cd4b1c61a01364012c43ea0bd6ed4d10bdcb8041e2ab
        version: 0.4.2
      entries:
      - name: proactive-node-scaling-operator.v0.4.2
        version: 0.4.2
      - name: proactive-node-scaling-operator.v0.4.0
        version: 0.4.0
      - name: proactive-node-scaling-operator.v0.3.10
        version: 0.3.10
      - name: proactive-node-scaling-operator.v0.3.9
        version: 0.3.9
      - name: proactive-node-scaling-operator.v0.3.8
        version: 0.3.8
      - name: proactive-node-scaling-operator.v0.3.7
        version: 0.3.7
      - name: proactive-node-scaling-operator.v0.3.6
        version: 0.3.6
      - name: proactive-node-scaling-operator.v0.3.5
        version: 0.3.5
      - name: proactive-node-scaling-operator.v0.3.4
        version: 0.3.4
      - name: proactive-node-scaling-operator.v0.3.3
        version: 0.3.3
      - name: proactive-node-scaling-operator.v0.3.2
        version: 0.3.2
      - name: proactive-node-scaling-operator.v0.3.1
        version: 0.3.1
      - name: proactive-node-scaling-operator.v0.3.0
        version: 0.3.0
      - name: proactive-node-scaling-operator.v0.2.0
        version: 0.2.0
      - name: proactive-node-scaling-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: proactive-node-scaling-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Oracle
      provider-url: https://www.oracle.com/cloud/
    name: oci-ccm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oci-ccm-operator.v0.0.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICloudManager",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicloudmanager",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicloudmanager",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicloudmanager"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "loadBalancer": {
                    "securityList": "SecurityListId",
                    "securityListManagementMode": "Frontend",
                    "subnet": "SubnetId"
                  },
                  "vcn": "VcnId"
                }
              },
              {
                "apiVersion": "ccm.oraclecloud.com/v1alpha1",
                "kind": "OCICsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "oci-ccm-operator",
                    "app.kubernetes.io/instance": "ocicsidriver",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ocicsidriver",
                    "app.kubernetes.io/part-of": "oci-ccm-operator"
                  },
                  "name": "ocicsidriver"
                },
                "spec": {
                  "compartment": "CompartmentId",
                  "driverNamespace": "kube-system",
                  "vcn": "VcnId"
                }
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider, Storage
          containerImage: phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
          createdAt: "2023-07-28T23:42:18Z"
          operators.openshift.io/infrastructure-features: '["csi"]'
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Oracle
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: OCICloudManager
            name: ocicloudmanagers.ccm.oraclecloud.com
            version: v1alpha1
          - kind: OCICsiDriver
            name: ocicsidrivers.ccm.oraclecloud.com
            version: v1alpha1
        description: OCI Kubernetes Cloud Controller Manager implementation (or out-of-tree
          cloud-provider) for Oracle Cloud Infrastructure (OCI).
        displayName: oci-ccm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OCI
        - CCM
        - OpenShift
        - Kubernetes
        links:
        - name: Oracle
          url: https://www.oracle.com/cloud/
        - name: OCI Cloud Controller Manager (CCM)
          url: https://github.com/oracle/oci-cloud-controller-manager
        maintainers:
        - email: david.d.campbell@oracle.com
          name: David Campbell
        - email: alex.blinov@oracle.com
          name: Alex Blinov
        maturity: alpha
        provider:
          name: Oracle
          url: https://www.oracle.com/cloud/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - phx.ocir.io/ax1eoqvpvbtg/oci-ccm-operator:0.0.9
        - quay.io/openshift-community-operators/oci-ccm-operator@sha256:eaf63c6f9ee9e9c3562d6ced4cff091c1e401daa826ee3a92b321a5cc8f13da2
        version: 0.0.9
      entries:
      - name: oci-ccm-operator.v0.0.9
        version: 0.0.9
      name: alpha
    defaultChannel: alpha
    packageName: oci-ccm-operator
    provider:
      name: Oracle
      url: https://www.oracle.com/cloud/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-kogito-operator.8.0.5-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
          description: IBM BAMOE Kogito Operator for deployment and management of
            Kogito services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/kogito-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          IBM BAMOE Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a IBM BAMOE Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy IBM BAMOE Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** IBM BAMOE Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the IBM BAMOE Kogito Operator.
        displayName: IBM BAMOE Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        - name: Source
          url: https://github.com/kiegroup/kogito-operator
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: beta
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator-bundle@sha256:e0e7fc3a69575199a6fcf93985451dd66a5eb7d1ab8caecf059d7d52f3953dc4
        - registry.redhat.io/ibm-bamoe/bamoe-kogito-rhel8-operator@sha256:8a99fcd217502b7e850af1e1097de58b6cb19d1e786394811a8e075c4d871279
        version: 8.0.5-3
      entries:
      - name: bamoe-kogito-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-kogito-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-kogito-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-kogito-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-kogito-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-kogito-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-kogito-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-kogito-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-kogito-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-kogito-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-kogito-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-kogito-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x
    defaultChannel: 8.x
    packageName: bamoe-kogito-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solution-Soft
      provider-url: www.solution-soft.com
    name: timemachine-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: timemachine-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tm.solution-soft.com/v1alpha1",
                "kind": "TimeMachine",
                "metadata": {
                  "name": "timemachine-sample"
                },
                "spec": {
                  "cluster_provider": "ibm",
                  "aws": {
                    "shared_volume_name": ""
                  },
                  "ibm": {
                    "persistentvolume": {
                      "storageclass": "ibmc-file-silver-gid"
                    }
                  },
                  "tm_lichost": "52.152.148.40",
                  "tm_licpass": "qwaszx",
                  "tm_licport": "57777",
                  "imagePullSecrets": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
          createdAt: "2024-02-05T09:46:17Z"
          description: Time Machine Operator enables time travel of OpenShift pods,
            without the need to change the system clock on any of the cluster nodes
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/timemachine-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: '>=0.0.4 <1.0.17'
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: SolutionSoft Systems, Inc. - support@solution-soft.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Time Machine
            displayName: Time Machine
            kind: TimeMachine
            name: timemachines.tm.solution-soft.com
            version: v1alpha1
        description: |-
          Time Machine operator enables time travel of OpenShift pods/deployments.

          To time travel a pod where your application is, you just need to install Time Machine operator and create a Time Machine custom resource in the same namespace, without the need of changing the container images of the application you’re using, or adding containers to the pod with your application.

          Time travelling is done on the namespace level where all the (configured) target pods/deployments see the same virtual time.
        displayName: Time Machine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - time shift
        - time travel
        - virtual clocks
        - time machine
        - virtual time
        - solution-soft
        links:
        - name: Timemachine Operator
          url: https://catalog.redhat.com/software/operators/detail/5e98727a6c5dcb34dfbb19e3
        maintainers:
        - email: dtrajkovic@solution-soft.com
          name: Dragan Trajkovic
        maturity: alpha
        provider:
          name: Solution-Soft
          url: www.solution-soft.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0c304c11986790f446604319dfc29db494b15f5f9fa13f2c436688ec29c92dbe
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator-bundle-rhmp@sha256:d90f6bce87b783a80b1b81275cb61162c7c2a7898aa67ea85e3223deb1dfcb4b
        - registry.connect.redhat.com/rhel-timemachine-container/timemachine-operator@sha256:ba030e8c3463d352bc8f1d4ac8baa097c6ecebd22720406c5e87ac4d059cdc70
        - registry.connect.redhat.com/rhel-timemachine-container/virtual-clock-testing-tool@sha256:a81a823328b0ee795c40668c89a08bcf545c5192d9eed0146071e3721a0a8372
        version: 1.0.17
      entries:
      - name: timemachine-operator.v1.0.17
        version: 1.0.17
      - name: timemachine-operator.v1.0.16
        version: 1.0.16
      name: alpha
    defaultChannel: alpha
    packageName: timemachine-operator-rhmp
    provider:
      name: Solution-Soft
      url: www.solution-soft.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftxray-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: xray-operator.v2.1.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftXray",
                "metadata": {
                  "name": "openshiftxray"
                },
                "spec": {
                  "xray": {
                    "common": {
                      "xrayUserId": "1000721035",
                      "xrayGroupId": "1000721035"
                    },
                    "analysis": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-analysis"
                      },
                      "name": "xray-analysis",
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "database": {
                      "password": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "global": {
                      "postgresqlTlsSecret": null
                    },
                    "indexer": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-indexer"
                      },
                      "name": "xray-indexer",
                      "podManagementPolicy": "Parallel",
                      "updateStrategy": "RollingUpdate"
                    },
                    "persist": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-persist"
                      },
                      "name": "xray-persist",
                      "persistence": {
                        "size": "10Gi"
                      },
                      "podManagementPolicy": "Parallel",
                      "preStartCommand": null,
                      "updateStrategy": "RollingUpdate"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "rabbitmq": {
                      "enabled": true,
                      "replicaCount": 1,
                      "rbac": {
                        "create": true
                      },
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-rabbitmq"
                      },
                      "podSecurityContext": {
                        "fsGroup": 1000721001,
                        "runAsUser": 1000721001
                      },
                      "auth": {
                        "username": "xray",
                        "password": "xray",
                        "erlangCookie": "XRAYRABBITMQCLUSTER",
                        "tls": {
                          "enabled": false,
                          "failIfNoPeerCert": true,
                          "sslOptionsVerify": "verify_peer",
                          "caCertificate": "",
                          "serverCertificate": "",
                          "serverKey": ""
                        }
                      }
                    },
                    "rabbitmq-ha": {
                      "enabled": false,
                      "image": {
                        "repository": "registry.connect.redhat.com/jfrog/xray-rabbitmq"
                      },
                      "rabbitmqUsername": "guest",
                      "rabbitmqPassword": "guest",
                      "managementUsername": "guest",
                      "managementPassword": "guest",
                      "initContainer": {
                        "enabled": false
                      },
                      "securityContext": {
                        "fsGroup": 1000721035,
                        "runAsUser": 1000721035,
                        "runAsGroup": 1000721035
                      },
                      "replicaCount": 1
                    },
                    "replicaCount": 1,
                    "unifiedUpgradeAllowed":  "true",
                    "router": {
                      "image": {
                        "imagePullPolicy": "IfNotPresent",
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-router"
                      },
                      "name": "router"
                    },
                    "server": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/xray-server"
                      },
                      "name": "xray-server",
                      "podManagementPolicy": "Parallel",
                      "replicaCount": 1,
                      "updateStrategy": "RollingUpdate"
                    },
                    "xray": {
                      "consoleLog": false,
                      "jfrogUrl": "OVERRIDE",
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
          createdAt: 2020-05-22 00:00:00+00:00
          description: JFrog Xray Enterprise deploys Xray continuous security scanner
            into Openshift (Requires Jfrog Artifactory)
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Xray Instances
            displayName: Xray
            kind: OpenshiftXray
            name: openshiftxrays.charts.helm.k8s.io
            version: v1
        description: "## Overview Openshift Operator to deploy JFrog Xray Continuous
          Security scanner into your Openshift cluster. NOTE: Artifactory is required
          for JFrog Xray to connect with and scan artifacts against.\n### Major version
          v2.0.4 Breaking change notification!!\nThis release has major bug fix for
          Rabbitmq Pod. Please update to version 2.0.4 of the operator.\n### Major
          version v2.0.0 Breaking change notification!! Xray is dependent upon Rabbitmq.
          The version of Rabbitmq that this chart uses has significantly changed from
          version 1.1.8 of the operator. For full details on how to migrate please
          visit: `https://github.com/jfrog/charts/blob/master/stable/xray/RABBITMQ_MIGRATION_NOTES.md`\nIf
          you are using operator version < 2.1.0 there is a issue with upgrade where
          xray version is not updated if you update operator version.As workaround
          you can update the Custom Resource OpenshiftXray which has hardcored image
          tag value.After upgrading to version 2.1.0 you can remove the image tags
          parameters from your CR so that future upgrade works without workaround.\n###
          Security Context Constraints To deploy this helm chart you will need to
          be a cluster admin w/ access to the anyuid scc and add the operator service
          account to the anyuid scc. ``` oc adm policy add-scc-to-user anyuid -z xray-operator
          ``` ``` oc adm policy add-scc-to-user anyuid -z openshiftxray ``` ``` oc
          adm policy add-scc-to-user anyuid -z openshiftxray-rabbitmq-ha ``` ## Usage
          An external DB is required. The operator will not deploy a DB but will require
          you to specify the configuration values to connect to it. Search for JFrog
          and click JFrog Xray Operator to install. Go to the Installed Operators.
          Wait for the JFrog Xray Operator to complete the installation. Open the
          Operator and click on the provided API: Xray Click Create New Instance and
          provide the following parameters for your DB configuration: ``` JFROG_URL
          DATABASE_URL DATABASE_USER DATABASE_PASSWORD ``` JFROG_URL is the external
          ip or DNS of your Artifactory to connect Xray to. Artifactory is required
          to use this operator. You can set your JFROG_URL to the service name of
          your Artifactory Nginx: ``` oc get svc -n my_namespace | grep nginx ```
          DATABASE_URL must be a Postgresql URL in the format: `` postgres://postgres-postgresql:5432/xraydb?sslmode=disable
          `` DATABASE_USER and DATABASE_PASSWORD must supply a valid user on Postgresql.
          Click Create for Xray to deploy into OpenShift. Open Artifactory in a web
          browser to complete the onboarding wizard for Xray! ### Air gap environments
          To use Xray Operator in an air gap environment you will need to download
          the images as image streams into your Openshift air gap cluster manually.
          Use the image overrides to then specify the image stream locations that
          are local to your cluster. Next you will need to setup the database of index
          data for xray to use when scanning artifacts. Follow the link below for
          instructions on setup: https://www.jfrog.com/confluence/display/JFROG/Configuring+Xray#ConfiguringXray-SynchronizingtheDatabase "
        displayName: JFrog Xray Continuous Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Xray Continuous Security
          url: https://jfrog.com/xray/
        - name: JFrog Xray Continuous Security Video
          url: https://youtu.be/wGzyKlQklKQ
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/xray-persist@sha256:93fae0c124919a57d5162e32e88d3456ef6254889e454f89c21c89e69e6ae323
        - registry.connect.redhat.com/jfrog/xray-router@sha256:019235289e17d0e22dcd23a1505f71a78710827dfab338d8fa980f64c14f162f
        - registry.connect.redhat.com/jfrog/init@sha256:3f27e44b44c3556c50318110dbd542e34ef1061e28d7f0e0e7b8baa73c5f3e87
        - registry.connect.redhat.com/jfrog/xray-analysis@sha256:bcd8211f1bc753bfd7735165ce39b476e4efff198bffa877883620b41fcb33b5
        - registry.connect.redhat.com/jfrog/xray-operator-bundle@sha256:1a950c41ce2ae204d6bcba959b2d1a3df2d994bf1fefce4c7095c4ca20a5fd6d
        - registry.connect.redhat.com/jfrog/xray-indexer@sha256:7f40d2dfe4a35b25d16dfcbfa04b4a58f227c132d4c3c86345e7b2bd19a97311
        - registry.connect.redhat.com/jfrog/xray-operator@sha256:da7f6977170876b335875f4074ddfc0266b665351249295ce3c53f0e95891cc2
        - registry.connect.redhat.com/jfrog/xray-rabbitmq@sha256:459d7d5e78192a6f45a5527e8fbaf1d8b30299b10ad4c42515d3d2bb5fcc75f4
        - registry.connect.redhat.com/jfrog/xray-server@sha256:c5584c4e9528f0199e840b2b1ad1144220850ca9595ebf3e270bc6f86de4a71a
        version: 2.1.6
      entries:
      - name: xray-operator.v2.1.6
        version: 2.1.6
      name: alpha
    defaultChannel: alpha
    packageName: openshiftxray-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-iam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-iam-operator.v4.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "Authentication",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-iam-operator",
                    "app.kubernetes.io/managed-by": "ibm-iam-operator",
                    "app.kubernetes.io/name": "ibm-iam-operator"
                  },
                  "name": "example-authentication"
                },
                "spec": {
                  "auditService": {
                    "imageName": "auditService no longer used - ignore",
                    "imageRegistry": "auditService no longer used - ignore",
                    "imageTag": "auditService no longer used - ignore"
                  },
                  "authService": {
                    "imageName": "icp-platform-auth",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "ldapsCACert": "platform-auth-ldaps-ca-cert",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "650Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "400Mi",
                        "memory": "350Mi"
                      }
                    },
                    "routerCertSecret": "saml-auth-secret"
                  },
                  "clientRegistration": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "158Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "config": {
                    "authUniqueHosts": "internal-ip1 internal-ip2 mycluster.icp",
                    "bootstrapUserId": "kubeadmin",
                    "claimsMap": "name=\"givenName\" family_name=\"givenName\" given_name=\"givenName\" preferred_username=\"displayName\" display_name=\"displayName\"",
                    "claimsSupported": "name,family_name,display_name,given_name,preferred_username",
                    "clusterCADomain": "mycluster.icp",
                    "clusterExternalAddress": "10.0.0.1",
                    "clusterInternalAddress": "10.0.0.1",
                    "clusterName": "mycluster",
                    "defaultAdminPassword": "password",
                    "defaultAdminUser": "cpadmin",
                    "defaultLogin": "",
                    "enableImpersonation": false,
                    "fipsEnabled": true,
                    "ibmCloudSaas": false,
                    "icpPort": 8443,
                    "installType": "fresh",
                    "isOpenshiftEnv": true,
                    "nonceEnabled": true,
                    "oidcIssuerURL": "https://127.0.0.1:443/idauth/oidc/endpoint/OP",
                    "openshiftPort": 443,
                    "preferredLogin": "",
                    "providerIssuerURL": "",
                    "roksEnabled": true,
                    "roksURL": "https://roks.domain.name:443",
                    "roksUserPrefix": "changeme",
                    "saasClientRedirectUrl": "",
                    "scimAdminPassword": "scimpassword",
                    "scimAdminUser": "scimadmin",
                    "scopeClaim": "profile=\"name,family_name,display_name,given_name,preferred_username\"",
                    "wlpClientID": "4444be3a738841016ab76d71b650e836",
                    "wlpClientRegistrationSecret": "f1362ca4d20b8389af2d1ea68042c9af",
                    "wlpClientSecret": "aa73bf39752053bf723d1143fb4cf8a2",
                    "xframeDomain": "",
                    "zenFrontDoor": false
                  },
                  "identityManager": {
                    "imageName": "icp-identity-manager",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "masterNodesList": "10.0.0.1",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "identityProvider": {
                    "imageName": "icp-identity-provider",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "ephemeral-storage": "550Mi",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "50m",
                        "ephemeral-storage": "300Mi",
                        "memory": "150Mi"
                      }
                    }
                  },
                  "initMongodb": {
                    "imageName": "im-initcontainer",
                    "imageRegistry": "registry.marketplace.redhat.com/cpopen/cpfs",
                    "imageTag": "4.11.0",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "178Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "operatorVersion": "0.14.1",
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
          createdAt: "2025-02-20T00:03:25Z"
          description: The IAM operator provides a simple Kubernetes CRD-Based API
            to manage the lifecycle of IAM services. With this operator, you can simply
            deploy and upgrade the IAM services
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.11.0
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-iam-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Authentication is the Schema for the authentications API
            displayName: Authentication
            kind: Authentication
            name: authentications.operator.ibm.com
            version: v1alpha1
          - description: Client is the Schema for the clients API
            displayName: Client
            kind: Client
            name: clients.oidc.security.ibm.com
            version: v1
        description: |-
          The `ibm-iam-operator` installs the IBM Cloud Platform Common Services Identity and access management (IAM) service.

          **Important:** Do not install this operator directly. Install this operator only by using the IBM Common Service Operator. For more information about installing this operator and other Common Services operators, see [Installer documentation](http://ibm.biz/cpcs_opinstall). Additionally, you can exit this panel and navigate to the IBM Common Services tile in OperatorHub to learn more about the operator. If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak to learn more about how to install and use the operator service. For more information about IBM Cloud Paks, see [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).

          You can use the `ibm-iam-operator` to install the authentication and authorization services for the IBM Cloud Platform Common Services.

          With these services, you can configure security for IBM Cloud Platform Common Services, IBM Certified Containers (IBM products), or IBM Cloud Paks that are installed.

          For more information about the available IBM Cloud Platform Common Services, see the [IBM Knowledge Center](http://ibm.biz/cpcsdocs).

          ## Supported platforms

           - Red Hat OpenShift Container Platform 4.3 or newer installed on one of the following platforms:

              - Linux x86_64
              - Linux on Power (ppc64le)
              - Linux on IBM Z and LinuxONE

          ## Prerequisites

          Before you install this operator, you need to first install the operator dependencies and prerequisites:

          - For the list of operator dependencies, see the IBM Knowledge Center [Common Services dependencies documentation](http://ibm.biz/cpcs_opdependencies).

          - For the list of prerequisites for installing the operator, see the IBM Knowledge Center [Preparing to install services documentation](http://ibm.biz/cpcs_opinstprereq).

          ## Documentation

          To install the operator by using the IBM Common Services Operator, follow the installation and configuration instructions that are in the IBM Knowledge Center.

          - If you are using the operator as part of an IBM Cloud Pak, see the documentation for that IBM Cloud Pak [IBM Cloud Paks that use Common Services](http://ibm.biz/cpcs_cloudpaks).
          - If you are using the operator with an IBM Containerized Software, see the IBM Cloud Platform Common Services Knowledge Center [Installer documentation](http://ibm.biz/cpcs_opinstall).
        displayName: IBM IM Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: stable
        provider:
          name: IBM
        relatedImages:
        - registry.connect.redhat.com/cpopen/ibm-iam-operator@sha256:d7e2507481b280118997ec9fe8062a9d56ed0867265732ecb5c89ef6fccd98fc
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-platform-auth@sha256:52ce08ac2bce7c02dc30e2c0aa74df4bde8f156debccc59960f02b4b0040ffea
        - registry.marketplace.redhat.com/cpopen/cpfs/iam-custom-hostname@sha256:b8886b9b54e6ee4c6b2981ef092e54348a0286577d9924cf136cee881f5bf1f4
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-manager@sha256:0f46b288d060a15a48ecad8f328a8046bf232b8629c3d79da6a82ea712b363f4
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-bundle@sha256:98cf2a6c7b7af5d7cbfbc61ce140201272b30b12edabaec2b2d4287b0690d89e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator-catalog@sha256:f289ac0d44803ff12d41c3659e6fdf04c404e0b3b233b6bc44eec671abb1b47e
        - registry.marketplace.redhat.com/cpopen/ibm-iam-operator@sha256:0e7a8257632f74c9672a58413a6ec5f44b8ceea8124408f931e385a352fbdabd
        - registry.marketplace.redhat.com/cpopen/cpfs/icp-identity-provider@sha256:97b9ecef8caef156a0eb073b6eb2a34b63209b44511f5694a243f11285d88662
        - registry.marketplace.redhat.com/cpopen/cpfs/im-initcontainer@sha256:7cec55c00827b747f23c2f1703d99e1ca5b30a0c445f8ad0471b25b0c9cca57c
        version: 4.11.0
      entries:
      - name: ibm-iam-operator.v4.11.0
        version: 4.11.0
      name: v4.10
    defaultChannel: v4.10
    packageName: ibm-iam-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: VASTData
      provider-url: https://www.vastdata.com
    name: vast-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vast-csi-operator.v2.6.3
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"storage.vastdata.com/v1","kind":"VastCSIDriver","metadata":{"name":"csi.vastdata.com"},"spec":{"applySecurityContextConstraints":true,"attachRequired":true,"controller":{"dnsPolicy":"Default","extraArgs":{"csiAttacher":[],"csiProvisioner":[],"csiResizer":[],"csiSnapshotter":[]},"nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"resources":{"csiAttacher":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiProvisioner":{"limits":{"memory":"400Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiResizer":{"limits":{"memory":"300Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiSnapshotter":{"limits":{"memory":"200Mi"},"requests":{"cpu":"10m","memory":"20Mi"}},"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}}},"runOnControlPlane":false,"runOnMaster":false,"tolerations":[]},"deletionViewPolicy":null,"deletionVipPool":null,"dontUseTrashApi":false,"driverType":"nfs","endpoint":null,"image":{"csiAttacher":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94"},"csiNodeDriverRegistrar":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac"},"csiProvisioner":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385"},"csiResizer":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43"},"csiSnapshotter":{"imagePullPolicy":"IfNotPresent","repository":"registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c"},"csiVastPlugin":{"imagePullPolicy":"IfNotPresent","repository":"registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156"}},"imagePullSecrets":[],"kubeletPath":"/var/lib/kubelet","logLevel":5,"node":{"dnsPolicy":"Default","nodeAffinity":{},"nodeSelector":{},"podAffinity":{},"podAntiAffinity":{},"propagateHostMountOptions":true,"resources":{"csiVastPlugin":{"limits":{"memory":"400Mi"},"requests":{"cpu":"100m","memory":"50Mi"}},"nodeDriverRegistrar":{"limits":{"memory":"100Mi"},"requests":{"cpu":"10m","memory":"20Mi"}}},"tolerations":[]},"numWorkers":10,"operationRetryIntervalMax":60,"operationRetryIntervalStart":10,"operationTimeout":15,"secretName":null,"sslCertsSecretName":null,"truncateVolumeName":64,"useLocalIpForMount":null,"verifySsl":false}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastStorage","metadata":{"name":"vastdata-filesystem"},"spec":{"allowVolumeExpansion":true,"blockingClones":false,"clusterName":"","createSnapshotClass":true,"driverType":"nfs","ephemeralVolumeNameFormat":"csi:{namespace}:{name}:{id}","fsType":"","hostEncryption":null,"mountOptions":[],"provisioner":"","qosPolicy":"","reclaimPolicy":"Delete","secretName":"","secretNamespace":"","setDefaultStorageClass":false,"snapshotClass":{"deletionPolicy":"Delete","setDefaultSnapshotClass":false,"snapshotNameFormat":"csi:{namespace}:{name}:{id}"},"storagePath":"","subsystem":"","tenantName":"","viewPolicy":"","vipPool":"","vipPoolFQDN":null,"volumeGroup":"","volumeNameFormat":"csi:{namespace}:{name}:{id}"}},{"apiVersion":"storage.vastdata.com/v1","kind":"VastCluster","metadata":{"name":"vastcluster-sample"},"spec":{"endpoint":null,"passphrase":null,"password":null,"sslCert":null,"tenant":null,"token":null,"username":null}}]'
          capabilities: Seamless Upgrades
          categories: Storage
          containerImage: registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
          createdAt: "2024-07-01T11:59:59Z"
          description: |-
            VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
            The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
            ### Installation
            Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: vast-csi
          operators.operatorframework.io/builder: operator-sdk-v1.3.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/vast-data/vast-csi
          support: VastData Infra Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VastCSIDriver is a common specification for VAST CSI Controller
              and VAST CSI Node plugins, typically intended for creation in a single
              instance.
            displayName: VastCSIDriver
            kind: VastCSIDriver
            name: vastcsidrivers.storage.vastdata.com
            version: v1
          - description: VastCluster encapsulates the connection details required
              for accessing a VAST cluster. This information is ultimately stored
              in a Kubernetes Secret.
            displayName: VastCluster
            kind: VastCluster
            name: vastclusters.storage.vastdata.com
            version: v1
          - description: Represents VAST storage. Each VastStorage instance generates
              a corresponding storage class that can be used in PVCs to create workloads.
            displayName: VastStorage
            kind: VastStorage
            name: vaststorages.storage.vastdata.com
            version: v1
        description: |-
          VAST CSI Operator supports seamless installation and upgrade of VAST Data's CSI Driver.
          The VAST CSI Driver allows container orchestration frameworks such as Kubernetes to dynamically provision, manage and scale volumes on the [VAST Data Platform](https://www.vastdata.com/platform/overview). VAST provides scalable, reliable, and fast persistent storage that can be accessed remotely by any Kubernetes application container.
          ### Installation
          Refer to the VASTData Operator for Kubernetes [official documentation](https://support.vastdata.com/s/document-item?bundleId=vast-csi-driver-2.6-administrator-s-guide&topicId=deploying-vast-csi-driver/deploying-vast-csi-driver-on-openshift-with-vast-csi-operator.html&_LANG=enus).
        displayName: VAST CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - VAST
        - VASTData
        - CSI
        - Multi Cluster
        - AI
        - High-Performance Computing
        links:
        - name: VAST CSI Plugin
          url: https://github.com/vast-data/vast-csi
        maintainers:
        - email: volodymyr.boiko@vastdata.com
          name: VAST Infrastructure Team
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: VASTData
          url: https://www.vastdata.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:405a14e1aa702f7ea133cea459e8395fe40a6125c088c55569e696d48e1bd385
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:2e04046334baf9be425bb0fa1d04c2d1720d770825eedbdbcdb10d430da4ad8c
        - registry.connect.redhat.com/vastdataorg/csi@sha256:7169f279d0dd7130535d74df5b938ed6b7944bfd4fe52d242ceee5178f2ea156
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f25af73ee708ff9c82595ae99493cdef9295bd96953366cddf36305f82555dac
        - registry.k8s.io/sig-storage/csi-resizer@sha256:a541e6cc2d8b011bb21b1d4ffec6b090e85270cce6276ee302d86153eec0af43
        - registry.connect.redhat.com/vastdataorg/csi-operator@sha256:ba7f30a42e327004ecbc81ff5ccc453bf3046a29bab5785b9f8a9746e5c6a3fd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:b4d611100ece2f9bc980d1cb19c2285b8868da261e3b1ee8f45448ab5512ab94
        - registry.connect.redhat.com/vastdataorg/csi-operator-bundle@sha256:925cc412650ec3a32ea4a55ffe65d16b76916d51f0b57ff62182b4ce12c5d2d8
        version: 2.6.3
      entries:
      - name: vast-csi-operator.v2.6.3
        version: 2.6.3
      name: stable
    defaultChannel: stable
    packageName: vast-csi-operator
    provider:
      name: VASTData
      url: https://www.vastdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/file-integrity-operator
    name: file-integrity-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: file-integrity-operator.v0.1.13
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.13'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator
          url: https://file-integrity-operator.domain
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator-metadata@sha256:7fb2750941193d64f8cca278ff7d81d3ca5e29a2ad0568dff7ddae4396548bb3
        - registry.redhat.io/openshift4/file-integrity-rhel8-operator@sha256:a65f1244636649734763164a5d6155a6f6a7b79b3950ecc03fb3c58188583599
        version: 0.1.13
      entries:
      - name: file-integrity-operator.v0.1.13
        version: 0.1.13
      name: "4.7"
    - currentCSV: file-integrity-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:57e796822bcd705974c0efcd6c190ce39eac2e70b9d6a0c27f69d85dfe966db2
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:c3e34e6198eadb48c007cd949d8be38e26ef7d15d2f9047a8c3c89c03201e4ea
        version: 0.1.32
      entries:
      - name: file-integrity-operator.v0.1.32
        version: 0.1.32
      name: release-0.1
    - currentCSV: file-integrity-operator.v1.3.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          createdAt: "2024-11-21T18:02:08Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <1.3.6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:1e7781ff09c5cfb87ede98b96d31e58b760bedbf7a23f41778654faba3507591
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:f7039c6aaaa3e60dc4a3965cc1341c5217837e4b4030eb943771967e0b206578
        version: 1.3.6
      entries:
      - name: file-integrity-operator.v1.3.6
        version: 1.3.6
      - name: file-integrity-operator.v1.3.5
        version: 1.3.5
      - name: file-integrity-operator.v1.3.4
        version: 1.3.4
      - name: file-integrity-operator.v1.3.3
        version: 1.3.3
      name: stable
    - currentCSV: file-integrity-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fileintegrity.openshift.io/v1alpha1",
                "kind": "FileIntegrity",
                "metadata": {
                  "name": "example-fileintegrity",
                  "namespace": "openshift-file-integrity"
                },
                "spec": {
                  "config": {},
                  "debug": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.6 <1.0.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-file-integrity
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/file-integrity-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FileIntegrity is the Schema for the fileintegrities API
            displayName: File Integrity
            kind: FileIntegrity
            name: fileintegrities.fileintegrity.openshift.io
            version: v1alpha1
          - description: FileIntegrityNodeStatus defines the status of a specific
              node
            kind: FileIntegrityNodeStatus
            name: fileintegritynodestatuses.fileintegrity.openshift.io
            version: v1alpha1
        description: An operator that manages file integrity checks on nodes.
        displayName: File Integrity Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Integrity
        - security
        links:
        - name: File Integrity Operator upstream repo
          url: https://github.com/openshift/file-integrity-operator
        - name: File Integrity Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/file_integrity_operator/file-integrity-operator-understanding.html
        maintainers:
        - email: mrogers@redhat.com
          name: Matt Rogers
        maturity: alpha
        provider:
          name: Red Hat
          url: https://github.com/openshift/file-integrity-operator
        relatedImages:
        - registry.redhat.io/compliance/openshift-file-integrity-operator-bundle@sha256:f234e8af232f2d3a9472307ae82834f96756841b8d60b8c42ac4f9ca3fd17545
        - registry.redhat.io/compliance/openshift-file-integrity-rhel8-operator@sha256:dd1dfeccf59214a0de653c3a4c4dbbc8c41640954318d13386cda73891660e4f
        version: 1.0.0
      entries:
      - name: file-integrity-operator.v1.0.0
        version: 1.0.0
      name: v1
    defaultChannel: stable
    packageName: file-integrity-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/file-integrity-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-lambda-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-lambda-controller.v1.9.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Alias",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "CodeSigningConfig",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "EventSourceMapping",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "lambda.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
          createdAt: "2025-09-19T19:23:07Z"
          description: AWS Lambda controller is a service controller for managing
            Lambda resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alias represents the state of an AWS lambda Alias resource.
            displayName: Alias
            kind: Alias
            name: aliases.lambda.services.k8s.aws
            version: v1alpha1
          - description: CodeSigningConfig represents the state of an AWS lambda CodeSigningConfig
              resource.
            displayName: CodeSigningConfig
            kind: CodeSigningConfig
            name: codesigningconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: EventSourceMapping represents the state of an AWS lambda
              EventSourceMapping resource.
            displayName: EventSourceMapping
            kind: EventSourceMapping
            name: eventsourcemappings.lambda.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS lambda Function resource.
            displayName: Function
            kind: Function
            name: functions.lambda.services.k8s.aws
            version: v1alpha1
          - description: FunctionURLConfig represents the state of an AWS lambda FunctionURLConfig
              resource.
            displayName: FunctionURLConfig
            kind: FunctionURLConfig
            name: functionurlconfigs.lambda.services.k8s.aws
            version: v1alpha1
          - description: LayerVersion represents the state of an AWS lambda LayerVersion
              resource.
            displayName: LayerVersion
            kind: LayerVersion
            name: layerversions.lambda.services.k8s.aws
            version: v1alpha1
          - description: Version represents the state of an AWS lambda Version resource.
            displayName: Version
            kind: Version
            name: versions.lambda.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Lambda resources in AWS from within your Kubernetes cluster.

          **About Amazon Lambda**

          Lambda is a compute service that lets you run code without provisioning or managing servers. Lambda runs your code on a high-availability compute infrastructure and performs all of the administration of the compute resources, including server and operating system maintenance, capacity provisioning and automatic scaling, code monitoring and logging. With Lambda, you can run code for virtually any type of application or backend service. All you need to do is supply your code in one of the [languages that Lambda supports](https://docs.aws.amazon.com/lambda/latest/dg/lambda-runtimes.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Lambda
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lambda
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Lambda Developer Resources
          url: https://aws.amazon.com/lambda/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: lambda maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-lambda-controller@sha256:9b2943aa770ebd5638927b19f05f629575e1f8fef129f8520e01c2238771d7dc
        - public.ecr.aws/aws-controllers-k8s/lambda-controller:1.9.1
        version: 1.9.1
      entries:
      - name: ack-lambda-controller.v1.9.1
        version: 1.9.1
      - name: ack-lambda-controller.v1.9.0
        version: 1.9.0
      - name: ack-lambda-controller.v1.8.0
        version: 1.8.0
      - name: ack-lambda-controller.v1.7.7
        version: 1.7.7
      - name: ack-lambda-controller.v1.7.6
        version: 1.7.6
      - name: ack-lambda-controller.v1.7.5
        version: 1.7.5
      - name: ack-lambda-controller.v1.7.4
        version: 1.7.4
      - name: ack-lambda-controller.v1.7.2
        version: 1.7.2
      - name: ack-lambda-controller.v1.7.1
        version: 1.7.1
      - name: ack-lambda-controller.v1.7.0
        version: 1.7.0
      - name: ack-lambda-controller.v1.6.3
        version: 1.6.3
      - name: ack-lambda-controller.v1.6.2
        version: 1.6.2
      - name: ack-lambda-controller.v1.6.1
        version: 1.6.1
      - name: ack-lambda-controller.v1.6.0
        version: 1.6.0
      - name: ack-lambda-controller.v1.5.8
        version: 1.5.8
      - name: ack-lambda-controller.v1.5.7
        version: 1.5.7
      - name: ack-lambda-controller.v1.5.6
        version: 1.5.6
      - name: ack-lambda-controller.v1.5.5
        version: 1.5.5
      - name: ack-lambda-controller.v1.5.4
        version: 1.5.4
      - name: ack-lambda-controller.v1.5.3
        version: 1.5.3
      - name: ack-lambda-controller.v1.5.2
        version: 1.5.2
      - name: ack-lambda-controller.v1.5.1
        version: 1.5.1
      - name: ack-lambda-controller.v1.5.0
        version: 1.5.0
      - name: ack-lambda-controller.v1.4.2
        version: 1.4.2
      - name: ack-lambda-controller.v1.4.0
        version: 1.4.0
      - name: ack-lambda-controller.v1.3.4
        version: 1.3.4
      - name: ack-lambda-controller.v1.3.2
        version: 1.3.2
      - name: ack-lambda-controller.v1.3.1
        version: 1.3.1
      - name: ack-lambda-controller.v1.3.0
        version: 1.3.0
      - name: ack-lambda-controller.v1.2.1
        version: 1.2.1
      - name: ack-lambda-controller.v1.2.0
        version: 1.2.0
      - name: ack-lambda-controller.v1.1.1
        version: 1.1.1
      - name: ack-lambda-controller.v1.1.0
        version: 1.1.0
      - name: ack-lambda-controller.v1.0.1
        version: 1.0.1
      - name: ack-lambda-controller.v1.0.0
        version: 1.0.0
      - name: ack-lambda-controller.v0.1.7
        version: 0.1.7
      - name: ack-lambda-controller.v0.1.6
        version: 0.1.6
      - name: ack-lambda-controller.v0.1.5
        version: 0.1.5
      - name: ack-lambda-controller.v0.1.3
        version: 0.1.3
      - name: ack-lambda-controller.v0.1.2
        version: 0.1.2
      - name: ack-lambda-controller.v0.1.1
        version: 0.1.1
      - name: ack-lambda-controller.v0.1.0
        version: 0.1.0
      - name: ack-lambda-controller.v0.0.17
        version: 0.0.17
      - name: ack-lambda-controller.v0.0.16
        version: 0.0.16
      - name: ack-lambda-controller.v0.0.15
        version: 0.0.15
      - name: ack-lambda-controller.v0.0.14
        version: 0.0.14
      - name: ack-lambda-controller.v0.0.13
        version: 0.0.13
      - name: ack-lambda-controller.v0.0.12
        version: 0.0.12
      - name: ack-lambda-controller.v0.0.11
        version: 0.0.11
      - name: ack-lambda-controller.v0.0.10
        version: 0.0.10
      - name: ack-lambda-controller.v0.0.9
        version: 0.0.9
      - name: ack-lambda-controller.v0.0.8
        version: 0.0.8
      - name: ack-lambda-controller.v0.0.7
        version: 0.0.7
      - name: ack-lambda-controller.v0.0.6
        version: 0.0.6
      - name: ack-lambda-controller.v0.0.5
        version: 0.0.5
      - name: ack-lambda-controller.v0.0.4
        version: 0.0.4
      - name: ack-lambda-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-lambda-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: tempo-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tempo-operator.v0.18.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoMonolithic",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "jaegerui": {
                    "enabled": true,
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      }
                    },
                    "route": {
                      "enabled": true
                    }
                  },
                  "resources": {
                    "limits": {
                      "cpu": "2",
                      "memory": "2Gi"
                    }
                  },
                  "storage": {
                    "traces": {
                      "backend": "memory"
                    }
                  }
                }
              },
              {
                "apiVersion": "tempo.grafana.com/v1alpha1",
                "kind": "TempoStack",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "resources": {
                    "total": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2Gi"
                      }
                    }
                  },
                  "storage": {
                    "secret": {
                      "name": "my-storage-secret",
                      "type": "s3"
                    }
                  },
                  "storageSize": "1Gi",
                  "template": {
                    "queryFrontend": {
                      "jaegerQuery": {
                        "enabled": true,
                        "ingress": {
                          "type": "route"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
          createdAt: 22 Sep 2025, 11:15
          description: Create and manage deployments of Tempo, a high-scale distributed
            tracing backend.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.6.0 <0.18.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-tempo-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/tempo-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: TempoMonolithic manages a Tempo deployment in monolithic
              mode.
            displayName: TempoMonolithic
            kind: TempoMonolithic
            name: tempomonolithics.tempo.grafana.com
            version: v1alpha1
          - description: TempoStack manages a Tempo deployment in microservices mode.
            displayName: TempoStack
            kind: TempoStack
            name: tempostacks.tempo.grafana.com
            version: v1alpha1
        description: |-
          Red Hat OpenShift distributed tracing platform based on Tempo. Tempo is an open-source, easy-to-use, and highly scalable distributed tracing backend. It provides observability for microservices architectures by allowing developers to track requests as they flow through distributed systems. Tempo is optimized to handle large volumes of trace data and is designed to be highly performant even under heavy loads.
          It can ingest common open source tracing protocols including Jaeger, Zipkin, and OpenTelemetry and requires only object storage to operate.

          The Tempo Operator supports Tempo deployments in Microservices mode (`TempoStack` CR) and Monolithic mode (`TempoMonolithic` CR).

          ### Core capabilities
          Tempo is used for monitoring and troubleshooting microservices-based distributed systems, including:
          * Distributed transaction monitoring
          * Root cause analysis
          * Performance / latency optimization

          ### Operator features
          * **Resource Limits** - Specify overall resource requests and limits in the `TempoStack` CR; the operator assigns fractions of it to each component
          * **AuthN and AuthZ** - Supports OpenID Control (OIDC) and role-based access control (RBAC)
          * **Managed upgrades** - Updating the operator will automatically update all managed Tempo clusters
          * **Multitenancy** - Multiple tenants can send traces to the same Tempo cluster
          * **mTLS** - Communication between the Tempo components can be secured via mTLS
          * **Jaeger UI** - Traces can be visualized in Jaeger UI and exposed via Ingress or OpenShift Route
          * **Observability** - The operator and `TempoStack` operands expose telemetry (metrics, traces) and integrate with Prometheus `ServiceMonitor` and `PrometheusRule`

          ### Before you start
          `TempoStack` requires object storage to store its traces. `TempoMonolithic` can store traces in-memory, in a Persistent Volume and in object storage.
          Please ensure that your system has a compatible object storage solution that is supported, such as OpenShift Data Foundation, Minio, AWS S3, Azure Storage, or Google Cloud Storage.

          ### Examples
          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples).

          ### Support & Troubleshooting
          Tempo Operator is available and supported as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html/distributed_tracing) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Tempo Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tempo
        - tracing
        - observability
        - monitoring
        - database
        links:
        - name: Tempo Operator
          url: https://github.com/grafana/tempo-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/tempo-gateway-rhel8@sha256:095ec83219660bb1955758b94a9431b3add572da4dc2cba61b42b59fa6b4bba0
        - registry.redhat.io/rhosdt/tempo-jaeger-query-rhel8@sha256:53b62fea39ce211842026e8365d61f579b8e01e8587a1ce639423d72f3b44f45
        - registry.redhat.io/rhosdt/tempo-query-rhel8@sha256:5b89188bc587a81af3ecdca3caabd6e48356e9ba4f9070860f5160544ce30eb9
        - registry.redhat.io/rhosdt/tempo-rhel8-operator@sha256:9bfa5a2ffb09b501f001a4275379210621f06620adcf6454a9327c65538872f5
        - registry.redhat.io/rhosdt/tempo-rhel8@sha256:530a0ef3aaea3f07d6210a683a771c416cea8ce5accac8f5f751dcd7232cb341
        - registry.redhat.io/rhosdt/tempo-operator-bundle@sha256:bcc3521d20a1662d4032901f46e3bd4744c9910b060f52bdf10c6a8dfea1276b
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:6bcfa981a7a75c242a16afa6dab36bf6e3b8a1442fa863875919ad551f48eceb
        - registry.redhat.io/rhosdt/tempo-gateway-opa-rhel8@sha256:22e876837d00371d24feac3f989ead3d4f034400b292f043b18f3839bfde564a
        version: 0.18.0-1
      entries:
      - name: tempo-operator.v0.18.0-1
        version: 0.18.0-1
      - name: tempo-operator.v0.16.0-2
        version: 0.16.0-2
      - name: tempo-operator.v0.16.0-1
        version: 0.16.0-1
      - name: tempo-operator.v0.15.4-1
        version: 0.15.4-1
      - name: tempo-operator.v0.15.3-1
        version: 0.15.3-1
      - name: tempo-operator.v0.14.1-2
        version: 0.14.1-2
      - name: tempo-operator.v0.14.1-1
        version: 0.14.1-1
      - name: tempo-operator.v0.13.0-1-0.1730817419.p
        version: 0.13.0-1+0.1730817419.p
      - name: tempo-operator.v0.13.0-1
        version: 0.13.0-1
      - name: tempo-operator.v0.10.0-8
        version: 0.10.0-8
      - name: tempo-operator.v0.10.0-7
        version: 0.10.0-7
      - name: tempo-operator.v0.10.0-6
        version: 0.10.0-6
      - name: tempo-operator.v0.8.0-3
        version: 0.8.0-3
      - name: tempo-operator.v0.8.0-2
        version: 0.8.0-2
      - name: tempo-operator.v0.6.0-2
        version: 0.6.0-2
      - name: tempo-operator.v0.3.1-3
        version: 0.3.1-3
      - name: tempo-operator.v0.1.0-6
        version: 0.1.0-6
      - name: tempo-operator.v0.1.0-6-0.1687199947.p
        version: 0.1.0-6+0.1687199947.p
      name: stable
    defaultChannel: stable
    packageName: tempo-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://www.redhat.com
    name: sap-commerce-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sap-commerce-operator.v0.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisApp",
                "metadata": {
                  "name": "hybrisapp-sample"
                },
                "spec": {
                  "aJPServicePort": 30000,
                  "apachejvmRouteName": "",
                  "baseImageName": "hybris-base",
                  "baseImageTag": "v0.2",
                  "hybrisANTTaskNames": "",
                  "podHealthyProbePath": "/platform/init",
                  "sourceRepoLocalPropertiesOverride": "",
                  "sourceRepoRef": "main",
                  "sourceRepoURL": "https://github.com/redhat-sap/sap-commerce-app-example.git",
                  "startupProbeFailureThreshold": 100,
                  "startupProbePeriodSecond": 60
                }
              },
              {
                "apiVersion": "hybris.hybris.org/v1alpha1",
                "kind": "HybrisBase",
                "metadata": {
                  "name": "hybrisbase-sample"
                },
                "spec": {
                  "URL": "http://nexus.redhat.com/nexus/service/local/repositories/releases/content/org/hybris/hybris/1.0/hybris-1.0.tar.gz",
                  "buildSourceRepo": "https://github.com/redhat-sap/sap-commerce-operator",
                  "buildSourceRepoBranch": "main",
                  "imageName": "hybris-base",
                  "imageTag": "v0.2",
                  "jdkURL": "https://github.com/SAP/SapMachine/releases/download/sapmachine-11.0.5/sapmachine-jdk-11.0.5-1.x86_64.rpm",
                  "password": "admin123",
                  "username": "admin"
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: quay.io/redhat-sap-cop/hybris-operator:0.0.4
          description: Builds and runs SAP Commerce in OpenShift
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-sap/sap-commerce-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HybrisApp is the Schema for the hybrisapps API
            displayName: Hybris Application
            kind: HybrisApp
            name: hybrisapps.hybris.hybris.org
            version: v1alpha1
          - description: HybrisBase is the Schema for the Hybris Base API
            displayName: Hybris Base
            kind: HybrisBase
            name: hybrisbases.hybris.hybris.org
            version: v1alpha1
        description: |-
          Builds and runs SAP Commerce in OpenShift

          * The operator provides APIs to initialize SAP Commerce with the provided SAP Commerce Server. Though there are no restrictions from the operator on the base package to be used from SAP, the default SAP base image provides a server preconfigured with an application server based on Apache Tomcat. Both HSQLDB and Tomcat are included in the SAP Commerce package, and are designed to be used for testing, development, and demonstration purposes. The custom classes created for various business needs along with an external RDBMS configuration can be used in a HybrisApp creation to execute SAP Commerce in a production setup.

          * The operator also enables the use of SAP Commerce in a clustered model. A SAP Commerce Cluster is a number of individual SAP Commerce installations using a common set of data on one database. A cluster configuration lets you balance the load between the nodes to scale SAP Commerce in a way that is transparent to the user.

          * The clustering offers few communication protocols as options (UDP: both multicast or unicast, or JGroups). In order to support multicast for SAP Commerce clustering the corresponding OpenShift namespace has to be enabled for multicast usage.

          For more details on SAP Commerce consult the product documentation: [https://help.sap.com/viewer/product/SAP_COMMERCE](https://help.sap.com/viewer/product/SAP_COMMERCE)
        displayName: SAP Commerce Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sap-commerce-operator
        links:
        - name: SAP Commerce Operator
          url: https://github.com/redhat-sap/sap-commerce-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat, Inc.
          url: https://www.redhat.com
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/sap-commerce-operator@sha256:9b6a3c9c7a3af1a91578fa06c202d128fc7bfff3ae21300a4382cffb11def17a
        - quay.io/redhat-sap-cop/hybris-operator:0.0.4
        version: 0.0.4
      entries:
      - name: sap-commerce-operator.v0.0.4
        version: 0.0.4
      - name: sap-commerce-operator.v0.0.3
        version: 0.0.3
      - name: sap-commerce-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: sap-commerce-operator
    provider:
      name: Red Hat, Inc.
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Omar Gaye IBM
      provider-url: omarg.net
    name: memcached-operator-ogaye
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: memcached-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.omarg.net/v1alpha1",
                "kind": "Memcached",
                "metadata": {
                  "name": "memcached-sample"
                },
                "spec": {
                  "foo": "bar"
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.6.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Memcached is the Schema for the memcacheds API
            displayName: Memcached
            kind: Memcached
            name: memcacheds.cache.omarg.net
            version: v1alpha1
        description: Omar memcached operator
        displayName: memcached-operator-ogaye
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Omar
        - Memcached
        links:
        - name: Memcached Operator OGaye
          url: https://memcached-operator.domain
        maturity: alpha
        provider:
          name: Omar Gaye IBM
          url: omarg.net
        relatedImages:
        - registry.connect.redhat.com/ogaye/ogaye@sha256:f35a21fb3936b259aabe27b453612616fca4e6327c9042361b4d00ef332905c0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9211b70edbb93d718428e1c9a992e5fc74a20cb213f8a0592ffe340c362890fe
        - quay.io/ogaye/memcached-operator@sha256:5340ac11ce676dd3204bf3807a4c2333a8dddcbcec918db0aeafbc977163c1c6
        version: 0.0.1
      entries:
      - name: memcached-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: memcached-operator-ogaye
    provider:
      name: Omar Gaye IBM
      url: omarg.net
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE
      provider-url: ""
    name: hpe-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-csi-operator.v3.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "storage.hpe.com/v1",
                "kind": "HPECSIDriver",
                "metadata": {
                  "name": "hpecsidriver-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "csp": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  },
                  "disable": {
                    "alletra6000": false,
                    "alletra9000": false,
                    "alletraStorageMP": false,
                    "b10000FileService": false,
                    "nimble": false,
                    "primera": false
                  },
                  "disableHostDeletion": false,
                  "disableNodeConfiguration": false,
                  "disableNodeConformance": false,
                  "disableNodeGetVolumeStats": false,
                  "disableNodeMonitor": false,
                  "disablePreInstallHooks": false,
                  "imagePullPolicy": "IfNotPresent",
                  "images": {
                    "b10000FileServiceCSP": "quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013",
                    "csiAttacher": "registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da",
                    "csiControllerDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiExtensions": "quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89",
                    "csiNodeDriver": "quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676",
                    "csiNodeDriverRegistrar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97",
                    "csiProvisioner": "registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9",
                    "csiResizer": "registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4",
                    "csiSnapshotter": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769",
                    "csiVolumeGroupProvisioner": "quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399",
                    "csiVolumeGroupSnapshotter": "quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8",
                    "csiVolumeMutator": "quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed",
                    "nfsProvisioner": "quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002",
                    "nimbleCSP": "quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1",
                    "primera3parCSP": "quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf"
                  },
                  "iscsi": {
                    "chapSecretName": ""
                  },
                  "kubeletRootDir": "/var/lib/kubelet",
                  "logLevel": "info",
                  "maxVolumesPerNode": 100,
                  "node": {
                    "affinity": {},
                    "labels": {},
                    "nodeSelector": {},
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
          createdAt: "2025-09-25T16:28:16Z"
          description: A Container Storage Interface (CSI) driver for HPE storage
            platforms. The CSI driver allows you to use HPE storage with your preferred
            container orchestrator.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <3.0.1'
          operatorhub.io/ui-metadata-max-k8s-version: "1.33"
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/hpe-storage/co-deployments
          support: HPE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE CSI Driver for Kubernetes
            displayName: HPECSIDriver
            kind: HPECSIDriver
            name: hpecsidrivers.storage.hpe.com
            version: v1
        description: |
          The HPE CSI Operator for OpenShift deploys the HPE CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE storage systems.
          The HPE CSI Driver for Kubernetes leverages HPE storage platforms to provide scalable and persistent storage for stateful applications.
          Please refer to our CSI driver [documentation](https://scod.hpedev.io/csi_driver/index.html) for supported HPE platforms and full list of supported CSI features.
          ## Installation
          Refer to the HPE CSI Operator for OpenShift [documentation](https://scod.hpedev.io/csi_driver/deployment.html#operator).
        displayName: HPE CSI Operator for OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE CSI
        - HPE Storage
        - CSI
        links:
        - name: Documentation
          url: https://scod.hpedev.io/csi_driver/index.html
        maintainers:
        - email: dev-hi-containers@hpe.com
          name: HPE Storage Containers Team
        maturity: stable
        minKubeVersion: 1.26.0
        provider:
          name: HPE
        relatedImages:
        - quay.io/hpestorage/alletrastoragemp-b10000-nfs-csp@sha256:6a1855a2ef1a83b886bb1f45b596e8e420e5e13c84bbf446c8c202eaf27a2013
        - registry.connect.redhat.com/hpestorage/csi-driver-operator-bundle@sha256:39fedc5650ff2fef6ed1d47a5745851101b7dd8f6ce1f8b72ac3a78bf3c1fd6a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - quay.io/hpestorage/alletra-9000-primera-and-3par-csp@sha256:8af6e87075c47ba8d9714d55848ab2a0fef62b0768681e30d9e63ed7f396b0bf
        - quay.io/hpestorage/nfs-provisioner@sha256:5bf5952453ecca28e26724df85ebd5c8f0b7479ae4cb683d4a09c415da00e002
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8ddd178ba5d08973f1607f9b84619b58320948de494b31c9d7cd5375b316d6d4
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - quay.io/hpestorage/alletra-6000-and-nimble-csp@sha256:5480ffdebd74fdc547bc4a11b4434abac7bbfee3cee349f1ac893fbad84bb8d1
        - quay.io/hpestorage/csi-extensions@sha256:27074aaa8df3b867cfa3d12318612c45f8e541d750c27db5e080734592160c89
        - quay.io/hpestorage/csi-driver@sha256:fffd5a243dc0f7e1027f69754fce77e5df04cb0a71b808f22a3b3a09b392c676
        - quay.io/hpestorage/volume-group-provisioner@sha256:0a5f39f52113b963978a695bb8e5a5a3c4d387a02eb4d4a9647b4bc804de7399
        - registry.connect.redhat.com/hpestorage/csi-driver-operator@sha256:29e92f68cf0d7691a6c56876e7dde0d69a8cb3ff6ae6c9371c604e61cba16898
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - quay.io/hpestorage/volume-group-snapshotter@sha256:177d9abf8fbd3ac26ca4c12ca4b4adde20ebab308c79f10b04632f0171746fa8
        - quay.io/hpestorage/volume-mutator@sha256:bf6beac1af5edcd3c9774f0dcc9579b79ef9cc8e4b74a4238082ab9aa443b5ed
        version: 3.0.1
      entries:
      - name: hpe-csi-operator.v3.0.1
        version: 3.0.1
      name: stable
    defaultChannel: stable
    packageName: hpe-csi-operator
    provider:
      name: HPE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataStax
      provider-url: ""
    name: cass-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cass-operator.v1.23.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cassandra.datastax.com/v1beta1",
                "kind": "CassandraDatacenter",
                "metadata": {
                  "name": "dc1"
                },
                "spec": {
                  "clusterName": "development",
                  "config": {
                    "cassandra-yaml": {
                      "authenticator": "PasswordAuthenticator",
                      "authorizer": "CassandraAuthorizer",
                      "num_tokens": 16,
                      "role_manager": "CassandraRoleManager"
                    },
                    "jvm-server-options": {
                      "initial_heap_size": "1G",
                      "max_heap_size": "1G"
                    }
                  },
                  "managementApiAuth": {
                    "insecure": {}
                  },
                  "podTemplateSpec": {
                    "spec": {
                      "containers": [
                        {
                          "name": "cassandra"
                        }
                      ],
                      "securityContext": {}
                    }
                  },
                  "racks": [
                    {
                      "name": "rack1"
                    },
                    {
                      "name": "rack2"
                    },
                    {
                      "name": "rack3"
                    }
                  ],
                  "resources": {
                    "requests": {
                      "cpu": "1000m",
                      "memory": "2Gi"
                    }
                  },
                  "serverType": "cassandra",
                  "serverVersion": "4.1.2",
                  "size": 3,
                  "storageConfig": {
                    "cassandraDataVolumeClaimSpec": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "resources": {
                        "requests": {
                          "storage": "10Gi"
                        }
                      },
                      "storageClassName": "server-storage"
                    }
                  }
                }
              },
              {
                "apiVersion": "control.k8ssandra.io/v1alpha1",
                "kind": "CassandraTask",
                "metadata": {
                  "name": "example-cleanup"
                },
                "spec": {
                  "datacenter": {
                    "name": "dc2",
                    "namespace": "cass-operator"
                  },
                  "jobs": [
                    {
                      "command": "cleanup",
                      "name": "cleanup-run"
                    }
                  ]
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database,Big Data,OpenShift Optional
          certified: "true"
          containerImage: cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
          createdAt: "2025-03-20"
          description: |
            Simple provisioning, turn-key operations, and automated remediation of Apache Cassandra clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/k8ssandra/cass-operator
          support: DataStax
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CassandraDatacenter is the Schema for the cassandradatacenters
              API
            displayName: Cassandra Datacenter
            kind: CassandraDatacenter
            name: cassandradatacenters.cassandra.datastax.com
            version: v1beta1
          - description: CassandraTask is the Schema for the cassandrajobs API
            displayName: Cassandra Task
            kind: CassandraTask
            name: cassandratasks.control.k8ssandra.io
            version: v1alpha1
        description: "## Apache Cassandra\n\nThe Apache Cassandra database is the
          right choice when you need scalability and\nhigh availability without compromising
          performance. Linear scalability and\nproven fault-tolerance on commodity
          hardware or cloud infrastructure make it the\nperfect platform for mission-critical
          data. Cassandra's support for replicating\nacross multiple datacenters is
          best-in-class, providing lower latency for your\nusers and the peace of
          mind of knowing that you can survive regional outages.\n\n## DataStax Enterprise\n\nThe
          most advanced distribution of Apache Cassandra™ on the market, with the\nenterprise
          functionality needed for serious production systems and backed up and\nsupported
          by the best distributed-experts in the world. It's one platform for\nall
          types of applications anywhere, any cloud, any model: key-value, graph,\ntabular,
          JSON.\n\nDataStax Enterprise is a fully integrated and optimized database,
          with graph,\nanalytics, and search included, all with a unified security
          model. Simply put,\nit's the only database capable of meeting today's demanding
          requirements\n\n## Operator Details\n\n`cass-operator` is designed as a
          modular operator for Apache Cassandra and\nderived  distributions. Apache
          Cassandra is a distributed database consisting of\nmultiple nodes working
          in concert to store data and process queries along a\nnumber of fault domains.
          `cass-operator` has the deployment of a Cassandra\ncluster around the logical
          domain of a datacenter with the `CassandraDatacenter`\ncustom resource.\n
          \   \nUpon submission of one of these resources it handles provisioning
          the underlying\nstateful sets (analogous to C\\* logical racks), services,
          and configuration.\nAdditionally through monitoring pod state via Kubernetes
          callbacks it handles day to day\noperations such as restarting failed processes,
          scaling clusters up, and deploying\nconfiguration changes in a rolling,
          non-disruptive, fashion.\n    \nThis operator is designed to be `Namespace`
          scoped. A single Kubernetes cluster may\nbe running multiple instances of
          this operator, in separate namespaces, to support\na number of C\\* clusters
          and environments. Configuration is simple with the usage of\nYAML based
          overrides in the Custom Resource paired with an `init` container.\n    \nIn
          C\\* clusters ordering and timing of certain operations are important to
          keep the system\nevenly distributed. `cass-operator` takes advantage of
          a sidecar process within the\nmain container to handle the orchestration
          of starting our main C* process."
        displayName: DataStax Kubernetes Operator for Apache Cassandra
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - k8ssandra
        - cass-operator
        - DataStax Enterprise
        - Apache Cassandra
        - DSE
        - C*
        - Cassandra
        links:
        - name: DataStax
          url: https://datastax.com
        - name: Operator GitHub Repo
          url: https://github.com/k8ssandra/cass-operator
        - name: Config Builder GitHub Repo
          url: https://github.com/datastax/cass-config-builder
        - name: Documentation [1]
          url: https://github.com/k8ssandra/cass-operator/tree/master/docs/user
        - name: Documentation [2]
          url: https://docs.datastax.com/en/cass-operator/doc/cass-operator/cassOperatorTOC.html
        maintainers:
        - email: cass-operator@datastax.com
          name: DataStax Operator Team
        maturity: stable
        minKubeVersion: 1.21.0
        provider:
          name: DataStax
        relatedImages:
        - cr.k8ssandra.io/k8ssandra/cass-operator:v1.23.2
        - quay.io/community-operator-pipeline-prod/cass-operator-community@sha256:428effade29fcb3e8981e84f7df945f9bc49fe1a9b56217573db8b387706c26b
        version: 1.23.2
      entries:
      - name: cass-operator.v1.23.2
        version: 1.23.2
      - name: cass-operator.v1.23.1
        version: 1.23.1
      - name: cass-operator.v1.23.0
        version: 1.23.0
      - name: cass-operator.v1.21.1
        version: 1.21.1
      - name: cass-operator.v1.19.1
        version: 1.19.1
      - name: cass-operator.v1.17.1
        version: 1.17.1
      - name: cass-operator.v1.16.0
        version: 1.16.0
      - name: cass-operator.v1.14.0
        version: 1.14.0
      - name: cass-operator.v1.13.0
        version: 1.13.0
      - name: cass-operator.v1.10.4
        version: 1.10.4
      - name: cass-operator.v1.10.1
        version: 1.10.1
      name: stable
    defaultChannel: stable
    packageName: cass-operator-community
    provider:
      name: DataStax
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Seldon Technologies
      provider-url: ""
    name: seldon-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: seldon-operator.v1.14.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machinelearning.seldon.io/v1",
                "kind": "SeldonDeployment",
                "metadata": {
                  "labels": {
                    "app": "seldon",
                    "app.kubernetes.io/instance": "seldon1",
                    "app.kubernetes.io/name": "seldon",
                    "app.kubernetes.io/version": "v0.5"
                  },
                  "name": "seldon-model"
                },
                "spec": {
                  "name": "test-deployment",
                  "predictors": [
                    {
                      "componentSpecs": [
                        {
                          "spec": {
                            "containers": [
                              {
                                "image": "seldonio/mock_classifier:1.6.0",
                                "name": "classifier"
                              }
                            ]
                          }
                        }
                      ],
                      "graph": {
                        "children": [],
                        "name": "classifier",
                        "type": "MODEL"
                      },
                      "name": "example",
                      "replicas": 1
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning
          certified: "false"
          containerImage: docker.io/seldonio/seldon-core-operator:1.14.1
          createdAt: "2019-05-21 15:00:00"
          description: The Seldon operator for management, monitoring and operations
            of machine learning systems through the Seldon Engine. Once installed,
            the Seldon Operator provides multiple functions which facilitate the productisation,
            monitoring and maintenance of machine learning systems at scale.
          olm.skipRange: <1.13.1
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/SeldonIO/seldon-core
          support: Clive Cox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A seldon engine deployment
            displayName: Seldon Deployment
            kind: SeldonDeployment
            name: seldondeployments.machinelearning.seldon.io
            version: v1
        description: "The Seldon operator enables for native operation of production
          machine learning workloads, including monitoring and operations of language-agnostic
          models with the benefits of real-time metrics and log analysis.\n   \n##
          Overview\nSeldon Core is an open source platform for deploying machine learning
          models on a Kubernetes cluster.\n\n* Deploy machine learning models in the
          cloud or on-premise.\n* Get metrics and ensure proper governance and compliance
          for your running machine learning models.\n* Create powerful inference graphs
          made up of multiple components.\n* Provide a consistent serving layer for
          models built using heterogeneous ML toolkits.\n\nYou can get started by
          following the guides in our documentation at https://docs.seldon.io/projects/seldon-core/en/latest/workflow/README.html\n"
        displayName: Seldon Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mlops
        - aiops
        - production
        - monitoring
        links:
        - name: Website
          url: https://www.seldon.io/
        - name: Documentation
          url: https://docs.seldon.io/projects/seldon-core/en/latest/
        maintainers:
        - email: hello@seldon.io
          name: Seldon Technologies
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Seldon Technologies
        relatedImages:
        - docker.io/seldonio/seldon-core-operator:1.14.1
        - quay.io/openshift-community-operators/seldon-operator@sha256:864e87ef21e81624010a1cbcee6b17c8c816dcb0a699c66badd32eec3baf39fd
        - ""
        version: 1.14.1
      entries:
      - name: seldon-operator.v1.14.1
        version: 1.14.1
      name: stable
    defaultChannel: stable
    packageName: seldon-operator
    provider:
      name: Seldon Technologies
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microsafe
      provider-url: https://microsafe.com.mx/
    name: operator-aims-mobile
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-aims-mobile.v0.0.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.microsafe.com.mx/v1",
                "kind": "AimsMobile",
                "metadata": {
                  "name": "aimsmobile-sample"
                },
                "spec": {
                  "adminApp": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "NODE_ENV": "production",
                      "VITE_EXPRESS_HOST": "aims-mob-service",
                      "VITE_EXPRESS_PORT": "8080",
                      "VITE_EXPRESS_PROTOCOL": "http",
                      "VITE_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "VITE_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "VITE_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "VITE_VERSION": "1.0.0"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0",
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-admin",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-mob-admin-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8083,
                      "targetPort": 8080
                    }
                  },
                  "dbService": {
                    "autoscaling": {
                      "enabled": false
                    },
                    "config": {
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port"
                    },
                    "image": "quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9",
                    "ingress": {
                      "enabled": false
                    },
                    "name": "aims-db-service",
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "cpu": "300m",
                        "memory": "256Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "128Mi"
                      }
                    },
                    "secretName": "aims-db-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 3000,
                      "targetPort": 3000
                    }
                  },
                  "mainService": {
                    "annotations": {
                      "instrumentation.opentelemetry.io/inject-nodejs": "true"
                    },
                    "autoscaling": {
                      "enabled": true,
                      "maxReplicas": 2,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "config": {
                      "AIMS1_API_URL": "http://aims-db-service:3000",
                      "EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_HOST": "aims-mob-admin",
                      "FRONTEND_EXPRESS_PORT": "4000",
                      "FRONTEND_EXPRESS_PROTOCOL": "http",
                      "FRONTEND_IP": "aims-mob-admin",
                      "FRONTEND_KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "FRONTEND_KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "FRONTEND_KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "FRONTEND_PORT": "80",
                      "FRONTEND_PROTOCOL": "http",
                      "FRONTEND_VERSION": "1.0.0",
                      "KEYCLOAK_CLIENT_ID": "rhbk-aims.prod.id",
                      "KEYCLOAK_REALM": "rhbk-aims.prod.realm",
                      "KEYCLOAK_URL": "rhbk-aims.prod.url",
                      "MSSQL_HOST": "mssql-database.prod.host",
                      "MSSQL_PORT": "mssql-database.prod.port",
                      "NODE_ENV": "production",
                      "OTLP_LOG_EXPORTER_URL": "opentelemetry-logs.prod.url",
                      "OTLP_METRIC_EXPORTER_URL": "opentelemetry-metrics.prod.url",
                      "OTLP_TRACE_EXPORTER_URL": "opentelemetry-traces.prod.url",
                      "REMEDY_API_URL": "remedy-aims.prod.svc",
                      "REMEDY_USER": "RMDYUSR"
                    },
                    "ingress": {
                      "enabled": true,
                      "isOpenshiftRoute": true
                    },
                    "name": "aims-mob-service",
                    "proxyContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949",
                      "name": "aims-mob-service-proxy",
                      "resources": {
                        "limits": {
                          "cpu": "2000m",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    },
                    "replicas": 2,
                    "secretName": "aims-mob-service-secrets",
                    "service": {
                      "enabled": true,
                      "port": 8085,
                      "targetPort": 8081
                    },
                    "serviceContainer": {
                      "image": "quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17",
                      "name": "aims-mob-service-main",
                      "resources": {
                        "limits": {
                          "cpu": "2500m",
                          "memory": "2048Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "128Mi"
                        }
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": false,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools, Integration & Delivery
          createdAt: "2025-09-25T19:23:26Z"
          description: Este operador despliega los servicios AIMS Mobile.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/network-aware: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          instrumentation.opentelemetry.io/inject-nodejs: "true"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: Valid Subscription Microsafe
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://quay.io/repository/microsafe_app_ban/operador-aims-mobile-private
          support: Microsafe
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AimsMobile
            name: aimsmobiles.cache.microsafe.com.mx
            version: v1
        description: Este operador despliega la aplicación AIMS Mobile.
        displayName: AIMS Mobile Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aims mobile
        - aims
        links:
        - name: AIMS Mobile Operator
          url: https://microsafe.com.mx/
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Microsafe
          url: https://microsafe.com.mx/
        relatedImages:
        - registry.connect.redhat.com/microsafe/operador-aims-mobile-bundle@sha256:1572f715c13d852c2a824b726912c4e5cc49aea43d0d91b6fd673f7198b0d038
        - quay.io/microsafe_app_ban/aims-db-service-private@sha256:bcb4d8624626436526549427f9f169cff714b39cbcc530dd2b3470c6712abaf9
        - quay.io/microsafe_app_ban/aims-mob-admin-private@sha256:8e333f900885b0d3f6442c8ffcb43bd93540b8c38ad5bf17cbbd78929e0b1be0
        - quay.io/microsafe_app_ban/aims-mob-proxy-private@sha256:29b82de987b28c23f17d5f2176877c5aecea1a0ef182bb3a63612290ff469949
        - quay.io/microsafe_app_ban/aims-mob-service-private@sha256:f9586200106aa4de259f5cae2d22d13dc147a76dfd06e4f4bea09797071dad17
        - quay.io/microsafe_app_ban/operador-aims-mobile-private@sha256:731f2d731c79ade02f4b6c39b380590833cacddfd795b703cec10ad55f730fad
        version: 0.0.10
      entries:
      - name: operator-aims-mobile.v0.0.10
        version: 0.0.10
      - name: operator-aims-mobile.v0.0.9
        version: 0.0.9
      - name: operator-aims-mobile.v0.0.8
        version: 0.0.8
      name: alpha
    defaultChannel: alpha
    packageName: operator-aims-mobile
    provider:
      name: Microsafe
      url: https://microsafe.com.mx/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v23.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "isva-sample"
                },
                "spec": {
                  "image": "icr.io/isva/verify-access-wrp:10.0.5.0",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator:23.3.0
          createdAt: Mon Mar  20 00:59:34 UTC 2023
          description: The IBM Security Verify Access Operator manages the lifecycle
            of IBM Security Verify Access worker containers.
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Security Verify Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Security Verify Access.

          The IBM Security Verify Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Security Verify Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Security Verify Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - quay.io/openshift-community-operators/ibm-security-verify-access-operator@sha256:ec3970c7a6a5ee3ac863dfe6f16234fe19cd7ea4a541fd0cbae9a8183d0763a2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - icr.io/isva/verify-access-operator:23.3.0
        version: 23.3.0
      entries:
      - name: ibm-security-verify-access-operator.v23.3.0
        version: 23.3.0
      - name: ibm-security-verify-access-operator.v22.10.0
        version: 22.10.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: windows-machine-config-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windows-machine-config-operator.v10.18.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          createdAt: "2025-05-14T20:50:15Z"
          description: An operator that enables Windows container workloads on OCP
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=10.17.0 <10.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-windows-machine-config-operator
          operators.openshift.io/valid-subscription: '["Red Hat OpenShift support
            for Windows Containers"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/windows-machine-config-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          ### Introduction
          The Windows Machine Config Operator configures Windows Machines into nodes, enabling Windows container workloads to
          be run on OCP clusters. Windows instances can be added either by creating a [MachineSet](https://docs.openshift.com/container-platform/4.18/machine_management/creating_machinesets/creating-machineset-aws.html#machine-api-overview_creating-machineset-aws),
          or by specifying existing instances through a [ConfigMap](https://docs.openshift.com/container-platform/4.18/windows_containers/byoh-windows-instance.html)
          The operator completes all the necessary steps to configure the Windows instance so that it can join the cluster as a worker node.

          Usage of this operator requires an OpenShift Container Platform for Windows Containers subscription. Users looking to
          deploy Windows containers workloads in production clusters should acquire a subscription before attempting to
          install this operator. Users without a subscription can try the community operator, a distribution which lacks
          official support.

          ### Pre-requisites
          * A Red Hat OpenShift Container Platform for Windows Containers [subscription](https://access.redhat.com/support/policy/updates/openshift#windows)
          * OCP 4.18 cluster running on Azure, AWS, GCP or vSphere configured with hybrid OVN Kubernetes networking
          * [WMCO prerequisites](https://docs.openshift.com/container-platform/4.18/windows_containers/understanding-windows-container-workloads.html#wmco-prerequisites__understanding-windows-container-workloads)

          ### Usage
          Once the `openshift-windows-machine-config-operator` namespace has been created, a secret must be created containing
          the private key that will be used to access the Windows VMs:
          ```
          # Create secret containing the private key in the openshift-windows-machine-config-operator namespace
          oc create secret generic cloud-private-key --from-file=private-key.pem=/path/to/key -n openshift-windows-machine-config-operator
          ```
          We strongly recommend not using the same
          [private key](https://docs.openshift.com/container-platform/4.18/installing/installing_azure/installing-azure-default.html#ssh-agent-using_installing-azure-default)
          used when installing the cluster

          Below is an example of a vSphere Windows MachineSet which can create Windows Machines that the WMCO can react upon.
          Please note that the windows-user-data secret will be created by the WMCO lazily when it is configuring the first
          Windows Machine. After that, the windows-user-data will be available for the subsequent MachineSets to be consumed.
          It might take around 10 minutes for the Windows VM to be configured so that it joins the cluster. Please note that
          the MachineSet should have following labels:
          * `machine.openshift.io/os-id: Windows`
          * `machine.openshift.io/cluster-api-machine-role: worker`
          * `machine.openshift.io/cluster-api-machine-type: worker`

          The following label must be added to the Machine spec within the MachineSet spec:
          * `node-role.kubernetes.io/worker: ""`

          Not having these labels will result in the Windows node not being marked as a worker.

          `<infrastructureID>` should be replaced with the output of:
          ```
          oc get -o jsonpath='{.status.infrastructureName}{"\n"}' infrastructure cluster
          ```

          The following template variables need to be replaced as follows with values from your vSphere environment:
          * `<Windows_VM_template>`: template name
          * `<VM Network Name>`: network name
          * `<vCenter DataCenter Name>`: datacenter name
          * `<Path to VM Folder in vCenter>`: path where your OpenShift cluster is running
          * `<vCenter Datastore Name>`: datastore name
          * `<vCenter Server FQDN/IP>`: IP address or FQDN of the vCenter server

          Please note that on vSphere, Windows Machine names cannot be more than 15 characters long. The MachineSet name therefore
          cannot be more than 9 characters long, due to the way Machine names are generated from it.
          ```
          apiVersion: machine.openshift.io/v1beta1
          kind: MachineSet
          metadata:
            labels:
              machine.openshift.io/cluster-api-cluster: <infrastructureID>
            name: winworker
            namespace: openshift-machine-api
          spec:
            replicas: 1
            selector:
              matchLabels:
                machine.openshift.io/cluster-api-cluster: <infrastructureID>
                machine.openshift.io/cluster-api-machineset: winworker
            template:
              metadata:
                labels:
                  machine.openshift.io/cluster-api-cluster: <infrastructureID>
                  machine.openshift.io/cluster-api-machine-role: worker
                  machine.openshift.io/cluster-api-machine-type: worker
                  machine.openshift.io/cluster-api-machineset: winworker
                  machine.openshift.io/os-id: Windows
              spec:
                metadata:
                  labels:
                    node-role.kubernetes.io/worker: ""
                providerSpec:
                  value:
                    apiVersion: vsphereprovider.openshift.io/v1beta1
                    credentialsSecret:
                      name: vsphere-cloud-credentials
                      namespace: openshift-machine-api
                    diskGiB: 128
                    kind: VSphereMachineProviderSpec
                    memoryMiB: 16384
                    metadata:
                      creationTimestamp: null
                    network:
                      devices:
                      - networkName: "<VM Network Name>"
                    numCPUs: 4
                    numCoresPerSocket: 1
                    snapshot: ""
                    template: <Windows_VM_template
                    userDataSecret:
                      name: windows-user-data
                    workspace:
                      datacenter: <vCenter DataCenter Name>
                      datastore: <vCenter Datastore Name>
                      folder: <Path to VM Folder in vCenter> # e.g. /DC/vm/ocp45-2tdrm
                      server: <vCenter Server FQDN/IP>
          ```
          Example MachineSet for other cloud providers:
          - [AWS](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-aws.html)
          - [Azure](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-azure.html)
          - [GCP](https://docs.openshift.com/container-platform/4.18/windows_containers/creating_windows_machinesets/creating-windows-machineset-gcp.html)

          ### Limitations
          #### DeploymentConfigs
          Windows Nodes do not support workloads created via DeploymentConfigs. Please use a normal Deployment, or other method to
          deploy workloads.

          ### Reporting issues
          Support for this distribution of WMCO requires a Red Hat OpenShfit subscription. Support should be requested through the Red Hat Customer Portal.
          Please read through the [troubleshooting document](https://docs.openshift.com/container-platform/4.18/support/troubleshooting/troubleshooting-windows-container-workload-issues.html)
          before opening a support case.
        displayName: Windows Machine Config Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - windows
        maintainers:
        - email: team-winc@redhat.com
          name: Red Hat, Windows Container Support for OpenShift
        maturity: stable
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4-wincw/windows-machine-config-rhel9-operator@sha256:f391499545a2981a3c3492766bbaae6a72a82db07d88251e85fda5d672a108e0
        - registry.redhat.io/openshift4-wincw/windows-machine-config-operator-bundle@sha256:4ebf4ace7dcaf0a935b83314b2a514baffca704b2f36cec5cd61dcf13e459946
        version: 10.18.1
      entries:
      - name: windows-machine-config-operator.v10.18.1
        version: 10.18.1
      - name: windows-machine-config-operator.v10.18.0
        version: 10.18.0
      name: stable
    defaultChannel: stable
    packageName: windows-machine-config-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: web-terminal
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: web-terminal.v1.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
            ]
          capabilities: Basic Install
          categories: Developer Tools
          certified: "false"
          containerImage: registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
          createdAt: "2021-10-26T07:24:32Z"
          description: Start a web terminal in your browser with common CLI tools
            for interacting with the cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "true"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/redhat-developer/web-terminal-operator/
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          required:
          - kind: DevWorkspaceRouting
            name: devworkspaceroutings.controller.devfile.io
            version: v1alpha1
          - kind: DevWorkspace
            name: devworkspaces.workspace.devfile.io
            version: v1alpha1
        description: |
          Start a web terminal in your browser with common CLI tools for interacting with
          the cluster.

          **Note:** The Web Terminal Operator integrates with the OpenShift Console in
          OpenShift 4.5.3 and higher to simplify web terminal instance creation and
          automate OpenShift login. In earlier versions of OpenShift, the operator can
          be installed but web terminals will have to be created and accessed manually.

          ## Description
          The Web Terminal Operator leverages the
          [DevWorkspace Operator](https://github.com/devfile/devworkspace-operator)
          to provision enviroments which support common cloud CLI tools. When this
          operator is installed, the DevWorkspace Operator will be installed as a
          dependency.

          ## How to Install
          Press the **Install** button, choose the upgrade strategy, and wait for the
          **Installed** Operator status.

          When the operator is installed, you will see a terminal button appear on the
          top right of the console after refreshing the OpenShift console window.

          ## How to Uninstall
          The Web Terminal Operator requires manual steps to fully uninstall the operator.
          As the Web Terminal Operator is designed as a way to access the OpenShift
          cluster, web terminal instances store user credentials. To avoid exposing these
          credentials to unwanted parties, the operator deploys webhooks and finalizers
          that aren't removed when the operator is uninstalled. See the
          [uninstall guide](https://docs.openshift.com/container-platform/latest/web_console/web_terminal/uninstalling-web-terminal.html)
          for more details.

          ## Documentation
          Documentation for this Operator is available at https://docs.openshift.com/container-platform/latest/web\_console/web\_terminal/installing-web-terminal.html
        displayName: Web Terminal
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - workspace
        - devtools
        - developer
        - ide
        - terminal
        links:
        - name: Web Terminal Repo
          url: https://github.com/redhat-developer/web-terminal-operator/
        maintainers:
        - email: dakwon@redhat.com
          name: David Kwon
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/web-terminal/web-terminal-exec-rhel9@sha256:6e58f2d6a1619a06be05807ff5407924087b7b27611628f9baf999733b67645b
        - registry.redhat.io/web-terminal/web-terminal-operator-bundle@sha256:fb969c8b8ab53ab6639795086efad6870dd212f04c94934b75910598efcf3ca3
        - registry.redhat.io/web-terminal/web-terminal-rhel9-operator@sha256:68396a490566a3c9e82ad2b96270a39218f33204bc02cdc836fb9acd54d184f3
        - registry.redhat.io/web-terminal/web-terminal-tooling-rhel9@sha256:0b133afa920b5180a3c3abe3dd5c73a9cfc252a71346978f94adcb659d683404
        version: 1.13.0
      entries:
      - name: web-terminal.v1.13.0
        version: 1.13.0
      - name: web-terminal.v1.12.1
        version: 1.12.1
      - name: web-terminal.v1.12.1-0.1745393748.p
        version: 1.12.1+0.1745393748.p
      - name: web-terminal.v1.12.1-0.1743657317.p
        version: 1.12.1+0.1743657317.p
      - name: web-terminal.v1.12.1-0.1742484227.p
        version: 1.12.1+0.1742484227.p
      - name: web-terminal.v1.11.1-0.1747215995.p
        version: 1.11.1+0.1747215995.p
      - name: web-terminal.v1.11.1
        version: 1.11.1
      - name: web-terminal.v1.11.0
        version: 1.11.0
      - name: web-terminal.v1.10.1
        version: 1.10.1
      - name: web-terminal.v1.10.1-0.1740684238.p
        version: 1.10.1+0.1740684238.p
      - name: web-terminal.v1.10.0-0.1720435222.p
        version: 1.10.0+0.1720435222.p
      - name: web-terminal.v1.10.0-0.1732652667.p
        version: 1.10.0+0.1732652667.p
      - name: web-terminal.v1.10.0-0.1731481377.p
        version: 1.10.0+0.1731481377.p
      - name: web-terminal.v1.10.0-0.1727169028.p
        version: 1.10.0+0.1727169028.p
      - name: web-terminal.v1.10.0
        version: 1.10.0
      - name: web-terminal.v1.10.0-0.1720402943.p
        version: 1.10.0+0.1720402943.p
      - name: web-terminal.v1.9.0-0.1708477317.p
        version: 1.9.0+0.1708477317.p
      - name: web-terminal.v1.9.0
        version: 1.9.0
      - name: web-terminal.v1.8.0-0.1708477299.p
        version: 1.8.0+0.1708477299.p
      - name: web-terminal.v1.8.0-0.1701199376.p
        version: 1.8.0+0.1701199376.p
      - name: web-terminal.v1.8.0-0.1692219820.p
        version: 1.8.0+0.1692219820.p
      - name: web-terminal.v1.8.0
        version: 1.8.0
      - name: web-terminal.v1.7.0-0.1684429884.p
        version: 1.7.0+0.1684429884.p
      - name: web-terminal.v1.7.0-0.1708477265.p
        version: 1.7.0+0.1708477265.p
      - name: web-terminal.v1.7.0
        version: 1.7.0
      - name: web-terminal.v1.7.0-0.1681197295.p
        version: 1.7.0+0.1681197295.p
      - name: web-terminal.v1.7.0-0.1682321121.p
        version: 1.7.0+0.1682321121.p
      - name: web-terminal.v1.7.0-0.1692219820.p
        version: 1.7.0+0.1692219820.p
      - name: web-terminal.v1.6.0
        version: 1.6.0
      - name: web-terminal.v1.6.0-0.1692219820.p
        version: 1.6.0+0.1692219820.p
      - name: web-terminal.v1.5.1-0.1661829403.p
        version: 1.5.1+0.1661829403.p
      - name: web-terminal.v1.5.1
        version: 1.5.1
      - name: web-terminal.v1.5.0-0.1657220207.p
        version: 1.5.0+0.1657220207.p
      - name: web-terminal.v1.5.0
        version: 1.5.0
      - name: web-terminal.v1.4.0
        version: 1.4.0
      - name: web-terminal.v1.3.0
        version: 1.3.0
      name: fast
    defaultChannel: fast
    packageName: web-terminal
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: servicemeshoperator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: "1.0"
    - currentCSV: servicemeshoperator.v2.6.10
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"maistra.io/v2\",\n    \"kind\":
            \"ServiceMeshControlPlane\",\n    \"metadata\": {\n      \"name\": \"basic\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"version\": \"v2.6\",\n      \"policy\": {\n         \"type\":
            \"Istiod\"\n      },\n      \"telemetry\": {\n         \"type\": \"Istiod\"\n
            \     },\n      \"addons\": {\n        \"prometheus\": {\n          \"enabled\":
            true\n        },\n        \"kiali\": {\n          \"enabled\": true\n
            \       },\n        \"grafana\": {\n          \"enabled\": true\n        }\n
            \     }\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n    \"kind\":
            \"ServiceMeshMemberRoll\",\n    \"metadata\": {\n      \"name\": \"default\",\n
            \     \"namespace\": \"control-plane-namespace\"\n    },\n    \"spec\":
            {\n      \"members\": [\n        \"your-project\",\n        \"another-of-your-projects\"
            \n      ]\n    }\n  },\n  {\n    \"apiVersion\": \"maistra.io/v1\",\n
            \   \"kind\": \"ServiceMeshMember\",\n    \"metadata\": {\n      \"name\":
            \"default\",\n      \"namespace\": \"application-namespace\"\n    },\n
            \   \"spec\": {\n      \"controlPlaneRef\": {\n        \"name\": \"basic\",\n
            \       \"namespace\": \"control-plane-namespace\"\n      }\n    }\n  }\n]"
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
          createdAt: 2025-08-11T09:32:44CEST
          description: The OpenShift Service Mesh 2 Operator enables you to install,
            configure, and manage an instance of Red Hat OpenShift Service Mesh 2.6
            and earlier releases. OpenShift Service Mesh 2 is based on the open source
            Istio project.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2.6.10-0'
          operators.openshift.io/infrastructure-features: '["Disconnected","fips"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/maistra/istio-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: An Istio control plane installation
            displayName: Istio Service Mesh Control Plane
            kind: ServiceMeshControlPlane
            name: servicemeshcontrolplanes.maistra.io
            version: v2
          - description: Marks the containing namespace as a member of the referenced
              Service Mesh
            displayName: Istio Service Mesh Member
            kind: ServiceMeshMember
            name: servicemeshmembers.maistra.io
            version: v1
          - description: A list of namespaces in Service Mesh
            displayName: Istio Service Mesh Member Roll
            kind: ServiceMeshMemberRoll
            name: servicemeshmemberrolls.maistra.io
            version: v1
        description: |-
          Red Hat OpenShift Service Mesh 2 is a platform that provides behavioral insight and operational control over a service mesh, providing a uniform way to connect, secure, and monitor microservice applications.

          ### Overview

          Red Hat OpenShift Service Mesh 2, based on the open source [Istio](https://istio.io/) project, adds a transparent layer on existing
          distributed applications without requiring any changes to the service code. You add Red Hat OpenShift Service Mesh 2
          support to services by deploying a special sidecar proxy throughout your environment that intercepts all network
          communication between microservices. You configure and manage the service mesh using the control plane features.

          Red Hat OpenShift Service Mesh 2 provides an easy way to create a network of deployed services that provides discovery,
          load balancing, service-to-service authentication, failure recovery, metrics, and monitoring. A service mesh also
          provides more complex operational functionality, including A/B testing, canary releases, rate limiting, access
          control, and end-to-end authentication.

          ### Core Capabilities

          Red Hat OpenShift Service Mesh 2 supports uniform application of a number of key capabilities across a network of services:

          + **Traffic Management** - Control the flow of traffic and API calls between services, make calls more reliable,
            and make the network more robust in the face of adverse conditions.

          + **Service Identity and Security** - Provide services in the mesh with a verifiable identity and provide the
            ability to protect service traffic as it flows over networks of varying degrees of trustworthiness.

          + **Policy Enforcement** - Apply organizational policy to the interaction between services, ensure access policies
            are enforced and resources are fairly distributed among consumers. Policy changes are made by configuring the
            mesh, not by changing application code.

          + **Telemetry** - Gain understanding of the dependencies between services and the nature and flow of traffic between
            them, providing the ability to quickly identify issues.

          ### Joining Projects Into a Mesh

          Once an instance of Red Hat OpenShift Service Mesh 2 has been installed, it will only exercise control over services within its own
          project.  Other projects may be added into the mesh using one of two methods:

          1. A **ServiceMeshMember** resource may be created in other projects to add those projects into the mesh.  The
            **ServiceMeshMember** specifies a reference to the **ServiceMeshControlPlane** object that was used to install
            the control plane.  The user creating the **ServiceMeshMember** resource must have permission to *use* the
            **ServiceMeshControlPlane** object.  The adminstrator for the project containing the control plane can grant
            individual users or groups the *use* permissions.

          2. A **ServiceMeshMemberRoll** resource may be created in the project containing the control plane.  This resource
            contains a single *members* list of all the projects that should belong in the mesh.  The resource must be named
            *default*.  The user creating the resource must have *edit* or *admin* permissions for all projects in the
            *members* list.

          ### More Information

          * [Documentation](https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html)
          * [Bugs](https://issues.redhat.com/projects/OSSM)
        displayName: Red Hat OpenShift Service Mesh 2
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - istio
        - maistra
        - servicemesh
        links:
        - name: Red Hat OpenShift Service Mesh 2
          url: https://docs.openshift.com/container-platform/latest/service_mesh/v2x/servicemesh-release-notes.html
        - name: Istio
          url: https://istio.io/
        - name: Operator Source Code
          url: https://github.com/Maistra/istio-operator
        - name: Bugs
          url: https://issues.redhat.com/projects/OSSM
        maintainers:
        - email: istio-feedback@redhat.com
          name: Red Hat, OpenShift Service Mesh
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:c77183e50a34d33b5e2fede60350f54b18ab3536c3189e331569d5df862d9979
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:515c5d9e6700fae981b0ea3807436b250acb1534e9ecf50043e5df8474e66dce
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:6a420228e684596ce2b900614a874f3e1284a6f2bbbba4c5a71ade158867affe
        - registry.redhat.io/openshift-service-mesh/istio-rhel8-operator@sha256:22c81aa0538d48bb865134dc35db63e58983aae9b3f948c7e3bc70f376e67474
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:4ed63a5440753b1b22d44191246c82e7671e3f545c1dbdae4cb4fad163ca3a46
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:1ca80927334c6af8e386de8d44704f568b4b35d3f7afbe785a999d0fa4efd0bf
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:d16a60204d266bf4cc266c005fabf175b533b827500f171013884a713c88857a
        - registry.redhat.io/openshift4/ose-prometheus@sha256:423ae2e81cfe660b7217f7662754f366a91d457d6dd1589b4fc40733896eb9d6
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:b241452cce5c9af5aa0f12d1db3a7094cc004ce355988b8a5207091810a6f84b
        - registry.redhat.io/openshift-service-mesh/istio-operator-bundle@sha256:de03cab8e92e3c24a7a966ff291bf6904f0b9331993d3217bf7b8c67b510d075
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:e734ca310e09f737b4f06f3ecfeb9ac12971c63884fdcc27e3bebef9d5b18722
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel9@sha256:c1cadf40bdb9511d612a0257c12eb3a9738823592ebb6e32a4f51b3a8dfe08db
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:b134578e0677963197e353d8acd0ac3330df1b6d2752aa41cb072e86b5b4b91b
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:57eb7ba03f24781501a7ceed2e160ff95317869af6c2caf8f9a83990e6a6c7b9
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader@sha256:fc500a282c6668ff987a632efbbd19ecc3aebdc021f831a7b386e360a5c8324d
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:3dc1bddb070fd530ce6251666989b9a849de90591718a578da4a9900f98d93ec
        - registry.redhat.io/openshift-service-mesh/pilot-rhel8@sha256:c34b4cd9a82fef40708a25f0568f6dc632878affba35d5d631a48ec31da08cd3
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6b4623c416fc85bd8e38904c16494fa3c4aeae25c90dc0da9763a0a87141bc04
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d95762c62087072cadee5948c706619f09202a349e1310f0ed81bfb42f11203f
        - registry.redhat.io/openshift-service-mesh/ratelimit-rhel8@sha256:a3d65983d46dcf47b94f26e7265ed32082208ce08abcf3587a9a4e15c19b274a
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:a36d5db6097749c55e946bba3b8ddd13a84b1f6f68b62d0ccbad5a803e9b46fb
        - registry.redhat.io/openshift-service-mesh/grafana-rhel8@sha256:6fb7d50b188dc27ffb9a396edabbfb21767fe438344fd70d4754d1ec7cf27957
        - registry.redhat.io/openshift-service-mesh/istio-cni-rhel8@sha256:4669724b430683bf2e895c5cf941c2d99536117815999d38131e5b6f162bbef0
        - registry.redhat.io/openshift-service-mesh/istio-must-gather-rhel8@sha256:3d773d16fed72278949b9e3b54dc59e346e95341dc94b715bcfc328d3c9e83b2
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-service-mesh/prometheus-rhel8@sha256:71e193a2d403bb06a7a76686b376985315aa0c32f495b8b9dacd73c7ccce2af3
        version: 2.6.10-0
      entries:
      - name: servicemeshoperator.v2.6.10
        version: 2.6.10-0
      - name: servicemeshoperator.v2.6.9
        version: 2.6.9-0
      - name: servicemeshoperator.v2.6.8
        version: 2.6.8-0
      - name: servicemeshoperator.v2.6.7
        version: 2.6.7-0
      - name: servicemeshoperator.v2.6.6
        version: 2.6.6-0
      - name: servicemeshoperator.v2.6.5
        version: 2.6.5-0
      - name: servicemeshoperator.v2.6.4
        version: 2.6.4-0
      - name: servicemeshoperator.v2.6.3
        version: 2.6.3-0
      - name: servicemeshoperator.v2.6.2
        version: 2.6.2-0
      - name: servicemeshoperator.v2.6.1
        version: 2.6.1-0
      - name: servicemeshoperator.v2.6.0
        version: 2.6.0-0
      - name: servicemeshoperator.v2.5.2
        version: 2.5.2-0
      - name: servicemeshoperator.v2.5.1
        version: 2.5.1-0
      - name: servicemeshoperator.v2.5.0
        version: 2.5.0-0
      - name: servicemeshoperator.v2.4.5
        version: 2.4.5-0
      - name: servicemeshoperator.v2.4.4
        version: 2.4.4-0
      - name: servicemeshoperator.v2.4.3
        version: 2.4.3-0
      - name: servicemeshoperator.v2.4.2
        version: 2.4.2-0
      - name: servicemeshoperator.v2.4.1
        version: 2.4.1-0
      - name: servicemeshoperator.v2.4.0
        version: 2.4.0-0
      - name: servicemeshoperator.v2.3.3
        version: 2.3.3-0
      - name: servicemeshoperator.v2.3.2
        version: 2.3.2-0
      - name: servicemeshoperator.v2.3.1
        version: 2.3.1-0
      - name: servicemeshoperator.v2.3.0
        version: 2.3.0-0
      - name: servicemeshoperator.v2.2.3
        version: 2.2.3-0
      - name: servicemeshoperator.v2.2.2
        version: 2.2.2-0
      - name: servicemeshoperator.v2.2.1
        version: 2.2.1-0
      - name: servicemeshoperator.v2.2.0
        version: 2.2.0-0
      - name: servicemeshoperator.v2.1.3
        version: 2.1.3-0
      - name: servicemeshoperator.v2.1.2
        version: 2.1.2-0
      - name: servicemeshoperator.v2.1.1
        version: 2.1.1-0
      - name: servicemeshoperator.v2.1.0
        version: 2.1.0-0
      - name: servicemeshoperator.v2.0.8
        version: 2.0.8-0
      name: stable
    defaultChannel: stable
    packageName: servicemeshoperator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://redhat.com/
    name: openstack-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openstack-operator.v1.0.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.openstack.org/v1beta1",
                "kind": "OpenStack",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "openstack-operator",
                    "app.kubernetes.io/instance": "openstack",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "openstack",
                    "app.kubernetes.io/part-of": "openstack-operator"
                  },
                  "name": "openstack",
                  "namespace": "openstack-operators"
                },
                "spec": {
                  "operatorOverrides": [
                    {
                      "controllerManager": {
                        "resources": {
                          "limits": {
                            "cpu": "600m",
                            "memory": "2Gi"
                          }
                        }
                      },
                      "name": "infra",
                      "replicas": 1
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-02T13:33:09Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.0.14
          operatorframework.io/initialization-resource: '{"apiVersion":"operator.openstack.org/v1beta1","kind":"OpenStack","metadata":{"name":"openstack","namespace":"openstack-operators"},"spec":{}}'
          operatorframework.io/suggested-namespace: openstack-operators
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OpenStack is the Schema for the openstacks API
            displayName: Open Stack
            kind: OpenStack
            name: openstacks.operator.openstack.org
            version: v1beta1
        description: Install and configure OpenStack
        displayName: OpenStack
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenStack
        links:
        - name: Openstack Operator
          url: https://github.com/openstack-k8s-operators/
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://redhat.com/
        relatedImages:
        - registry.redhat.io/rhoso/openstack-multipathd-rhel9@sha256:6ff8ae176f8a633bcd78b1f8407924c74bb2f317eba16b90e7f0f9e53af74e8e
        - registry.redhat.io/rhoso/openstack-nova-api-rhel9@sha256:4116206f1fbe60fdd9458a5edee848571c94e2b0512d9c2cfcfd6890d60d4395
        - registry.redhat.io/rhoso/openstack-designate-mdns-rhel9@sha256:ce43dd1a30afd4753aa11d32df2b24ed4bf3d67847044aaae8f3e5c8901caf05
        - registry.redhat.io/ubi9/httpd-24@sha256:29382518403a4361c48fccc057163dc4a9f43ee159432754fa63c76485b223f5
        - registry.redhat.io/rhoso/openstack-heat-api-cfn-rhel9@sha256:2789929ae3edd7c68579f34457d3071273e49a66f462043cd9c46e75bdda282a
        - registry.redhat.io/rhoso/openstack-swift-object-rhel9@sha256:8fc2ab8caa9c7c35d94ef51c303763ea71e40b9c795c4555c17e6194344419b3
        - registry.redhat.io/rhoso/openstack-manila-api-rhel9@sha256:e4dd3d5b66c5216b45a1515a975b85fd2ac4ef698a9350dd16e01417da462b32
        - registry.redhat.io/rhoso/openstack-barbican-worker-rhel9@sha256:485525116fb3c22d8c60330dd87ebd3854cd71ded36bf88bcf18cee5957e22a3
        - registry.redhat.io/rhoso/openstack-ovn-base-rhel9@sha256:a3b8ec7ad1073facb8876acc787549a1a68b9bf81bebdc127f13eb6385e2d1da
        - registry.redhat.io/rhoso-operators/placement-rhel9-operator@sha256:a16bd4fb929eb5bcf96f2aefcb5a7abf4143f4f9ed24e53de93f4a3caa89a507
        - registry.redhat.io/rhoso-operators/openstack-baremetal-agent-rhel9@sha256:08675b16d89d7931b1ef7b96e1e884c88e8e7fbfbe79a944f9213c52aba3ba83
        - registry.redhat.io/rhoso-operators/glance-rhel9-operator@sha256:ee22cf417b7771930a442c2302277e6d37f368f5ae551fee9a50dbbe41c8fe3a
        - registry.redhat.io/rhoso-operators/barbican-rhel9-operator@sha256:b3b769e0dc32c12c5cc9578c2cac2622b973177ca14697bc1042d21d473e4e84
        - registry.redhat.io/rhoso-operators/manila-rhel9-operator@sha256:d9e69eb140b8ff76d5bf49deab8d700639ccdfd1ab5c8c7667bc2bbb50f3d60b
        - registry.redhat.io/rhoso/openstack-designate-worker-rhel9@sha256:a6324a8a72722822df7692b8df0420388a1b4aecfff3da7e91da4afe4d9e2260
        - registry.redhat.io/rhoso/openstack-neutron-sriov-agent-rhel9@sha256:118d5f9430192ab48cf159ee2f7912f08e935ad56ef3364f18b3680210bb3e52
        - registry.redhat.io/rhoso/openstack-ironic-api-rhel9@sha256:50ccfe7d9ba22a50e0c1c8159cade6f4b1ac13ad8f5770f20c7be18fba58c8c3
        - registry.redhat.io/rhoso/openstack-nova-conductor-rhel9@sha256:c67ca0e6c8126f56c6a83632b6a96bd4a35006f5f414a8779467b81864b31d12
        - registry.redhat.io/rhoso/openstack-designate-api-rhel9@sha256:b76ca7820553285bb4ee2bfb08e39aedbfe8c16b4edb478c06418a1e335a64ab
        - registry.redhat.io/rhoso/openstack-iscsid-rhel9@sha256:508bd67021c271a10bb55ab4af033dc4e931708674b9c5ba94f72c07cd0c8ee2
        - registry.redhat.io/rhoso/openstack-ironic-pxe-rhel9@sha256:42e35ded74033f1f07287441917829cc934930a5df696c33cd8957be3c10fa41
        - registry.redhat.io/rhoso-operators/nova-rhel9-operator@sha256:3156dde249e49b40bed71092fb5213b112d43b663ba8c556e355aed524bec822
        - registry.redhat.io/rhoso-operators/telemetry-rhel9-operator@sha256:0635515a5fed12f17de4cc5ac5f9567e3b3546fc2f69730c7beab25cf722f538
        - registry.redhat.io/rhoso-operators/ironic-rhel9-operator@sha256:dd67ffe3578ce619748735c1a5cd2a57d704d0c36c1b841d513ea36479aa4682
        - registry.redhat.io/rhoso/openstack-barbican-keystone-listener-rhel9@sha256:cd8d4f0a94b4fe080ebf8e459f091db6637ceec6d83ac23cf04d66bd03c71e76
        - registry.redhat.io/rhoso/openstack-cron-rhel9@sha256:f09e724c69cec2c3b685783f9263b26c280a57b5909a723341416b342d18f9a4
        - registry.redhat.io/rhoso-operators/prometheus-podman-exporter-rhel9@sha256:f33e7506db6dfdf530d3e2442baaf360d212bb1a47a7db887dba967ed0655cc2
        - registry.redhat.io/rhoso/openstack-ironic-neutron-agent-rhel9@sha256:0abaa8f437025d2e7006c59c546ff6935556d707b57993719faef94142e1c43d
        - registry.redhat.io/rhoso/openstack-nova-scheduler-rhel9@sha256:f0e5d3a693c9998ca3839b3ec7dfd5da0f9cf139e0b5b6913e92a10d5a13e6bc
        - registry.redhat.io/rhoso/openstack-ceilometer-notification-rhel9@sha256:bd2b5f50a2bef4e77e3daa89aca51df6e0e55c3784886296d92c97616cd566d1
        - registry.redhat.io/rhoso/openstack-heat-api-rhel9@sha256:ccbc41b552bb34d223c9615a4e3b56bb516bc5bd8a694f048c15063ba45fca9f
        - registry.redhat.io/rhoso/openstack-aodh-api-rhel9@sha256:4adc09cc1d922b387320da467de89d3f41a9c49b46d9b728458fec36505ec940
        - registry.redhat.io/rhoso/edpm-hardened-uefi-rhel9@sha256:c18ec7023fbcc4ab5183f689adc1f01a626a4923c15c61a37388f99904db7bc3
        - registry.redhat.io/rhoso/openstack-netutils-rhel9@sha256:b02ab1617922a2560bf739172bc2ce0f6b1fb3ef55054bd17e4b16bdf9512b4b
        - registry.redhat.io/rhoso/openstack-glance-api-rhel9@sha256:b1c0b09bb9aafe3bc31f43f8a99a92dde8d78cf90e781dd14300580986d8c980
        - registry.redhat.io/rhoso/openstack-cinder-volume-rhel9@sha256:8a09e80dac94fb4b3d35033e91812f95f0acac66532d9abc88aa1fb715d21822
        - registry.redhat.io/rhoso/openstack-neutron-metadata-agent-ovn-rhel9@sha256:20323b151474703df610b6c24ccb9c748265196bbc70a7b5ece1b801ff540cd9
        - registry.redhat.io/rhoso-operators/mysqld-exporter-rhel9@sha256:b49e966cdd588071e0e1f2791ec6c62310078a7d39ea77f83843d3708a07c2b5
        - registry.redhat.io/rhoso/openstack-manila-scheduler-rhel9@sha256:8ce47f7fd8660167b3b292659d4fc58cf34ed671d5779693d24b71a9fddc1e7c
        - registry.redhat.io/rhoso/openstack-barbican-api-rhel9@sha256:bc942204843c45c6cb7ea841af8611f04efb003de7eeccc6813934aad556da73
        - registry.redhat.io/rhoso/openstack-nova-compute-rhel9@sha256:6264075310231af358803b53d5a09250dab32d63fc5966aa8fb957e5f56096a8
        - registry.redhat.io/rhoso/openstack-neutron-dhcp-agent-rhel9@sha256:cc9223ef705bcd5d0dcabbf71ad9f6e540ae7f003a6afd13161301f842dfa5fa
        - registry.redhat.io/rhoso/openstack-aodh-notifier-rhel9@sha256:d2e89f27aac9afa504411ae6a6163442cd717f8c560e505a7da8e3fd9e2cff95
        - registry.redhat.io/rhoso/openstack-mariadb-rhel9@sha256:6115a7458cfa412c1659242fcec64b63f0a5ca414ecf3c5f890f0afec0bca069
        - registry.redhat.io/rhoso/openstack-designate-backend-bind9-rhel9@sha256:82cca999357b0ab25bcc081bf802fab831587cf661db806af4ca453e0a626743
        - registry.redhat.io/rhoso-operators/openstack-rhel9-operator@sha256:64e2ad312f757101f083eedcb17e0b6c05cf6ebf6ed9e7db2e33e84c6aca9898
        - registry.redhat.io/rhoso/openstack-ironic-inspector-rhel9@sha256:88357e83e8a4bbc3c5a37f0420ad75556821834a975cd8a1cf5fee56de2a4f99
        - registry.redhat.io/openshift4/ose-prometheus-node-exporter-rhel9@sha256:479c51de2fc8a3b3858d71bbe4eeafbd5b43263472346f9dde348115082894cc
        - registry.redhat.io/rhoso-operators/openstack-must-gather-rhel9@sha256:60b4355f4dbef3ad64103357086b10a5c0af126dd23798d6a9d725646f1e13b4
        - registry.redhat.io/rhoso/openstack-aodh-evaluator-rhel9@sha256:48c9192acf157df5b8f0827e35f0380d50cd4562676aba3c0d7791f92328af02
        - registry.redhat.io/rhoso/openstack-cinder-scheduler-rhel9@sha256:2079ffe9b620e9277573ac796bf7e393b520631130e5bbb6f5ab51d872c943f3
        - registry.redhat.io/rhoso/openstack-octavia-housekeeping-rhel9@sha256:0a79259e398af746d0e8a6dda1aae5ee976c6bf8991b17f4a5da0d07dca1e36c
        - registry.redhat.io/rhoso/openstack-ovn-northd-rhel9@sha256:668f742b2a18579bc8cfc610a862e43c5b877413cf65eabbe0ee36b3c7869446
        - registry.redhat.io/rhoso/openstack-octavia-worker-rhel9@sha256:303e35926e43fe403673e18a25c2e815547c28a5f8c0f09d470cd4148a21c528
        - registry.redhat.io/rhoso/openstack-rsyslog-rhel9@sha256:272e729accc2b64b8fcc0ece4ae5aa9ce6d454e34a25a8a8e2b99ac56b73361e
        - registry.redhat.io/rhoso/openstack-ironic-conductor-rhel9@sha256:538da9fdeec546d794d192361d0a76f8b4e7682b6eb6d9cea371d8a52d603800
        - registry.redhat.io/rhoso/openstack-octavia-api-rhel9@sha256:7b8d5927162ddab512096beb02e837ba51f6353c5187f3de72eb3db6e56ab6e7
        - registry.redhat.io/rhoso-operators/heat-rhel9-operator@sha256:583abb81a49db85949f7a4b846591ad8532afa79440ad7e2937ee00eacd972b3
        - registry.redhat.io/rhoso/openstack-ceilometer-compute-rhel9@sha256:296f7f6c6bdcd7f5d07b8470308a4c71055f29a17c48d2ebb03b7f54a0e927ff
        - registry.redhat.io/openshift4/ose-kube-state-metrics-rhel9@sha256:db6074b8aeedfd7603ae3adbe67071115f54e7f01a24997f1cd8fa8958aeddce
        - registry.redhat.io/rhoso-operators/horizon-rhel9-operator@sha256:9a2a13635588c6cc9949b272e1a1708c5fb53f19b05b8310ace703df5958ac63
        - registry.redhat.io/rhoso/openstack-manila-share-rhel9@sha256:277174c885998e6f5b867d099664f6fa28f974bffb524903ff75283035cd8dc9
        - registry.redhat.io/rhoso/openstack-neutron-ovn-agent-rhel9@sha256:bed1956c66b0e70064b51e34faac64036f77c7ecb14f46ceac4bc9ef52f0b9e0
        - registry.redhat.io/rhoso-operators/openstack-operator-bundle@sha256:945271aec89541d6d8a749c41af2fff3fe24e41fd50e935e3944451c940bf940
        - registry.redhat.io/rhoso/openstack-ovn-bgp-agent-rhel9@sha256:a404c3416ab65e7134f438362e887529be474639e725e18f3542e86973edac87
        - registry.redhat.io/rhoso/openstack-keystone-rhel9@sha256:ac076c81247d774a9aeee1b63540ca6e18058213cf6aad10a6fd1f535a394ef4
        - registry.redhat.io/rhoso/openstack-heat-engine-rhel9@sha256:151e04b3e5a408aadef3a17954255acc3733d09961878a135e177188457447f0
        - registry.redhat.io/rhoso-operators/infra-rhel9-operator@sha256:a2dc031a6ad2726d155e77cafb676f8e35a86e963b97ccdceed0da137e67ba0d
        - registry.redhat.io/rhoso/openstack-cinder-backup-rhel9@sha256:910693233ee3a1f3832ebe744f67a5139249523456de4abbd731c9d2f544b7c3
        - registry.redhat.io/rhoso-operators/ovn-rhel9-operator@sha256:66cd8add6078b63a36879c974db9e9f358593bb0651aa39fe85153d25404ae27
        - registry.redhat.io/rhoso/openstack-tempest-all-rhel9@sha256:16b1fbfd36ce9a27106bf09c414eaa93f14c9a6b11839068976c2fdb537d2ba7
        - registry.redhat.io/rhoso/openstack-ovn-sb-db-server-rhel9@sha256:8ba4f8ed49e214f421e94133da0f17b54ce99a52bbbd367d0aef357de844ea01
        - registry.redhat.io/rhoso/openstack-nova-novncproxy-rhel9@sha256:04881867f9b24403bf6ef051eee3902a17e8476eee894ef579c0d5d130d2510d
        - registry.redhat.io/rhoso/openstack-ovn-nb-db-server-rhel9@sha256:8c2ce1027779dc66d67a0e0d03e48b226d3278208cae0b2b81b35ad1b1f3ae72
        - registry.redhat.io/rhoso/openstack-swift-container-rhel9@sha256:95fec49643fc18f402b8ba8057f2c35f214bf6138c5f0d7bbe1b96cbd78ef8d6
        - registry.redhat.io/rhoso/openstack-ceilometer-ipmi-rhel9@sha256:2fb5723c68d0f2c3ae28b86e606ccae4b80f3bd9b16c01dd6c9adcae629db7f1
        - registry.redhat.io/rhoso/openstack-placement-api-rhel9@sha256:0aa5e138947bf236f31c81bfb6f641d95af63812c06dca43141721b383c7e08e
        - registry.redhat.io/rhoso/openstack-designate-producer-rhel9@sha256:04af4230f815a6e3fde22003cf382eebd9cbc3c3ec4056795f4474abc1e65fea
        - registry.redhat.io/rhoso-operators/openstack-network-exporter-rhel9@sha256:8e8e775e37c28cb11b7aae9199405b85d5862dce195d165ca598fa1dc89bf914
        - registry.redhat.io/rhoso/openstack-frr-rhel9@sha256:f6516277149d1560c0bf2ac74a99cfb4ac2c84f9f090ab231b6d12bdb824de3b
        - registry.redhat.io/rhoso/openstack-aodh-listener-rhel9@sha256:037845be6787f610e479345206b0fcfb2f1dc430540816bc5a54a1d26da2acbd
        - registry.redhat.io/rhoso/openstack-rabbitmq-rhel9@sha256:56bc2d2b201225fb98db43ea9ca68cceb4842f8f0f1a96ab3ab62ab785651d12
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:2cc2308a6d2ac74ac820f042f93ebff78c7742541e92aabc9450aca792e1fe83
        - registry.redhat.io/rhoso/ironic-python-agent-rhel9@sha256:2ccd8e22a242059ca5c4e2bce853c36a71c36fdf40e8e36ab183a8218b0cc87b
        - registry.redhat.io/rhoso/openstack-octavia-health-manager-rhel9@sha256:8065851058019757b56f86ca4f030d44f169100eb2e37e08e0587a7b1e4d2e72
        - registry.redhat.io/rhoso-operators/test-rhel9-operator@sha256:df7566483abb18aa3c45d8bc4cd91531c7eb5b63345e511151a7cdca2dbcef17
        - registry.redhat.io/rhoso-operators/designate-rhel9-operator@sha256:d446af622bd32666eb4eb5f9b7f38f0c33ac81251a2230db9d5baf71081b9d6a
        - registry.redhat.io/rhoso-operators/mariadb-rhel9-operator@sha256:907e3a9d15941440d5289d7e448d9f0351053fc20fcb8cb3b9db0a84debd876f
        - registry.redhat.io/rhoso-operators/ee-openstack-ansible-ee-rhel9@sha256:61e729fb89c4872837cbe9cb1e3cea402b61f0da240751302605df487721f0aa
        - registry.redhat.io/rhoso-operators/keystone-rhel9-operator@sha256:879d4b1f9e720f8581b3a55c788a97bbe7b52f9108df3d5dce2e9bcb4200df6d
        - registry.redhat.io/rhoso-operators/cinder-rhel9-operator@sha256:12b8dd8fd3683250365eff6f07339cb976ee29da80642e591c10e9fa5ba627c9
        - registry.redhat.io/rhoso/openstack-ceilometer-central-rhel9@sha256:e6d9f47406b7e3ff2a7f2256088cc64f043b8e497382b9f569eb90182598bbe9
        - registry.redhat.io/rhoso/openstack-swift-proxy-server-rhel9@sha256:858551fd1c519797cae0e146e9c05412ed6e66dac01d0e68e62dd2d4bd335391
        - registry.redhat.io/rhoso-operators/rabbitmq-cluster-rhel9-operator@sha256:e095e6cbda841b9c726c1ec0839a98964b894cf25ae9228052d314b1ee2f161d
        - registry.redhat.io/rhoso/openstack-memcached-rhel9@sha256:3e704b9664c65be39e83c2ed802722ffeaf969d8e571147a2be128d5c56331c1
        - registry.redhat.io/rhoso/openstack-neutron-server-rhel9@sha256:2d3afde3201aa18b192ad0fb079c1e4f4f673565604f1af2b79aba8a7bc88bf2
        - registry.redhat.io/rhoso/openstack-cinder-api-rhel9@sha256:72a674caeeeecce3b60217f3607ad5b00651fa3d52f2f5d905f6e6395638dbc8
        - registry.redhat.io/rhoso-operators/octavia-rhel9-operator@sha256:aa0f00f8e03dfdb547e53cf611581e9cc091eaf049558121c2bc5d702e556bcb
        - registry.redhat.io/rhoso/openstack-openstackclient-rhel9@sha256:2952521d4b1799fd32ff773ef2d457a3abc354b639cf7f1e0f7f6615ca35ac8f
        - registry.redhat.io/rhoso-operators/sg-core-rhel9@sha256:b56923f869ff57396e6822133dfc3b21007c3df652cfe962c7ad3028214d988c
        - registry.redhat.io/rhoso/openstack-horizon-rhel9@sha256:82d0e6d24733afd305996e0f4a118df3e48621c40c9669b01f48669dbf40d868
        - registry.redhat.io/rhoso/openstack-redis-rhel9@sha256:c6dbeacbbd5aed9d13e540f1c987a574ea45747983c280446e5afadcf7c33b37
        - registry.redhat.io/rhoso-operators/neutron-rhel9-operator@sha256:30b6a812c47db9e6691211185c304c3e8fc6a8b9ba7ca79acf59c5ec07fb9fb0
        - registry.redhat.io/rhoso/openstack-ovn-controller-rhel9@sha256:3c42b7e0305c508d74602b9378143deb07a569443bba35850cdb7b6e3f5202e2
        - registry.redhat.io/rhoso-operators/openstack-baremetal-rhel9-operator@sha256:03916e507b88943c0e00cedeb3cf20fa0787ccc762e47f38d305ff4d1b125b12
        - registry.redhat.io/rhoso/openstack-unbound-rhel9@sha256:7f17a719e8e49d457ff16437709c2b4eca5b9c2f58972d1e351c5c1c1ac38962
        - registry.redhat.io/rhoso/openstack-designate-central-rhel9@sha256:d28d8d91d8bf75a1b77a5d650c7f2650a1791e2a1b0341450fa46978c955b92f
        - registry.redhat.io/rhoso-operators/swift-rhel9-operator@sha256:c99ef14338290d63e2bcbf129e953b434aea382e47072812bae31b1ea838e7df
        - registry.redhat.io/openshift-power-monitoring/kepler-rhel9@sha256:8e9e5bfeed61aaeee85bf5eefc3390556b8d4cbfc797a82cabe95bc0367126eb
        - registry.redhat.io/rhoso/openstack-swift-account-rhel9@sha256:31e727f37db92fcc21319c44396576a03a118ba25d403715650162f7c0ead31e
        version: 1.0.14
      entries:
      - name: openstack-operator.v1.0.14
        version: 1.0.14
      - name: openstack-operator.v1.0.13
        version: 1.0.13
      - name: openstack-operator.v1.0.12
        version: 1.0.12
      - name: openstack-operator.v1.0.11
        version: 1.0.11
      - name: openstack-operator.v1.0.10
        version: 1.0.10
      - name: openstack-operator.v1.0.9
        version: 1.0.9
      - name: openstack-operator.v1.0.8
        version: 1.0.8
      - name: openstack-operator.v1.0.7
        version: 1.0.7
      name: stable-v1.0
    defaultChannel: stable-v1.0
    packageName: openstack-operator
    provider:
      name: Red Hat Inc.
      url: https://redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.16
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
          createdAt: "2025-09-23 17:13:29"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.16-3,hyperconverged-cluster-operator-test-rhel9:v4.18.16-1,kubevirt-common-instancetypes-test-rhel9:v4.18.16-1,kubevirt-tekton-tasks-test-rhel9:v4.18.16-1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:86d21c945452fb380a895aa96630726fdba73b4e10549aa8ae353c7cac37fe44
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:aa24c4a937c9c1e52c30f229af5b26ce413b684375801d0f5f1ef2a50e8502e2
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:397156fa413fc0910d767f95867126acf7590ddc6546376bb096c086cdab011d
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:6542fcf51604d3ba95ec0d42aa7bc19bdef940e1959808efc98f170a68a7d01c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:82dc19865ed64c129b5214b413f5f8da16d90da532c40d3e02b28f767d72f3c5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:a67782645abd4a869ca8db132953f5f4da72c853fe16f9521c77f241a0703bfa
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:134cb74301a9f9f704d35dea38ffa73aa3ee0ecf43b7fb3e73a89c9b432af9d4
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:6cbb8785d06e58af9b8fc1a66d0d5989a1b490b04d0191252cbb16ace14a3cdd
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:424ea484cd63823b4585e630ffc4053c32309e3f7dc24a212582afc60064e503
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:01082e3cdbac6fa50bec206401660dedd832a71f1f9e52c27281587f1f0122ba
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:a4b7120550d915cb74f03dc3530082f3a0d6850809062469f59934d8518479b9
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:0db7542ab7d67866046b1dde2f3ec8a04c47caa9c622ea9832dc59d127b4e919
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:fb9321f06c842dcb628f9082094fb697146e8740130c889bee81ba6941177303
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:cfba9b634c5a8871511ae6c1644e63560fa51d8b2729eb0db207b89ea9a1b3e9
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:0384141b5c99df83759f69c88895bc23f946da385d2c6db0fc75f80c0853a77c
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:2cffcd87d0a42748ffdbc63707751eb8212ce34ac33d9225690f04ce696178cb
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:42e5ad70d114edd15c6e4d13d48399f572ce71ead1f45a9d94de2262a77754db
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:3da5b2baa67e468fe72efcdb4caeeb4f3384365719d5edb5fff6656edcd74141
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:c43b9a6ac10a1d61eda6d292e707658301a71f61751002f58eb052489d5e612f
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:521b8f630c112e9d8eb4d5686ec7a6907fff8cf491628b99d6aa4a10253b3d78
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:d5d60e012a55a2126d33dac888d4e3d5026ca1380f75a8f263a85e6f7c998baa
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:350bc99751e2cac633b98c633b6c83e60bfd55dde4301282ff8eb61f85ea8d49
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:e21e94b2778c786d2b6aa0c107df267055b1f7d36c589a5900d313b3fe3404c6
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:b91f86bcf2ee712c2d05e240d77215cbd1ae91887e7dca2980f9dadd9390d00b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:8ab2c1ea24981ca684505c2f9e050cd938efbcbdedabd74897e0c6c0d5c5a433
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:68b318dd7fed09f60f6df215c9812b38791707b4e9af1de27acf91365766808f
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:c3a694c846fc85709e756a617d2581c06062d482b0bba88983192f1efbf8cb21
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:7e60c4e1fb9a0e10b039887747287cc06ca156f813d88c7421f3950eaf07bf31
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:710d24725c6aa3b8282d32a310d3e107b6a2e57708425d5e3acde9e82e33591f
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:deaf31fb3ac137556fedec55c44abb71d01c6927aaa21e5d23627ce139f12107
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:8c3f276434dfb7462c556f7be3f93a0c857a463e2fa2befcaef3a7c0859c7368
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:6037e0abd3bdfeffefe91142c88eb6df842cba7c5f3726a64bf7c656a4dc8040
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:e51dfca6233c67b1b5372c682a2fc6e0fe54f2beef86e4c9451aabc80017a5fb
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e3af0269b0a3225b2b5d43cce57a40aef6788830e8af8f4330bd36e4aaf8ec97
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:1c4bcdb3e2074b3f99c4dd0c57a264330c7989b8564efcc50bfe0e086d487f5c
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:27d1ea63ca9dc4fe073a28793f574da3a991a4c36e8da55fc2c7889cede187ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:9f8ab3e7e76ccad5a794de310ed0ba39e070e00647e42cf2b9a1bcdd8c377f3f
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:e0d239237d8052c36aff1509d65c00909b8563ce14341b7eb7fb79e9b84deace
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:af0b3d4305dcc5ccf00569deddf122aabf444bda0d3f0c2a5c4fad7e0c4117e2
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:071b2bf8f6470a6f87c3804eac3a10a947376469e42a55e4f19d3b2fc98c7e95
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:8339ad1cbefafbb91be533f1980ffb0e22114d52fd5c11c1b06cc55d9554c9f4
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:f514eb6d468fdbc6f5bc2b60cafdaff19ea55cb9a9c60f22b9db466f5b8f9000
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:c839c0eef642ac854e04129b007bdd529f9b380840e052fc306ecc0d62e4d055
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:d661d52528a89d87764e16b5d1dcb6ed328075477453e69c6cb9814e2a2d0ea9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:21c02925a871875a8eb70efda72f79d4c287da5cb9c0a4fc19de9ebfcb06e496
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:ae92990aa7ee3cf0bb98549ba78882c032afc8ed323fa83d1e8de24951b09d11
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:ecbdb919859e032ba4f2fe3af4dc79e87a5ad28545942d9d75c734aa814b7b23
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:e6507ebc8d71561772e573691b160f60ee6819b3f6404d5f0db123f0e4bc17e4
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:250cb53bbe119c5609a99ad32e97fc23572af9c08490e0f5212cf0977f555827
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:e9d85d461e281aa141a491ebc24d373bbb6a5ad200a8e40910497eeaf9df361e
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:7eb05621f7cfe8162b08bff8a7411d6b05959affd7dc0b7729b4ccf1d90bb005
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:55996948c6d66abfafc630d7f6fe4b67201af0b61613a39e1747218393fb5e01
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:7f61b66615c9b22ac0e7e6ded50fb5a4f0a71ca228d34340977d82c8eb8322d4
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:09abeaf5bd5a7ac9f272759932e8442828efe6935dcd3b365eb42b7655c1de18
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:d32a745d70478945bb25f04ca85c17dbc1fefa0a791522e42591a6876f085029
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:db86fa7625d770a7eb8a4af977499f4569023992d33caac0caf8e3b71871f574
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:f5ebfeb2d0bcb0e86999d4efdd08d03c34c90daa2523b1382e02f9d4930ded67
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:e3f9e0c1b53a54354246d9c934ea48c038a547845d6a75c7bfac5e82f780ee8a
        version: 4.18.16
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.16
        version: 4.18.16
      - name: kubevirt-hyperconverged-operator.v4.18.15
        version: 4.18.15
      - name: kubevirt-hyperconverged-operator.v4.18.14
        version: 4.18.14
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.12
        version: 4.18.12
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.10
        version: 4.18.10
      - name: kubevirt-hyperconverged-operator.v4.18.9
        version: 4.18.9
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.7
        version: 4.18.7
      - name: kubevirt-hyperconverged-operator.v4.18.6
        version: 4.18.6
      - name: kubevirt-hyperconverged-operator.v4.18.5
        version: 4.18.5
      - name: kubevirt-hyperconverged-operator.v4.18.4
        version: 4.18.4
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      name: candidate
    - currentCSV: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
          createdAt: "2024-08-12 07:46:45"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.14.1 <4.16.0'
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.99.0-73,kubevirt-common-instancetypes-test-rhel9:v4.99.0-32,kubevirt-tekton-tasks-test-rhel9:v4.99.0-31
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/mtq-operator-rhel9@sha256:033bf7a6ef8ced1ac7275a9ddea8b16a10340a0fb0fb39619d89354d661f05b6
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:e39732ebd3a4452fa7a678aeea5e64a7973fd72802eff193b16680c2f0129cb1
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:09a30b59c906ddc4604d70f18210bcebef1f90f8bf965422b0caba363cd88989
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:6baf67f3f410c071cc26f675d86edbbae5eaeb3410888cf622fad384b43b80ea
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:e7145f016642b02bd651e21aa19edc7f93d0d1c2da3c6c7eed5d1174cf23619c
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9822f7a1590d830ccd64cc0729d2e30a2e04fb2b022f54498e1f1abff3748c2c
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:15179f0826831f051fdb867067be7c5528f01f442ae421b8782ae0cd1ed1b73d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:d74c629e415d177069a0570fb8db9252d2b193d2ec7972c844686ffd14033f62
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:e6d9a6c41433360abd42d1118f5af3df9592d6d36af5f65082ae214cc8513c1b
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:185a83c994aced74546fe630cd71656f1e4112cb15be75f41dfcb1b9ed0966bd
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:820a9e6bd0329df5c541ef0b67b3c7cdd9f81be33de87c5b5e5e43777bea197e
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:84679ffe8ce86c674818531897d122f5568d10c7e1e2ff919185d5771e5b4c39
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:cffec516ab9ddc50cab77339d0cb3239ef4e7fd43bd2afdc73370aa487ea5704
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:07435335fdbf6178bc00cc02b93cc0adb42784339151b2d9c31dc8da98b8bc6a
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:4c66a26b9f74d83a9bc60edb40ffd9d6ff2382d20500677d4567f8e141c3ddf0
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:42a34caa18b5f5452b9d5603602ea74b1ee59a8bbb52bddf18901fecc27aed8a
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:06cc71822684c3ed094f327553451daf076d6bb97065c314da1a7d2adbbef3d2
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:30c669770b0b992a3c84bb81c892e70a24fb14af187e8a4ee19cb3e0dd56ddbb
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:41ee360ac43ee69ccee62d4ac3d2f07c8a8589e5b85c659b47018dd36c2b932f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:6b7beed3cba9093b3aeb3109f23ec38c20f5568915dff774ad5e9e6bb5db3a58
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:0769512df0fa1325d76a5f10a1a704cf60fc65f62f57b157ea53bc9ba682851f
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:37e6b4fa5ff436572412ec0bdb1edc4a49e62c994e056c93578605f6b74adbcd
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:4cff43cc145361c74c357055d51060ccc27e6bca58ccc3b5f5b983c7b8545114
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:b410508ecc71b0a3bf4825793efde4b07085e2243d8b3585da852fd1940c4be1
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:c5e8f44f647f593b00d3427172932d7aa694379d124209b9c66b4c1ffe3d4116
        - registry.redhat.io/container-native-virtualization/mtq-controller-rhel9@sha256:ad5938ebd5f07196ce5c35232a1c3b0a653301a83929ce84c3d1cb4dd1b91816
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:4b3ac5ec2fafef0b7702a02847996f0903ec8d4262ccd03e8fa74885265ea4e9
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:3a45cd2f3ea25f356167ce4fa20894a2a0706b4bcffa42c806f6017812053466
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:dec430b74887e9b42d639f14c07d9f58a4881cac24e949dab98f96af150f1958
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:6193e569d4ab43806ac7b4870ef3634a5de3687527d6d190e59a9f735fa4778c
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:b59e34cdb6214eacb9dc176a7b33125d7c1871f6b81696d59646294bfaf749e7
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:d6dfd062df3d0d49f7efb928b835b9c8e0bcfb514a653ba2896a0152f24a0a21
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:442e5cbcc63961642b662544032240bffb3863227324128a67f0bb6a05a1c364
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:f4e7876fa8b27a8082efa99455d9bf7b0c479cff1e4c8a26136bf63092699c0b
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:88f02cbebd6c3a2add6e0f8ca744e42bbd182f5a07d4fac59cf5cf07a4edfd45
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:ac33de246ee42bbfe735d7349d85c6df7e583e7599dea7dd7a2235bd0e59845f
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:8e19d99232be4c37e857a4931627fb0aed721893e27b0d0e522fcf13bf67c1c9
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:f33ac4d7477a730f273a462f171aed35dea38374d5ea7ef0bae2d7f74413a757
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:83c03929cf11d7202e6c6cb9dd767bd4d7473f4fc1d562b79719ceb3d8382e68
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c61732be8024b4ede45eec8f786e51b198632ee5bfb0c2f5af84f37544584283
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:18dfb38223647e11020e3ef1093a695454b90cc50922f80afb612212b7d13309
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:9bfdaf1cca6eac15b63a4ea6b7de28f2328516c5cf8bd5727935dd729e45657d
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:cc190d7c44b20f10ae61709fd2f5f7c0b248df73b768eb696d5a6cec90a34e59
        - registry.redhat.io/container-native-virtualization/mtq-lock-server-rhel9@sha256:01cc6797800ccfb99453858ab36c0100c5ca343f1f71f975914b2437d1f757ba
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:d7e8e075a08b148ddfa91b44bb9cae1ef823756760279c4bb6d8489bc7e0eb2c
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:84b199e98e283c863d52a7b00044c20cb56ad563494adbec2229af265bf3f7c5
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:af5d2a2d4f0b6f0e1a13f1029073bb56980d8a70503b21628f18e8986e9b531c
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:758a9f6309a367d3f31cc03dd89356a8232a0fbd277570c01ebfa1fa30089a8d
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:5b2c64218342a5744182fc06f550bcdcf954e931275a9d1505ef0bbfbd0f8044
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:5f23fc8e9bc386622460c23b9644f58648d9f748f5670927a02888cd36c99a21
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:2e5d5a2a70cda1f414f22b1c713d77d068c5fb16b1487a572019a9aa7130e6be
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:5ba79dd6bdf9e3778d3e301919b37e45cfc1eaf72fda77eba6588788829ba29b
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:811e8e7e9872408794ef36229c42c6c437b1a4ace76c7134d91bb2331fe7815b
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:a32214d166ee692a4cae917fe97e0535a4a696155e1c878ae5335c77dadef0df
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:c14cd037d14489321629908fc6b30248f7a20d866619177a165d72cfc1b6b7ef
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:bf30b6c7173ebb6a4592246af16db87d7d391182f52286b559f8bad8f4b6cc5b
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:8c64b526fd42a17c909bf45eb795121a81d502424a9cf11f8087118751333894
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:bf2ec130d1f3f32314d740037313f49f874a271e8b9a7f6cd2b93698a8423e8d
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:4bdafc4404d5cfd844d05f32d88c262fba56dc931bfd5522b78c25d4ff253883
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:63389195cf584a07ddbc2efec90cf6deaf737b91314b67db2b9f53dcc7acb57c
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:2564af054f52a8bb340e0270379c72da18658662a2a5bd26ac7bd979ddfccb56
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:fa0478fbf91248166f0eed8ec5f8dbb5b31d26031438376d9301836b6a069643
        version: 4.99.0-0.1723448771
      entries:
      - name: kubevirt-hyperconverged-operator.v4.99.0-0.1723448771
        version: 4.99.0-0.1723448771
      name: dev-preview
    - currentCSV: kubevirt-hyperconverged-operator.v4.18.13
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "true"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
          createdAt: "2025-08-17 16:56:34"
          description: Creates and maintains an OpenShift Virtualization Deployment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.7 <4.18.0'
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"openshift-cnv"},"spec":{}}'
          operatorframework.io/suggested-namespace: openshift-cnv
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware","fips","sno"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: Red Hat
          test-images-nvrs: kubevirt-ssp-operator-test-rhel9:v4.18.13-6,hyperconverged-cluster-operator-test-rhel9:v4.18.13-4,kubevirt-common-instancetypes-test-rhel9:v4.18.13-3,kubevirt-tekton-tasks-test-rhel9:v4.18.13-2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of OpenShift Virtualization
            displayName: OpenShift Virtualization Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          # Requirements
          Your cluster must be installed on bare metal infrastructure with Red Hat Enterprise Linux CoreOS workers.

          # Details
          **OpenShift Virtualization** extends Red Hat OpenShift Container Platform, allowing you to host and manage virtualized workloads on the same platform as container-based workloads. From the OpenShift Container Platform web console, you can import a VMware virtual machine from vSphere, create new or clone existing VMs, perform live migrations between nodes, and more. You can use OpenShift Virtualization to manage both Linux and Windows VMs.

          The technology behind OpenShift Virtualization is developed in the [KubeVirt](https://kubevirt.io) open source community. The KubeVirt project extends [Kubernetes](https://kubernetes.io) by adding additional virtualization resource types through [Custom Resource Definitions](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/) (CRDs). Administrators can use Custom Resource Definitions to manage `VirtualMachine` resources alongside all other resources that Kubernetes provides.
        displayName: OpenShift Virtualization
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        - CNV
        - Container-native virtualization
        - Container native virtualization
        - Virt
        - Virtual
        links:
        - name: Source Code
          url: https://github.com/kubevirt
        - name: OpenShift Virtualization
          url: https://www.openshift.com/learn/topics/virtualization/
        - name: KubeVirt Project
          url: https://kubevirt.io
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/container-native-virtualization/vm-network-latency-checkup-rhel9@sha256:a4d0418e49929e186e37f20f64c46f8eca20e314f6da4deb55f0b5407686c2e2
        - registry.redhat.io/openshift4/ose-csi-livenessprobe-rhel9@sha256:9fbf91cd986fe4064b99d6eb4bde7a79f58966763765a5001d9ae75ed22655f5
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-create-datavolume-rhel9@sha256:710dd1bcf4e171943d5260bd16121abd75058bbeb80bc8c237d24aadca99585c
        - registry.redhat.io/container-native-virtualization/virt-handler-rhel9@sha256:396eec2f40ad14e9348b229498313e478903125dbb8c6f7c5f3dc3dad359d18f
        - registry.redhat.io/container-native-virtualization/virt-controller-rhel9@sha256:36a46082ca4ce5fd4784a9613d49c30585883bccf7433dc2ecd8de787eace937
        - registry.redhat.io/container-native-virtualization/wasp-agent-rhel9@sha256:c73b90403d587fa6993e736fd84c08ed39381ac506502c3711da99dd4ea39b4b
        - registry.redhat.io/container-native-virtualization/cluster-network-addons-operator-rhel9@sha256:47d3263ddc6d80e844f8240502df81c0b71c2ed85a29a0d4afac7b5c98379091
        - registry.redhat.io/container-native-virtualization/virt-cdi-controller-rhel9@sha256:6c39b35d0142e12757091831e2f9900c5f8ebcb1498e0eb75b62c86ee2e7238c
        - registry.redhat.io/container-native-virtualization/cnv-containernetworking-plugins-rhel9@sha256:8a69cfd1bcf5745779a019499632141ecdbc4b2d9a4858c66d28df11087359f3
        - registry.redhat.io/container-native-virtualization/aaq-server-rhel9@sha256:b7771fb216f69af28f49c615fa050bde00d27067acdb505578d104ce1337febd
        - registry.redhat.io/openshift4/ose-coredns-rhel9@sha256:21ad0c18f57b964b86ced3498ac0c36a6daf8a60bdeece77a0b41b168dbd6586
        - registry.redhat.io/container-native-virtualization/vm-console-proxy-rhel9@sha256:877fe3539fe1a4d9f8cf25bfe2c1bb847ff9413693848ad539c26442472c1d3e
        - registry.redhat.io/container-native-virtualization/aaq-controller-rhel9@sha256:6ceaacb96b29fbd5a07dac9e8aeb038b76d53f6dc30188b88068755f35da68a9
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-operator-rhel9@sha256:e4666fc81b5fb739151a3ef226c6f214d50a6dd5fc8b08c2512c3e8ddf42a1f6
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadserver-rhel9@sha256:4f806c32d077c6021033733674ed5701bd45086f9fdbfe90985813e2efc27ee9
        - registry.redhat.io/container-native-virtualization/virt-cdi-cloner-rhel9@sha256:1e73a912388b8b3de05e669128d30dd3081d5747932dc10287d613e19f59a49b
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar-rhel9@sha256:ca5969eb64a112ae5c12ae3b6752b6858dfb999d78d61daffd61db5e07a7cdb9
        - registry.redhat.io/container-native-virtualization/kubemacpool-rhel9@sha256:b0a4d68244205f839c0bb2c5cd4f4f305537ce57b42d24d0b46fa4fe2fe5cbba
        - registry.redhat.io/container-native-virtualization/virt-cdi-apiserver-rhel9@sha256:960ab26612067d50aae973b63ffe482649c90ff8f42c2dde698ef548422460cd
        - registry.redhat.io/container-native-virtualization/virt-exportserver-rhel9@sha256:9600fd87e8331989db167870e83618bfb5f273e34e28c2006fa72b49aceb86e0
        - registry.redhat.io/container-native-virtualization/virt-cdi-uploadproxy-rhel9@sha256:d90907bbc0d2810d592316f609bc8fd0f551de4c37ac02c1c03ccc014cdef056
        - registry.redhat.io/container-native-virtualization/kubesecondarydns-rhel9@sha256:c78c9ee196047ada7ecb614a42cbbfe3dec934534942ee5cacf34cf030e50378
        - registry.redhat.io/container-native-virtualization/virt-cdi-operator-rhel9@sha256:f54b97303ba8579a2ef38d971035117efbab46d0eece22a759bc2fba98d11511
        - registry.redhat.io/container-native-virtualization/hco-bundle-registry-rhel9@sha256:1d984a34589472f1c554bb4a53b6df470104affacd341eccd5fa1e334a7d8854
        - registry.redhat.io/container-native-virtualization/virt-cdi-importer-rhel9@sha256:28d40c603ff685d922149ae30dc77ba1d9bc9af0ce2ab6d9038ff8272940ca08
        - registry.redhat.io/container-native-virtualization/kubevirt-ipam-controller-rhel9@sha256:c399069fc6a0d3b3e7093ab6cc2d540d05a04a68213a24a7a500a83d084246df
        - registry.redhat.io/container-native-virtualization/kubevirt-template-validator-rhel9@sha256:3c2706973b2bca2fbbf88d5af1de0b3c53ef1c7e5029b9a3ef42f961795ad65c
        - registry.redhat.io/container-native-virtualization/kubevirt-common-instancetypes-rhel9@sha256:725286e0c9265a66e3f4b2a86fcfa7cbdda4e963863af9fb2af9e1226e10f34f
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-rhel9@sha256:3e546a40d6a3e2cce2567750f93260b3db0978477e8ce34f63cd5d195c58c195
        - registry.redhat.io/container-native-virtualization/kubevirt-ssp-operator-rhel9@sha256:3765263eb3b193d860699d3e348346b762eabca51ecf31f8f97db69f3ca5e4bd
        - registry.redhat.io/container-native-virtualization/hostpath-csi-driver-rhel9@sha256:8985f049a49cae810ffb59dd4c94afca632808d3c8ba40809c5447a382b63bad
        - registry.redhat.io/container-native-virtualization/virt-operator-rhel9@sha256:7095f50cb41030a9f49af2ad9a99d81407b5d8491c421c34934f788f02eb7528
        - registry.redhat.io/container-native-virtualization/sidecar-shim-rhel9@sha256:f501a890dd87d37eb24cdd86d5889eadfad1e454fcc782e8ca2454150e8dc494
        - registry.redhat.io/container-native-virtualization/libguestfs-tools-rhel9@sha256:04b135c11a1e4518febbad5ee56e118ca711ad3454db19f1824cda0d4d497742
        - registry.redhat.io/migration-toolkit-virtualization/mtv-rhv-populator-rhel8@sha256:ef0dca763d85023bf3b1e3be135177007b1fe5c56bc356a8cf7145b4dd3c669c
        - registry.redhat.io/container-native-virtualization/kubevirt-apiserver-proxy-rhel9@sha256:ebc17f70a654181527a8a1d39e069076af82343cab55531fd7a86eaa85b12fd2
        - registry.redhat.io/container-native-virtualization/virtio-win-rhel9@sha256:a9788ba6756715f286ede2fa04e460fd021b24f2a0df068c2aa626c89814a7b5
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter-rhel9@sha256:5794ed8b4af6cc37d278cd91f08f0418b89550269d69c9196772fcca1a4933e7
        - registry.redhat.io/container-native-virtualization/multus-dynamic-networks-rhel9@sha256:0c2c54d18dddf8ae7f2993dd42b3d07217f63266218c329119dda13e0c431fb6
        - registry.redhat.io/container-native-virtualization/virt-artifacts-server-rhel9@sha256:f15c753ca6687605d956d5200ec9104067c9bfe66e987dce7c32bc45ed79c57d
        - registry.redhat.io/container-native-virtualization/pr-helper-rhel9@sha256:ee7ae60e4ba4b1d13b0aecb053e3ccd22b62d2277d28413f394d32e6a3ffe139
        - registry.redhat.io/container-native-virtualization/bridge-marker-rhel9@sha256:b5a6c60e663b81cef29e9622af8f0eb62838be99ff51f576ca1b275c6e5a204a
        - registry.redhat.io/container-native-virtualization/kubevirt-api-lifecycle-automation-rhel9@sha256:4c58b6e5356866deb394d55e2f96e4d83489d24c988e8016ccddd8738b895b7b
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-cni-rhel9@sha256:88d6db3caa017e4c2d8114af1abd28df4d0751b01f291d42b84586378e0c9484
        - registry.redhat.io/container-native-virtualization/kubevirt-realtime-checkup-rhel9@sha256:0838f54f7ab4fbc04346ef67501fbca0b6c2097558bbd9c6fec2632ba0ce7bc8
        - registry.redhat.io/container-native-virtualization/virt-launcher-rhel9@sha256:9e46d9a0ca2e1678a02a52d14e2d8f0eb007b21ae105100106532cd12f317a41
        - registry.redhat.io/container-native-virtualization/ovs-cni-plugin-rhel9@sha256:a0821002f32c485c28d43dcc3afe5f447ecec4a567ef1b5f07b162f96245cbe6
        - registry.redhat.io/container-native-virtualization/virt-api-rhel9@sha256:fd17e015ad32d54f8464bc337c874b9131e0edca750adb4165161f0d382b8c3a
        - registry.redhat.io/container-native-virtualization/hostpath-provisioner-operator-rhel9@sha256:45c7a05260790a6ea78c809316dc58ed98b32cd94a884e82f24650051a27a163
        - registry.redhat.io/container-native-virtualization/passt-network-binding-plugin-sidecar-rhel9@sha256:2e6b6d7d04690a616eb210fe4b142372807bb9c59b6af6e7d6fa196f8dadff3e
        - registry.redhat.io/container-native-virtualization/hyperconverged-cluster-webhook-rhel9@sha256:8e69000f409c115b6af634c057106dbe894395708a694c3a4cb030b06e77407b
        - registry.redhat.io/container-native-virtualization/virt-exportproxy-rhel9@sha256:498282a9a9afceec8587eef445214f5db6ed6c09c2c3362049c4097a210e5aa5
        - registry.redhat.io/container-native-virtualization/aaq-operator-rhel9@sha256:11031e69287b4fafa77e2d54dc40d8c7de64e7ed8528abfca7a798b5aa8970b9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/ose-csi-external-provisioner-rhel9@sha256:faf31a311d34ea8bad50c83918d5ac1dc32277b9d8e356a866d166f97209e40d
        - registry.redhat.io/container-native-virtualization/cnv-must-gather-rhel9@sha256:509f4493d980ed49e616764bf9d9edad4a774919b3dd0d950324d4756f2c54f6
        - registry.redhat.io/container-native-virtualization/kubevirt-console-plugin-rhel9@sha256:61f96cac775c4f636f0685900c419a6771fe3f3b44648d93a176b4ab5efa9092
        - registry.redhat.io/container-native-virtualization/kubevirt-tekton-tasks-disk-virt-customize-rhel9@sha256:70219973c00c12519a1e5abd9fa3061669de8e7273ea1dcf9f72d40ae683d244
        - registry.redhat.io/container-native-virtualization/kubevirt-dpdk-checkup-rhel9@sha256:f0eba5e54eced4a4478d395b098b3dca201b7b374f9e6b7f4a5c1669953c0991
        - registry.redhat.io/container-native-virtualization/kubevirt-storage-checkup-rhel9@sha256:be1b6c7e71d25c45b0f4bc2ac8507a1c215e30e39f51bc9f907436f6bcc96b4b
        version: 4.18.13
      entries:
      - name: kubevirt-hyperconverged-operator.v4.18.13
        version: 4.18.13
      - name: kubevirt-hyperconverged-operator.v4.18.11
        version: 4.18.11
      - name: kubevirt-hyperconverged-operator.v4.18.8
        version: 4.18.8
      - name: kubevirt-hyperconverged-operator.v4.18.3
        version: 4.18.3
      - name: kubevirt-hyperconverged-operator.v4.18.2
        version: 4.18.2
      - name: kubevirt-hyperconverged-operator.v4.18.1
        version: 4.18.1
      - name: kubevirt-hyperconverged-operator.v4.18.0
        version: 4.18.0
      - name: kubevirt-hyperconverged-operator.v4.17.4
        version: 4.17.4
      - name: kubevirt-hyperconverged-operator.v4.17.3
        version: 4.17.3
      - name: kubevirt-hyperconverged-operator.v4.17.2
        version: 4.17.2
      - name: kubevirt-hyperconverged-operator.v4.17.1
        version: 4.17.1
      - name: kubevirt-hyperconverged-operator.v4.17.0
        version: 4.17.0
      - name: kubevirt-hyperconverged-operator.v4.16.3
        version: 4.16.3
      - name: kubevirt-hyperconverged-operator.v4.16.2
        version: 4.16.2
      - name: kubevirt-hyperconverged-operator.v4.16.1
        version: 4.16.1
      - name: kubevirt-hyperconverged-operator.v4.16.0
        version: 4.16.0
      - name: kubevirt-hyperconverged-operator.v4.15.2
        version: 4.15.2
      - name: kubevirt-hyperconverged-operator.v4.15.1
        version: 4.15.1
      - name: kubevirt-hyperconverged-operator.v4.15.0
        version: 4.15.0
      - name: kubevirt-hyperconverged-operator.v4.14.3
        version: 4.14.3
      - name: kubevirt-hyperconverged-operator.v4.14.2
        version: 4.14.2
      - name: kubevirt-hyperconverged-operator.v4.14.1
        version: 4.14.1
      - name: kubevirt-hyperconverged-operator.v4.14.0
        version: 4.14.0
      - name: kubevirt-hyperconverged-operator.v4.13.4
        version: 4.13.4
      - name: kubevirt-hyperconverged-operator.v4.13.3
        version: 4.13.3
      - name: kubevirt-hyperconverged-operator.v4.13.2
        version: 4.13.2
      - name: kubevirt-hyperconverged-operator.v4.13.1
        version: 4.13.1
      - name: kubevirt-hyperconverged-operator.v4.13.0
        version: 4.13.0
      - name: kubevirt-hyperconverged-operator.v4.12.2
        version: 4.12.2
      - name: kubevirt-hyperconverged-operator.v4.12.1
        version: 4.12.1
      - name: kubevirt-hyperconverged-operator.v4.12.0
        version: 4.12.0
      name: stable
    defaultChannel: stable
    packageName: kubevirt-hyperconverged
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fortanix Inc.
      provider-url: https://fortanix.com
    name: ccm-node-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ccm-node-agent-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.operatorhub.io/v1",
                "kind": "CcmNodeAgent",
                "metadata": {
                  "name": "ccmnodeagent-sample"
                },
                "spec": {
                  "ccmJoinToken": {
                    "secretName": "ccm-join-token",
                    "token": ""
                  },
                  "fullnameOverride": "",
                  "image": {
                    "ccmNodeAgent": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/ccm-node-agent-ubi7",
                      "tag": "latest"
                    },
                    "ccmNodeAgentInit": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "busybox",
                      "tag": "latest"
                    },
                    "sgxSoftwareInstaller": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "docker.io/fortanix/sgx-software-installer-ubi7",
                      "tag": "latest"
                    }
                  },
                  "log": {
                    "debug": false
                  },
                  "nameOverride": "",
                  "podAnnotations": {},
                  "ports": {
                    "containerPort": 9092,
                    "hostPort": 9092
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    },
                    "requests": {
                      "cpu": "100m",
                      "memory": "128Mi"
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true,
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          operators.operatorframework.io/builder: operator-sdk-v1.16.0-ocp
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CCM Node Agent software enables registration of the compute
              nodes to Fortanix CCM when installed in the Openshift Cluster Nodes<U+200B>.
            displayName: CCM Node Agent
            kind: CcmNodeAgent
            name: ccmnodeagents.charts.operatorhub.io
            version: v1
        description: "CCM Node Agent software enables registration of the compute
          nodes to Fortanix CCM when installed in the Openshift Cluster Nodes​. \nIt
          also assists with application attestation and visibility for Fortanix CCM.\n\n##
          About this Operator\nThis Operator is based on a Helm chart for CCM Node
          Agent. It enrolls all the nodes in the openshift cluster to a particular
          CCM account. You have to provide the base64 encode of join token of CCM
          account while creating CcmNodeAgent resource.\n\n## Prerequisites\nAll the
          worker nodes in the cluster must have SGX capabilities.\n\n## Installation\n*
          Choose a namespace to install the Operator. Operator will be available in
          all namespaces.\n* Select Approval Strategy: Available options are Automatic
          or Manual. Automatic approval will automatically upgrade the running instance
          of your Operator without human intervention when a new version is available.
          In case of Manual approval, when a new version is available, OLM will create
          an update request. You must manually approve the request to update the Operator
          to a new version.\n* Generate your CCM Account Join Token: Please log in
          to https://ccm.fortanix.com, and in the Management Console tab, click the
          + ENROLL NODE button. In the ENROLL NODE screen, a Join Token will be generated
          in the text box for \"Get a join token to register an SGX compute node\".
          This Join Token is used by the compute node to authenticate itself.\n* Base64
          encode the Join Token:\n```shell\necho -n <join-token> | openssl base64
          -A\n```\n* Create an instance of CcmNodeAgent by setting the `spec.ccmJoinToken.Token`
          to the base64 encoded CCM account join token.\n* After CcmNodeAgent resource
          is installed successfully you should be able to see all the worker nodes
          of the cluster in your CCM account under Compute Node Tab.\n\n## Getting
          Started\nNow to run your application inside SGX enclaves on your cluster
          follow the [documentation](https://support.fortanix.com/hc/en-us/articles/360043527431-User-s-Guide-Add-and-Edit-an-Application).\n"
        displayName: CCM Node Agent Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Dockerfile
        - Helm
        - CCM
        - Confidential Compute
        links:
        - name: CCM Node Agent Operator
          url: https://support.fortanix.com/hc/en-us/categories/360003107511-Confidential-Computing-Manager
        maintainers:
        - email: info@fortanix.com
          name: Fortanix Inc.
        maturity: alpha
        provider:
          name: Fortanix Inc.
          url: https://fortanix.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:62575404000cb77fa8d2284c87b31e39e667a55b0fb47821263b23b3d1135a7e
        - docker.io/fortanix/ccm-node-agent-operator@sha256:e2a3d52551198102ddbe7f4ebd9621edc357c57ddeff47b7ed9f2cf56a1bed39
        - registry.connect.redhat.com/fortanix/ccm-node-agent-operator-bundle@sha256:12746020eefd383963e08a86e3542608aed1999b4e520435e366a307a6a9eebc
        version: 0.0.2
      entries:
      - name: ccm-node-agent-operator.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ccm-node-agent-operator
    provider:
      name: Fortanix Inc.
      url: https://fortanix.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://www.medik8s.io/
    name: customized-user-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: customized-user-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediation",
                "metadata": {
                  "name": "customizeduserremediation-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationConfig",
                "metadata": {
                  "name": "customizeduserremediationconfig-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "customized-user-remediation.medik8s.io/v1alpha1",
                "kind": "CustomizedUserRemediationTemplate",
                "metadata": {
                  "name": "customizeduserremediationtemplate-sample",
                  "namespace": "openshift-workload-availability"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/customized-user-remediation-operator:v0.1.0
          createdAt: "2024-03-25 16:17:57"
          description: Customized User Remediation Operator will remediate unhealthy
            nodes by running a pre defined user script on the unhealthy node. Since
            the user has full control over the script content, is may be quite risky
            so apply discretion.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/customized-user-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CustomizedUserRemediationConfig is the Schema for the customizeduserremediationconfigs
              API
            displayName: Customized User Remediation Config
            kind: CustomizedUserRemediationConfig
            name: customizeduserremediationconfigs.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediation is the Schema for the customizeduserremediations
              API
            displayName: Customized User Remediation
            kind: CustomizedUserRemediation
            name: customizeduserremediations.customized-user-remediation.medik8s.io
            version: v1alpha1
          - description: CustomizedUserRemediationTemplate is the Schema for the customizeduserremediationtemplates
              API
            displayName: Customized User Remediation Template
            kind: CustomizedUserRemediationTemplate
            name: customizeduserremediationtemplates.customized-user-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Customized User Remediation (CUR) works with Node Health Check to remediate unhealthy nodes by running a user customized script on the unhealthy node.
          This operator does not preform any fencing on the remediated node so users should be careful with the scripts that are being used.

          ### Compatibility
          CUR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Flexibility - CUR is quite flexible sine it runs in the remediation process a user customized script.
        displayName: Customized User Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - remediation
        - auto-healing
        - recovery
        - high-availability
        - baremetal
        - script
        - CUR
        links:
        - name: Customized User Remediation
          url: https://customized-user-remediation.domain
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Medik8s
          url: https://www.medik8s.io/
        relatedImages:
        - quay.io/medik8s/customized-user-remediation-operator:v0.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/customized-user-remediation@sha256:dc517e2349fe993e3417b858865fb73b98b7b163c0b4231d4d7be1f11992bb78
        version: 0.1.0
      entries:
      - name: customized-user-remediation.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: customized-user-remediation
    provider:
      name: Medik8s
      url: https://www.medik8s.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Solace Corporation
      provider-url: www.solace.com
    name: pubsubplus-eventbroker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pubsubplus-eventbroker-operator.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pubsubplus.solace.com/v1beta1",
                "kind": "PubSubPlusEventBroker",
                "metadata": {
                  "name": "non-ha-standalone-example"
                },
                "spec": {
                  "redundancy": false
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          certified: "true"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.10
          containerImage: registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
          createdAt: "2025-07-31T19:48:36Z"
          description: The Solace PubSub+ Event Broker Operator deploys and manages
            the lifecycle of PubSub+ Event Brokers
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
          support: Solace Products
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PubSub+ Event Broker
            displayName: PubSub+ Event Broker
            kind: PubSubPlusEventBroker
            name: pubsubpluseventbrokers.pubsubplus.solace.com
            version: v1beta1
        description: |
          ## Solace PubSub+ Platform
          Solace [PubSub+ Platform](https://solace.com/products/platform/) is a complete event streaming and management platform
          for real-time enterprises. The [PubSub+ Software Event Broker](https://solace.com/products/event-broker/software/) efficiently streams
          event-driven information between applications, IoT devices, and user interfaces running in the cloud, on-premises,
          and in hybrid environments using open APIs and protocols like AMQP, JMS, MQTT, REST and WebSocket.
          It can be installed into a variety of public and private clouds, PaaS, and on-premises environments.
          Event brokers in multiple locations can be linked together in an [Event Mesh](https://solace.com/what-is-an-event-mesh/)
          to dynamically share events across the distributed enterprise.

          ## Solace PubSub+ Event Broker Operator
          Solace Pubsub+ Event Broker Operator automatically deploys and manages PubSub+ Software Event Brokers on Kubernetes and OpenShift environments.

          Features include:
          * Initial deployment configuration: Production-ready HA, non-HA or for developers
          * Rolling update for configuration changes or broker upgrades
          * Prometheus monitoring support

          ## Getting started
          Refer to the [Solace PubSub+ Event Broker Operator Quick Start](https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/README.md)
        displayName: Solace PubSub+ Event Broker Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - solace
        - pubsubplus
        - pubsub+
        - pubsub
        - messaging
        - advanced event broker
        - event broker
        - event mesh
        - event streaming
        - data streaming
        - event integration
        - middleware
        links:
        - name: Operator QuickStart
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart
        - name: Operator Documentation
          url: https://github.com/SolaceProducts/pubsubplus-kubernetes-quickstart/blob/main/docs/EventBrokerOperatorUserGuide.md
        - name: Solace.com
          url: https://solace.com
        - name: Understanding Solace Technology
          url: https://solace.com/products/platform/
        - name: Solace PubSub+ Technical Documentation
          url: https://docs.solace.com/
        - name: Solace Developer Portal
          url: https://www.solace.dev/
        - name: Solace Community
          url: https://solace.community/
        maintainers:
        - email: support@solace.com
          name: Solace R&D
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Solace Corporation
          url: www.solace.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/pubsubplus-eventbroker-operator@sha256:2761ebb89dace09a1c1ccad92c28c877f3a6f75a674191ffc3788be7fa549926
        - registry.connect.redhat.com/solace/pubsubplus-eventbroker-operator:1.4.0
        version: 1.4.0
      entries:
      - name: pubsubplus-eventbroker-operator.v1.4.0
        version: 1.4.0
      - name: pubsubplus-eventbroker-operator.v1.3.0
        version: 1.3.0
      - name: pubsubplus-eventbroker-operator.v1.2.0
        version: 1.2.0
      - name: pubsubplus-eventbroker-operator.v1.1.2
        version: 1.1.2
      - name: pubsubplus-eventbroker-operator.v1.1.1
        version: 1.1.1
      - name: pubsubplus-eventbroker-operator.v1.1.0
        version: 1.1.0
      - name: pubsubplus-eventbroker-operator.v1.0.3
        version: 1.0.3
      - name: pubsubplus-eventbroker-operator.v1.0.2
        version: 1.0.2
      - name: pubsubplus-eventbroker-operator.v1.0.1
        version: 1.0.1
      - name: pubsubplus-eventbroker-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pubsubplus-eventbroker-operator
    provider:
      name: Solace Corporation
      url: www.solace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Rebaca Technologies Pvt Ltd
      provider-url: https://www.rebaca.com/telecom/solutions
    name: abot-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abot-operator.2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "abot.com/v2",
                "kind": "Abot",
                "metadata": {
                  "name": "abot"
                },
                "spec": {
                  "abotanalytics": {
                    "image": "registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807",
                    "resources": {
                      "limits": {
                        "cpu": "4000m",
                        "memory": "6000Mi"
                      },
                      "requests": {
                        "cpu": "2000m",
                        "memory": "4000Mi"
                      }
                    }
                  },
                  "abotbdd": {
                    "image": "registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed",
                    "resources": {
                      "limits": {
                        "cpu": "2000m",
                        "memory": "2000Mi"
                      },
                      "requests": {
                        "cpu": "1000m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotgrafana": {
                    "image": "registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotgui": {
                    "image": "registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3",
                    "resources": {
                      "limits": {
                        "cpu": "1000m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "600m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "abotkibana": {
                    "image": "registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotlogstash": {
                    "image": "registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "abotmongo": {
                    "image": "registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba",
                    "resources": {
                      "limits": {
                        "cpu": "200m",
                        "memory": "500Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "300Mi"
                      }
                    }
                  },
                  "abotnginx": {
                    "baseimage": "registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c",
                    "image": "registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "abotrabbitmq": {
                    "image": "registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30",
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "abotsidecar": {
                    "image": "registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733",
                    "resources": {
                      "limits": {
                        "cpu": "400m",
                        "memory": "1500Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "elasticsearch": {
                    "image": "registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae",
                    "replicas": 3,
                    "resources": {
                      "limits": {
                        "cpu": "800m",
                        "memory": "3000Mi"
                      },
                      "requests": {
                        "cpu": "400m",
                        "memory": "1000Mi"
                      }
                    },
                    "volumeClaim": {
                      "storage": "10Gi"
                    }
                  },
                  "imagePullPolicy": "IfNotPresent",
                  "initContainers": {
                    "image": "registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac",
                    "resources": {
                      "limits": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "protocol": "https",
                  "pvc": {
                    "abotvol": "50Gi",
                    "analyticsvol": "50Gi",
                    "databak": "10Gi",
                    "mongovol": "10Gi",
                    "prometheusvol": "10Gi"
                  },
                  "replicas": 1,
                  "service": {
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "name": "abot"
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-02-15T09:21:28Z"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/abot-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Abot
            name: abots.abot.com
            version: v2
        description: A Helm based operator for deploying Abot application in Openshift
          cluster. The ABot Test Bench is a test automation tool for 4G, 5G and ORAN
          Telecom networks that follow specifications and use standards driven design
          and compliance as specified by the 3GPP consortium. It provides a platform
          for a user to quickly check for protocol compliance and test for load characteristics
          of a particular system under test.
        displayName: ABot-Operator-v2.0.0
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - 5G
        - 4G
        - ORAN
        - test tool
        - compliance
        - interop
        links:
        - name: Abot Operator
          url: https://abot-operator.domain
        maintainers:
        - email: suchismita.nag@rebaca.com
          name: Suchismita Nag
        - email: samir@rebaca.com
          name: Samir Chatterjee
        maturity: alpha
        provider:
          name: Rebaca Technologies Pvt Ltd
          url: https://www.rebaca.com/telecom/solutions
        relatedImages:
        - registry.connect.redhat.com/rebaca/abot-operator-bundle@sha256:35f90c08438e4e5aadd01de20bc82e59454c394c0e1d24ade6c56d6e02053bd4
        - registry.connect.redhat.com/rebaca/abot-baseimage@sha256:ca96dd8b4fea37a5314ff04c232f8238e1250f9bb2177dba4ccd27d30c9fcb3c
        - registry.redhat.io/rhacm2/prometheus-rhel8@sha256:df960b0960d576b9fc584507f40b6aeb5180b3c12188e161883c6988d8e74733
        - registry.connect.redhat.com/rebaca/abot-analytics@sha256:e19acdd4ebbe981372b02eb4d675b1d77cc01362437d4195bf581b5970339807
        - registry.connect.redhat.com/rebaca/abot-mongodb@sha256:1daca868a143bab74e04ce8dd949a031a1052ea8415af994a0d3ed0f59c25aba
        - registry.connect.redhat.com/rebaca/abot-operator@sha256:e3b3c32d12223097a9d0eb466c3f49fbdab519a675123c7e69597b53ab4c79fe
        - registry.connect.redhat.com/elastic/elasticsearch@sha256:f47fd7df80ab05635e31d733f6969d38205f9f81cf95da8594a047f3d07c59ae
        - registry.connect.redhat.com/elastic/logstash@sha256:dbc25dc5c31dfee08e86221af9f549b50075b2ccf7f255779c5f022c54cd892e
        - registry.access.redhat.com/ubi8/ubi@sha256:627867e53ad6846afba2dfbf5cef1d54c868a9025633ef0afd546278d4654eac
        - registry.connect.redhat.com/rebaca/abot-gui@sha256:25ec796b6d11b1e1f82632b0a3035205836bdd67f8c7ddc0ba9e98358e3476a3
        - registry.connect.redhat.com/rebaca/abot-bdd@sha256:30597d8b39d24fd243c12316212874d3cc9180d2eae92d976ddae78dd42239ed
        - registry.connect.redhat.com/rebaca/abot-nginx@sha256:7a93c8d9be7a46b5571537b70e52e2644d117825a41dfadeb6da3a88e54a2ed2
        - registry.connect.redhat.com/rebaca/abot-rabbitmq@sha256:87d0d462e1408df781f17bd3feb169deddb6ba13602ab988f792f347acfe2e30
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:50f1b4c65541b56b85f42bc9b3127bc7bc9faad1974067ce11481de7596e4c6e
        - registry.redhat.io/rhel8/grafana@sha256:d894cd34f2d8cac30e0fab69bd869f4b0f258b5086a8fd7ead0c5886ce2c5f7a
        - registry.connect.redhat.com/elastic/kibana@sha256:2be3542d50f51ad786ead958a399f26852339e9417279688aad194c6794336ce
        version: 2.0.0
      entries:
      - name: abot-operator.2.0.0
        version: 2.0.0
      name: alpha
    defaultChannel: alpha
    packageName: abot-operator-rhmp
    provider:
      name: Rebaca Technologies Pvt Ltd
      url: https://www.rebaca.com/telecom/solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Community
      provider-url: ""
    name: runtime-component-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runtime-component.v0.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
          createdAt: "2022-03-25T14:00:00Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: <0.8.1
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          This advanced Operator is capable of deploying any runtime component image with consistent, production-grade QoS. It enables enterprise architects to govern the way their applications get deployed & managed in the cluster, while dramatically reducing the learning curve for developers to deploy into Kubernetes - allowing them to focus on writing the code!
          Here are some key features:

          #### Application Lifecyle
          You can deploy your runtime component container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the Operator will watch for any updates and will re-deploy the modified image.

          #### Custom RBAC
          This Operator is capable of using a custom ServiceAccount from the caller, allowing it to follow RBAC restrictions. By default it creates a ServiceAccount if one is not specified, which can also be bound with specific roles.

          #### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as: labels, annotations, and environment variables from a ConfigMap, a Secret or a value.

          #### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configuration, such as TLS settings, are also easily enabled.

          #### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto scaling to create (and delete) instances based on resource consumption.

          #### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you.
          This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart.
          We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.

          #### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating kubernetes Secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to its required services without any intervention or interruption.

          #### Exposing metrics to Prometheus
          The Runtime Component Operator exposes the runtime container's metrics via the Prometheus Operator.
          You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `spec.monitoring` field to control configurations such as the poll interval and security credentials.

          #### Easily mount logs and transaction directories
          If you need to mount the logs and transaction data from your runtime component to an external volume such as NFS (or any storage supported in your cluster), simply add the following (to specify the volume size and the location to persist) to your RuntimeComponent CR:
          ``` storage: size: 2Gi mountPath: "/logs" ```

          #### Integration with OpenShift Serverless
          Deploy your serverless runtime component using a single toggle.  The Operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing your pod to automatically scale to 0 when it is idle.

          #### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and how they are connected.

          See our [**documentation**](https://github.com/application-stacks/runtime-component-operator/tree/main/doc/) for more information.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://github.com/application-stacks/runtime-component-operator/tree/main/doc
        - name: Repository
          url: https://github.com/application-stacks/runtime-component-operator
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: beta
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:4603bf04e53ce948bbf0ee5d239d5e265b40598044ef9e1f4539f4e466e59cb2
        - registry.connect.redhat.com/ibm/runtime-component-operator@sha256:d37b3f62575c9ad1ffcb22c89e2c3886275c326bf27aa49fb1808acba475853d
        - registry.connect.redhat.com/ibm/open-liberty-samples@sha256:8ea2d3405ff2829d93c5dda4dab5d695ea8ead34e804aaf6e39ea84f53a15ee4
        version: 0.8.1
      entries:
      - name: runtime-component.v0.8.1
        version: 0.8.1
      - name: runtime-component.v0.8.0
        version: 0.8.0
      name: beta2
    - currentCSV: runtime-component.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
          createdAt: "2023-08-28T16:46:50Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          olm.skipRange: '>=0.8.0 <1.2.2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.24.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:a178716e62ab8a752bcb9023ad9631af2dbc72357266cd13b43398e21f230828
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:6bf52d536e94d6aa6eaa3ca9639e0fe9f5e71dbd891c50362939dd6aca519d02
        - icr.io/appcafe/runtime-component-operator@sha256:1b818f1a9c27baab6a39ae42d86f78bac085d772cae5cd97b357d0ee5799cf85
        version: 1.2.2
      entries:
      - name: runtime-component.v1.2.2
        version: 1.2.2
      name: v1.2
    - currentCSV: runtime-component.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
          createdAt: "2024-08-01T20:26:42Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.3.3'
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:c5288fea5b2ef8125f7b1ff7f7633eccd4e52d14ac9cfc5e473c2e317018ca3d
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:422a7a00edc338497c0a91890b09204b8505ef0e377bb9b5944ace0102055e91
        - icr.io/appcafe/runtime-component-operator@sha256:682c5cb3e24566bf2933b1d4a4f6bf520d9b546edcedc1ee3214ec1ac4d48629
        version: 1.3.3
      entries:
      - name: runtime-component.v1.3.3
        version: 1.3.3
      name: v1.3
    - currentCSV: runtime-component.v1.4.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
          createdAt: "2025-07-18T20:38:10Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.4.4'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - icr.io/appcafe/runtime-component-operator@sha256:6f527c5173421c28d175b42d8c690eb63c4d1c5092c31d5350fd7a5e93661eaa
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:704925128f81d4ffadc4dcfaa69fc8011cf3d0f431d24222a8d7a03bad675331
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:494f54bd1476465a4226cb62f0d97cd59cd85b55ff955073e176bda4bd2afbd3
        version: 1.4.4
      entries:
      - name: runtime-component.v1.4.4
        version: 1.4.4
      name: v1.4
    - currentCSV: runtime-component.v1.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "manageTLS": true,
                  "replicas": 1,
                  "service": {
                    "port": 9443
                  },
                  "serviceAccount": {
                    "mountToken": true
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeComponent",
                "metadata": {
                  "name": "runtimecomponent-sample"
                },
                "spec": {
                  "applicationImage": "icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8",
                  "expose": true,
                  "replicas": 1,
                  "service": {
                    "port": 9080
                  }
                }
              },
              {
                "apiVersion": "rc.app.stacks/v1beta2",
                "kind": "RuntimeOperation",
                "metadata": {
                  "name": "runtimeoperation-sample"
                },
                "spec": {
                  "command": [
                    "./your_script.sh"
                  ],
                  "containerName": "app",
                  "podName": "Specify_Pod_Name_Here"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Application Runtime
          certified: "true"
          containerImage: icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
          createdAt: "2025-09-15T18:08:27Z"
          description: Deploys any runtime component with dynamic and auto-tuning
            configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.8.0 <1.5.0'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/application-stacks/runtime-component-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1
          - description: Represents the deployment of a runtime component
            displayName: RuntimeComponent
            kind: RuntimeComponent
            name: runtimecomponents.rc.app.stacks
            version: v1beta2
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1
          - description: Day-2 operation to execute on an instance of runtime component
            displayName: RuntimeOperation
            kind: RuntimeOperation
            name: runtimeoperations.rc.app.stacks
            version: v1beta2
        description: |
          Runtime Component Operator allows you to deploy and manage any runtime components securely and easily on Red Hat OpenShift as well as other Kubernetes-based platforms in a consistent way. You can also perform day-2 operations using the operator.
          ## Documentation
          See our user guide [here](https://ibm.biz/rco-docs). If you are **upgrading from versions 0.8.x or below**, make sure to review the documentation, prior to the upgrade, on [behavioural changes](https://ibm.biz/rco-upgrade-v1) that could impact your applications.
          ## Supported platforms
          Kubernetes platform installed on one of the following platforms:
          - Linux&reg; x86_64 (amd64)
          - Linux&reg; on IBM&reg; Z (s390x)
          - Linux&reg; on Power&reg; (ppc64le)
          ## Details
          Key features provided by the operator:
          ### Integration with Certificate Managers
          The [cert-manager APIs](https://cert-manager.io/) when available on the cluster will be used to generate certificates for the application. Otherwise, on Red Hat OpenShift, the operator will generate certificates using OpenShift's Certificate Manager. The operator will automatically provision TLS certificates for applications' pods and they are automatically refreshed when the certificates are updated. Optionally, you can bring your own (BYO) certificate authority (CA) or Issuer to generate certificates to secure your applications.
          ### Automatically restrict network communication
          Network policies are created for each application by default to limit incoming traffic to pods in the same namespace that are part of the same application. Only the ports configured by the service are allowed. The network policy can be configured to allow either namespaces and/or pods with certain labels. On OpenShift, the operator automatically configures network policy to allow traffic from ingress, when the application is exposed, and from the monitoring stack.
          ### Exposing metrics to Prometheus
          Expose the application's metrics via the Prometheus Operator. You can pick between a basic mode, where you simply specify the label that Prometheus is watching to scrape the metrics from the container, or you can specify the full `ServiceMonitor` spec embedded into the RuntimeComponent's `.spec.monitoring` field to control configurations such as poll interval and security credentials.
          ### Easily mount logs and transaction directories
          Do you need to mount the logs and transaction data from your application to an external volume such as NFS (or any storage supported in your cluster)? Simply specify the volume size and the location to persist and the operator takes care of the rest. For example, add the following configuration into the RuntimeComponent's `.spec.storage` field :
          ``` size: 2Gi mountPath: "/logs" ```
          ### Service Binding
          Your runtime components can expose services by a simple toggle. We take care of the heavy lifting such as creating Kubernetes secrets with information other services can use to bind. We also keep the bindable information synchronized, so your applications can dynamically reconnect to their required services without any intervention or interruption.
          ### Integration with Knative (OpenShift Serverless)
          Deploy your serverless runtime component using a single toggle. The operator will convert all of its generated resources into [Knative](https://knative.dev) resources, allowing the application to automatically scale to 0 when it is idle.
          ### Application Lifecycle
          You can deploy your application container by either pointing to a container image, or an OpenShift ImageStream. When using an ImageStream the operator will watch for any updates and will automatically deploy the new image.
          ### Custom RBAC
          This Operator is capable of using a custom Kubernetes service account from the caller, allowing it to follow RBAC restrictions. By default, it creates a service account if one is not specified, which can also be bound with specific roles.
          ### Environment Configuration
          You can configure a variety of artifacts with your deployment, such as labels, annotations, and environment variables from a ConfigMap, a Secret, or a value.
          ### Routing
          Expose your application to external users via a single toggle to create a Route on OpenShift or an Ingress on other Kubernetes environments. Advanced configurations, such as for TLS, are also easily enabled. Renewed certificates are automatically made available to the runtime component.
          ### High Availability via Horizontal Pod Autoscaling
          Run multiple instances of your application for high availability. Either specify a static number of replicas or easily configure horizontal auto-scaling to create (and delete) instances based on resource consumption.
          ### Persistence and advanced storage
          Enable persistence for your application by specifying simple requirements: just tell us the size of the storage and where you would like it to be mounted and we will create and manage that storage for you. This toggles a StatefulSet resource instead of a Deployment resource, so your container can recover transactions and state upon a pod restart. We offer an advanced mode where you can specify a built-in PersistentVolumeClaim, allowing you to configure many details of the persistent volume, such as its storage class and access mode.
          ### Integration with OpenShift's Topology UI
          We set the corresponding labels to support OpenShift's Developer Topology UI, which allows you to visualize your entire set of deployments and services and how they are connected.
        displayName: Runtime Component
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Runtime Component
        - Application Runtime
        - Application
        - Node
        - Swift
        - MicroProfile
        - Spring
        - Java
        - Runtime
        - Open Source
        - Application Stacks
        - RCO
        links:
        - name: Documentation
          url: https://ibm.biz/rco-docs
        - name: Repository
          url: https://ibm.biz/rco-repo
        maintainers:
        - email: arturdzm@ibm.com
          name: Artur Dzmitryieu
        - email: leojc@ca.ibm.com
          name: Leo Christy Jesuraj
        maturity: stable
        provider:
          name: Community
        relatedImages:
        - registry.connect.redhat.com/ibm/runtime-component-operator-bundle@sha256:8da482ee4c94b2546bf2b83345b1e7df6e70dcc63f68d8a2e839f10138951d2e
        - icr.io/appcafe/open-liberty/samples/getting-started@sha256:1862d437f8a3b85bdfb455f91dba3c039268ff8811ed9084d81a02956ce8d6b8
        - icr.io/appcafe/runtime-component-operator@sha256:58bfceaed2fc828bce1c0e4e21a583d25f6d72f88abdc78df2510867a4f2020f
        version: 1.5.0
      entries:
      - name: runtime-component.v1.5.0
        version: 1.5.0
      name: v1.5
    defaultChannel: v1.5
    packageName: runtime-component-operator-certified
    provider:
      name: Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/managed-by: ibm-spectrum-scale-csi-operator
      app.kubernetes.io/name: ibm-spectrum-scale-csi-operator
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-spectrum-scale-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-spectrum-scale-csi-operator.v2.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "CSIScaleOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/managed-by": "ibm-spectrum-scale-csi-operator",
                    "app.kubernetes.io/name": "ibm-spectrum-scale-csi-operator",
                    "release": "ibm-spectrum-scale-csi-operator"
                  },
                  "name": "ibm-spectrum-scale-csi",
                  "namespace": "ibm-spectrum-scale-csi-driver"
                },
                "spec": {
                  "attacherNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "clusters": [
                    {
                      "id": "\u003cPrimary Cluster ID\u003e",
                      "primary": {
                        "primaryFs": "\u003c Primary Filesystem \u003e"
                      },
                      "restApi": [
                        {
                          "guiHost": "\u003c Primary cluster GUI IP/Hostname \u003e"
                        }
                      ],
                      "secrets": "secret1",
                      "secureSslMode": false
                    }
                  ],
                  "pluginNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "provisionerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "resizerNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ],
                  "snapshotterNodeSelector": [
                    {
                      "key": "scale",
                      "value": "true"
                    }
                  ]
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
          createdAt: Thu May 5 04:42:15 IST 2022
          description: An operator for deploying and managing the IBM Spectrum Scale
            CSI Driver.
          operators.operatorframework.io/builder: operator-sdk-v1.14.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/IBM/ibm-spectrum-scale-csi/
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIScaleOperator is the Schema for the csiscaleoperators
              API
            displayName: IBM CSI Spectrum Scale Driver
            kind: CSIScaleOperator
            name: csiscaleoperators.csi.ibm.com
            version: v1
        description: |
          The IBM Spectrum Scale CSI Operator for Kubernetes installs, manages,
          upgrades the IBM Spectrum Scale CSI Driver on OpenShift and Kubernetes
          for provisioning of persistent volumes on IBM Spectrum Scale.

          ### Documentation Documentation for IBM Spectrum Scale CSI Driver and Operator can be found at [IBM Documentation](https://www.ibm.com/docs/en/spectrum-scale-csi).

          ### Getting help
          IBM Spectrum Scale CSI driver is part of the IBM Spectrum Scale offering.
          Please follow the [IBM support procedure](https://www.ibm.com/mysupport/s/?language=en_US) for any issues with the driver.

          For non-urgent issues, suggestions, recommendations feel free to open an
          issue in [github](https://github.com/IBM/ibm-spectrum-scale-csi). Issues will be addressed as team availability permits.

          **Note:** Custom Resource example represents the minimum settings needed to run
          the operator. If your environment needs more advanced settings (e.g. remote clusters,
          node mapping, image pull secret, Taint and Toleration etc.) please refer to  the sample [Custom Resource](https://raw.githubusercontent.com/IBM/ibm-spectrum-scale-csi/v2.6.0/operator/config/samples/csiscaleoperators.csi.ibm.com_cr.yaml).
        displayName: IBM Spectrum Scale CSI Plugin Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - GPFS
        - Spectrum Scale
        - CSI Driver
        links:
        - name: IBM CSI Spectrum Scale Documentation
          url: https://www.ibm.com/docs/en/spectrum-scale-csi
        - name: CSI Developer Documentation
          url: https://kubernetes-csi.github.io/docs/
        maintainers:
        - email: madhu.punjabi@in.ibm.com
          name: Madhu Thorat
        - email: amdabhad@in.ibm.com
          name: Amol Dabhade
        - email: Nitish.Kumar4@ibm.com
          name: Nitish Kumar
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - quay.io/openshift-community-operators/ibm-spectrum-scale-csi-operator@sha256:01d24dfa0d6b6146230459f8f7e74f2749d7fff3c4d75c40de386e014f9367d6
        - quay.io/ibm-spectrum-scale/ibm-spectrum-scale-csi-operator@sha256:b62ad588fbc338090f405740fd8bceaf59b3267cdb532c60ae965643ccae4d5d
        version: 2.6.0
      entries:
      - name: ibm-spectrum-scale-csi-operator.v2.6.0
        version: 2.6.0
      - name: ibm-spectrum-scale-csi-operator.v2.5.1
        version: 2.5.1
      - name: ibm-spectrum-scale-csi-operator.v2.5.0
        version: 2.5.0
      - name: ibm-spectrum-scale-csi-operator.v2.4.0
        version: 2.4.0
      - name: ibm-spectrum-scale-csi-operator.v2.3.1
        version: 2.3.1
      - name: ibm-spectrum-scale-csi-operator.v2.3.0
        version: 2.3.0
      name: stable
    defaultChannel: stable
    packageName: ibm-spectrum-scale-csi-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MayaData
      provider-url: ""
    name: openebs
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: openebsoperator.v3.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "openebs.io/v1alpha1",
                "kind": "OpenEBSInstallTemplate",
                "metadata": {
                  "name": "oebs",
                  "namespace": "openebs"
                },
                "spec": {
                  "rbac": {
                    "create": true,
                    "pspEnabled": false,
                    "kyvernoEnabled": false
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": "openebs-maya-operator"
                  },
                  "imagePullSecrets": [],
                  "release": {
                    "version": "3.0.0"
                  },
                  "legacy": {
                    "enabled": false
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": ""
                  },
                  "apiserver": {
                    "enabled": true,
                    "image": "openebs/m-apiserver",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "ports": {
                      "externalPort": 5656,
                      "internalPort": 5656
                    },
                    "sparse": {
                      "enabled": "false"
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "defaultStorageConfig": {
                    "enabled": "true"
                  },
                  "varDirectoryPath": {
                    "baseDir": "/var/openebs"
                  },
                  "provisioner": {
                    "enabled": true,
                    "image": "openebs/openebs-k8s-provisioner",
                    "imageTag": "2.12.2",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "patchJivaNodeAffinity": "enabled",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "localprovisioner": {
                    "enabled": true,
                    "image": "openebs/provisioner-localpv",
                    "imageTag": "3.0.0",
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "enableDeviceClass": true,
                    "enableHostpathClass": true,
                    "basePath": "/var/openebs/local",
                    "waitForBDBindTimeoutRetryCount": "12",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "snapshotOperator": {
                    "enabled": true,
                    "controller": {
                      "image": "openebs/snapshot-controller",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "provisioner": {
                      "image": "openebs/snapshot-provisioner",
                      "imageTag": "2.12.2",
                      "resources": {}
                    },
                    "replicas": 1,
                    "enableLeaderElection": true,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    }
                  },
                  "ndm": {
                    "enabled": true,
                    "image": "openebs/node-disk-manager",
                    "imageTag": "1.7.0",
                    "sparse": {
                      "path": "/var/openebs/sparse",
                      "size": "10737418240",
                      "count": "0"
                    },
                    "filters": {
                      "enableOsDiskExcludeFilter": true,
                      "osDiskExcludePaths": "/,/etc/hosts,/boot",
                      "enableVendorFilter": true,
                      "excludeVendors": "CLOUDBYT,OpenEBS",
                      "enablePathFilter": true,
                      "includePaths": "",
                      "excludePaths": "/dev/loop,/dev/fd0,/dev/sr0,/dev/ram,/dev/dm-,/dev/md,/dev/rbd,/dev/zd"
                    },
                    "probes": {
                      "enableSeachest": false
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "resources": {}
                  },
                  "ndmOperator": {
                    "enabled": true,
                    "image": "openebs/node-disk-operator",
                    "imageTag": "1.7.0",
                    "replicas": 1,
                    "upgradeStrategy": "Recreate",
                    "nodeSelector": {},
                    "tolerations": [],
                    "healthCheck": {
                      "initialDelaySeconds": 15,
                      "periodSeconds": 20
                    },
                    "readinessCheck": {
                      "initialDelaySeconds": 5,
                      "periodSeconds": 10
                    },
                    "resources": {}
                  },
                  "ndmExporter": {
                    "enabled": false,
                    "image": {
                      "registry": null,
                      "repository": "openebs/node-disk-exporter",
                      "pullPolicy": "IfNotPresent",
                      "tag": "1.7.0"
                    },
                    "nodeExporter": {
                      "name": "ndm-node-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-node-exporter"
                      },
                      "metricsPort": 9101
                    },
                    "clusterExporter": {
                      "name": "ndm-cluster-exporter",
                      "podLabels": {
                        "name": "openebs-ndm-cluster-exporter"
                      },
                      "metricsPort": 9100
                    }
                  },
                  "webhook": {
                    "enabled": true,
                    "image": "openebs/admission-server",
                    "imageTag": "2.12.2",
                    "failurePolicy": "Fail",
                    "replicas": 1,
                    "healthCheck": {
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60
                    },
                    "nodeSelector": {},
                    "tolerations": [],
                    "affinity": {},
                    "hostNetwork": false,
                    "resources": {}
                  },
                  "helper": {
                    "image": "openebs/linux-utils",
                    "helperImageTag": "3.0.0"
                  },
                  "featureGates": {
                    "enabled": true,
                    "GPTBasedUUID": {
                      "enabled": true,
                      "featureGateFlag": "GPTBasedUUID"
                    },
                    "APIService": {
                      "enabled": false,
                      "featureGateFlag": "APIService",
                      "address": "0.0.0.0:9115"
                    },
                    "UseOSDisk": {
                      "enabled": false,
                      "featureGateFlag": "UseOSDisk"
                    },
                    "ChangeDetection": {
                      "enabled": false,
                      "featureGateFlag": "ChangeDetection"
                    }
                  },
                  "crd": {
                    "enableInstall": true
                  },
                  "policies": {
                    "monitoring": {
                      "enabled": true,
                      "image": "openebs/m-exporter",
                      "imageTag": "2.12.2"
                    }
                  },
                  "analytics": {
                    "enabled": true,
                    "pingInterval": "24h"
                  },
                  "jiva": {
                    "image": "openebs/jiva",
                    "imageTag": "2.12.2",
                    "replicas": 3,
                    "defaultStoragePath": "/var/openebs",
                    "enabled": false,
                    "openebsLocalpv": {
                      "enabled": false
                    },
                    "localpv-provisioner": {
                      "openebsNDM": {
                        "enabled": false
                      }
                    }
                  },
                  "cstor": {
                    "pool": {
                      "image": "openebs/cstor-pool",
                      "imageTag": "2.12.2"
                    },
                    "poolMgmt": {
                      "image": "openebs/cstor-pool-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "target": {
                      "image": "openebs/cstor-istgt",
                      "imageTag": "2.12.2"
                    },
                    "volumeMgmt": {
                      "image": "openebs/cstor-volume-mgmt",
                      "imageTag": "2.12.2"
                    },
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "openebs-ndm": {
                    "enabled": false
                  },
                  "localpv-provisioner": {
                    "enabled": false,
                    "openebsNDM": {
                      "enabled": false
                    }
                  },
                  "zfs-localpv": {
                    "enabled": false
                  },
                  "lvm-localpv": {
                    "enabled": false
                  },
                  "nfs-provisioner": {
                    "enabled": false
                  },
                  "cleanup": {
                    "image": {
                      "registry": "",
                      "repository": "bitnami/kubectl",
                      "tag": "",
                      "imagePullSecrets": []
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: index.docker.io/openebs/helm-operator:v0.0.9
          createdAt: "2021-11-02T02:56:58Z"
          description: Creates and maintains OpenEBS Control Plane deployments
          repository: https://github.com/openebs/helm-operator
          support: https://slack.openebs.io/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a OpenEBS Install Operator
            displayName: OpenEBS Install Template
            kind: OpenEBSInstallTemplate
            name: openebsinstalltemplates.openebs.io
            version: v1alpha1
        description: |
          **OpenEBS** is a leading container attached storage solution that enables the use of containers for mission-critical, persistent workloads and for other stateful workloads such as logging or Prometheus for example.
          OpenEBS itself is deployed as just another container on your host and enables storage services that can be designated on a per pod, application, cluster or container level, including:
          * Data persistence across nodes
          * Synchronization of data across availability zones and cloud providers
          * A common layer whether you are running on the cloud, or your bare metal
          * Integration with Kubernetes, so developer and application intent flows into OpenEBS
          * Management of tiering to and from S3 and other targets.
          ## OpenEBS Operator
          OpenEBS primarily provides container attached block storage (iSCSI volumes) by leveraging/aggregating the storage on the nodes, with the storage controller itself running as a container. Different storage engines (Jiva & cStor) are supported, with tools available to dynamically provision Kubernetes Local PVs. The volumes are dynamically provisioned via PersistentVolumeClaims and are managed by a control plane component called "maya", which also runs as a deployment in the K8s cluster. In addition to maya, a typical OpenEBS installation comprises several other resources, which aid with various functionalities, ranging from snapshotting to disk management. All these components are described briefly below:
          * **Maya-API-Server** - A storage orchestrator which integrates into Kubernetes workflow to help provision
            and manage OpenEBS Jiva & cStor (storage engine) volumes

          * **Dynamic-OpenEBS-Provisioner** - A Kubernetes external storage provisioner that utilizes APIs exposed by maya-apiserver
            to perform provision & delete operations of Jiva & cStor volumes

          * **Dynamic-LocalPV-Provisioner** - A dynamic provisioner for Kubernetes Local PVs
          * **OpenEBS-Snapshot-Operator** - A Kubernetes snapshot controller that creates & restores OpenEBS volume snapshots
          * **Node-Disk-Manager** - A disk management controller which identifies available disks, maintains inventory, and
            dynamically attaches/detaches disks to backend storage pods

          The helm-based OpenEBS Operator eases the setup of all the above mentioned components, with a simple custom resource provided to define the install options, thereby enabling applications to start using the OpenEBS storageclasses in their PVCs. The OpenEBSInstallTemplate CR can be used to specify start-up parameters & also update/overwrite the definitions post install.
          ## Pre-Requisites
          Before installing OpenEBS control plane, perform the following steps to ensure successful deployment of the Node-Disk-Manager & the OpenEBS volume replicas, respectively.
          * Configure the OpenEBS service account on the openshift-operators namespace/project to use the privileged security context constraint.

            **Note**: The serviceaccount name is same as the one specified in the `spec.serviceAccount.name` field of the OpenEBSInstallTemplate CR.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:openebs-maya-operator
            ```

          * Configure the default service account on the namespace/project in which the volume replicas are deployed to use privileged
            security context constraint.


            ```
            oc adm policy add-scc-to-user privileged system:serviceaccount:openshift-operators:default
            ```

          ## Getting Started
          * Try the quickstart [guide](https://github.com/openebs/helm-operator/blob/master/olm/README.md)
          * To learn how to contribute, please read the [contribution guide](https://github.com/openebs/helm-operator/blob/master/CONTRIBUTING.md)
          * OpenEBS welcomes your feedback and contributions in any form possible. [Join our Community](https://openebs.org/community)

          ## License
          OpenEBS is distributed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt)
        displayName: OpenEBS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - OpenEBS
        - ContainerAttachedStorage
        links:
        - name: OpenEBS Website
          url: https://openebs.io
        - name: Operator Source Code
          url: https://github.com/openebs/helm-operator
        - name: Install Instructions
          url: https://github.com/openebs/helm-operator/blob/master/olm/README.md
        maintainers:
        - email: kiran.mova@openebs.io
          name: kmova
        maturity: alpha
        minKubeVersion: 1.12.0
        provider:
          name: MayaData
        relatedImages:
        - index.docker.io/openebs/helm-operator:v0.0.9
        - quay.io/openshift-community-operators/openebs@sha256:ff179bdd6e63b8e2cb1ce3b96d5e03d20c1a2410296fed6fb331bf5bc2e6b14d
        version: 3.0.0
      entries:
      - name: openebsoperator.v3.0.0
        version: 3.0.0
      - name: openebsoperator.v2.12.2
        version: 2.12.2
      name: alpha
    defaultChannel: alpha
    packageName: openebs
    provider:
      name: MayaData
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ember-csi-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ember-csi-community-operator.v0.9.9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ember-csi.io/v1alpha1",
                "kind": "EmberStorageBackend",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "config": {
                    "envVars": {
                      "X_CSI_PERSISTENCE_CONFIG": {
                        "storage": "crd"
                      },
                      "X_CSI_DEBUG_MODE": "",
                      "X_CSI_DEFAULT_MOUNT_FS": "ext4",
                      "X_CSI_EMBER_CONFIG": {
                        "grpc_workers": 30,
                        "disable_logs": false,
                        "plugin_name": "",
                        "slow_operations": true,
                        "disabled": [],
                        "enable_probe": false,
                        "debug": false,
                        "user_id": "ember_csi.io",
                        "project_id": "ember_csi.io"
                      },
                      "X_CSI_BACKEND_CONFIG": {
                        "enable_unsupported_driver": false,
                        "driver": "ACCESSIscsi",
                        "name": "",
                        "driver__ACCESSIscsi__vrts_lun_sparse": true,
                        "driver__ACCESSIscsi__vrts_target_config": "/etc/cinder/vrts_target.xml",
                        "driver__AS13000__as13000_ipsan_pools__transform_csv": "Pool0",
                        "driver__AS13000__as13000_meta_pool": "",
                        "driver__AS13000__as13000_token_available_time": 3300,
                        "driver__Acs5000ISCSI__acs5000_copy_interval": 5,
                        "driver__Acs5000ISCSI__acs5000_target": 0,
                        "driver__Acs5000ISCSI__acs5000_volpool_name__transform_csv": "pool01",
                        "driver__DPLFC__dpl_pool": "",
                        "driver__DPLFC__dpl_port": 8357,
                        "driver__DPLISCSI__dpl_pool": "",
                        "driver__DPLISCSI__dpl_port": 8357,
                        "driver__DSWARE__dsware_rest_url": "",
                        "driver__DSWARE__dsware_storage_pools": "",
                        "driver__DSWARE__manager_ips__transform_csv_kvs": "",
                        "driver__Datera__backend_availability_zone": "",
                        "driver__Datera__chap_password": "",
                        "driver__Datera__chap_username": "",
                        "driver__Datera__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__Datera__driver_client_cert": "",
                        "driver__Datera__driver_client_cert_key": "",
                        "driver__Datera__driver_data_namespace": "",
                        "driver__Datera__driver_ssl_cert_path": "",
                        "driver__Datera__driver_ssl_cert_verify": false,
                        "driver__Datera__driver_use_ssl": false,
                        "driver__Datera__filter_function": "",
                        "driver__Datera__goodness_function": "",
                        "driver__Datera__iscsi_iotype": "fileio",
                        "driver__Datera__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__Datera__iscsi_target_flags": "",
                        "driver__Datera__iscsi_write_cache": "on",
                        "driver__Datera__num_shell_tries": 3,
                        "driver__Datera__num_volume_device_scan_tries": 3,
                        "driver__Datera__report_discard_supported": false,
                        "driver__Datera__storage_protocol": "iscsi",
                        "driver__Datera__target_helper": "tgtadm",
                        "driver__Datera__target_ip_address": "$my_ip",
                        "driver__Datera__target_port": 3260,
                        "driver__Datera__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__Datera__target_protocol": "iscsi",
                        "driver__Datera__trace_flags__transform_csv": "",
                        "driver__Datera__use_chap_auth": false,
                        "driver__Datera__volume_backend_name": "",
                        "driver__Datera__volume_clear": "zero",
                        "driver__Datera__volume_clear_ionice": "",
                        "driver__Datera__volume_clear_size": 0,
                        "driver__Datera__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__Datera__volume_copy_bps_limit": 0,
                        "driver__Datera__volume_dd_blocksize": "1M",
                        "driver__Datera__volumes_dir": "$state_path/volumes",
                        "driver__FJDXFC__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FJDXISCSI__cinder_eternus_config_file": "/etc/cinder/cinder_fujitsu_eternus_dx.xml",
                        "driver__FlashSystemFC__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemFC__flashsystem_multihostmap_enabled": true,
                        "driver__FlashSystemISCSI__flashsystem_connection_protocol": "FC",
                        "driver__FlashSystemISCSI__flashsystem_multihostmap_enabled": true,
                        "driver__GPFS__gpfs_images_dir": "",
                        "driver__GPFS__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFS__gpfs_max_clone_depth": 0,
                        "driver__GPFS__gpfs_mount_point_base": "",
                        "driver__GPFS__gpfs_sparse_volumes": true,
                        "driver__GPFS__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_hosts__transform_csv": "",
                        "driver__GPFSRemote__gpfs_hosts_key_file": "$state_path/ssh_known_hosts",
                        "driver__GPFSRemote__gpfs_images_dir": "",
                        "driver__GPFSRemote__gpfs_images_share_mode__transform_empty_none": "",
                        "driver__GPFSRemote__gpfs_max_clone_depth": 0,
                        "driver__GPFSRemote__gpfs_mount_point_base": "",
                        "driver__GPFSRemote__gpfs_private_key": "",
                        "driver__GPFSRemote__gpfs_sparse_volumes": true,
                        "driver__GPFSRemote__gpfs_ssh_port": 22,
                        "driver__GPFSRemote__gpfs_storage_pool": "system",
                        "driver__GPFSRemote__gpfs_strict_host_key_policy": false,
                        "driver__GPFSRemote__gpfs_user_login": "root",
                        "driver__GPFSRemote__gpfs_user_password": "",
                        "driver__HBSDFC__backend_availability_zone": "",
                        "driver__HBSDFC__chap_password": "",
                        "driver__HBSDFC__chap_username": "",
                        "driver__HBSDFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDFC__driver_client_cert": "",
                        "driver__HBSDFC__driver_client_cert_key": "",
                        "driver__HBSDFC__driver_data_namespace": "",
                        "driver__HBSDFC__driver_ssl_cert_path": "",
                        "driver__HBSDFC__driver_ssl_cert_verify": false,
                        "driver__HBSDFC__driver_use_ssl": false,
                        "driver__HBSDFC__filter_function": "",
                        "driver__HBSDFC__goodness_function": "",
                        "driver__HBSDFC__iscsi_iotype": "fileio",
                        "driver__HBSDFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDFC__iscsi_target_flags": "",
                        "driver__HBSDFC__iscsi_write_cache": "on",
                        "driver__HBSDFC__num_shell_tries": 3,
                        "driver__HBSDFC__num_volume_device_scan_tries": 3,
                        "driver__HBSDFC__report_discard_supported": false,
                        "driver__HBSDFC__storage_protocol": "iscsi",
                        "driver__HBSDFC__target_helper": "tgtadm",
                        "driver__HBSDFC__target_ip_address": "$my_ip",
                        "driver__HBSDFC__target_port": 3260,
                        "driver__HBSDFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDFC__target_protocol": "iscsi",
                        "driver__HBSDFC__trace_flags__transform_csv": "",
                        "driver__HBSDFC__use_chap_auth": false,
                        "driver__HBSDFC__volume_backend_name": "",
                        "driver__HBSDFC__volume_clear": "zero",
                        "driver__HBSDFC__volume_clear_ionice": "",
                        "driver__HBSDFC__volume_clear_size": 0,
                        "driver__HBSDFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDFC__volume_copy_bps_limit": 0,
                        "driver__HBSDFC__volume_dd_blocksize": "1M",
                        "driver__HBSDFC__volumes_dir": "$state_path/volumes",
                        "driver__HBSDISCSI__backend_availability_zone": "",
                        "driver__HBSDISCSI__chap_password": "",
                        "driver__HBSDISCSI__chap_username": "",
                        "driver__HBSDISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__HBSDISCSI__driver_client_cert": "",
                        "driver__HBSDISCSI__driver_client_cert_key": "",
                        "driver__HBSDISCSI__driver_data_namespace": "",
                        "driver__HBSDISCSI__driver_ssl_cert_path": "",
                        "driver__HBSDISCSI__driver_ssl_cert_verify": false,
                        "driver__HBSDISCSI__driver_use_ssl": false,
                        "driver__HBSDISCSI__filter_function": "",
                        "driver__HBSDISCSI__goodness_function": "",
                        "driver__HBSDISCSI__iscsi_iotype": "fileio",
                        "driver__HBSDISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__HBSDISCSI__iscsi_target_flags": "",
                        "driver__HBSDISCSI__iscsi_write_cache": "on",
                        "driver__HBSDISCSI__num_shell_tries": 3,
                        "driver__HBSDISCSI__num_volume_device_scan_tries": 3,
                        "driver__HBSDISCSI__report_discard_supported": false,
                        "driver__HBSDISCSI__storage_protocol": "iscsi",
                        "driver__HBSDISCSI__target_helper": "tgtadm",
                        "driver__HBSDISCSI__target_ip_address": "$my_ip",
                        "driver__HBSDISCSI__target_port": 3260,
                        "driver__HBSDISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__HBSDISCSI__target_protocol": "iscsi",
                        "driver__HBSDISCSI__trace_flags__transform_csv": "",
                        "driver__HBSDISCSI__use_chap_auth": false,
                        "driver__HBSDISCSI__volume_backend_name": "",
                        "driver__HBSDISCSI__volume_clear": "zero",
                        "driver__HBSDISCSI__volume_clear_ionice": "",
                        "driver__HBSDISCSI__volume_clear_size": 0,
                        "driver__HBSDISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__HBSDISCSI__volume_copy_bps_limit": 0,
                        "driver__HBSDISCSI__volume_dd_blocksize": "1M",
                        "driver__HBSDISCSI__volumes_dir": "$state_path/volumes",
                        "driver__HPE3PARFC__hpe3par_api_url": "",
                        "driver__HPE3PARFC__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARFC__hpe3par_cpg_snap": "",
                        "driver__HPE3PARFC__hpe3par_debug": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARFC__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARFC__hpe3par_password": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARFC__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARFC__hpe3par_target_nsp": "",
                        "driver__HPE3PARFC__hpe3par_username": "",
                        "driver__HPE3PARFC__san_ip": "",
                        "driver__HPE3PARFC__san_login": "admin",
                        "driver__HPE3PARFC__san_password": "",
                        "driver__HPE3PARFC__san_private_key": "",
                        "driver__HPE3PARFC__san_ssh_port": 22,
                        "driver__HPE3PARFC__ssh_conn_timeout": 30,
                        "driver__HPE3PARFC__target_ip_address": "$my_ip",
                        "driver__HPE3PARFC__target_port": 3260,
                        "driver__HPE3PARFC__unique_fqdn_network": true,
                        "driver__HPE3PARISCSI__hpe3par_api_url": "",
                        "driver__HPE3PARISCSI__hpe3par_cpg__transform_csv": "OpenStack",
                        "driver__HPE3PARISCSI__hpe3par_cpg_snap": "",
                        "driver__HPE3PARISCSI__hpe3par_debug": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_chap_enabled": false,
                        "driver__HPE3PARISCSI__hpe3par_iscsi_ips__transform_csv": "",
                        "driver__HPE3PARISCSI__hpe3par_password": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_expiration": "",
                        "driver__HPE3PARISCSI__hpe3par_snapshot_retention": "",
                        "driver__HPE3PARISCSI__hpe3par_target_nsp": "",
                        "driver__HPE3PARISCSI__hpe3par_username": "",
                        "driver__HPE3PARISCSI__san_ip": "",
                        "driver__HPE3PARISCSI__san_login": "admin",
                        "driver__HPE3PARISCSI__san_password": "",
                        "driver__HPE3PARISCSI__san_private_key": "",
                        "driver__HPE3PARISCSI__san_ssh_port": 22,
                        "driver__HPE3PARISCSI__ssh_conn_timeout": 30,
                        "driver__HPE3PARISCSI__target_ip_address": "$my_ip",
                        "driver__HPE3PARISCSI__target_port": 3260,
                        "driver__HPE3PARISCSI__unique_fqdn_network": true,
                        "driver__HPMSAFC__driver_ssl_cert_path": "",
                        "driver__HPMSAFC__driver_ssl_cert_verify": false,
                        "driver__HPMSAFC__driver_use_ssl": false,
                        "driver__HPMSAFC__hpmsa_pool_name": "A",
                        "driver__HPMSAFC__hpmsa_pool_type": "virtual",
                        "driver__HPMSAFC__san_ip": "",
                        "driver__HPMSAFC__san_login": "admin",
                        "driver__HPMSAFC__san_password": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_path": "",
                        "driver__HPMSAISCSI__driver_ssl_cert_verify": false,
                        "driver__HPMSAISCSI__driver_use_ssl": false,
                        "driver__HPMSAISCSI__hpmsa_iscsi_ips__transform_csv": "",
                        "driver__HPMSAISCSI__hpmsa_pool_name": "A",
                        "driver__HPMSAISCSI__hpmsa_pool_type": "virtual",
                        "driver__HPMSAISCSI__san_ip": "",
                        "driver__HPMSAISCSI__san_login": "admin",
                        "driver__HPMSAISCSI__san_password": "",
                        "driver__HuaweiFC__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiFC__hypermetro_devices": "",
                        "driver__HuaweiFC__metro_domain_name": "",
                        "driver__HuaweiFC__metro_san_address": "",
                        "driver__HuaweiFC__metro_san_password": "",
                        "driver__HuaweiFC__metro_san_user": "",
                        "driver__HuaweiFC__metro_storage_pools": "",
                        "driver__HuaweiISCSI__cinder_huawei_conf_file": "/etc/cinder/cinder_huawei_conf.xml",
                        "driver__HuaweiISCSI__hypermetro_devices": "",
                        "driver__HuaweiISCSI__metro_domain_name": "",
                        "driver__HuaweiISCSI__metro_san_address": "",
                        "driver__HuaweiISCSI__metro_san_password": "",
                        "driver__HuaweiISCSI__metro_san_user": "",
                        "driver__HuaweiISCSI__metro_storage_pools": "",
                        "driver__IBMStorage__chap": "disabled",
                        "driver__IBMStorage__connection_type": "iscsi",
                        "driver__IBMStorage__management_ips": "",
                        "driver__IBMStorage__proxy": "cinder.volume.drivers.ibm.ibm_storage.proxy.IBMStorageProxy",
                        "driver__InStorageMCSFC__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSFC__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSFC__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSFC__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSFC__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSFC__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSFC__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSFC__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSFC__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSFC__instorage_san_secondary_ip": "",
                        "driver__InStorageMCSISCSI__instorage_mcs_allow_tenant_qos": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_rate": 50,
                        "driver__InStorageMCSISCSI__instorage_mcs_localcopy_timeout": 120,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_autoexpand": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_compression": false,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_grainsize": 256,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_intier": true,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_iogrp": "0",
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_rsize": 2,
                        "driver__InStorageMCSISCSI__instorage_mcs_vol_warning": 0,
                        "driver__InStorageMCSISCSI__instorage_mcs_volpool_name__transform_csv": "volpool",
                        "driver__InStorageMCSISCSI__instorage_san_secondary_ip": "",
                        "driver__InfiniboxVolume__chap_password": "",
                        "driver__InfiniboxVolume__chap_username": "",
                        "driver__InfiniboxVolume__infinidat_iscsi_netspaces__transform_csv": "",
                        "driver__InfiniboxVolume__infinidat_pool_name": "",
                        "driver__InfiniboxVolume__infinidat_storage_protocol": "fc",
                        "driver__InfiniboxVolume__infinidat_use_compression": false,
                        "driver__InfiniboxVolume__num_volume_device_scan_tries": 3,
                        "driver__InfiniboxVolume__san_ip": "",
                        "driver__InfiniboxVolume__san_login": "admin",
                        "driver__InfiniboxVolume__san_password": "",
                        "driver__InfiniboxVolume__san_thin_provision": true,
                        "driver__InfiniboxVolume__use_chap_auth": false,
                        "driver__InfiniboxVolume__volume_dd_blocksize": "1M",
                        "driver__InfortrendCLIFC__infortrend_cli_cache": false,
                        "driver__InfortrendCLIFC__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIFC__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIFC__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIFC__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIFC__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIFC__java_path": "/usr/bin/java",
                        "driver__InfortrendCLIISCSI__infortrend_cli_cache": false,
                        "driver__InfortrendCLIISCSI__infortrend_cli_max_retries": 5,
                        "driver__InfortrendCLIISCSI__infortrend_cli_path": "/opt/bin/Infortrend/raidcmd_ESDS10.jar",
                        "driver__InfortrendCLIISCSI__infortrend_cli_timeout": 60,
                        "driver__InfortrendCLIISCSI__infortrend_iqn_prefix": "iqn.2002-10.com.infortrend",
                        "driver__InfortrendCLIISCSI__infortrend_pools_name__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_a_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__infortrend_slots_b_channels_id__transform_csv": "",
                        "driver__InfortrendCLIISCSI__java_path": "/usr/bin/java",
                        "driver__JovianISCSI__backend_availability_zone": "",
                        "driver__JovianISCSI__chap_password": "",
                        "driver__JovianISCSI__chap_username": "",
                        "driver__JovianISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__JovianISCSI__driver_client_cert": "",
                        "driver__JovianISCSI__driver_client_cert_key": "",
                        "driver__JovianISCSI__driver_data_namespace": "",
                        "driver__JovianISCSI__driver_ssl_cert_path": "",
                        "driver__JovianISCSI__driver_ssl_cert_verify": false,
                        "driver__JovianISCSI__driver_use_ssl": false,
                        "driver__JovianISCSI__filter_function": "",
                        "driver__JovianISCSI__goodness_function": "",
                        "driver__JovianISCSI__iscsi_iotype": "fileio",
                        "driver__JovianISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__JovianISCSI__iscsi_target_flags": "",
                        "driver__JovianISCSI__iscsi_write_cache": "on",
                        "driver__JovianISCSI__num_shell_tries": 3,
                        "driver__JovianISCSI__num_volume_device_scan_tries": 3,
                        "driver__JovianISCSI__report_discard_supported": false,
                        "driver__JovianISCSI__storage_protocol": "iscsi",
                        "driver__JovianISCSI__target_helper": "tgtadm",
                        "driver__JovianISCSI__target_ip_address": "$my_ip",
                        "driver__JovianISCSI__target_port": 3260,
                        "driver__JovianISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__JovianISCSI__target_protocol": "iscsi",
                        "driver__JovianISCSI__trace_flags__transform_csv": "",
                        "driver__JovianISCSI__use_chap_auth": false,
                        "driver__JovianISCSI__volume_backend_name": "",
                        "driver__JovianISCSI__volume_clear": "zero",
                        "driver__JovianISCSI__volume_clear_ionice": "",
                        "driver__JovianISCSI__volume_clear_size": 0,
                        "driver__JovianISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__JovianISCSI__volume_copy_bps_limit": 0,
                        "driver__JovianISCSI__volume_dd_blocksize": "1M",
                        "driver__JovianISCSI__volumes_dir": "$state_path/volumes",
                        "driver__KaminarioISCSI__disable_discovery": false,
                        "driver__KaminarioISCSI__san_ip": "",
                        "driver__KaminarioISCSI__san_login": "admin",
                        "driver__KaminarioISCSI__san_password": "",
                        "driver__KaminarioISCSI__unique_fqdn_network": true,
                        "driver__KaminarioISCSI__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__iscsi_iotype": "fileio",
                        "driver__LVMVolume__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__LVMVolume__iscsi_target_flags": "",
                        "driver__LVMVolume__iscsi_write_cache": "on",
                        "driver__LVMVolume__lvm_conf_file": "/etc/cinder/lvm.conf",
                        "driver__LVMVolume__lvm_mirrors": 0,
                        "driver__LVMVolume__lvm_suppress_fd_warnings": false,
                        "driver__LVMVolume__lvm_type": "auto",
                        "driver__LVMVolume__nvmet_port_id": 1,
                        "driver__LVMVolume__scst_target_driver": "iscsi",
                        "driver__LVMVolume__scst_target_iqn_name": "",
                        "driver__LVMVolume__spdk_max_queue_depth": 64,
                        "driver__LVMVolume__spdk_rpc_ip": "",
                        "driver__LVMVolume__spdk_rpc_password": "",
                        "driver__LVMVolume__spdk_rpc_port": 8000,
                        "driver__LVMVolume__spdk_rpc_username": "",
                        "driver__LVMVolume__target_helper": "tgtadm",
                        "driver__LVMVolume__target_ip_address": "$my_ip",
                        "driver__LVMVolume__target_port": 3260,
                        "driver__LVMVolume__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__LVMVolume__target_protocol": "iscsi",
                        "driver__LVMVolume__volume_clear": "zero",
                        "driver__LVMVolume__volume_clear_size": 0,
                        "driver__LVMVolume__volume_dd_blocksize": "1M",
                        "driver__LVMVolume__volume_group": "cinder-volumes",
                        "driver__LVMVolume__volumes_dir": "$state_path/volumes",
                        "driver__LenovoFC__lenovo_pool_name": "A",
                        "driver__LenovoFC__lenovo_pool_type": "virtual",
                        "driver__LenovoISCSI__lenovo_iscsi_ips__transform_csv": "",
                        "driver__LenovoISCSI__lenovo_pool_name": "A",
                        "driver__LenovoISCSI__lenovo_pool_type": "virtual",
                        "driver__LinstorDrbd__linstor_autoplace_count": 0,
                        "driver__LinstorDrbd__linstor_controller_diskless": true,
                        "driver__LinstorDrbd__linstor_default_blocksize": 4096,
                        "driver__LinstorDrbd__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorDrbd__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorDrbd__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorDrbd__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__LinstorIscsi__linstor_autoplace_count": 0,
                        "driver__LinstorIscsi__linstor_controller_diskless": true,
                        "driver__LinstorIscsi__linstor_default_blocksize": 4096,
                        "driver__LinstorIscsi__linstor_default_storage_pool_name": "DfltStorPool",
                        "driver__LinstorIscsi__linstor_default_uri": "linstor://localhost",
                        "driver__LinstorIscsi__linstor_default_volume_group_name": "drbd-vg",
                        "driver__LinstorIscsi__linstor_volume_downsize_factor__transform_string_float": "4096",
                        "driver__MStorageFC__nec_actual_free_capacity": false,
                        "driver__MStorageFC__nec_auto_accesscontrol": true,
                        "driver__MStorageFC__nec_backend_max_ld_count": 1024,
                        "driver__MStorageFC__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_backup_pools__transform_csv": "",
                        "driver__MStorageFC__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageFC__nec_diskarray_name": "",
                        "driver__MStorageFC__nec_ismcli_fip": "",
                        "driver__MStorageFC__nec_ismcli_password": "",
                        "driver__MStorageFC__nec_ismcli_privkey": "",
                        "driver__MStorageFC__nec_ismcli_user": "",
                        "driver__MStorageFC__nec_ismview_alloptimize": false,
                        "driver__MStorageFC__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageFC__nec_ldname_format": "LX:%s",
                        "driver__MStorageFC__nec_ldset": "",
                        "driver__MStorageFC__nec_pools__transform_csv": "",
                        "driver__MStorageFC__nec_queryconfig_view": false,
                        "driver__MStorageFC__nec_ssh_pool_port_number": 22,
                        "driver__MStorageFC__nec_unpairthread_timeout": 3600,
                        "driver__MStorageISCSI__nec_actual_free_capacity": false,
                        "driver__MStorageISCSI__nec_auto_accesscontrol": true,
                        "driver__MStorageISCSI__nec_backend_max_ld_count": 1024,
                        "driver__MStorageISCSI__nec_backup_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_backup_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_cv_ldname_format": "LX:__ControlVolume_%xh",
                        "driver__MStorageISCSI__nec_diskarray_name": "",
                        "driver__MStorageISCSI__nec_ismcli_fip": "",
                        "driver__MStorageISCSI__nec_ismcli_password": "",
                        "driver__MStorageISCSI__nec_ismcli_privkey": "",
                        "driver__MStorageISCSI__nec_ismcli_user": "",
                        "driver__MStorageISCSI__nec_ismview_alloptimize": false,
                        "driver__MStorageISCSI__nec_ismview_dir": "/tmp/nec/cinder",
                        "driver__MStorageISCSI__nec_ldname_format": "LX:%s",
                        "driver__MStorageISCSI__nec_ldset": "",
                        "driver__MStorageISCSI__nec_pools__transform_csv": "",
                        "driver__MStorageISCSI__nec_queryconfig_view": false,
                        "driver__MStorageISCSI__nec_ssh_pool_port_number": 22,
                        "driver__MStorageISCSI__nec_unpairthread_timeout": 3600,
                        "driver__MacroSANFC__backend_availability_zone": "",
                        "driver__MacroSANFC__chap_password": "",
                        "driver__MacroSANFC__chap_username": "",
                        "driver__MacroSANFC__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANFC__driver_client_cert": "",
                        "driver__MacroSANFC__driver_client_cert_key": "",
                        "driver__MacroSANFC__driver_data_namespace": "",
                        "driver__MacroSANFC__driver_ssl_cert_path": "",
                        "driver__MacroSANFC__driver_ssl_cert_verify": false,
                        "driver__MacroSANFC__driver_use_ssl": false,
                        "driver__MacroSANFC__filter_function": "",
                        "driver__MacroSANFC__goodness_function": "",
                        "driver__MacroSANFC__iscsi_iotype": "fileio",
                        "driver__MacroSANFC__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANFC__iscsi_target_flags": "",
                        "driver__MacroSANFC__iscsi_write_cache": "on",
                        "driver__MacroSANFC__num_shell_tries": 3,
                        "driver__MacroSANFC__num_volume_device_scan_tries": 3,
                        "driver__MacroSANFC__report_discard_supported": false,
                        "driver__MacroSANFC__storage_protocol": "iscsi",
                        "driver__MacroSANFC__target_helper": "tgtadm",
                        "driver__MacroSANFC__target_ip_address": "$my_ip",
                        "driver__MacroSANFC__target_port": 3260,
                        "driver__MacroSANFC__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANFC__target_protocol": "iscsi",
                        "driver__MacroSANFC__trace_flags__transform_csv": "",
                        "driver__MacroSANFC__use_chap_auth": false,
                        "driver__MacroSANFC__volume_backend_name": "",
                        "driver__MacroSANFC__volume_clear": "zero",
                        "driver__MacroSANFC__volume_clear_ionice": "",
                        "driver__MacroSANFC__volume_clear_size": 0,
                        "driver__MacroSANFC__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANFC__volume_copy_bps_limit": 0,
                        "driver__MacroSANFC__volume_dd_blocksize": "1M",
                        "driver__MacroSANFC__volumes_dir": "$state_path/volumes",
                        "driver__MacroSANISCSI__backend_availability_zone": "",
                        "driver__MacroSANISCSI__chap_password": "",
                        "driver__MacroSANISCSI__chap_username": "",
                        "driver__MacroSANISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__MacroSANISCSI__driver_client_cert": "",
                        "driver__MacroSANISCSI__driver_client_cert_key": "",
                        "driver__MacroSANISCSI__driver_data_namespace": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_path": "",
                        "driver__MacroSANISCSI__driver_ssl_cert_verify": false,
                        "driver__MacroSANISCSI__driver_use_ssl": false,
                        "driver__MacroSANISCSI__filter_function": "",
                        "driver__MacroSANISCSI__goodness_function": "",
                        "driver__MacroSANISCSI__iscsi_iotype": "fileio",
                        "driver__MacroSANISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__MacroSANISCSI__iscsi_target_flags": "",
                        "driver__MacroSANISCSI__iscsi_write_cache": "on",
                        "driver__MacroSANISCSI__num_shell_tries": 3,
                        "driver__MacroSANISCSI__num_volume_device_scan_tries": 3,
                        "driver__MacroSANISCSI__report_discard_supported": false,
                        "driver__MacroSANISCSI__storage_protocol": "iscsi",
                        "driver__MacroSANISCSI__target_helper": "tgtadm",
                        "driver__MacroSANISCSI__target_ip_address": "$my_ip",
                        "driver__MacroSANISCSI__target_port": 3260,
                        "driver__MacroSANISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__MacroSANISCSI__target_protocol": "iscsi",
                        "driver__MacroSANISCSI__trace_flags__transform_csv": "",
                        "driver__MacroSANISCSI__use_chap_auth": false,
                        "driver__MacroSANISCSI__volume_backend_name": "",
                        "driver__MacroSANISCSI__volume_clear": "zero",
                        "driver__MacroSANISCSI__volume_clear_ionice": "",
                        "driver__MacroSANISCSI__volume_clear_size": 0,
                        "driver__MacroSANISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__MacroSANISCSI__volume_copy_bps_limit": 0,
                        "driver__MacroSANISCSI__volume_dd_blocksize": "1M",
                        "driver__MacroSANISCSI__volumes_dir": "$state_path/volumes",
                        "driver__NetAppCmodeFibreChannel__netapp_vserver": "",
                        "driver__NetAppCmodeISCSI__netapp_vserver": "",
                        "driver__NexentaISCSI__nexenta_blocksize": 4096,
                        "driver__NexentaISCSI__nexenta_dataset_compression": "on",
                        "driver__NexentaISCSI__nexenta_dataset_dedup": "off",
                        "driver__NexentaISCSI__nexenta_dataset_description": "",
                        "driver__NexentaISCSI__nexenta_folder": "",
                        "driver__NexentaISCSI__nexenta_group_snapshot_template": "group-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_host": "",
                        "driver__NexentaISCSI__nexenta_host_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_iscsi_target_host_group": "all",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_groups": "",
                        "driver__NexentaISCSI__nexenta_iscsi_target_portal_port": 3260,
                        "driver__NexentaISCSI__nexenta_iscsi_target_portals": "",
                        "driver__NexentaISCSI__nexenta_lu_writebackcache_disabled": false,
                        "driver__NexentaISCSI__nexenta_luns_per_target": 100,
                        "driver__NexentaISCSI__nexenta_ns5_blocksize": 32,
                        "driver__NexentaISCSI__nexenta_origin_snapshot_template": "origin-snapshot-%s",
                        "driver__NexentaISCSI__nexenta_rest_backoff_factor__transform_string_float": "0.5",
                        "driver__NexentaISCSI__nexenta_rest_connect_timeout__transform_string_float": "30",
                        "driver__NexentaISCSI__nexenta_rest_protocol": "auto",
                        "driver__NexentaISCSI__nexenta_rest_read_timeout__transform_string_float": "300",
                        "driver__NexentaISCSI__nexenta_rest_retry_count": 3,
                        "driver__NexentaISCSI__nexenta_sparse": false,
                        "driver__NexentaISCSI__nexenta_target_group_prefix": "cinder",
                        "driver__NexentaISCSI__nexenta_target_prefix": "iqn.1986-03.com.sun:02:cinder",
                        "driver__NexentaISCSI__nexenta_use_https": true,
                        "driver__NexentaISCSI__nexenta_volume": "cinder",
                        "driver__NexentaISCSI__nexenta_volume_group": "iscsi",
                        "driver__NimbleFC__nimble_pool_name": "default",
                        "driver__NimbleFC__nimble_subnet_label": "*",
                        "driver__NimbleFC__nimble_verify_cert_path": "",
                        "driver__NimbleFC__nimble_verify_certificate": false,
                        "driver__NimbleISCSI__nimble_pool_name": "default",
                        "driver__NimbleISCSI__nimble_subnet_label": "*",
                        "driver__NimbleISCSI__nimble_verify_cert_path": "",
                        "driver__NimbleISCSI__nimble_verify_certificate": false,
                        "driver__PVMEFC__driver_ssl_cert_path": "",
                        "driver__PVMEFC__driver_ssl_cert_verify": false,
                        "driver__PVMEFC__driver_use_ssl": false,
                        "driver__PVMEFC__pvme_pool_name": "A",
                        "driver__PVMEFC__san_ip": "",
                        "driver__PVMEFC__san_login": "admin",
                        "driver__PVMEFC__san_password": "",
                        "driver__PVMEISCSI__driver_ssl_cert_path": "",
                        "driver__PVMEISCSI__driver_ssl_cert_verify": false,
                        "driver__PVMEISCSI__driver_use_ssl": false,
                        "driver__PVMEISCSI__pvme_iscsi_ips__transform_csv": "",
                        "driver__PVMEISCSI__pvme_pool_name": "A",
                        "driver__PVMEISCSI__san_ip": "",
                        "driver__PVMEISCSI__san_login": "admin",
                        "driver__PVMEISCSI__san_password": "",
                        "driver__PowerFlex__powerflex_allow_migration_during_rebuild": false,
                        "driver__PowerFlex__powerflex_allow_non_padded_volumes": false,
                        "driver__PowerFlex__powerflex_max_over_subscription_ratio__transform_string_float": "10.0",
                        "driver__PowerFlex__powerflex_rest_server_port": 443,
                        "driver__PowerFlex__powerflex_round_volume_capacity": true,
                        "driver__PowerFlex__powerflex_server_api_version": "",
                        "driver__PowerFlex__powerflex_storage_pools": "",
                        "driver__PowerFlex__powerflex_unmap_volume_before_deletion": false,
                        "driver__PowerMaxFC__driver_ssl_cert_verify": false,
                        "driver__PowerMaxFC__initiator_check": false,
                        "driver__PowerMaxFC__interval": 3,
                        "driver__PowerMaxFC__load_balance": false,
                        "driver__PowerMaxFC__load_balance_real_time": false,
                        "driver__PowerMaxFC__load_data_format": "Avg",
                        "driver__PowerMaxFC__load_look_back": 60,
                        "driver__PowerMaxFC__load_look_back_real_time": 1,
                        "driver__PowerMaxFC__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__port_load_metric": "PercentBusy",
                        "driver__PowerMaxFC__powermax_array": "",
                        "driver__PowerMaxFC__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxFC__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxFC__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxFC__powermax_service_level": "",
                        "driver__PowerMaxFC__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxFC__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxFC__powermax_srp": "",
                        "driver__PowerMaxFC__retries": 200,
                        "driver__PowerMaxFC__san_ip": "",
                        "driver__PowerMaxFC__san_login": "admin",
                        "driver__PowerMaxFC__san_password": "",
                        "driver__PowerMaxFC__u4p_failover_autofailback": true,
                        "driver__PowerMaxFC__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxFC__u4p_failover_retries": 3,
                        "driver__PowerMaxFC__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxFC__u4p_failover_timeout": 20,
                        "driver__PowerMaxFC__vmax_workload": "",
                        "driver__PowerMaxISCSI__chap_password": "",
                        "driver__PowerMaxISCSI__chap_username": "",
                        "driver__PowerMaxISCSI__driver_ssl_cert_verify": false,
                        "driver__PowerMaxISCSI__initiator_check": false,
                        "driver__PowerMaxISCSI__interval": 3,
                        "driver__PowerMaxISCSI__load_balance": false,
                        "driver__PowerMaxISCSI__load_balance_real_time": false,
                        "driver__PowerMaxISCSI__load_data_format": "Avg",
                        "driver__PowerMaxISCSI__load_look_back": 60,
                        "driver__PowerMaxISCSI__load_look_back_real_time": 1,
                        "driver__PowerMaxISCSI__port_group_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__port_load_metric": "PercentBusy",
                        "driver__PowerMaxISCSI__powermax_array": "",
                        "driver__PowerMaxISCSI__powermax_array_tag_list__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_port_group_name_template": "portGroupName",
                        "driver__PowerMaxISCSI__powermax_port_groups__transform_csv": "",
                        "driver__PowerMaxISCSI__powermax_service_level": "",
                        "driver__PowerMaxISCSI__powermax_short_host_name_template": "shortHostName",
                        "driver__PowerMaxISCSI__powermax_snapvx_unlink_limit": 3,
                        "driver__PowerMaxISCSI__powermax_srp": "",
                        "driver__PowerMaxISCSI__retries": 200,
                        "driver__PowerMaxISCSI__san_ip": "",
                        "driver__PowerMaxISCSI__san_login": "admin",
                        "driver__PowerMaxISCSI__san_password": "",
                        "driver__PowerMaxISCSI__u4p_failover_autofailback": true,
                        "driver__PowerMaxISCSI__u4p_failover_backoff_factor": 1,
                        "driver__PowerMaxISCSI__u4p_failover_retries": 3,
                        "driver__PowerMaxISCSI__u4p_failover_target__transform_csv_kvs": "",
                        "driver__PowerMaxISCSI__u4p_failover_timeout": 20,
                        "driver__PowerMaxISCSI__use_chap_auth": false,
                        "driver__PowerMaxISCSI__vmax_workload": "",
                        "driver__PowerStore__powerstore_appliances__transform_csv": "",
                        "driver__PowerStore__powerstore_ports__transform_csv": "",
                        "driver__PureFC__driver_ssl_cert_path": "",
                        "driver__PureFC__driver_ssl_cert_verify": false,
                        "driver__PureFC__pure_api_token": "",
                        "driver__PureFC__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureFC__pure_eradicate_on_delete": false,
                        "driver__PureFC__pure_host_personality__transform_empty_none": "",
                        "driver__PureFC__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureFC__pure_replica_interval_default": 3600,
                        "driver__PureFC__pure_replica_retention_long_term_default": 7,
                        "driver__PureFC__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureFC__pure_replica_retention_short_term_default": 14400,
                        "driver__PureFC__pure_replication_pg_name": "cinder-group",
                        "driver__PureFC__pure_replication_pod_name": "cinder-pod",
                        "driver__PureFC__san_ip": "",
                        "driver__PureFC__use_chap_auth": false,
                        "driver__PureISCSI__driver_ssl_cert_path": "",
                        "driver__PureISCSI__driver_ssl_cert_verify": false,
                        "driver__PureISCSI__pure_api_token": "",
                        "driver__PureISCSI__pure_automatic_max_oversubscription_ratio": true,
                        "driver__PureISCSI__pure_eradicate_on_delete": false,
                        "driver__PureISCSI__pure_host_personality__transform_empty_none": "",
                        "driver__PureISCSI__pure_iscsi_cidr": "0.0.0.0/0",
                        "driver__PureISCSI__pure_replica_interval_default": 3600,
                        "driver__PureISCSI__pure_replica_retention_long_term_default": 7,
                        "driver__PureISCSI__pure_replica_retention_long_term_per_day_default": 3,
                        "driver__PureISCSI__pure_replica_retention_short_term_default": 14400,
                        "driver__PureISCSI__pure_replication_pg_name": "cinder-group",
                        "driver__PureISCSI__pure_replication_pod_name": "cinder-pod",
                        "driver__PureISCSI__san_ip": "",
                        "driver__PureISCSI__use_chap_auth": false,
                        "driver__QnapISCSI__chap_password": "",
                        "driver__QnapISCSI__chap_username": "",
                        "driver__QnapISCSI__driver_ssl_cert_verify": false,
                        "driver__QnapISCSI__qnap_management_url": "",
                        "driver__QnapISCSI__qnap_poolname": "",
                        "driver__QnapISCSI__qnap_storage_protocol": "iscsi",
                        "driver__QnapISCSI__san_login": "admin",
                        "driver__QnapISCSI__san_password": "",
                        "driver__QnapISCSI__target_ip_address": "$my_ip",
                        "driver__QnapISCSI__use_chap_auth": false,
                        "driver__Quobyte__quobyte_client_cfg": "",
                        "driver__Quobyte__quobyte_mount_point_base": "$state_path/mnt",
                        "driver__Quobyte__quobyte_overlay_volumes": false,
                        "driver__Quobyte__quobyte_qcow2_volumes": true,
                        "driver__Quobyte__quobyte_sparsed_volumes": true,
                        "driver__Quobyte__quobyte_volume_from_snapshot_cache": false,
                        "driver__Quobyte__quobyte_volume_url": "",
                        "driver__RBD__deferred_deletion_delay": 0,
                        "driver__RBD__deferred_deletion_purge_interval": 60,
                        "driver__RBD__enable_deferred_deletion": false,
                        "driver__RBD__rados_connect_timeout": -1,
                        "driver__RBD__rados_connection_interval": 5,
                        "driver__RBD__rados_connection_retries": 3,
                        "driver__RBD__rbd_ceph_conf": "",
                        "driver__RBD__rbd_cluster_name": "ceph",
                        "driver__RBD__rbd_exclusive_cinder_pool": false,
                        "driver__RBD__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBD__rbd_keyring_conf": "",
                        "driver__RBD__rbd_max_clone_depth": 5,
                        "driver__RBD__rbd_pool": "rbd",
                        "driver__RBD__rbd_secret_uuid": "",
                        "driver__RBD__rbd_store_chunk_size": 4,
                        "driver__RBD__rbd_user": "",
                        "driver__RBD__report_dynamic_total_capacity": true,
                        "driver__RBD__volume_dd_blocksize": "1M",
                        "driver__RBDISCSI__deferred_deletion_delay": 0,
                        "driver__RBDISCSI__deferred_deletion_purge_interval": 60,
                        "driver__RBDISCSI__driver_ssl_cert_verify": false,
                        "driver__RBDISCSI__enable_deferred_deletion": false,
                        "driver__RBDISCSI__rados_connect_timeout": -1,
                        "driver__RBDISCSI__rados_connection_interval": 5,
                        "driver__RBDISCSI__rados_connection_retries": 3,
                        "driver__RBDISCSI__rbd_ceph_conf": "",
                        "driver__RBDISCSI__rbd_cluster_name": "ceph",
                        "driver__RBDISCSI__rbd_exclusive_cinder_pool": false,
                        "driver__RBDISCSI__rbd_flatten_volume_from_snapshot": false,
                        "driver__RBDISCSI__rbd_iscsi_api_debug": false,
                        "driver__RBDISCSI__rbd_iscsi_api_password": "",
                        "driver__RBDISCSI__rbd_iscsi_api_url": "",
                        "driver__RBDISCSI__rbd_iscsi_api_user": "",
                        "driver__RBDISCSI__rbd_iscsi_target_iqn": "",
                        "driver__RBDISCSI__rbd_max_clone_depth": 5,
                        "driver__RBDISCSI__rbd_pool": "rbd",
                        "driver__RBDISCSI__rbd_secret_uuid": "",
                        "driver__RBDISCSI__rbd_store_chunk_size": 4,
                        "driver__RBDISCSI__rbd_user": "",
                        "driver__RBDISCSI__report_dynamic_total_capacity": true,
                        "driver__RBDISCSI__suppress_requests_ssl_warnings": false,
                        "driver__RBDISCSI__volume_dd_blocksize": "1M",
                        "driver__RSD__podm_password": "",
                        "driver__RSD__podm_url": "",
                        "driver__RSD__podm_username": "",
                        "driver__SCFC__dell_api_async_rest_timeout": 15,
                        "driver__SCFC__dell_api_sync_rest_timeout": 30,
                        "driver__SCFC__dell_sc_api_port": 3033,
                        "driver__SCFC__dell_sc_server_folder": "openstack",
                        "driver__SCFC__dell_sc_ssn": 64702,
                        "driver__SCFC__dell_sc_verify_cert": false,
                        "driver__SCFC__dell_sc_volume_folder": "openstack",
                        "driver__SCFC__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCFC__excluded_domain_ips__transform_csv": "",
                        "driver__SCFC__included_domain_ips__transform_csv": "",
                        "driver__SCFC__secondary_san_ip": "",
                        "driver__SCFC__secondary_san_login": "Admin",
                        "driver__SCFC__secondary_san_password": "",
                        "driver__SCFC__secondary_sc_api_port": 3033,
                        "driver__SCISCSI__dell_api_async_rest_timeout": 15,
                        "driver__SCISCSI__dell_api_sync_rest_timeout": 30,
                        "driver__SCISCSI__dell_sc_api_port": 3033,
                        "driver__SCISCSI__dell_sc_server_folder": "openstack",
                        "driver__SCISCSI__dell_sc_ssn": 64702,
                        "driver__SCISCSI__dell_sc_verify_cert": false,
                        "driver__SCISCSI__dell_sc_volume_folder": "openstack",
                        "driver__SCISCSI__dell_server_os": "Red Hat Linux 6.x",
                        "driver__SCISCSI__excluded_domain_ips__transform_csv": "",
                        "driver__SCISCSI__included_domain_ips__transform_csv": "",
                        "driver__SCISCSI__secondary_san_ip": "",
                        "driver__SCISCSI__secondary_san_login": "Admin",
                        "driver__SCISCSI__secondary_san_password": "",
                        "driver__SCISCSI__secondary_sc_api_port": 3033,
                        "driver__SdsISCSI__backend_availability_zone": "",
                        "driver__SdsISCSI__chap_password": "",
                        "driver__SdsISCSI__chap_username": "",
                        "driver__SdsISCSI__chiscsi_conf": "/etc/chelsio-iscsi/chiscsi.conf",
                        "driver__SdsISCSI__driver_client_cert": "",
                        "driver__SdsISCSI__driver_client_cert_key": "",
                        "driver__SdsISCSI__driver_data_namespace": "",
                        "driver__SdsISCSI__driver_ssl_cert_path": "",
                        "driver__SdsISCSI__driver_ssl_cert_verify": false,
                        "driver__SdsISCSI__driver_use_ssl": false,
                        "driver__SdsISCSI__filter_function": "",
                        "driver__SdsISCSI__goodness_function": "",
                        "driver__SdsISCSI__iscsi_iotype": "fileio",
                        "driver__SdsISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SdsISCSI__iscsi_target_flags": "",
                        "driver__SdsISCSI__iscsi_write_cache": "on",
                        "driver__SdsISCSI__num_shell_tries": 3,
                        "driver__SdsISCSI__num_volume_device_scan_tries": 3,
                        "driver__SdsISCSI__report_discard_supported": false,
                        "driver__SdsISCSI__storage_protocol": "iscsi",
                        "driver__SdsISCSI__target_helper": "tgtadm",
                        "driver__SdsISCSI__target_ip_address": "$my_ip",
                        "driver__SdsISCSI__target_port": 3260,
                        "driver__SdsISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SdsISCSI__target_protocol": "iscsi",
                        "driver__SdsISCSI__trace_flags__transform_csv": "",
                        "driver__SdsISCSI__use_chap_auth": false,
                        "driver__SdsISCSI__volume_backend_name": "",
                        "driver__SdsISCSI__volume_clear": "zero",
                        "driver__SdsISCSI__volume_clear_ionice": "",
                        "driver__SdsISCSI__volume_clear_size": 0,
                        "driver__SdsISCSI__volume_copy_blkio_cgroup_name": "cinder-volume-copy",
                        "driver__SdsISCSI__volume_copy_bps_limit": 0,
                        "driver__SdsISCSI__volume_dd_blocksize": "1M",
                        "driver__SdsISCSI__volumes_dir": "$state_path/volumes",
                        "driver__SolidFire__driver_ssl_cert_verify": false,
                        "driver__SolidFire__san_ip": "",
                        "driver__SolidFire__san_login": "admin",
                        "driver__SolidFire__san_password": "",
                        "driver__SolidFire__sf_account_prefix": "",
                        "driver__SolidFire__sf_allow_tenant_qos": false,
                        "driver__SolidFire__sf_api_port": 443,
                        "driver__SolidFire__sf_api_request_timeout": 30,
                        "driver__SolidFire__sf_cluster_pairing_timeout": 60,
                        "driver__SolidFire__sf_emulate_512": true,
                        "driver__SolidFire__sf_enable_vag": false,
                        "driver__SolidFire__sf_provisioning_calc": "maxProvisionedSpace",
                        "driver__SolidFire__sf_svip": "",
                        "driver__SolidFire__sf_volume_clone_timeout": 600,
                        "driver__SolidFire__sf_volume_create_timeout": 60,
                        "driver__SolidFire__sf_volume_pairing_timeout": 3600,
                        "driver__SolidFire__sf_volume_prefix": "UUID-",
                        "driver__StorPool__storpool_replication": 3,
                        "driver__StorPool__storpool_template": "",
                        "driver__StorwizeSVCFC__cycle_period_seconds": 300,
                        "driver__StorwizeSVCFC__storwize_peer_pool": "",
                        "driver__StorwizeSVCFC__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCFC__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCFC__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCFC__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCFC__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCFC__storwize_svc_multipath_enabled": false,
                        "driver__StorwizeSVCFC__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCFC__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCFC__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCFC__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCFC__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCFC__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCFC__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCFC__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCFC__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__StorwizeSVCISCSI__cycle_period_seconds": 300,
                        "driver__StorwizeSVCISCSI__storwize_peer_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_preferred_host_site__transform_csv_kvs": "",
                        "driver__StorwizeSVCISCSI__storwize_san_secondary_ip": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_allow_tenant_qos": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_rate": 50,
                        "driver__StorwizeSVCISCSI__storwize_svc_flashcopy_timeout": 120,
                        "driver__StorwizeSVCISCSI__storwize_svc_iscsi_chap_enabled": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_mirror_pool": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_retain_aux_volume": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_stretched_cluster_partner": "",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_autoexpand": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_compression": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_easytier": true,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_grainsize": 256,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_iogrp": "0",
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_nofmtdisk": false,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_rsize": 2,
                        "driver__StorwizeSVCISCSI__storwize_svc_vol_warning": 0,
                        "driver__StorwizeSVCISCSI__storwize_svc_volpool_name__transform_csv": "volpool",
                        "driver__SynoISCSI__chap_password": "",
                        "driver__SynoISCSI__chap_username": "",
                        "driver__SynoISCSI__driver_use_ssl": false,
                        "driver__SynoISCSI__iscsi_secondary_ip_addresses__transform_csv": "",
                        "driver__SynoISCSI__synology_admin_port": 5000,
                        "driver__SynoISCSI__synology_device_id": "",
                        "driver__SynoISCSI__synology_one_time_pass": "",
                        "driver__SynoISCSI__synology_password": "",
                        "driver__SynoISCSI__synology_pool_name": "",
                        "driver__SynoISCSI__synology_ssl_verify": true,
                        "driver__SynoISCSI__synology_username": "admin",
                        "driver__SynoISCSI__target_ip_address": "$my_ip",
                        "driver__SynoISCSI__target_port": 3260,
                        "driver__SynoISCSI__target_prefix": "iqn.2010-10.org.openstack:",
                        "driver__SynoISCSI__target_protocol": "iscsi",
                        "driver__SynoISCSI__use_chap_auth": false,
                        "driver__Unity__remove_empty_host": false,
                        "driver__Unity__unity_io_ports__transform_csv": "",
                        "driver__Unity__unity_storage_pool_names__transform_csv": "",
                        "driver__VNX__default_timeout": 31536000,
                        "driver__VNX__destroy_empty_storage_group": false,
                        "driver__VNX__force_delete_lun_in_storagegroup": true,
                        "driver__VNX__ignore_pool_full_threshold": false,
                        "driver__VNX__initiator_auto_deregistration": false,
                        "driver__VNX__initiator_auto_registration": false,
                        "driver__VNX__io_port_list__transform_csv": "",
                        "driver__VNX__iscsi_initiators": "",
                        "driver__VNX__max_luns_per_storage_group": 255,
                        "driver__VNX__naviseccli_path": "",
                        "driver__VNX__storage_vnx_authentication_type": "global",
                        "driver__VNX__storage_vnx_pool_names__transform_csv": "",
                        "driver__VNX__storage_vnx_security_file_dir": "",
                        "driver__VNX__vnx_async_migrate": true,
                        "driver__VZStorage__vzstorage_default_volume_format": "raw",
                        "driver__VZStorage__vzstorage_mount_options__transform_csv": "",
                        "driver__VZStorage__vzstorage_mount_point_base": "$state_path/mnt",
                        "driver__VZStorage__vzstorage_shares_config": "/etc/cinder/vzstorage_shares",
                        "driver__VZStorage__vzstorage_sparsed_volumes": true,
                        "driver__VZStorage__vzstorage_used_ratio__transform_string_float": "0.95",
                        "driver__WindowsISCSI__windows_iscsi_lun_path": "C:\\iSCSIVirtualDisks",
                        "driver__WindowsSmbfs__smbfs_default_volume_format": "vhd",
                        "driver__WindowsSmbfs__smbfs_mount_point_base": "C:\\OpenStack\\_mnt",
                        "driver__WindowsSmbfs__smbfs_pool_mappings__transform_csv_kvs": "",
                        "driver__WindowsSmbfs__smbfs_shares_config": "C:\\OpenStack\\smbfs_shares.txt",
                        "driver__XtremIOFC__driver_ssl_cert_path": "",
                        "driver__XtremIOFC__driver_ssl_cert_verify": false,
                        "driver__XtremIOFC__san_ip": "",
                        "driver__XtremIOFC__san_login": "admin",
                        "driver__XtremIOFC__san_password": "",
                        "driver__XtremIOFC__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOFC__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOFC__xtremio_clean_unused_ig": false,
                        "driver__XtremIOFC__xtremio_cluster_name": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_path": "",
                        "driver__XtremIOISCSI__driver_ssl_cert_verify": false,
                        "driver__XtremIOISCSI__san_ip": "",
                        "driver__XtremIOISCSI__san_login": "admin",
                        "driver__XtremIOISCSI__san_password": "",
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_count": 5,
                        "driver__XtremIOISCSI__xtremio_array_busy_retry_interval": 5,
                        "driver__XtremIOISCSI__xtremio_clean_unused_ig": false,
                        "driver__XtremIOISCSI__xtremio_cluster_name": "",
                        "driver__ZadaraVPSAISCSI__zadara_access_key": "",
                        "driver__ZadaraVPSAISCSI__zadara_default_snap_policy": false,
                        "driver__ZadaraVPSAISCSI__zadara_ssl_cert_verify": true,
                        "driver__ZadaraVPSAISCSI__zadara_use_iser": true,
                        "driver__ZadaraVPSAISCSI__zadara_vol_encrypt": false,
                        "driver__ZadaraVPSAISCSI__zadara_vol_name_template": "OS_%s",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_host": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_poolname": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_port": "",
                        "driver__ZadaraVPSAISCSI__zadara_vpsa_use_ssl": false
                      }
                    },
                    "sysFiles": {
                      "name": ""
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "false"
          containerImage: quay.io/embercsi/ember-csi-operator:latest
          createdAt: 2021-02-22:17:47:00
          description: Multi-vendor CSI plugin supporting 80+ storage drivers
          repository: https://github.com/embercsi/ember-csi-operator
          support: http://readthedocs.org/projects/ember-csi/
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an Ember CSI storage backend
            displayName: Storage Backends
            kind: EmberStorageBackend
            name: embercsis.ember-csi.io
            version: v1alpha1
        description: |
          Ember-CSI is a multi-vendor CSI plugin driver supporting
          over 80 storage drivers in a single plugin to provide block and mount
          storage to container orchestration systems.

          The operator adds support for management of multiple Ember CSI instances
          within a cluster and simplifies storage backend deployments.

          ### Supported drivers
          Ember-CSI includes a good number of storage drivers, but due to limitation on hardware availability only a small number of them have been validated at one point or another. In alphabetical order they are:

          * HPE3PARFC
          * HPE3PARISCSI
          * KaminarioISCSI
          * LVMVolume
          * PowerMaxFC
          * PowerMaxISCSI
          * PureFC
          * PureISCSI
          * QnapISCSI
          * RBD
          * SolidFire
          * SynoISCSI
          * XtremIOFC
          * XtremIOISCSI

          ### Getting started
          Please have a look at our quick start documentation on how to get started:
          [github.com/embercsi/ember-csi-operator#quick-start](https://github.com/embercsi/ember-csi-operator#quick-start).
          We recommend to use the user friendly "Edit Form" when creating a new
          EmberStorageBackend. You'll find the link in the upper right corner of the
          editor.

          ### Documentation
          Documentation on Ember CSI itself can be found at [docs.ember-csi.io](http://docs.ember-csi.io/).

          ### Getting help
          Please contact us if you encounter any issues while using Ember CSI.
          Get in touch through the #ember-csi channel on Freenode, also accessible through
          [Kiwi IRC](https://kiwiirc.com/nextclient/irc.freenode.net/ember-csi).

          You can also create an issue on the [Github project](https://github.com/embercsi/ember-csi-operator)
          for bugs, enhancements, or other requests.
        displayName: Ember CSI Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Ember-CSI
        - CSI
        links:
        - name: Learn more about the project
          url: http://ember-csi.io/
        - name: Documentation
          url: http://readthedocs.org/projects/ember-csi/
        - name: Ember-CSI Source Code
          url: https://github.com/embercsi/ember-csi
        - name: Ember-CSI Operator Source Code
          url: https://github.com/embercsi/ember-csi-operator
        maintainers:
        - email: geguileo@redhat.com
          name: Gorka Eguileor
        - email: cschwede@redhat.com
          name: Christian Schwede
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/ember-csi-community-operator@sha256:5dce9b3018cfb300cd651c7980353cedc6617383d5c525e6a4a39ffb8e105774
        - quay.io/embercsi/ember-csi-operator:latest
        version: 0.9.9
      entries:
      - name: ember-csi-community-operator.v0.9.9
        version: 0.9.9
      - name: ember-csi-community-operator.v0.9.8
        version: 0.9.8
      - name: ember-csi-community-operator.v0.9.7
        version: 0.9.7
      name: beta
    defaultChannel: beta
    packageName: ember-csi-community-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: samsung
      provider-url: ""
    name: socmmd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: socmmd.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "CMMD",
                "metadata": {
                  "name": "my-app",
                  "namespace": "my-namespace"
                },
                "spec": {
                  "allocateMode": "auto",
                  "enable": true,
                  "payload": {
                    "apiVersion": "apps/v1",
                    "kind": "Deployment",
                    "metadata": {
                      "name": "my-app"
                    },
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "name": "my-app"
                        }
                      },
                      "template": {
                        "metadata": {
                          "labels": {
                            "name": "my-app"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "quay.io/my-account/my-app:latest",
                              "name": "my-app-01"
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "cmmd.samsung.com/v1",
                "kind": "Label",
                "metadata": {
                  "labels": {
                    "machineconfiguration.openshift.io/role": "cmmd"
                  },
                  "name": "cmmd-node-label",
                  "namespace": "socmmd"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          console.openshift.io/plugins: '["socmmd-console-plugin"]'
          containerImage: quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
          createdAt: "2024-08-01T05:45:38Z"
          description: Samsung Operator For CMM-D
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "cmmd.samsung.com/v1",
              "kind": "Label",
              "metadata": {
                "name": "cmmd-node-label",
                "namespace": "socmmd",
                "labels": {}
              },
              "spec": {}
            }
          operatorframework.io/suggested-namespace: socmmd-system
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
          support: samsung
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CMMD is the Schema for the cmmds API.
            displayName: CMMD
            kind: CMMD
            name: cmmds.cmmd.samsung.com
            version: v1
          - description: Label is the Schema for the labels API.
            displayName: Node Label
            kind: Label
            name: labels.cmmd.samsung.com
            version: v1
        description: Samsung Operator For CMM-D
        displayName: socmmd
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - samsung
        - memory
        - extend
        - cmm-d
        links:
        - name: Samsung Operator For CMM-D
          url: https://github.com/OpenMPDK/SMDK/tree/main/lib/redhat/socmmd
        maintainers:
        - email: junhyeok.im@samsung.com
          name: junhyeok-im
        maturity: stable
        provider:
          name: samsung
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/socmmd@sha256:5adf9706daf49b35cfcf74b6fb07b056f87714b87047803b797a56f138c77486
        - quay.io/samsung_sag/socmmd@sha256:2ef361fb15f0fec99e0a8f1ef83ad4dd14f701d579c900cce084c4a62cf805f8
        version: 0.0.2
      entries:
      - name: socmmd.v0.0.2
        version: 0.0.2
      - name: socmmd.v0.0.1
        version: 0.0.1
      name: stable
    defaultChannel: stable
    packageName: socmmd
    provider:
      name: samsung
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: must-gather-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: must-gather-operator.v1.1.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "MustGather",
                "metadata": {
                  "name": "example-mustgather"
                },
                "spec": {
                  "caseID": "02527285",
                  "caseManagementAccountSecretRef": {
                    "name": "case-management-creds"
                  },
                  "serviceAccountRef": {
                    "name": "must-gather-admin"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security, Monitoring
          certified: "false"
          containerImage: quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
          createdAt: "2021-12-02T17:13:53Z"
          description: This operator provides a facility to easily upload must-gather
            reports to a Red Hat case.
          operatorframework.io/suggested-namespace: global-load-balancer-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/must-gather-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: MustGather
            name: mustgathers.redhatcop.redhat.io
            version: v1alpha1
        description: |-
          The must gather operator helps collecting must gather information on a cluster and uploading it to a case.
          To use the operator a cluster administrator can create the following must gather CR:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: example-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
          ```

          this request will collect the standard must gather info and upload it to case `#02527285` using the credentials found in the `caseManagementCreds` secret.

          A more complex example:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: MustGather
          metadata:
            name: full-mustgather
          spec:
            caseID: '02527285'
            caseManagementAccountSecretRef:
              name: case-management-creds
            serviceAccountRef:
              name: must-gather-admin
            mustGatherImages:
            - quay.io/kubevirt/must-gather:latest
            - quay.io/ocs-dev/ocs-must-gather
          ```

          in this example we are using a specific service account (which must have cluster admin permissions as per must-gather requirements) and we are specifying a couple of additional must gather images to be run for the `kubevirt` and `ocs` subsystem. If not specified serviceAccountRef.Name will default to `default`. Also the standard must gather image: `quay.io/openshift/origin-must-gather:latest` is always added by default.
        displayName: Must Gather Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - case_management
        - support
        - must_gather
        links:
        - name: Must Gather Operator
          url: https://github.com/redhat-cop/must-gather-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/must-gather-operator
        maintainers:
        - email: rspazzol@redhat.com'
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/redhat-cop/must-gather-operator@sha256:aebef122cdb8409f5a5d70d3cdc585e02272704d57e103553a84bc6fc238e646
        - quay.io/openshift-community-operators/must-gather-operator@sha256:a71e5e25aaa92f03334776fee87316809bf2a46dfb34c497d63b228a194982f2
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.1.2
      entries:
      - name: must-gather-operator.v1.1.2
        version: 1.1.2
      - name: must-gather-operator.v1.1.1
        version: 1.1.1
      - name: must-gather-operator.v1.1.0
        version: 1.1.0
      - name: must-gather-operator.v1.0.6
        version: 1.0.6
      - name: must-gather-operator.v1.0.5
        version: 1.0.5
      - name: must-gather-operator.v1.0.4
        version: 1.0.4
      - name: must-gather-operator.v1.0.3
        version: 1.0.3
      - name: must-gather-operator.v1.0.2
        version: 1.0.2
      - name: must-gather-operator.v1.0.1
        version: 1.0.1
      - name: must-gather-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: must-gather-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: nbde-tang-server
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nbde-tang-server.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          containerImage: registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
          createdAt: "2025-04-24T18:06:07Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/openshift/nbde-tang-server
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tang Server allows definition of the NBDE Tang Server deployment
            displayName: Tang Server
            kind: TangServer
            name: nbde.openshift.io
            version: v1alpha1
        description: NBDE Tang Server operator provides Network-Bound Disk Encryption
          in OpenShift
        displayName: NBDE Tang Server
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - NBDE
        - PBD
        - Installation
        - Network
        links:
        - name: NBDE Tang Server
          url: https://github.com/openshift/nbde-tang-server
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:c15928ecc2f0ff370f567a2ca909b7a71e2becaff0af649275396e59232b7b8c
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-operator-bundle@sha256:79bf610b144cbdf4404a1153f1c395179ce1743274eec7102c862aae4f45890a
        - registry.redhat.io/nbde-tang-server/nbde-tang-server-rhel9-operator@sha256:cdfb97d26f9dfda6e5c9a5955ebac560a87348dbd7f49b041204903f52c542b2
        version: 1.1.1
      entries:
      - name: nbde-tang-server.v1.1.1
        version: 1.1.1
      - name: nbde-tang-server.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: nbde-tang-server
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-prometheus-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-prometheus-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:6f0e07eb583201703f50ce9b3bf47e7aa63630074afcb588902caf05ff3ba2bf
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:5959c0c359fb2c47a66a10bc89e6600ed5e6b5c2ea881357e5f3fa5ac1836cbd
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:35fae43ffc92770e13919f081195b35d8dd293f223fb9e577d75ae53f156305a
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:e135baa6b7eefcaf70780b6833acb689c5aef076b6280ff29c02949fbb33b06e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:6a447f455b8a3a3a8390787b36fe1d59a7d8bbf834f66a63fb0023f0557796e4
        version: 4.17.13-rhodf
      entries:
      - name: odf-prometheus-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-prometheus-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-prometheus-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-prometheus-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-prometheus-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-prometheus-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-prometheus-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-prometheus-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-prometheus-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-prometheus-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-prometheus-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-prometheus-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-prometheus-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-prometheus-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-prometheus-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
          createdAt: "2024-03-01T23:43:00Z"
          description: Manage the full lifecycle of configuring and managing Prometheus
            and Alertmanager servers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/operator-type: non-standalone
          repository: https://github.com/prometheus-operator/prometheus-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          The Prometheus Operator for Kubernetes provides easy monitoring definitions for Kubernetes services and deployment and management of Prometheus instances.

          Once installed, the Prometheus Operator provides the following features:

          * **Create/Destroy**: Easily launch a Prometheus instance for your Kubernetes namespace, a specific application or team easily using the Operator.
          * **Simple Configuration**: Configure the fundamentals of Prometheus like versions, persistence, retention policies, and replicas from a native Kubernetes resource.
          * **Target Services via Labels**: Automatically generate monitoring target configurations based on familiar Kubernetes label queries; no need to learn a Prometheus specific configuration language.

          ### Other Supported Features

          **High availability**

          Multiple instances are run across failure zones and data is replicated. This keeps your monitoring available during an outage, when you need it most.

          **Updates via automated operations**

          New Prometheus versions are deployed using a rolling update with no downtime, making it easy to stay up to date.

          **Handles the dynamic nature of containers**

          Alerting rules are attached to groups of containers instead of individual instances, which is ideal for the highly dynamic nature of container deployment.
        displayName: Prometheus Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - prometheus
        - monitoring
        - tsdb
        - alerting
        - observability
        links:
        - name: Prometheus
          url: https://www.prometheus.io/
        - name: Documentation
          url: https://github.com/prometheus-operator/prometheus-operator/tree/master/Documentation
        - name: Prometheus Operator
          url: https://github.com/prometheus-operator/prometheus-operator
        maintainers:
        - email: openshift-operators@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.16.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9-operator@sha256:e24bbaab46c98fe940b3138cefcfd3fda69380920ca6f6d7f56bb5dd93373f71
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        - registry.redhat.io/odf4/odf-prometheus-operator-bundle@sha256:64c4cc3155e8ea6630bc7fecd392a41d71193cfb20f832b6334cd4795cf0735e
        - registry.redhat.io/openshift4/ose-prometheus-alertmanager-rhel9@sha256:7916e81066baac027fb575a1edaab3c98a00505c73c4ac3c2530c111c11f0030
        - registry.redhat.io/openshift4/ose-prometheus-config-reloader-rhel9@sha256:4352d0aecaac2efc3a15bf87925c08f7f85ec7baac9eebc29787f05af047514e
        version: 4.18.11-rhodf
      entries:
      - name: odf-prometheus-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-prometheus-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-prometheus-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-prometheus-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-prometheus-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-prometheus-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-prometheus-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-prometheus-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-prometheus-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-prometheus-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-prometheus-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-prometheus-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-prometheus-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: security-profiles-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: security-profiles-operator.v0.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "AppArmorProfile",
                "metadata": {
                  "annotations": {
                    "description": "Block writing to any files in the disk."
                  },
                  "name": "test-profile"
                },
                "spec": {
                  "abstract": {
                    "filesystem": {
                      "readOnlyPaths": [],
                      "readWritePaths": [],
                      "writeOnlyPaths": []
                    }
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "SecurityProfilesOperatorDaemon",
                "metadata": {
                  "name": "spod",
                  "namespace": "security-profiles-operator"
                },
                "spec": {
                  "enableAppArmor": false,
                  "enableLogEnricher": false,
                  "enableSelinux": false
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileRecording",
                "metadata": {
                  "name": "test-recording"
                },
                "spec": {
                  "kind": "SeccompProfile",
                  "podSelector": {
                    "matchLabels": {
                      "app": "alpine"
                    }
                  },
                  "recorder": "bpf"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha1",
                "kind": "ProfileBinding",
                "metadata": {
                  "name": "profile-binding"
                },
                "spec": {
                  "image": "nginx:1.19.1",
                  "profileRef": {
                    "kind": "SeccompProfile",
                    "name": "profile-allow-unsafe"
                  }
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "RawSelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "policy": "(blockinherit container)\n(allow process var_log_t ( dir ( open read getattr lock search ioctl add_name remove_name write ))) \n(allow process var_log_t ( file ( getattr read write append ioctl lock map open create  ))) \n(allow process var_log_t ( sock_file ( getattr read write append open  ))) \n"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Blocks all syscalls."
                  },
                  "name": "profile-block-all"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ERRNO"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls whilst logging their use. Similar to running as unconfined in terms of enforcement."
                  },
                  "name": "profile-complain-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_LOG"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "UNSAFE: Allows all syscalls. Similar to running as unconfined as it provides no enforcement."
                  },
                  "name": "profile-allow-unsafe"
                },
                "spec": {
                  "defaultAction": "SCMP_ACT_ALLOW"
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1beta1",
                "kind": "SeccompProfile",
                "metadata": {
                  "annotations": {
                    "description": "Enables complain mode whilst blocking high-risk syscalls. Some essential syscalls are allowed to decrease log noise."
                  },
                  "name": "profile-complain-block-high-risk"
                },
                "spec": {
                  "architectures": [
                    "SCMP_ARCH_X86_64",
                    "SCMP_ARCH_PPC64LE"
                  ],
                  "defaultAction": "SCMP_ACT_LOG",
                  "syscalls": [
                    {
                      "action": "SCMP_ACT_ALLOW",
                      "names": [
                        "exit",
                        "exit_group",
                        "futex",
                        "nanosleep"
                      ]
                    },
                    {
                      "action": "SCMP_ACT_ERRNO",
                      "names": [
                        "acct",
                        "add_key",
                        "bpf",
                        "clock_adjtime",
                        "clock_settime",
                        "create_module",
                        "delete_module",
                        "finit_module",
                        "get_kernel_syms",
                        "get_mempolicy",
                        "init_module",
                        "ioperm",
                        "iopl",
                        "kcmp",
                        "kexec_file_load",
                        "kexec_load",
                        "keyctl",
                        "lookup_dcookie",
                        "mbind",
                        "mount",
                        "move_pages",
                        "name_to_handle_at",
                        "nfsservctl",
                        "open_by_handle_at",
                        "perf_event_open",
                        "personality",
                        "pivot_root",
                        "process_vm_readv",
                        "process_vm_writev",
                        "ptrace",
                        "query_module",
                        "quotactl",
                        "reboot",
                        "request_key",
                        "set_mempolicy",
                        "setns",
                        "settimeofday",
                        "stime",
                        "swapoff",
                        "swapon",
                        "_sysctl",
                        "sysfs",
                        "umount2",
                        "umount",
                        "unshare",
                        "uselib",
                        "userfaultfd",
                        "ustat",
                        "vm86old",
                        "vm86"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "security-profiles-operator.x-k8s.io/v1alpha2",
                "kind": "SelinuxProfile",
                "metadata": {
                  "name": "errorlogger"
                },
                "spec": {
                  "allow": {
                    "var_log_t": {
                      "dir": [
                        "open",
                        "read",
                        "getattr",
                        "lock",
                        "search",
                        "ioctl",
                        "add_name",
                        "remove_name",
                        "write"
                      ],
                      "file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "ioctl",
                        "lock",
                        "map",
                        "open",
                        "create"
                      ],
                      "sock_file": [
                        "getattr",
                        "read",
                        "write",
                        "append",
                        "open"
                      ]
                    }
                  },
                  "inherit": [
                    {
                      "name": "container"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          createdAt: "2025-02-25T13:34:54Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.1 <0.9.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-security-profiles
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ProfileBinding is the Schema for the profilebindings API.
            displayName: Profile Binding
            kind: ProfileBinding
            name: profilebindings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: ProfileRecording is the Schema for the profilerecordings
              API.
            displayName: Profile Recording
            kind: ProfileRecording
            name: profilerecordings.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SeccompProfile is a cluster level specification for a seccomp
              profile. See https://github.com/opencontainers/runtime-spec/blob/master/config-linux.md#seccomp
            displayName: Seccomp Profile
            kind: SeccompProfile
            name: seccompprofiles.security-profiles-operator.x-k8s.io
            version: v1beta1
          - description: SecurityProfileNodeStatus is a per-node status of a security
              profile
            displayName: Security Profile Node Status
            kind: SecurityProfileNodeStatus
            name: securityprofilenodestatuses.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - description: SecurityProfilesOperatorDaemon is the Schema to configure
              the spod deployment.
            displayName: Security Profiles Operator Daemon
            kind: SecurityProfilesOperatorDaemon
            name: securityprofilesoperatordaemons.security-profiles-operator.x-k8s.io
            version: v1alpha1
          - kind: RawSelinuxProfile
            name: rawselinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: SelinuxProfile is the Schema for the selinuxprofiles API.
            displayName: Selinux Profile
            kind: SelinuxProfile
            name: selinuxprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha2
          - description: AppArmorProfile is a cluster level specification for an AppArmor
              profile.
            displayName: App Armor Profile
            kind: AppArmorProfile
            name: apparmorprofiles.security-profiles-operator.x-k8s.io
            version: v1alpha1
        description: SPO is an operator which aims to make it easier for users to
          use SELinux, seccomp and AppArmor in Kubernetes clusters
        displayName: Security Profiles Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - selinux
        - seccomp
        - apparmor
        - ebpf
        links:
        - name: Security Profiles Operator
          url: https://github.com/openshift/security-profiles-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-security-profiles-operator-bundle@sha256:bd5bdc7085f3e9ce8494488f29b4d83a29796c47d536106cd6e14a7b67ab07ee
        - registry.redhat.io/compliance/openshift-security-profiles-rhel8-operator@sha256:828274d71c65ef820e6a8b652e3f102624d09a0609fb00bae7b32dc62ff37fd1
        - registry.redhat.io/compliance/openshift-selinuxd-rhel8@sha256:81273ff149482109417064b40d333edb131481c05547e4ffa6485475460a48ba
        - registry.redhat.io/compliance/openshift-selinuxd-rhel9@sha256:3278d69814a8d0d1e31dcdef454c24828aadbb468a3d5d3e15910c76481765bd
        version: 0.9.0
      entries:
      - name: security-profiles-operator.v0.9.0
        version: 0.9.0
      - name: security-profiles-operator.v0.8.6
        version: 0.8.6
      - name: security-profiles-operator.v0.8.5
        version: 0.8.5
      - name: security-profiles-operator.v0.8.4
        version: 0.8.4
      - name: security-profiles-operator.v0.8.2
        version: 0.8.2
      name: release-alpha-rhel-8
    defaultChannel: release-alpha-rhel-8
    packageName: security-profiles-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: 3scale-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: 3scale-operator.v0.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                },
                "status": {
                  "deployments": {
                    "ready": [
                      "deployment x ready"
                    ]
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                },
                "status": {
                  "adminId": 1,
                  "tenantId": 2
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ApplicationAuth",
                "metadata": {
                  "name": "applicationauth-sample"
                },
                "spec": {
                  "applicationCRName": "application-cr-name",
                  "authSecretRef": {
                    "name": "auth-secret-reference"
                  },
                  "generateSecret": true
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                },
                "status": {}
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "productCRName": "product1-sample",
                  "production": true
                },
                "status": {}
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=0.11.0 <0.12.5'
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          rht_backend_redis_requirements: 6.2.0
          rht_mysql_requirements: 8.0.0
          rht_postgres_requirements: 10.0.0
          rht_system_redis_requirements: 6.2.0
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: ApplicationAuth is the Schema for the applicationauths API
            displayName: Application Auth
            kind: ApplicationAuth
            name: applicationauths.capabilities.3scale.net
            version: v1beta1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-amp2/manticore-rhel9@sha256:ccf6baf5f2e9494a34739a79abe0f46e36d8ab9dc81c426f7a3a8964e8f39859
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/mysql-80@sha256:ee9aca9096473da004f9140305742eb7b9a77ed81ce55d5a6503388cccab030f
        - registry.redhat.io/3scale-amp2/zync-rhel9@sha256:df9d62897ff286165772150a8b5e6952d2ce3b52a8afa07bd693a2d95c2c67ff
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/3scale-amp2/3scale-operator-bundle@sha256:934fd9123d1da778b9f8375854f2a61167f0efbc76371edfdc9962f7626e694a
        - registry.redhat.io/3scale-amp2/backend-rhel8@sha256:eb2288de900dc42ead755b28cdb2509b25ecc8dd256781beaa4a2a86e7cbe532
        - registry.redhat.io/3scale-amp2/3scale-rhel9-operator@sha256:fcea6a77d0a48f35f4c3ce63e72dde104aac6b580f340950b311eb418fff5ec9
        - registry.redhat.io/3scale-amp2/apicast-gateway-rhel8@sha256:cd1c4e0a2259d244eee0c552755d80a4dc751aa3cb0d3a69a5342c3135a0216a
        - registry.redhat.io/rhel9/memcached@sha256:8738110ef16c9e35deeb07c0577ffe09d6ef029761b504a8d20253fa715da394
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:a95d09fc3b224f550dec3de3d23fd2dbfc0a220dc869b4ad9559ee2f85327daa
        - registry.redhat.io/3scale-amp2/system-rhel8@sha256:4c72dfeb1919a1de4fbc47f4611d0c830ef89cf9f17255ad55a9b370618e0e8f
        version: 0.12.5
      entries:
      - name: 3scale-operator.v0.12.5
        version: 0.12.5
      - name: 3scale-operator.v0.12.4
        version: 0.12.4
      - name: 3scale-operator.v0.12.3
        version: 0.12.3
      name: threescale-2.15
    - currentCSV: 3scale-operator.v0.11.8-mas
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManager",
                "metadata": {
                  "name": "apimanager-sample"
                },
                "spec": {
                  "wildcardDomain": "example.com"
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerBackup",
                "metadata": {
                  "name": "apimanagerbackup-sample"
                },
                "spec": {
                  "backupDestination": {
                    "persistentVolumeClaim": {
                      "volumeName": "mypreexistingvolume"
                    }
                  }
                }
              },
              {
                "apiVersion": "apps.3scale.net/v1alpha1",
                "kind": "APIManagerRestore",
                "metadata": {
                  "name": "apimanagerrestore-sample"
                },
                "spec": {
                  "restoreSource": {
                    "persistentVolumeClaim": {
                      "claimSource": {
                        "claimName": "mybackeduppvc"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1alpha1",
                "kind": "Tenant",
                "metadata": {
                  "name": "tenant-sample"
                },
                "spec": {
                  "email": "admin@example.com",
                  "masterCredentialsRef": {
                    "name": "system-seed"
                  },
                  "organizationName": "Example.com",
                  "passwordCredentialsRef": {
                    "name": "ecorp-admin-secret"
                  },
                  "systemMasterUrl": "https://master.example.com",
                  "tenantSecretRef": {
                    "name": "ecorp-tenant-secret",
                    "namespace": "operator-test"
                  },
                  "username": "admin"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ActiveDoc",
                "metadata": {
                  "name": "activedoc-from-url"
                },
                "spec": {
                  "activeDocOpenAPIRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.json"
                  },
                  "name": "Operated ActiveDoc From URL"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Application",
                "metadata": {
                  "name": "application-sample"
                },
                "spec": {
                  "accountCR": {
                    "name": "developeraccount-sample"
                  },
                  "applicationPlanName": "plan01",
                  "description": "testing application ",
                  "name": "testApp",
                  "productCR": {
                    "name": "product-sample"
                  },
                  "suspend": false
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Backend",
                "metadata": {
                  "name": "backend1-sample"
                },
                "spec": {
                  "name": "Operated Backend 1",
                  "privateBaseURL": "https://api.example.com",
                  "systemName": "backend1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "CustomPolicyDefinition",
                "metadata": {
                  "name": "custompolicydefinition-sample"
                },
                "spec": {
                  "name": "MyCustomPolicy",
                  "schema": {
                    "$schema": "http://json-schema.org/draft-07/schema#",
                    "configuration": {
                      "properties": {
                        "someAttr": {
                          "description": "Some attribute",
                          "type": "integer"
                        }
                      },
                      "type": "object"
                    },
                    "name": "MyCustomPolicy",
                    "summary": "some summary",
                    "version": "0.0.1"
                  },
                  "version": "0.0.1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperAccount",
                "metadata": {
                  "name": "developeraccount-simple-sample"
                },
                "spec": {
                  "orgName": "Ecorp"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "DeveloperUser",
                "metadata": {
                  "name": "developeruser-admin-sample"
                },
                "spec": {
                  "developerAccountRef": {
                    "name": "developeraccount1"
                  },
                  "email": "myusername2@example.com",
                  "passwordCredentialsRef": {
                    "name": "mysecret"
                  },
                  "role": "admin",
                  "username": "myusername2"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "OpenAPI",
                "metadata": {
                  "name": "openapi-from-url"
                },
                "spec": {
                  "openapiRef": {
                    "url": "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/master/examples/v3.0/petstore.yaml"
                  }
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "Product",
                "metadata": {
                  "name": "product1-sample"
                },
                "spec": {
                  "name": "OperatedProduct 1"
                }
              },
              {
                "apiVersion": "capabilities.3scale.net/v1beta1",
                "kind": "ProxyConfigPromote",
                "metadata": {
                  "name": "proxyconfigpromote-sample"
                },
                "spec": {
                  "ProductCRName": "product1-sample",
                  "Production": true
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
          createdAt: "2019-05-30T22:40:00Z"
          description: 3scale Operator to provision 3scale and publish/manage API
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Integration", "Red
            Hat 3scale API Management"]'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/3scale/3scale-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ActiveDoc is the Schema for the activedocs API
            displayName: Active Doc
            kind: ActiveDoc
            name: activedocs.capabilities.3scale.net
            version: v1beta1
          - description: APIManagerBackup represents an APIManager backup
            displayName: APIManagerBackup
            kind: APIManagerBackup
            name: apimanagerbackups.apps.3scale.net
            version: v1alpha1
          - description: APIManagerRestore represents an APIManager restore
            displayName: APIManagerRestore
            kind: APIManagerRestore
            name: apimanagerrestores.apps.3scale.net
            version: v1alpha1
          - description: APIManager is the Schema for the apimanagers API
            displayName: APIManager
            kind: APIManager
            name: apimanagers.apps.3scale.net
            version: v1alpha1
          - description: Application is the Schema for the applications API
            displayName: Application
            kind: Application
            name: applications.capabilities.3scale.net
            version: v1beta1
          - description: Backend is the Schema for the backends API
            displayName: 3scale Backend
            kind: Backend
            name: backends.capabilities.3scale.net
            version: v1beta1
          - description: CustomPolicyDefinition is the Schema for the custompolicydefinitions
              API
            displayName: Custom Policy Definition
            kind: CustomPolicyDefinition
            name: custompolicydefinitions.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperAccount is the Schema for the developeraccounts
              API
            displayName: Developer Account
            kind: DeveloperAccount
            name: developeraccounts.capabilities.3scale.net
            version: v1beta1
          - description: DeveloperUser is the Schema for the developerusers API
            displayName: Developer User
            kind: DeveloperUser
            name: developerusers.capabilities.3scale.net
            version: v1beta1
          - description: OpenAPI is the Schema for the openapis API
            displayName: Open API
            kind: OpenAPI
            name: openapis.capabilities.3scale.net
            version: v1beta1
          - description: Product is the Schema for the products API
            displayName: 3scale Product
            kind: Product
            name: products.capabilities.3scale.net
            version: v1beta1
          - description: ProxyConfigPromote is the Schema for the proxyconfigpromotes
              API
            displayName: Proxy Config Promote
            kind: ProxyConfigPromote
            name: proxyconfigpromotes.capabilities.3scale.net
            version: v1beta1
          - description: Tenant is the Schema for the tenants API
            displayName: Tenant
            kind: Tenant
            name: tenants.capabilities.3scale.net
            version: v1alpha1
        description: |-
          The 3scale Operator creates and maintains the Red Hat 3scale API Management on [OpenShift](https://www.openshift.com/) in various deployment configurations.

          [3scale API Management](https://www.redhat.com/en/technologies/jboss-middleware/3scale) makes it easy to manage your APIs.
          Share, secure, distribute, control, and monetize your APIs on an infrastructure platform built for performance, customer control, and future growth.

          ### Supported Features
          * **Installer** A way to install a 3scale API Management solution, providing configurability options at the time of installation
          * **Upgrade** Upgrade from previously installed 3scale API Management solution
          * **Reconcilliation** Tunable CRD parameters after 3scale API Management solution is installed
          * **Capabilities** Ability to define 3scale API definitions and set them into a 3scale API Management solution

          ### Documentation
          [3scale api management](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management)
          [Deploying 3scale using the operator](https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.11/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator)

          ### Getting help
          If you encounter any issues while using 3scale operator, you can create an issue on our [Github repo](https://github.com/3scale/3scale-operator) for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find using 3scale Operator
          * Fixing issues by opening [Pull Requests](https://github.com/3scale/3scale-operator/pulls)
          * Talking about 3scale Operator

          All bugs, tasks or enhancements are tracked as [GitHub issues](https://github.com/3scale/3scale-operator/issues).

          ### License
          3scale Operator is licensed under the [Apache 2.0 license](https://github.com/3scale/3scale-operator/blob/master/LICENSE)
        displayName: Red Hat Integration - 3scale
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - 3scale
        - API
        links:
        - name: GitHub
          url: https://github.com/3scale/3scale-operator
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_3scale_api_management/2.12/html/installing_3scale/install-threescale-on-openshift-guide#deploying-threescale-using-the-operator
        maintainers:
        - email: eastizle+3scaleoperator@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mstoklus@redhat.com
          name: Michal Stokluska
        - email: aucunnin@redhat.com
          name: Austin Cunningham
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/3scale-mas/3scale-rhel7-operator@sha256:0a6673eae2f0e8d95b919b0243e44d2c0383d13e2e616ac8d3f80742d496d292
        - registry.redhat.io/3scale-mas/memcached-rhel7@sha256:9c2aba45971c924fd55b2aab50ee8e6861a9defd3808b38ff30316727af31f9a
        - registry.redhat.io/rhscl/postgresql-10-rhel7@sha256:4f807ee6e885e966853f70e1c40bbc179dc1f994b401ac868b7f2b921f9d17a0
        - registry.redhat.io/rhel8/mysql-80@sha256:7c05e61c0f4f5dad6aeda136e9a1af2f04f4de3c51004e997d4eacae3778da17
        - registry.redhat.io/3scale-mas/3scale-operator-bundle@sha256:a6c2eff926feb11da27eb9972cb86ee0ef9ca0b013afb6fe6af4f72439d557ef
        - registry.redhat.io/3scale-mas/zync-rhel9@sha256:419eb09d1b165363d5b90df01c1fa10f1f91f03649a8758620ffa49276685feb
        - registry.redhat.io/3scale-mas/apicast-gateway-rhel8@sha256:a03a1ae7f01259192b8d7b4ecee6ddfcdc0cd9e2d24a8c2306accb93ee767135
        - registry.redhat.io/3scale-mas/backend-rhel8@sha256:7d07caf7b6c5a195ba4706e82f3d28c3793afab99b1a6d5a7020547c48534cdc
        - registry.redhat.io/3scale-mas/searchd-rhel7@sha256:1426e4bd4aa0eb28b479900ffc02621203886280d24997cd11818387ff30d9df
        - registry.redhat.io/3scale-mas/system-rhel7@sha256:b0fd85a4994502a3a2ae9ba6a7610c38af64b6cc897efac12feb4b115a484f77
        - registry.redhat.io/openshift4/ose-cli@sha256:1a8caae40be21f23ba3cb48b75809a73333907033a6041927aa568b5c9290f3c
        - registry.redhat.io/rhel8/redis-6@sha256:22431acc2c991b6ee6b4de0219366dee6ea745122a49a0175d5fb58ed7d9205f
        version: 0.11.8-mas
      entries:
      - name: 3scale-operator.v0.11.8-mas
        version: 0.11.8-mas
      - name: 3scale-operator.v0.11.7-mas
        version: 0.11.7-mas
      - name: 3scale-operator.v0.11.6-mas
        version: 0.11.6-mas
      - name: 3scale-operator.v0.11.5-mas
        version: 0.11.5-mas
      - name: 3scale-operator.v0.11.3-mas
        version: 0.11.3-mas
      - name: 3scale-operator.v0.11.2-mas
        version: 0.11.2-mas
      - name: 3scale-operator.v0.11.1-mas
        version: 0.11.1-mas
      - name: 3scale-operator.v0.11.0-mas
        version: 0.11.0-mas
      - name: 3scale-operator.v0.10.0-mas
        version: 0.10.0-mas
      name: threescale-mas
    defaultChannel: threescale-2.15
    packageName: 3scale-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CloudBees
      provider-url: https://www.cloudbees.com/products/continuous-integration
    name: cloudbees-ci-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloudbees-ci.v3.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.cloudbees.com/v1alpha1",
                "kind": "CloudBeesCI",
                "metadata": {
                  "name": "cloudbeesci-sample"
                },
                "spec": {
                  "Agents": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-agent:2.289.3.2"
                    },
                    "ImagePullSecrets": null,
                    "SeparateNamespace": {
                      "Create": false,
                      "Enabled": false,
                      "Name": null
                    }
                  },
                  "Hibernation": {
                    "Enabled": false,
                    "Image": {
                      "dockerImage": "cloudbees/managed-master-hibernation-monitor:247.c5dfce00a179",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "NodeSelector": {},
                    "Tolerations": []
                  },
                  "Master": {
                    "Enabled": true,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-core-mm:2.289.3.2"
                    },
                    "JavaOpts": null,
                    "OperationsCenterNamespace": null
                  },
                  "NetworkPolicy": {
                    "Enabled": false,
                    "JMXSelectors": [],
                    "ingressControllerSelector": []
                  },
                  "OperationsCenter": {
                    "AgentListenerPort": 50000,
                    "Annotations": {},
                    "CSRF": {
                      "ProxyCompatibility": false
                    },
                    "CasC": {
                      "ConfigMapName": "oc-casc-bundle",
                      "Enabled": false
                    },
                    "ContainerPort": 8080,
                    "ContextPath": null,
                    "Enabled": true,
                    "ExtraConfigMaps": [],
                    "ExtraContainers": [],
                    "ExtraGroovyConfiguration": {},
                    "ExtraVolumeMounts": [],
                    "ExtraVolumes": [],
                    "HealthProbeLivenessFailureThreshold": 12,
                    "HealthProbes": true,
                    "HostName": null,
                    "Image": {
                      "dockerImage": "cloudbees/cloudbees-cloud-core-oc:2.289.3.2",
                      "dockerPullPolicy": null
                    },
                    "ImagePullSecrets": null,
                    "Ingress": {
                      "Annotations": {
                        "kubernetes.io/tls-acme": "false"
                      },
                      "Class": "nginx",
                      "tls": {
                        "Enable": false,
                        "SecretName": null
                      }
                    },
                    "JavaOpts": null,
                    "JenkinsOpts": null,
                    "LoadBalancerIP": null,
                    "LoadBalancerSourceRanges": [
                      "0.0.0.0/0"
                    ],
                    "Name": "cjoc",
                    "NodeSelector": {},
                    "Platform": "standard",
                    "Protocol": "http",
                    "Resources": {
                      "Limits": {
                        "Cpu": 1,
                        "Memory": "2G"
                      },
                      "Requests": {
                        "Cpu": 1,
                        "Memory": "2G"
                      }
                    },
                    "Route": {
                      "tls": {
                        "CACertificate": null,
                        "Certificate": null,
                        "DestinationCACertificate": null,
                        "Enable": false,
                        "InsecureEdgeTerminationPolicy": "Redirect",
                        "Key": null,
                        "Termination": "edge"
                      }
                    },
                    "ServiceAgentListenerPort": 50000,
                    "ServiceAnnotations": {},
                    "ServicePort": 80,
                    "ServiceType": "ClusterIP",
                    "Tolerations": []
                  },
                  "Persistence": {
                    "AccessMode": "ReadWriteOnce",
                    "Annotations": {},
                    "Size": "20Gi",
                    "StorageClass": null
                  },
                  "PodSecurityPolicy": {
                    "Annotations": {},
                    "Enabled": false
                  },
                  "Subdomain": false,
                  "ingress-nginx": {
                    "Enabled": false,
                    "controller": {
                      "admissionWebhooks": {
                        "patch": {
                          "nodeSelector": {
                            "kubernetes.io/os": "linux"
                          }
                        }
                      },
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "nginx-ingress": {
                    "Enabled": false,
                    "controller": {
                      "ingressClass": "nginx",
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      },
                      "service": {
                        "externalTrafficPolicy": "Local"
                      }
                    },
                    "defaultBackend": {
                      "nodeSelector": {
                        "kubernetes.io/os": "linux"
                      }
                    }
                  },
                  "rbac": {
                    "agentsServiceAccountName": "jenkins-agents",
                    "hibernationMonitorServiceAccountName": "managed-master-hibernation-monitor",
                    "install": true,
                    "masterServiceAccountName": "jenkins",
                    "serviceAccountName": "cjoc"
                  },
                  "sidecarinjector": {
                    "Enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/cloudbees-ci-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: CloudBeesCI
            name: cloudbeescis.charts.cloudbees.com
            version: v1alpha1
        description: CloudBees CI
        displayName: CloudBees CI
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudbees
        - ci
        - enterprise
        - continuous
        - integration
        - jenkins
        - automation
        - pipeline
        - devops
        links:
        - name: CloudBees CI
          url: https://www.cloudbees.com/products/continuous-integration
        maintainers:
        - email: support@cloudbees.com
          name: Support
        - email: sales@cloudbees.com
          name: Sales
        maturity: alpha
        provider:
          name: CloudBees
          url: https://www.cloudbees.com/products/continuous-integration
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        - registry.connect.redhat.com/cloudbees/cloudbees-core-modern@sha256:895fcf14b337398585f51e4b6e1f46a4a17c01a153ca889bd4374d691d89b9c8
        - registry.marketplace.redhat.com/rhm/cloudbees/cloudbees-ci-operator@sha256:bd0bd72e9e7807b0fa307de8a93a56967221fe2dd181bac02383f0f9be9120e4
        version: 3.34.1
      entries:
      - name: cloudbees-ci.v3.34.1
        version: 3.34.1
      name: alpha
    defaultChannel: alpha
    packageName: cloudbees-ci-rhmp
    provider:
      name: CloudBees
      url: https://www.cloudbees.com/products/continuous-integration
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom, Inc.
      provider-url: https://broadcom.com
    name: uma-admin-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: uma-admin-operator.v2025.4.5-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.4.2.5,990005",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
          createdAt: "2025-04-16T06:55:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.4.5
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.4.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.4.2.5,990005"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:dbf68e729484ba11aaafbcf1a590aca7f233de10e71b01d78c319dc5b5c6c278
        - registry.connect.redhat.com/ca/uma-operator@sha256:4c2a7f68a47428f3e386107254f74fa8c34e477eb40fc17c173df92f3c66c559
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:c4918f259cf68ed3f55b054418242590547fa16142fd5063cd821906927822b3
        version: 2025.4.5-5
      entries:
      - name: uma-admin-operator.v2025.4.5-5
        version: 2025.4.5-5
      name: "2025.4"
    - currentCSV: uma-admin-operator.v2025.5.4-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.5.2.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
          createdAt: "2025-07-24T14:54:21Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.5.4
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.5.4
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.5.2.12,990012"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:514b953eaf0959bc120f80bdc0a2e19d8f059d3fe2883a4f6843a398470a56b1
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:b1b0f5689c396ac50805c1978219aa48448c32e8c8d2ae32b2d505274688f6cf
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:e9824dd4ca39accfc674b88f7a2b76fba3006ff856e78402348c720f2aa8a6b7
        version: 2025.5.4-12
      entries:
      - name: uma-admin-operator.v2025.5.4-12
        version: 2025.5.4-12
      name: "2025.5"
    - currentCSV: uma-admin-operator.v2025.7.3-21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.7.1.23,990023",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
          createdAt: "2025-08-23T09:00:37Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.3
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.7.3
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.7.1.23,990023"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:767d55dc7343406637dd16f1b3a73bfaec353e0b058143fd68df2474271e3b76
        - registry.connect.redhat.com/ca/uma-operator@sha256:188a2b8d2255ded6141a4b795f1b6d7bd98017b4fd57c9c1a1d38fc701358057
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:e30d81384573346a694a12d9ff249531fe3cbbcef0ed116406cffdd30cb9c07c
        version: 2025.7.3-21
      entries:
      - name: uma-admin-operator.v2025.7.3-21
        version: 2025.7.3-21
      name: "2025.7"
    - currentCSV: uma-admin-operator.v2025.8.2-9
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.8.1.9,990009",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
          createdAt: "2025-08-15T03:20:28Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.8.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.8.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "25.8.1.9,990009"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:760029188741c954842bcf237dee24544fe65f6290b07899c335136ac7bbb468
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:43c90bbe3b20e95e1f352cdad097d63820a1282deedeb769352cb100da43ecc3
        - registry.connect.redhat.com/ca/uma-operator@sha256:0cfb63c62258e54d67e7baca874003a04477f30bcf7f56861f565ad068070580
        version: 2025.8.2-9
      entries:
      - name: uma-admin-operator.v2025.8.2-9
        version: 2025.8.2-9
      name: "2025.8"
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: "2025.9"
    - currentCSV: uma-admin-operator.v2025.7.2-19
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "99.99.0.serp_F152813,19",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
          createdAt: "2025-07-30T08:07:39Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.7.2
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-admin-operator/2025.7.2
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               # tenantID is Cohort ID of DX SaaS tenant.
               # After logging into DX SaaS, Cohort ID is available at "Launch Pad -> Settings -> Connector Parameters"
               # tenantID is only required if Kubernetes/OpenShift events monitoring is enabled in UMA.
               tenantID: ""
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or OpenShift.
             type: Kubernetes
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 gcLoggingEnabled: false
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "800Mi"
                 dataReporter:
                   vertex:
                     ttl: 720
                     refreshInterval: 360
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:
               buildTag: "99.99.0.serp_F152813,19"
               # Property to enable Mutual TLS for UMA apis. When set to false TLS is used for the apis.
               mTLS: true

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 nodeAffinity: {}

                  # Node labels for UMA Pods
                  # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                  nodeSelector: {}

                  # Additional tolerations for UMA pods
                  # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                  # Default toleration used by UMA
                  tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               # make sure tenand id is provided above section
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-operator@sha256:7b93de818fa5160e72f878b0350d13fc1d5943eac95b0794d3d4f930cf21b59a
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:2b1cb609ec5abd21b12b3d3deb63809d8968ca43ea98de220b019faef2e7615a
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:2f853ad4fece4510e847e96fe8085b53c56ee342e1f387677ff8cef345298e72
        version: 2025.7.2-19
      entries:
      - name: uma-admin-operator.v2025.7.2-19
        version: 2025.7.2-19
      name: alpha
    - currentCSV: uma-admin-operator.v2025.9.0-8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.8,990008",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
          createdAt: "2025-09-11T11:16:06Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.0
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.0
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.8,990008"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:70c5237a251adb5f44a085d091ca6c24d5b313aa73f6545389a2b938740a9a20
        - registry.connect.redhat.com/ca/uma-operator@sha256:1cf3909dd2caeabc53209d6f8565dcb6c3562b9c6eb715dda8ba1953f264668e
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:037c2f96bf2c3588093a4e218109e354578384ae6eb63f88de7843b9152382a9
        version: 2025.9.0-8
      entries:
      - name: uma-admin-operator.v2025.9.0-8
        version: 2025.9.0-8
      name: beta
    - currentCSV: uma-admin-operator.v2025.9.1-12
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ca.broadcom.com/v1alpha1",
                "kind": "UniversalMonitoringAgentsAdmin",
                "metadata": {
                  "name": "uma-monitor"
                },
                "spec": {
                  "EnableResourceLimits": true,
                  "agentManager": {
                    "credential": "",
                    "credentialSecret": {
                      "enabled": false,
                      "key": "agentManager.credential",
                      "name": "caaiops-uma-credential"
                    },
                    "httpProxy": {
                      "host": "",
                      "password": "",
                      "port": "",
                      "username": ""
                    },
                    "url": "localhost:5001"
                  },
                  "architecture": "amd64",
                  "cloudmonitoring": {
                    "aws": {
                      "accessKey": "",
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "secretKey": "",
                      "servicesList": ""
                    }
                  },
                  "clusterName": "DevelopmentCluster",
                  "globalDeployment": {
                    "buildTag": "25.9.1.12,990012",
                    "caCert": "",
                    "clientCert": "",
                    "clientKey": "",
                    "daemonset": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": [
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/master"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/control-plane"
                        },
                        {
                          "effect": "NoSchedule",
                          "key": "node-role.kubernetes.io/storage"
                        },
                        {
                          "effect": "NoSchedule",
                          "operator": "Exists"
                        },
                        {
                          "effect": "NoExecute",
                          "operator": "Exists"
                        }
                      ]
                    },
                    "deployment": {
                      "nodeAffinity": {},
                      "nodeSelector": {},
                      "tolerations": []
                    },
                    "mixedMode": "true",
                    "serverCert": "",
                    "serverCollectorCert": null,
                    "serverCollectorKey": null,
                    "serverKey": "",
                    "umaAgentACCController": {
                      "enabled": true
                    },
                    "umaAgentACCUpgrade": {
                      "enabled": false
                    }
                  },
                  "monitor": {
                    "application": {
                      "autoattach": {
                        "dotnet": {
                          "enabled": true,
                          "propertiesOverride": ""
                        },
                        "dynamicPropertyResolution": {
                          "agentName": "",
                          "hostName": "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost",
                          "processName": ""
                        },
                        "filterType": "whitelist",
                        "java": {
                          "customJdkTools": {
                            "enabled": false,
                            "imageName": ""
                          },
                          "enabled": true,
                          "propertiesOverride": ""
                        }
                      },
                      "jmx": {
                        "enabled": true
                      },
                      "opentracing": {
                        "correlation": {
                          "enabled": false
                        },
                        "enabled": true,
                        "grpc": {
                          "hostport": ""
                        }
                      },
                      "zipkinTracing": {
                        "enabled": true,
                        "zipkinServer": {
                          "hostport": ""
                        }
                      }
                    },
                    "clusterPerformance": {
                      "dataReporter": {
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "nassClientEnabled": false,
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        },
                        "vertex": {
                          "refreshInterval": 480,
                          "ttl": 1440
                        }
                      },
                      "enabled": true,
                      "limits": {
                        "cpu": "2",
                        "memory": "3096Mi"
                      },
                      "node": {
                        "noscheduleTaint": {
                          "allowed": {
                            "tolerations": ""
                          },
                          "skipMetricAggregation": {
                            "enabled": true
                          }
                        }
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "1024Mi"
                      }
                    },
                    "container": {
                      "dockerstats": {
                        "collector": {
                          "limits": {
                            "cpu": "2",
                            "memory": "2048Mi"
                          },
                          "requests": {
                            "cpu": "200m",
                            "memory": "500Mi"
                          }
                        },
                        "daemonset": {
                          "apmia": {
                            "epagent": {
                              "port": 8889
                            },
                            "hostmonitoring": {
                              "linux": {
                                "groups": "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                              }
                            }
                          },
                          "enableAllContainerRuntimes": false,
                          "limits": {
                            "cpu": "2",
                            "memory": "1024Mi"
                          },
                          "privileged": true,
                          "requests": {
                            "cpu": "200m",
                            "memory": "300Mi"
                          }
                        },
                        "enabled": true
                      },
                      "prometheus": {
                        "backend": {
                          "custom": {
                            "promqlConfigMap": "custom-promql-config"
                          },
                          "enabled": false,
                          "endPoint": {
                            "configFiles": "",
                            "hostNameVerifierEnabled": true,
                            "metricAlias": "container_name=container,pod_name=pod",
                            "password": "",
                            "token": "",
                            "url": "",
                            "username": ""
                          },
                          "filter": {
                            "name": "",
                            "value": ""
                          }
                        },
                        "exporter": {
                          "enabled": true
                        },
                        "limits": {
                          "cpu": "2",
                          "memory": "1024Mi"
                        },
                        "requests": {
                          "cpu": "200m",
                          "memory": "300Mi"
                        }
                      }
                    },
                    "events": {
                      "clamp": {
                        "perDayLimit": 15000
                      },
                      "elasticDocTypeId": "itoa_events_apm",
                      "elasticDocTypeVersion": 1,
                      "elasticProduct": "ao",
                      "enabled": false,
                      "filter": {
                        "fieldselector": "type!=Normal",
                        "namespace": ""
                      },
                      "metricsEnabled": true
                    },
                    "ibmmqia": {
                      "enabled": false,
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      }
                    },
                    "kafka": {
                      "consumergroups": {
                        "filter": "",
                        "metrics": true
                      },
                      "debug": false,
                      "deployAsDaemonSet": false,
                      "enabled": false,
                      "jmx": {
                        "MBeans": {
                          "exclude": "",
                          "include": ""
                        },
                        "modules": "kafka,jvm,memory"
                      },
                      "limits": {
                        "cpu": "2",
                        "memory": "1024Mi"
                      },
                      "requests": {
                        "cpu": "200m",
                        "memory": "300Mi"
                      },
                      "sasl": {
                        "plain": {
                          "enabled": false,
                          "password": "",
                          "username": ""
                        }
                      },
                      "ssl": {
                        "client": {
                          "enabled": false,
                          "endpointIdentificationEnabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystoreKeyPassword": {
                            "value": ""
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        },
                        "jmx": {
                          "enabled": false,
                          "keystoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "keystorePassword": {
                            "value": ""
                          },
                          "truststoreKey": {
                            "value": "",
                            "valueFrom": {
                              "secretKeyRef": {
                                "key": "",
                                "name": ""
                              }
                            }
                          },
                          "truststorePassword": {
                            "value": ""
                          }
                        }
                      }
                    },
                    "node": {
                      "enabled": true
                    }
                  },
                  "prefix": "k8s",
                  "role": "admin",
                  "sustainabilityEvents": {
                    "enabled": false
                  },
                  "type": "Openshift"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime
          certified: "true"
          containerImage: registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
          createdAt: "2025-09-24T12:23:56Z"
          description: The DX APM Universal Monitoring Agent Admin For Multi-tenancy
            installs the UMA Admin deployment on a cluster. This deployment is required
            for the Cluster Administrator to view the DX APM cluster performance metrics.
            This UMA Admin capability provides the Cluster Administrator with complete
            cluster monitoring visibility. Installing the UMA Admin deployment requires
            admin privileges and the Application Administrator install the UMA Team
            capability.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2025.9.1
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/CA-Application-Performance-Management/certified-operators/tree/main/operators/uma-operator/2025.9.1
          support: Broadcom, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The Multitenancy Admin Universal Monitoring Agent for Kubernetes
              monitoring.
            displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
            kind: UniversalMonitoringAgentsAdmin
            name: universalmonitoringagentsadmins.ca.broadcom.com
            version: v1alpha1
        description: |
          The **DX APM Universal Monitoring Agent Admin For Multi-tenancy**  installs the UMA Admin deployment on a cluster. This deployment is required for the Cluster Administrator to view the DX APM cluster performance metrics. This UMA Admin capability provides the Cluster Administrator with complete cluster monitoring visibility. Installing the UMA Admin deployment requires admin privileges and the Application Administrator install the UMA Team capability.
          The Application Administrator for each application team that uses a DX APM tenant installs the UMA Team capability by using **DX APM Universal Monitoring Agent Team For Multi-tenancy** operator and specifying the namespaces the team owns or wants to monitor.

          Note:  Currently UMA multi-tenancy deployment does not support other UMA monitoring capabilities, such as Prometheus metric ingestion,  probe auto-attach(node.js, php), AWS Monitoring, and the UMA for HTTP Collector
          For more information: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html

          **Usage**:
          ```
           apiVersion: ca.broadcom.com/v1alpha1
           kind: UniversalMonitoringAgentsAdmin
           metadata:
             name: uma-monitor
           spec:
             agentManager:
               url: localhost:5001
               credential: ""
               credentialSecret:
                 enabled: false
                 name: "caaiops-uma-credential"
                 key: "agentManager.credential"
               httpProxy:
                 host:
                 port:
                 username:
                 password:
             # Name of the cluster.
             clusterName: DevelopmentCluster
             # Cluster type, either Kubernetes or Openshift.
             type: Openshift
             prefix: k8s
             openshift311Support: false
             # architecture This property defines the system architecture of the cluster where uma is being installed.
             # Example, if UMA is being installed on s390x machine set architecture as s390x.
             # supported values are amd64,s390x,arm64 Default: amd64
             architecture: amd64
             EnableResourceLimits: true
             role: admin

             monitor:
               application:
                 autoattach:
                   filterType: whitelist
                   # this feature leverages container metadata to resolve agent host name and agent name
                   dynamicPropertyResolution:
                     hostName: "{k8s_deployment_name},{k8s_daemonset_name},{k8s_deploymentconfig_name},{k8s_pod_name},ContainerHost"
                     agentName: ""
                     processName: ""
                   java:
                     enabled: true
                     propertiesOverride:
                     customJdkTools:
                       enabled: false
                       # Provide your custom created jdk tools jar image name below.
                       imageName: ""
                   dotnet:
                     enabled: true
                     propertiesOverride:
                 jmx:
                   enabled: true
                 opentracing:
                   enabled: true
                   grpc:
                     hostport: ""
                   correlation:
                     enabled: false
                 zipkinTracing:
                   enabled: true
                   zipkinServer:
                     hostport:
               events:
                 # When enabled, Kubernetes/Openshift events are published to DX SaaS.
                 enabled: false
                 # When enabled, Kubernetes/Openshift events are used to calculate the liveness and
                 # readiness probe status metrics.
                 metricsEnabled: true
                 filter:
                   # fieldselector property can be used to filter the Kubernetes events based on the value of one or more resource fields.
                   # Here are some examples of field selector queries: metadata.namespace!=default, metadata.name=my-service etc.
                   fieldselector: type!=Normal
                   namespace:
                 clamp:
                   perDayLimit: 15000
                 elasticDocTypeId: itoa_events_apm
                 elasticDocTypeVersion: 1
                 elasticProduct: ao
               container:
                 prometheus:
                   exporter:
                     enabled: true
                   backend:
                     enabled: false
                     endPoint:
                       url: ""
                       username: ""
                       password: ""
                       token: ""
                       # Property to enable strict hostnameVerifier for the Prometheus backend secure connection.
                       hostNameVerifierEnabled: true
                       configFiles: ""
                       metricAlias: container_name=container,pod_name=pod
                     filter:
                       name: ""
                       value: ""
                     custom:
                       promqlConfigMap: custom-promql-config
                   # The below properties define the resources for 'cluster-performance-prometheus' deployment that is responsible for
                   # Prometheus metrics ingestion.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"

                 dockerstats:
                   enabled: true
                   daemonset:
                     privileged: true
                     # This is used to enable all container runtime sockets like crio, containerd, docker etc.Default: false.
                     # By default, on openshift4: crio , openshift 3.11: docker and on k8s (eks, aks, gke and standalone k8s) : docker, containerd and eks specific sockets are enabled
                     enableAllContainerRuntimes: false
                     # Default epagent port of the UMA daemonset pod deployment (app-container-monitor)
                     apmia:
                       epagent:
                         port: 8889
                       hostmonitoring:
                         linux:
                           # List of linux groups monitored by host monitor. If it is empty, It will monitor all groups.
                           groups: "memory,cpu,network,diskio,hostresources,application,protocol,systemmanagement,performance,process,ipc,distributedsystem,topprocess,operatingsystem,storage,systemload,port"
                     # The below properties define the resources for 'app-container-monitor' daemonset.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "1024Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "300Mi"
                   collector:
                     # The below properties define the resources for 'collector' deployment.
                     limits:
                       cpu: "2"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "2048Mi"
                     requests:
                       cpu: "200m"
                       # Assign memory resources in 'Mi'(mebibytes) units only.
                       memory: "500Mi"
               clusterPerformance:
                 enabled: true
                 # The below properties define the resources for 'clusterinfo' deployment.
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "3096Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 dataReporter:
                   vertex:
                     # The vertex TTL (in minutes) used by the UMA for all the kubernetes/Openshift workload vertices.
                     ttl: 1440
                     # The vertex refresh interval(in minutes) at which UMA resends all the kubernetes/Openshift workload vertices.
                     refreshInterval: 480
                   # The below properties define the resources for 'container-monitor' deployment that is responsible for
                   # posting the K8s/OpenShift Cluster performance metrics, map data etc.
                   limits:
                     cpu: "2"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "1024Mi"
                   requests:
                     cpu: "200m"
                     # Assign memory resources in 'Mi'(mebibytes) units only.
                     memory: "300Mi"
                   nassClientEnabled: false
                 node:
                   noscheduleTaint:
                     skipMetricAggregation:
                       enabled: true
                     allowed:
                       tolerations: ""
               node:
                 enabled: true
               kafka:
                 enabled: false
                 consumergroups:
                   metrics: true
                   filter: ""
                 jmx:
                   modules: "kafka,jvm,memory"
                   MBeans:
                     # A semi-colon separated list of managed bean object names. JMX metrics will be automatically reported for matching beans.
                     include: ""
                     # A semi-colon separated list of managed bean object names. Matching beans will be ignored for Kafka JMX metrics reporting. This property overrides "include" configuration.
                     exclude: ""
                 deployAsDaemonSet: false
                 debug: false
                 ssl:
                   client:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                     keystoreKeyPassword:
                       value: ""
                     endpointIdentificationEnabled: false
                   jmx:
                     enabled: false
                     truststoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     truststorePassword:
                       value: ""
                     keystoreKey:
                       value: ""
                       valueFrom:
                         secretKeyRef:
                           name: ""
                           key: ""
                     keystorePassword:
                       value: ""
                 sasl:
                   plain:
                     enabled: false
                     username: ""
                     password: ""
                 # The below properties define the resources for 'apm-kafka-monitor' deployment or daemonset
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"
               ibmmqia:
                 enabled: false
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             cloudmonitoring:
               aws:
                 enabled: false
                 accessKey: ""
                 secretKey: ""
                 # services supported are rds,ec2,ebs,lambda,dynamodb,s3,sns,ecs,sqs,kinesis,clb,nlb,alb,redshift,elasticache,autoScaling,billing,apiGateway
                 # provide comma separated services to monitor
                 servicesList: ""
                 # The below properties define the resources for AWS monitoring deployment 'aws-monitor'
                 limits:
                   cpu: "2"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "1024Mi"
                 requests:
                   cpu: "200m"
                   # Assign memory resources in 'Mi'(mebibytes) units only.
                   memory: "300Mi"

             globalDeployment:

               # Below certificate properties are optional and can be configured by the users to use
               # corporate mandated CA, server and client TLS certificates. If not specified out of the box self-signed
               # TLS certificates are used by UMA for securing the apis.

               # Property to for specifying the base64 encoded CA certificate
               caCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate
               serverCert: ""
               # Property to for specifying the base64 encoded CA signed client certificate
               clientCert: ""
               # Property to for specifying the base64 encoded CA signed server certificate key
               serverKey: ""
               # Property to for specifying the base64 encoded CA signed client certificate key
               clientKey: ""
               mixedMode: "true"
               # Property to hold the build tag of the agent
               buildTag: "25.9.1.12,990012"
               # Property to for specifying the base64 encoded CA signed collector service server certificate
               serverCollectorCert:
               # Property to for specifying the base64 encoded CA signed collector service server certificate key
               serverCollectorKey:
               # Node affinity rules for UMA Pods
               # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
               daemonset:
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations:
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/master"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/control-plane"
                   - effect: "NoSchedule"
                     key: "node-role.kubernetes.io/storage"
                   - effect: "NoSchedule"
                     operator: "Exists"
                   - effect: "NoExecute"
                     operator: "Exists"
               deployment:
                 # Node affinity rules for UMA Pods
                 # Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity
                 nodeAffinity: {}

                 # Node labels for UMA Pods
                 # Ref: https://kubernetes.io/docs/user-guide/node-selection/
                 nodeSelector: {}

                 # Additional tolerations for UMA pods
                 # https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/
                 # Default toleration used by UMA
                 tolerations: []
               umaAgentACCController:
                 enabled: true
               umaAgentACCUpgrade:
                 enabled: false
             sustainabilityEvents:
               # turn sustainabilityEvents enabled to true for sending sustainability events to OI
               enabled: false
          ```
        displayName: DX APM Universal Monitoring Agent Admin For Multi-tenancy
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - uma-admin-operator
        links:
        - name: Uma Admin Operator
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-agents/SaaS/Universal-Monitoring-Agent/Install-the-Universal-Monitoring-Agent/Deploy-UMA-in-a-Multi-tenant-Environment.html
        - name: Intro to APM
          url: https://techdocs.broadcom.com/us/en/ca-enterprise-software/it-operations-management/dx-apm-saas/SaaS.html
        maintainers:
        - email: support@broadcom.com
          name: Broadcom, Inc.
        maturity: alpha
        provider:
          name: Broadcom, Inc.
          url: https://broadcom.com
        relatedImages:
        - registry.connect.redhat.com/ca/uma-admin-operator-bundle@sha256:5aa2e1b4f45b0751dfcb5c4c48d674460eff00194c6cdf0b135cb465fd4cef92
        - registry.connect.redhat.com/ca/uma-operator@sha256:1377779fca3c24c384cb4efee8581aee12199effefbbcd6a8f911b8d926b8938
        - registry.connect.redhat.com/ca/universalmonitoragent@sha256:814b955c35ef8d374801fd0f1cf7181a836aee14caaf1b1d458d44402475943f
        version: 2025.9.1-12
      entries:
      - name: uma-admin-operator.v2025.9.1-12
        version: 2025.9.1-12
      name: stable
    defaultChannel: stable
    packageName: uma-admin-operator
    provider:
      name: Broadcom, Inc.
      url: https://broadcom.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: VMware Tanzu
      provider-url: ""
    name: rabbitmq-cluster-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rabbitmq-cluster-operator.v2.16.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "rabbitmq.com/v1beta1",
                "kind": "RabbitmqCluster",
                "metadata": {
                  "name": "hello-world"
                },
                "spec": {
                  "override": {
                    "statefulset": {
                      "spec": {
                        "template": {
                          "spec": {
                            "containers": [],
                            "securityContext": {}
                          }
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Streaming & Messaging
          certified: "false"
          containerImage: quay.io/rabbitmqoperator/cluster-operator:2.16.0
          createdAt: 2025-08-12T09:34:33
          repository: https://github.com/rabbitmq/cluster-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Rabbitmq Cluster
            displayName: RabbitmqCluster
            kind: RabbitmqCluster
            name: rabbitmqclusters.rabbitmq.com
            version: v1beta1
        description: "## About the managed application\nRabbitMQ is an open-source
          message broker that originally implemented the\nAdvanced Message Queuing
          Protocol (AMQP). Nowadays features a plug-in\narchitecture supporting protocols
          such as Streaming Text Oriented Messaging\nProtocol (STOMP), MQ Telemetry
          Transport (MQTT), and others. RabbitMQ\nsupports native streaming and AMQP
          1.0.\n\nVisit [rabbitmq.com](rabbitmq.com) for more information.\n\n## About
          this Operator\nRabbitMQ Cluster Kubernetes Operator is developed and maintened
          by the VMware \nRabbitMQ team and provides a consistent and easy way to
          deploy RabbitMQ\nclusters to Kubernetes and run them, including \"day two\"
          (continuous)\noperations. RabbitMQ clusters deployed using the Operator
          can be used by\napplications running on Kubernetes or outside of Kubernetes.\n\nRead
          [RabbitMQ operator docs](https://www.rabbitmq.com/kubernetes/operator/operator-overview)\nfor
          more information.\n"
        displayName: RabbitMQ-cluster-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rabbitmq
        - messaging
        - amqp
        - amqp10
        - stomp
        - mqtt
        - mqtt5
        - queues
        - streaming
        links:
        - name: Documentation
          url: https://www.rabbitmq.com/kubernetes/operator/operator-overview.html
        - name: GitHub
          url: https://github.com/rabbitmq/cluster-operator
        maintainers:
        - email: rabbitmq-users@googlegroups.com
          name: VMware Tanzu
        maturity: stable
        minKubeVersion: 1.26.0
        nativeApis:
        - group: apps
          kind: Deployment
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: ReplicaSet
          version: v1
        - group: ""
          kind: Pod
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: ConfigMap
          version: v1
        provider:
          name: VMware Tanzu
        relatedImages:
        - quay.io/community-operator-pipeline-prod/rabbitmq-cluster-operator@sha256:681e51da33e3cd9d6a4b617dca4338827237b70d1e8065322bc4135846039c56
        - quay.io/rabbitmqoperator/cluster-operator:2.16.0
        version: 2.16.0
      entries:
      - name: rabbitmq-cluster-operator.v2.16.0
        version: 2.16.0
      - name: rabbitmq-cluster-operator.v2.15.0
        version: 2.15.0
      - name: rabbitmq-cluster-operator.v2.14.0
        version: 2.14.0
      - name: rabbitmq-cluster-operator.v2.13.0
        version: 2.13.0
      - name: rabbitmq-cluster-operator.v2.12.1
        version: 2.12.1
      - name: rabbitmq-cluster-operator.v2.10.0
        version: 2.10.0
      - name: rabbitmq-cluster-operator.v2.9.0
        version: 2.9.0
      - name: rabbitmq-cluster-operator.v2.8.0
        version: 2.8.0
      - name: rabbitmq-cluster-operator.v2.7.0
        version: 2.7.0
      - name: rabbitmq-cluster-operator.v2.6.0
        version: 2.6.0
      - name: rabbitmq-cluster-operator.v2.5.0
        version: 2.5.0
      - name: rabbitmq-cluster-operator.v2.4.0
        version: 2.4.0
      - name: rabbitmq-cluster-operator.v2.3.0
        version: 2.3.0
      - name: rabbitmq-cluster-operator.v2.2.0
        version: 2.2.0
      - name: rabbitmq-cluster-operator.v2.1.0
        version: 2.1.0
      - name: rabbitmq-cluster-operator.v2.0.0
        version: 2.0.0
      - name: rabbitmq-cluster-operator.v1.14.0
        version: 1.14.0
      name: stable
    defaultChannel: stable
    packageName: rabbitmq-cluster-operator
    provider:
      name: VMware Tanzu
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Stolostron
      provider-url: ""
    name: cluster-aas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-aas-operator.v0.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplate",
                "metadata": {
                  "name": "clustertemplate-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateInstance",
                "metadata": {
                  "name": "clustertemplateinstance-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "clustertemplate.openshift.io/v1alpha1",
                "kind": "ClusterTemplateQuota",
                "metadata": {
                  "name": "clustertemplatequota-sample"
                },
                "spec": {}
              }
            ]
          categories: Integration & Delivery,OpenShift Optional
          certified: "false"
          containerImage: quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
          createdAt: "2024-06-06T19:34:44Z"
          description: Easily install fully configured clusters with guard-rails.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/stolostron/cluster-templates-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents instance of a cluster
            displayName: Cluster template instance
            kind: ClusterTemplateInstance
            name: clustertemplateinstances.clustertemplate.openshift.io
            version: v1alpha1
          - description: Defines which ClusterTemplates can be used in a given namespace
            displayName: Cluster template quota
            kind: ClusterTemplateQuota
            name: clustertemplatequotas.clustertemplate.openshift.io
            version: v1alpha1
          - description: Template of a cluster - both installation and post-install
              setup are defined as ArgoCD application spec. Any application source
              is supported - typically a Helm chart
            displayName: Cluster template
            kind: ClusterTemplate
            name: clustertemplates.clustertemplate.openshift.io
            version: v1alpha1
          - kind: ClusterTemplateSetup
            name: clustertemplatesetup.clustertemplate.openshift.io
            version: v1alpha1
          - kind: Config
            name: config.clustertemplate.openshift.io
            version: v1alpha1
        description: |
          **Self-service clusters with guardrails!**
          Cluster as a service operator provides an easy way to define clusters as templates and allows creating instances of those templates even for non-privileged developer/devops engineers. Cluster templates operator also allows specifing quotas for the developer/devops engineers.

          ### User Interface
          The User interface is installed but is not enabled by default. To enable go to `Home` -> `Overview` -> in the status card click `Dynamic Plugins` -> click `View all` -> click the `pencil` icon -> pick `Enable` and hit `Save`. You can access the UI in the `All Clusters` perspective under `Infrastructure` -> `Cluster Templates`.

          ### Documentation
          Documentation can be found on our [website](https://github.com/stolostron/cluster-templates-operator).

          ### Getting help
          To report any issues or ask questions, create an issue on our [Github repo](https://github.com/stolostron/cluster-templates-operator/issues)
        displayName: Cluster as a service operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cluster
        - templates
        - hive
        - hypershift
        - selfservice
        - openshift
        - multicluster
        - quota
        - helm
        - argocd
        links:
        - name: Cluster as a service operator
          url: https://github.com/stolostron/cluster-templates-operator
        maintainers:
        - email: rawagner@redhat.com
          name: Rastislav Wagner
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Stolostron
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/cluster-aas-operator@sha256:aad6afc2f6f1f4e3b2c0d14b6c178d6f71ef318b0d3efc97faa5fdb736eaa721
        - quay.io/stolostron/cluster-templates-operator:2.11.0-0c3131154065397d37e3b66ae635d7ea88650bf9
        version: 0.1.5
      entries:
      - name: cluster-aas-operator.v0.1.5
        version: 0.1.5
      - name: cluster-aas-operator.v0.1.4
        version: 0.1.4
      - name: cluster-aas-operator.v0.1.3
        version: 0.1.3
      - name: cluster-aas-operator.v0.1.2
        version: 0.1.2
      - name: cluster-aas-operator.v0.1.1
        version: 0.1.1
      - name: cluster-aas-operator.v0.1.0
        version: 0.1.0
      - name: cluster-aas-operator.v0.0.6
        version: 0.0.6
      - name: cluster-aas-operator.v0.0.5
        version: 0.0.5
      - name: cluster-aas-operator.v0.0.4
        version: 0.0.4
      - name: cluster-aas-operator.v0.0.3
        version: 0.0.3
      - name: cluster-aas-operator.v0.0.2
        version: 0.0.2
      - name: cluster-aas-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cluster-aas-operator
    provider:
      name: Stolostron
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: www.redhat.com
    name: compliance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: compliance-operator.v0.1.32
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.32'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips"]'
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/4.7/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/compliance-openscap-rhel8@sha256:867c6d4205653d8a954d234cbaf470313009eb983d4f455fb45f49c4a7764422
        - registry.redhat.io/openshift4/compliance-rhel8-operator-metadata@sha256:1c06868bce1f5d7145ef658bfcc284c44011aa24a17166aa124547b38f2259e4
        - registry.redhat.io/openshift4/compliance-rhel8-operator@sha256:ed68a0574ddeaf7d676403cd843f28cb586894f35dbb3ca5e06a5264bd42f030
        - registry.redhat.io/openshift4/compliance-content-rhel8@sha256:4529b9bb32c1846a38e38363fa872713b1c1e6b26b34d887813432f97cff368c
        version: 0.1.32
      entries:
      - name: compliance-operator.v0.1.32
        version: 0.1.32
      name: "4.7"
    - currentCSV: compliance-operator.v0.1.61
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          olm.skipRange: '>=0.1.17 <0.1.61'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repo
          url: https://github.com/openshift/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/compliance-operator-understanding.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:8a2194d8d26dff03ec53ac0ab93c62c268e48675079a88db18bb14a977a391be
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:5578e2225365ac1c15e4c189c5c7a8d14fd3fd26189414b36028f7f0879bb1dc
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:c4bf5b2b20ff538adbc430b7ee993fbd7c291203a9810534005148304e3b169b
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:83ad3113964ec48c5cf28315d923adb6edb95e45960e2756665d617dcd9b2e47
        version: 0.1.61
      entries:
      - name: compliance-operator.v0.1.61
        version: 0.1.61
      name: release-0.1
    - currentCSV: compliance-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-node"
                },
                "spec": {
                  "content": "ssg-rhcos4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceScan",
                "metadata": {
                  "name": "example-compliancescan-platform"
                },
                "spec": {
                  "content": "ssg-ocp4-ds.xml",
                  "profile": "xccdf_org.ssgproject.content_profile_moderate",
                  "scanType": "Platform"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ComplianceSuite",
                "metadata": {
                  "name": "example-compliancesuite"
                },
                "spec": {
                  "autoApplyRemediations": false,
                  "scans": [
                    {
                      "content": "ssg-rhcos4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "workers-scan",
                      "nodeSelector": {
                        "node-role.kubernetes.io/worker": ""
                      },
                      "profile": "xccdf_org.ssgproject.content_profile_moderate"
                    },
                    {
                      "content": "ssg-ocp4-ds.xml",
                      "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4",
                      "name": "platform-scan",
                      "profile": "xccdf_org.ssgproject.content_profile_moderate",
                      "scanType": "Platform"
                    }
                  ],
                  "schedule": "0 1 * * *"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ProfileBundle",
                "metadata": {
                  "name": "ocp4"
                },
                "spec": {
                  "contentFile": "ssg-ocp4-ds.xml",
                  "contentImage": "registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "autoApplyRemediations": false,
                "debug": true,
                "kind": "ScanSetting",
                "metadata": {
                  "name": "my-companys-constraints"
                },
                "rawResultStorage": {
                  "rotation": 10,
                  "size": "2Gi"
                },
                "roles": [
                  "worker",
                  "master"
                ],
                "schedule": "0 1 * * *"
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "ScanSettingBinding",
                "metadata": {
                  "name": "nist-moderate"
                },
                "profiles": [
                  {
                    "apiGroup": "compliance.openshift.io/v1alpha1",
                    "kind": "Profile",
                    "name": "rhcos4-moderate"
                  }
                ],
                "settingsRef": {
                  "apiGroup": "compliance.openshift.io/v1alpha1",
                  "kind": "ScanSetting",
                  "name": "default"
                }
              },
              {
                "apiVersion": "compliance.openshift.io/v1alpha1",
                "kind": "TailoredProfile",
                "metadata": {
                  "name": "example-tailoredprofile"
                },
                "spec": {
                  "description": "Example of a tailoredProfile that extends OCP4 FedRAMP Moderate",
                  "disableRules": [
                    {
                      "name": "ocp4-file-permissions-node-config",
                      "rationale": "This breaks X application."
                    },
                    {
                      "name": "ocp4-account-disable-post-pw-expiration",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-accounts-no-uid-except-zero",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-chown",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmod",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchmodat",
                      "rationale": "testing this"
                    },
                    {
                      "name": "ocp4-audit-rules-dac-modification-fchown",
                      "rationale": "testing this"
                    }
                  ],
                  "extends": "ocp4-moderate",
                  "setValues": [
                    {
                      "name": "ocp4-var-selinux-state",
                      "rationale": "trolling dwalsh",
                      "value": "permissive"
                    }
                  ],
                  "title": "My little profile"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring,Security
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          must-gather-image: registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
          olm.skipRange: '>=0.1.17 <1.7.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-compliance
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ComplianceAsCode/compliance-operator
          support: Red Hat Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ComplianceCheckResult represent a result of a single compliance
              "test"
            kind: ComplianceCheckResult
            name: compliancecheckresults.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceRemediation represents a remediation that can be
              applied to the cluster to fix the found issues.
            displayName: Compliance Remediation
            kind: ComplianceRemediation
            name: complianceremediations.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceScan represents a scan with a certain configuration
              that will be applied to objects of a certain entity in the host. These
              could be nodes that apply to a certain nodeSelector, or the cluster
              itself.
            displayName: Compliance Scan
            kind: ComplianceScan
            name: compliancescans.compliance.openshift.io
            version: v1alpha1
          - description: ComplianceSuite represents a set of scans that will be applied
              to the cluster. These should help deployers achieve a certain compliance
              target.
            displayName: Compliance Suite
            kind: ComplianceSuite
            name: compliancesuites.compliance.openshift.io
            version: v1alpha1
          - description: ProfileBundle is the Schema for the profilebundles API
            displayName: Profile Bundle
            kind: ProfileBundle
            name: profilebundles.compliance.openshift.io
            version: v1alpha1
          - description: Profile is the Schema for the profiles API
            kind: Profile
            name: profiles.compliance.openshift.io
            version: v1alpha1
          - description: Rule is the Schema for the rules API
            kind: Rule
            name: rules.compliance.openshift.io
            version: v1alpha1
          - description: ScanSettingBinding is the Schema for the scansettingbindings
              API
            displayName: Scan Setting Binding
            kind: ScanSettingBinding
            name: scansettingbindings.compliance.openshift.io
            version: v1alpha1
          - description: ScanSetting is the Schema for the scansettings API
            kind: ScanSetting
            name: scansettings.compliance.openshift.io
            version: v1alpha1
          - description: TailoredProfile is the Schema for the tailoredprofiles API
            displayName: Tailored Profile
            kind: TailoredProfile
            name: tailoredprofiles.compliance.openshift.io
            version: v1alpha1
          - description: Variable describes a tunable in the XCCDF profile
            kind: Variable
            name: variables.compliance.openshift.io
            version: v1alpha1
        description: An operator which runs OpenSCAP and allows you to keep your cluster
          compliant with the security benchmark you need.
        displayName: Compliance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - compliance
        - openscap
        - audit
        links:
        - name: Compliance Operator upstream repository
          url: https://github.com/ComplianceAsCode/compliance-operator
        - name: Compliance Operator documentation
          url: https://docs.openshift.com/container-platform/latest/security/compliance_operator/co-overview.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: www.redhat.com
        relatedImages:
        - registry.redhat.io/compliance/openshift-compliance-openscap-rhel8@sha256:d0f89d82845c6245a18e7db345ac3be92b2863d57f55f63e21652086128920c9
        - registry.redhat.io/compliance/openshift-compliance-rhel8-operator@sha256:0a5629fd6f6b77cb36734eb663b847f51d45050532cb895c821e00416a1324e4
        - registry.redhat.io/compliance/openshift-compliance-operator-bundle@sha256:06e4abf4a94d44728b6bfa685b6ed027962b60fc2a0060a7256297857fd5f6fa
        - registry.redhat.io/compliance/openshift-compliance-content-rhel8@sha256:8d3257bd6a020df53f8d7a400ae877ae029c78ffcd8ec9e35270ee3ff5c8e509
        - registry.redhat.io/compliance/openshift-compliance-must-gather-rhel8@sha256:e6468ccc55d72539b7fb980d0bce5c072e85879d4e9db39974d53eaab4d32b25
        version: 1.7.1
      entries:
      - name: compliance-operator.v1.7.1
        version: 1.7.1
      - name: compliance-operator.v1.7.0
        version: 1.7.0
      - name: compliance-operator.v1.6.2
        version: 1.6.2
      - name: compliance-operator.v1.6.1
        version: 1.6.1
      - name: compliance-operator.v1.6.0
        version: 1.6.0
      - name: compliance-operator.v1.5.1
        version: 1.5.1
      - name: compliance-operator.v1.5.0
        version: 1.5.0
      - name: compliance-operator.v1.4.1
        version: 1.4.1
      - name: compliance-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: compliance-operator
    provider:
      name: Red Hat Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: https://www.percona.com/
    name: percona-postgresql-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:b1dc39979be1c347b123b2dd8070e4509179b7b499d4cf3fc7feb4552c6e2413
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        version: 2.7.0
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0
        version: 2.7.0
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1
        version: 2.6.1
      name: stable
    - currentCSV: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pgv2.percona.com/v2",
                "kind": "PerconaPGCluster",
                "metadata": {
                  "name": "cluster1"
                },
                "spec": {
                  "crVersion": "2.7.0",
                  "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da",
                  "imagePullPolicy": "Always",
                  "postgresVersion": 17,
                  "instances": [
                    {
                      "name": "instance1",
                      "replicas": 3,
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/data": "postgres"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      },
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      }
                    }
                  ],
                  "proxy": {
                    "pgBouncer": {
                      "replicas": 3,
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0",
                      "affinity": {
                        "podAntiAffinity": {
                          "preferredDuringSchedulingIgnoredDuringExecution": [
                            {
                              "weight": 1,
                              "podAffinityTerm": {
                                "labelSelector": {
                                  "matchLabels": {
                                    "postgres-operator.crunchydata.com/role": "pgbouncer"
                                  }
                                },
                                "topologyKey": "kubernetes.io/hostname"
                              }
                            }
                          ]
                        }
                      }
                    }
                  },
                  "backups": {
                    "pgbackrest": {
                      "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d",
                      "repoHost": {
                        "affinity": {
                          "podAntiAffinity": {
                            "preferredDuringSchedulingIgnoredDuringExecution": [
                              {
                                "weight": 1,
                                "podAffinityTerm": {
                                  "labelSelector": {
                                    "matchLabels": {
                                      "postgres-operator.crunchydata.com/data": "pgbackrest"
                                    }
                                  },
                                  "topologyKey": "kubernetes.io/hostname"
                                }
                              }
                            ]
                          }
                        }
                      },
                      "manual": {
                        "repoName": "repo1",
                        "options": [
                          "--type=full"
                        ]
                      },
                      "repos": [
                        {
                          "name": "repo1",
                          "schedules": {
                            "full": "0 0 * * 6"
                          },
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2",
                    "secret": "cluster1-pmm-secret",
                    "serverHost": "monitoring-service"
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          containerImage: registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
          createdAt: "2025-07-10T11:17:12.234Z"
          description: Percona Operator for PostgreSQL manages the lifecycle of Percona
            PostgreSQL cluster instances.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-postgresql-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-postgresql-operator
          support: percona.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PerconaPGCluster is the CRD that defines a Percona PG Cluster
            displayName: Percona PGCluster
            kind: PerconaPGCluster
            name: perconapgclusters.pgv2.percona.com
            version: v2
          - description: PerconaPGBackup is the CRD that defines a Percona PostgreSQL
              Backup
            displayName: Percona PGBackup
            kind: PerconaPGBackup
            name: perconapgbackups.pgv2.percona.com
            version: v2
          - description: PerconaPGRestore is the CRD that defines a Percona PostgreSQL
              Restore
            displayName: Percona PGRestore
            kind: PerconaPGRestore
            name: perconapgrestores.pgv2.percona.com
            version: v2
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API This Custom Resource requires enabling CrunchyBridgeClusters feature
              gate
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PerconaPGUpgrade is the Schema for the perconapgupgrades
              API
            displayName: Percona PGUpgrade
            kind: PerconaPGUpgrade
            name: perconapgupgrades.pgv2.percona.com
            version: v2
          - description: PGAdmin is the Schema for the pgadmins API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          ## Percona is Cloud Native


          Percona Operator for PostgreSQL automates and simplifies deploying and managing
          open source PostgreSQL clusters on Kubernetes. Percona Operator for PostgreSQL is based on Postgres Operator developed by Crunchy Data.

          Whether you need to get a simple PostgreSQL cluster up and running, need to deploy a high availability,
          fault tolerant cluster in production, or are running your own database-as-a-service, the Operator provides the essential
          features you need to keep your clusters healthy.

          Consult the
          [documentation](https://www.percona.com/doc/kubernetes-operator-for-postgresql/index.html)
          on the Percona Operator for PostgreSQL for complete details on capabilities
          and options.

          ### Supported Features

          * **PostgreSQL Cluster Provisioning**

          Create, Scale, & Delete PostgreSQL clusters with ease, while fully customizing your Pods and PostgreSQL configuration.

          * **High Availability**

              Safe, automated failover backed by a distributed consensus based high-availability solution. Uses Pod Anti-Affinity to help resiliency; you can configure how much
              enforced this can be. Failed primaries automatically heal, allowing for faster recovery time. Support for standby PostgreSQL clusters that work both within and
              across multiple Kubernetes clusters.

          * **Disaster Recovery**

              Backups and restores leverage the open source pgBackRest utility and includes support for full, incremental, and differential backups as well as efficient delta
              restores. Set how long you want your backups retained for. Works great with very large databases!

          * **Communication Security**

              Secure communication between your applications and data servers by enabling TLS for your PostgreSQL servers, including the ability to enforce all of your connections
              to use TLS.

          * **Monitoring**

              Track the health of your PostgreSQL clusters with Percona Monitoring and Management (PMM).

          * **PostgreSQL User Management**

              Quickly add and remove users from your PostgreSQL clusters with powerful commands. Manage password expiration policies or use your preferred PostgreSQL authentication
              scheme.

          * **Updates Management**

              Safely apply PostgreSQL updates with minimal availability impact to your PostgreSQL clusters.

          * **Advanced Replication Support**

              Choose between asynchronous replication and synchronous replication for workloads that are sensitive to losing transactions.

          * **Clone**

              Create new clusters from your existing clusters or backups.

          * **Connection Pooling**

              Use [pgBouncer](https://access.crunchydata.com/documentation/postgres-operator/latest/tutorial/pgbouncer/) for connection pooling

          * **Affinity and Tolerations**

              Have your PostgreSQL clusters deployed to Kubernetes Nodes of your preference with node affinity, or designate which nodes Kubernetes can schedule PostgreSQL instances
              to with Kubernetes  tolerations.

          * **Scheduled Backups**

              Choose the type of backup (full, incremental, differential) and how frequently you want it to occur on each PostgreSQL cluster.

          * **Backup to S3**

              Store your backups in Amazon S3 or any object storage system that supports the S3 protocol. The Operator can backup, restore, and create new clusters from these backups.

          * **Multi-Namespace Support**

              You can control how the Operator leverages Kubernetes Namespaces with several different deployment models:

              * Deploy the Operator and all PostgreSQL clusters to the same namespace
              * Deploy the Operator to one Namespace, and all PostgreSQL clusters to a different Namespace
              * Deploy the Operator to one Namespace, and have your PostgreSQL clusters managed across multiple Namespaces

          * **Full Customizability**

              The Operator not only makes it easy to get PostgreSQL up and running on Kubernetes-enabled platforms, but also allows you to further customize your deployments:

              * Selecting different storage classes for your primary, replica, and backup storage
              * Select your own container resources class for each PostgreSQL cluster deployment; differentiate between resources applied for primary and replica clusters
              * Use your own container image repository, including `imagePullSecrets` and private repositories  support
              * Customize your PostgreSQL configuration
              * Bring your own trusted certificate authority (CA) for use with the Operator API server
              * Override your PostgreSQL configuration for each cluster
              * Use your own custom images, re-define the image for each container separately
        displayName: Percona Operator for PostgreSQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - percona
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://docs.percona.com/percona-operator-for-postgresql/index.html
        - name: Github
          url: https://github.com/percona/percona-postgresql-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
          url: https://www.percona.com/
        relatedImages:
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:56534ca111d142ce0a59d34aff9b95ea229e2ba69fc4795f78acf4ef474e635d
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:cc5ce02f2dab691bea078d1c755661dfc04d34824dc310857104694552c127da
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2f74b1c4cf23d4fa95557b6b6ff40b2803095b8de2aa0dee2d33a5e5abcf02a0
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:d4b72d905f608244379b12960707f8ea190b9224c1b2723e3af017e260facbee
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:4c49d93231ba49046f32abdfe895bb02c37389ff24ea69972ce17809d2429037
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:babebf1e56af367f52119ed18bd9950e1584ad6e02095a7a6a7376b5f6756814
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:66c0a3c24930ad0e25ec54440f5cfcb145a3b4e23ccec8fb2de678fe440093b2
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:8acea1d25befc9a2d59f3969019ae42b75d3a84ed5ea1d8e7ab53a0a3c3c25fe
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:1d1c01cc53392a765459a6d912de1920e8802c2fe44d96aadf3aaa1399138481
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:5e0016fed626a1cebe4944e95e1002733b320dab7cdf34e58621431ce508a984
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:2d14c706b307d783f32161085c5ee4c06c7943f2eb5290419e46a1fc8f2769e9
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:02b4c6b0828d098126d665104c394fd2e83d87e98155c8fccc20e41c589161b7
        - registry.connect.redhat.com/percona/percona-postgresql-operator-containers@sha256:bd96ec33e2a4f44ddac387c847e21bf1813d52a1e0f9b541b359334bafe69568
        - registry.connect.redhat.com/percona/percona-postgresql-operator-bundle-marketplace@sha256:14fb2aa627ea7f38fba738cc4302cf1962660b10d3806c71161d8c7d64a8d060
        - registry.connect.redhat.com/percona/percona-postgresql-operator@sha256:986941a8c5f5d00a0c9cc7bd12acc9f78aa51fdcc98c7d0acddff05392d4b9a0
        version: 2.7.0-cw
      entries:
      - name: percona-postgresql-operator-certified-rhmp.v2.7.0-cw
        version: 2.7.0-cw
      - name: percona-postgresql-operator-certified-rhmp.v2.6.1-cw
        version: 2.6.1-cw
      name: stable-cw
    defaultChannel: stable
    packageName: percona-postgresql-operator-certified-rhmp
    provider:
      name: Percona
      url: https://www.percona.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Hitachi
      provider-url: ""
    name: hspc-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hspc-operator.v1.16.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.hitachi.com/v1",
                "kind": "HSPC",
                "metadata": {
                  "name": "hspc",
                  "namespace": "kube-system"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Storage
          containerImage: registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
          createdAt: "2025-08-26T06:53:32Z"
          description: An operator for managing Hitachi Storage Plug-in for Containers
            CSI driver
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          support: Hitachi
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HSPC is the Schema for the hspcs API
            displayName: HSPC
            kind: HSPC
            name: hspcs.csi.hitachi.com
            version: v1
        description: |-
          ## About
          Hitachi Storage Plug-in for Containers is a software plugin component for Kubernetes and OpenShift environments that is used to create and manage persistent block storage from Hitachi Virtual Storage Platform One storages.


          For full documentation, go to our [Product Documentation site](https://docs.hitachivantara.com/home) and refer to the reference guide for HSPC v3.16.1.

          ## Requirements

          ### Supported Driver Version

          * HSPC v3.16.1
        displayName: Hitachi Storage Plug-in for Containers
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Hitachi
        - Storage
        - Container Storage Interface
        - CSI
        links:
        - name: Product Documentation
          url: https://docs.hitachivantara.com/home
        maintainers:
        - email: Container-info@hitachivantara.com
          name: Hitachi
        maturity: stable
        provider:
          name: Hitachi
        relatedImages:
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-operator@sha256:2ae2786086617196f5352e7b78f445cfc1f3b32515138e8bcba7df9bc34d9b75
        - registry.k8s.io/sig-storage/csi-attacher@sha256:5aaefc24f315b182233c8b6146077f8c32e274d864cb03c632206e78bd0302da
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4a1e22167d03746e6748cdc41b6374262c666bc85999e90112edc236dfb2c95a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:57dba2ee519e49afacf899af7e265d977b02ec4c1f60c9b636ab0e575612dbfd
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.hitachivantara.com/hitachicsi-oci-oss/hspc-csi-driver@sha256:498bcfc3c56bf642aaa70aec4db61b8610af0a22e9e378dfba02356770566198
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:672e45d6a55678abc1d102de665b5cbd63848e75dc7896f238c8eaaf3c7d322f
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:a18e989a93722e43885120e90bc1d0da0740fcbf44bc10403572b368b9800606
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - registry.k8s.io/sig-storage/csi-resizer@sha256:5e7cbb63fd497fa913caa21fee1a69f727c220c6fa83c5f8bb0995e2ad73a474
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:49b94f975603d85a1820b72b1188e5b351d122011b3e5351f98c49d72719aa78
        - registry.k8s.io/sig-storage/csi-attacher@sha256:11b955fe4da278aa0e8ca9d6fd70758f2aec4b0c1e23168c665ca345260f1882
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:5244abbe87e01b35adeb8bb13882a74785df0c0619f8325c9e950395c3f72a97
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:bb057f866177d5f4139a1527e594499cbe0feeb67b63aaca8679dfdf0a6016f9
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:4c5a1b57e685b2631909b958487f65af7746361346fcd82a8635bea3ef14509d
        - registry.connect.redhat.com/hitachi/hspc-operator-bundle@sha256:2d5e1ed3782a992034bec29d85d24befed31ad5eaf7a21ea44acdd775f90cb73
        - registry.k8s.io/sig-storage/csi-resizer@sha256:8b9354b49c04564d9902e9b1ae6b323ba8f6ecd2d4a8e0bd4870192d1c586bd0
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:7b3d53cfefcb1c8836d5ac45f64dfdbf3fe6c813251b0b2ec79af7812cc53bfa
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:81d32545fb1b2c319bfeb36a70db4b4d3aea9c51c0574b209b8f64750b32d2f0
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        version: 1.16.1
      entries:
      - name: hspc-operator.v1.16.1
        version: 1.16.1
      name: stable
    defaultChannel: stable
    packageName: hspc-operator
    provider:
      name: Hitachi
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: operator-certification-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: operator-certification-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "certification.redhat.com/v1alpha1",
                "kind": "OperatorPipeline",
                "metadata": {
                  "name": "operatorpipeline-sample"
                },
                "spec": {
                  "applyCIPipeline": true,
                  "applyHostedPipeline": false,
                  "applyReleasePipeline": false,
                  "gitHubSecretName": "github-api-token",
                  "kubeconfigSecretName": "kubeconfig",
                  "operatorPipelinesRelease": "main",
                  "pyxisSecretName": "pyxis-api-secret"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-06-23T16:02:00Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io.bundle.channel.default.v1: alpha
          operators.operatorframework.io.bundle.channels: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/redhat-openshift-ecosystem/operator-certification-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: OperatorPipeline is the Schema for the operatorpipelines
              API
            displayName: Operator Pipeline
            kind: OperatorPipeline
            name: operatorpipelines.certification.redhat.com
            version: v1alpha1
        description: |-
          A Kubernetes operator to provision resources for the operator certification pipeline. This operator is installed in all namespaces which can support multi-tenant scenarios. **Note:** This operator should only be used by Red Hat partners attempting to certify their operator(s).


          **Requirements**


          The certification operator requires that you have the following tools installed, functional, and in your path.
          - [Install](https://docs.openshift.com/container-platform/4.8/cli_reference/openshift_cli/getting-started-cli.html#installing-openshift-cli) oc, the OpenShift CLI tool (tested with version 4.7.13)
          - [Install](https://tekton.dev/docs/cli/) tkn, the Tekton CLI tool (tested with version 0.19.1)
          - [Install](https://git-scm.com/downloads) git, the Git CLI tool (tested with 2.32.0)
          - The certification pipeline expects you to have the source files of your Operator bundle


          **Pre - Installation**


          The below steps exist for simplicity and to make the installation clearer. The operator watches all namespaces, so if secrets/configs/etc already exist in another namespace, feel free to use the existing namespace when following the operator installation steps.


          **Create a new namespace where the following secrets will be applied.**
          ```
          oc new-project oco
          ```


          **Add the kubeconfig secret which will be used to deploy the operator under test and run the certification checks.**
          * Open a terminal window
          * Set the `KUBECONFIG` environment variable
          ```
          export KUBECONFIG=/path/to/your/cluster/kubeconfig
          ```
          > *This kubeconfig will be used to deploy the Operator under test and run the certification checks.*
          ```
          oc create secret generic kubeconfig --from-file=kubeconfig=$KUBECONFIG
          ```


          **Configuring steps for submitting the results**
          - Add the github API token to the repo where the PR will be created
          ```
          oc create secret generic github-api-token --from-literal GITHUB_TOKEN=<github token>
          ```
          - Add RedHat Container API access key

            This API access key is specifically related to your unique partner account for Red Hat Connect portal. Instructions to obtain your API key can be found: [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/operator-cert-workflow.md#step-b---get-api-key)
          ```
          oc create secret generic pyxis-api-secret --from-literal pyxis_api_key=< API KEY >
          ```

          - Optional pipeline configurations can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#optional-configuration)


          **Execute the Pipeline (Development Iterations)**


          A pre-requisite to running a pipeline is that a `workspace-template.yaml` exists in the directory you want to execute the `tkn` commands from.

          To create a workspace-template.yaml
          ```
          cat <<EOF > workspace-template.yaml
          spec:
            accessModes:
              - ReadWriteOnce
            resources:
              requests:
                storage: 5Gi
          EOF
          ```

          There are multiple ways to execute the Pipeline which can be found [here](https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md#execute-the-pipeline-development-iterations)
        displayName: Operator Certification Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - certification
        - pipeline
        - openshift
        links:
        - name: Operator Certification CI Pipeline
          url: https://github.com/redhat-openshift-ecosystem/certification-releases/blob/main/4.9/ga/ci-pipeline.md
        - name: Red Hat OpenShift Pipelines
          url: https://github.com/openshift/tektoncd-operator
        maintainers:
        - email: certification-operator-devel@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.21.0
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/opdev/operator-certification-operator@sha256:7c84772f96650ff757c2552ba421355d54da41a5ed2155f72a14911b8a3f18ff
        - registry.connect.redhat.com/opdev/operator-certification-operator-bundle@sha256:c390f32b40449fb09477e288fe5347319b691d3e89e020163e32007e6679c685
        version: 1.3.1
      entries:
      - name: operator-certification-operator.v1.3.1
        version: 1.3.1
      name: alpha
    defaultChannel: alpha
    packageName: operator-certification-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: HPE Ezmeral
      provider-url: ""
    name: hpe-ezmeral-csi-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hpe-ezmeral-csi-operator.v1.0.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralCSIDriver",
                "metadata": {
                  "name": "hpeezmeralcsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1",
                  "pullPolicy": "Always"
                }
              },
              {
                "apiVersion": "ezmeral.hpe.com/v1",
                "kind": "HPEEzmeralNFSCSIDriver",
                "metadata": {
                  "name": "hpeezmeralnfscsidriver"
                },
                "spec": {
                  "controllerImage": "registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03",
                  "nodeImage": "registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a",
                  "pullPolicy": "Always"
                }
              }
            ]
          capabilities: Basic Install
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
          createdAt: "2025-06-25T11:51:05Z"
          description: A Container Storage Interface (CSI) driver for HPE Ezmeral
            Data Fabric platform. The CSI driver allows you to use HPE Ezmeral Data
            Fabric with your preferred container orchestrator.
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/mapr/mapr-csi
          support: HPE Ezmeral Data Fabric
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: HPE Ezmeral CSI Driver (FUSE)
            displayName: HPEEzmeralCSIDriver
            kind: HPEEzmeralCSIDriver
            name: hpeezmeralcsidrivers.ezmeral.hpe.com
            version: v1
          - description: HPE Ezmeral CSI Driver (NFS)
            displayName: HPEEzmeralNFSCSIDriver
            kind: HPEEzmeralNFSCSIDriver
            name: hpeezmeralnfscsidrivers.ezmeral.hpe.com
            version: v1
        description: |
          The HPE Ezmeral CSI Operator for Kubernetes packages, deploys, manages, upgrades the HPE Ezmeral CSI Driver on Kubernetes and OpenShift for dynamic provisioning of persistent volumes on HPE Ezmeral Data Fabric platform.
          The HPE Ezmeral CSI Driver for Kubernetes leverages HPE Ezmeral Data Fabric platform to provide scalable and persistent storage for stateful applications. Please refer to our CSI driver [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html) for full list of supported CSI features.
          ## Installation
          Refer to the HPE Ezmeral CSI Operator for Kubernetes [documentation](https://docs.datafabric.hpe.com/home/CSIdriver/csi_installing_operator.html)
        displayName: HPE Ezmeral Data Fabric CSI Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - HPE
        - HPE Ezmeral
        - HPE Ezmeral Data Fabric
        - Storage
        - FileSystem
        - BlockVolume
        links:
        - name: Documentation
          url: https://docs.datafabric.hpe.com/home/CSIdriver/csi_overview.html
        maintainers:
        - email: ezmeral-csi-support@hpe.com
          name: HPE Ezmeral Data Fabric
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: HPE Ezmeral
        relatedImages:
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator@sha256:c9c243c51cd760c2c2a66b489807a5e22053fcbeb94fa29afd10d9c91e2d3908
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ba961c2c2a29750c0132fe6dd6fa9f6001010afbc5f19b98add87b31b54bcf6
        - registry.connect.redhat.com/maprtech/csi-kdfplugin@sha256:ffd273ee2e608d2ef64fd3b2bcbb02d5f0a6c173ed96a15eaf5549ea45fea8d1
        - registry.connect.redhat.com/maprtech/csi-kdfprovisioner@sha256:e81b689b3bf92cd2c0076ed48c6aa186f7a1b914854f37ed4c08dc37720dbd03
        - registry.connect.redhat.com/maprtech/csi-nfsplugin@sha256:a5c954f62f1395bafe2d68454ac604bca4be97573a3482fcbd8e76b6361cc08a
        - registry.connect.redhat.com/maprtech/hpe-ezmeral-csi-operator-bundle@sha256:4629bd129addb09478e48964a62acee88f5dd2746875801bd19d2d55cadd9da2
        version: 1.0.18
      entries:
      - name: hpe-ezmeral-csi-operator.v1.0.18
        version: 1.0.18
      name: stable
    defaultChannel: stable
    packageName: hpe-ezmeral-csi-operator
    provider:
      name: HPE Ezmeral
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-test-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-test-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Testcore",
                "metadata": {
                  "name": "testcore-sample"
                },
                "spec": {
                  "agentController": {
                    "port": 8080
                  },
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "dataseqService": {
                    "ingressEnabled": true,
                    "port": 8080,
                    "providers": {
                      "loadCore": {
                        "enabled": true
                      }
                    }
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "ingress": {
                    "backwardsCompatibilityEnabled": true
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "systemMonitor": {
                    "enabled": false
                  },
                  "trafficController": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:35:34Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Testcore
            name: testcores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Test Core Operator
        displayName: Keysight WAP Test Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Test Core
          url: https://keysight-wap-test-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-test-core@sha256:9e2d81061abff49112f908e71dace03a748e283f32597aa22ad58471fff30caf
        - registry.connect.redhat.com/loadcore/keysight-wap-test-core-bundle@sha256:16c938699ab0d71138693641ba7dcce2f4304da1d357a67a057c4effb4300401
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-test-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-test-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ricardo Maraschini
      provider-url: https://github.com/ricardomaraschini/tagger
    name: tagger
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tagger.v2.1.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "tagger.dev/v1beta1",
                "kind": "Tag",
                "metadata": {
                  "name": "nginx"
                },
                "spec": {
                  "from": "docker.io/library/nginx:latest",
                  "generation": 0,
                  "mirror": false
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Developer Tools
          containerImage: quay.io/tagger/operator:v2.1.17
          createdAt: "2021-08-02T00:00:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/ricardomaraschini/tagger
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Tag is a Container Image representation as a Custom Resource.
            displayName: Tag
            kind: Tag
            name: tags.tagger.dev
            version: v1beta1
        description: |-
          A Container Image Orchestrator. Adds mirroring and Container Image manipulation capabilities
          to a Kubernetes cluster.

          Keeping track of all Container Images in use in a Kubernetes cluster is a complicated task.
          Container Images may come from numerous different Image Registries. In some cases, controlling
          how a stable version of a given Container Image looks escapes the user's authority. To add to
          this, Container Runtimes rely on remote registries (from the cluster's point of view) when
          obtaining Container Images, potentially making the process of pulling their blobs (manifests,
          config, and layers) slower.

          The notion of indexing Container Image versions by Tags is helpful. Still, it does not provide
          users with the right confidence to always use the intended Container Image – today's "latest"
          tag might not be tomorrow's "latest" tag. In addition to that, these Image Registries allow
          access to Container Images by their Manifest content's hash (i.e., usually sha256), which gives
          users the confidence at a cost in semantics.

          When releasing a new version of an application to Push and to Deploy are split into two
          distinct steps. Both the pusher and the puller need access to the same Image Registry, adding
          complexity.  Credentials are one example of the concerns. Other factors may pop up when
          running, for instance, in an air-gapped environment, where the cluster may not reach external
          Image Registries.

          Tagger aims to overcome these caveats by providing an image management abstraction layer. For
          instance, by providing a direct mapping between a Container Image tag (e.g., "latest") and its
          correspondent Manifest content's hash, users can then refer to the Container Image by its tag
          – and be sure to use that specific version. More than that, when allied with an Internal Image
          Registry, Tagger can also automatically mirror Container Images into the cluster.

          While using Tagger, Deployments can refer to Container Images by an arbitrarily defined name,
          such as "my-app", and Tagger will make sure that they use the right Container Image through its
          internal "tag to Manifest content's hash" mapping.

          For each new "release" of a given Container Image, Tagger creates a new Generation for it,
          making it easy to roll back to previously pushed "releases" of the same Container Image in case
          of problems.

          When integrated with an Internal Registry, Tagger allows users to push or pull Images directly
          without requiring an external Image Registry. It works as a layer between the user and the
          Internal Registry. Every time a new "release" of Container Images is pushed, all Deployments
          are updated automatically. Users don't need to know about the Internal Registry existence, if
          they are logged-in to the Kubernetes cluster, they can obtain old or provide new "Generations"
          for a Container Image.

          In summary, Tagger mirrors remote Container Images into a Kubernetes cluster indexing them in
          different Generations (allowing easy navigation through these multiple Generations), provides
          an interface allowing users to pull and push images directly to the Kubernetes cluster and
          provides full integration with Kubernetes Deployments (automatic triggers new rollouts on
          Container Image changes).

          By default Tagger won't be able to mirror until you provide it with a mirror registry config.
          After install you can configure the mirror by editing the Secret mirror-registry-config in the
          operator namespace. Follow below an example of a valid mirror-registry-config secret (you gonna
          have to provide your own address, username, password, etc).

          ```yaml
          apiVersion: v1
          kind: Secret
          metadata:
            name: mirror-registry-config
          data:
            address: cmVnaXN0cnkuaW8=
            username: YWRtaW4=
            password: d2hhdCB3ZXJlIHlvdSB0aGlua2luZz8K
            token: YW4gb3B0aW9uYWwgdG9rZW4gZm9yIHRva2VuIGJhc2VkIGF1dGg=
            insecure: dHJ1ZQ==
          ```

          To make pushing an pulling images easier you should go ahead and grab the `kubectl` plugin
          available [here](https://github.com/ricardomaraschini/tagger/releases). You can download a
          full version of the README with usage examples and configuration options
          [here](https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf).
        displayName: Tagger
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - tagger
        - mirroring
        - image
        links:
        - name: Documentation
          url: https://github.com/ricardomaraschini/tagger/releases/download/v2.1.17/README.pdf
        - name: Plugin (kubectl)
          url: https://github.com/ricardomaraschini/tagger/releases/tag/v2.1.17
        maintainers:
        - email: rmarasch@redhat.com
          name: Ricardo Maraschini
        maturity: alpha
        provider:
          name: Ricardo Maraschini
          url: https://github.com/ricardomaraschini/tagger
        relatedImages:
        - quay.io/openshift-community-operators/tagger@sha256:a528169d305dbc2bc69794437665e4f940ea708dfd957c58d69bc79316b68612
        - quay.io/tagger/operator:v2.1.17
        version: 2.1.17
      entries:
      - name: tagger.v2.1.17
        version: 2.1.17
      name: alpha
    defaultChannel: alpha
    packageName: tagger
    provider:
      name: Ricardo Maraschini
      url: https://github.com/ricardomaraschini/tagger
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: GitLab Inc
      provider-url: https://about.gitlab.com/
    name: gitlab-operator-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: stable
    - currentCSV: gitlab-operator-kubernetes.v2.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.gitlab.com/v1beta1",
                "kind": "GitLab",
                "metadata": {
                  "name": "gitlab",
                  "namespace": "gitlab-system"
                },
                "spec": {
                  "chart": {
                    "values": {
                      "gitlab": {
                        "gitlab-shell": {
                          "maxReplicas": 2,
                          "minReplicas": 1
                        },
                        "sidekiq": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1000M"
                            }
                          }
                        },
                        "webservice": {
                          "maxReplicas": 2,
                          "minReplicas": 1,
                          "resources": {
                            "requests": {
                              "cpu": "500m",
                              "memory": "1500M"
                            }
                          }
                        }
                      },
                      "global": {
                        "hosts": {
                          "domain": "example.com",
                          "hostSuffix": null
                        },
                        "ingress": {
                          "configureCertmanager": false,
                          "tls": {
                            "secretName": null
                          }
                        }
                      },
                      "installCertmanager": false,
                      "minio": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      },
                      "nginx-ingress": {
                        "controller": {
                          "electionID": "ingress-controller-leader",
                          "replicaCount": 1
                        }
                      },
                      "postgresql": {
                        "primary": {
                          "extendedConfiguration": "max_connections = 200"
                        }
                      },
                      "redis": {
                        "resources": {
                          "requests": {
                            "cpu": "100m"
                          }
                        }
                      }
                    },
                    "version": "9.4.1"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Developer Tools
          certified: "true"
          containerImage: registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
          createdAt: "2025-09-29T12:42:23Z"
          description: The GitLab operator is responsible for managing the full lifecycle
            of GitLab instances in your Kubernetes or Openshift container platforms.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          operatorframework.io/suggested-namespace: gitlab-system
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: unknown
          repository: https://gitlab.com/gitlab-org/cloud-native/gitlab-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GitLab is a complete DevOps platform, delivered in a single
              application
            displayName: GitLab
            kind: GitLab
            name: gitlabs.apps.gitlab.com
            version: v1beta1
        description: |
          # Overview

          **Installation using OLM is considered experimental.** GitLab does not support any issues related to instances deployed using OLM.
          For more information on potential issues with OLM, see [issue 241](https://gitlab.com/gitlab-org/cloud-native/gitlab-operator/-/issues/241).

          The GitLab operator is responsible for managing the full lifecycle of GitLab instances in your Kubernetes or Openshift container platforms.

          [Documentation](https://docs.gitlab.com/operator/)

          The operator, while new and still actively being developed, aims to:
          - ease installation and configuration of GitLab instances
          - offer seamless upgrades from version to version

          ## GitLab

          GitLab is a complete open-source DevOps platform, delivered as a single application, fundamentally changing the way Development, Security, and Ops teams collaborate and build software. From idea to production, GitLab helps teams improve cycle time from weeks to minutes, reduce development process costs and decrease time to market while increasing developer productivity.

          Built on Open Source, GitLab delivers new innovations and features on the same day of every month by leveraging contributions from a passionate, global community of thousands of developers and millions of users. Over 100,000 of the world’s most demanding organizations trust GitLab to deliver great software at new speeds.

          If you would like to enable advanced DevOps capabilities and activate enterprise features such as security, risk, and compliance capabilities, please contact our sales team to purchase an enterprise license.

          # Prerequisites

          Please visit [Prerequisites](https://docs.gitlab.com/operator/installation/#prerequisites) section of GitLab Operator Documentation.

          ## IngressClass

          Cluster-wide `IngressClass` should be created prior to Operator setup, as OLM does not currently support this object type:

          ```yaml
          apiVersion: networking.k8s.io/v1
          kind: IngressClass
          metadata:
            # Ensure this value matches `spec.chart.values.global.ingress.class`
            # in the GitLab CR on the next step.
            name: gitlab-nginx
          spec:
            controller: k8s.io/ingress-nginx
          ```
        displayName: GitLab
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - GitLab
        - DevOps
        - CI/CD
        - DAST
        - SAST
        links:
        - name: GitLab Docs
          url: https://docs.gitlab.com/
        - name: GitLab Operator Documentation
          url: https://docs.gitlab.com/operator/
        - name: GitLab and Kubernetes
          url: https://about.gitlab.com/solutions/kubernetes/
        - name: Gitlab Reference Architecture
          url: https://docs.gitlab.com/administration/reference_architectures/
        - name: GitLab Contact Sales
          url: https://about.gitlab.com/sales/
        maintainers:
        - email: distribution@gitlab.com
          name: GitLab Distribution Team
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: GitLab Inc
          url: https://about.gitlab.com/
        relatedImages:
        - registry.connect.redhat.com/gitlab/gitlab-operator-bundle@sha256:c4fc34be1dc7480a4c40e64ccbd75ef1e9eceeee46c49445200b55a3543b2317
        - registry.gitlab.com/gitlab-org/cloud-native/gitlab-operator@sha256:3d01778513befd1573c41bfd7dba9a34c073a334e7973c0b926992bf14cef0c9
        version: 2.4.1
      entries:
      - name: gitlab-operator-kubernetes.v2.4.1
        version: 2.4.1
      - name: gitlab-operator-kubernetes.v2.4.0
        version: 2.4.0
      - name: gitlab-operator-kubernetes.v2.3.2
        version: 2.3.2
      - name: gitlab-operator-kubernetes.v2.3.1
        version: 2.3.1
      - name: gitlab-operator-kubernetes.v2.3.0
        version: 2.3.0
      - name: gitlab-operator-kubernetes.v2.2.4
        version: 2.2.4
      - name: gitlab-operator-kubernetes.v2.2.3
        version: 2.2.3
      - name: gitlab-operator-kubernetes.v2.2.2
        version: 2.2.2
      - name: gitlab-operator-kubernetes.v2.2.1
        version: 2.2.1
      - name: gitlab-operator-kubernetes.v2.2.0
        version: 2.2.0
      - name: gitlab-operator-kubernetes.v2.1.2
        version: 2.1.2
      name: unstable
    defaultChannel: stable
    packageName: gitlab-operator-kubernetes
    provider:
      name: GitLab Inc
      url: https://about.gitlab.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Advanced Micro Devices, Inc.
      provider-url: ""
    name: amd-gpu-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: amd-gpu-operator.v1.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "amd.com/v1alpha1",
                "kind": "DeviceConfig",
                "metadata": {
                  "name": "test-deviceconfig",
                  "namespace": "openshift-amd-gpu"
                },
                "spec": {
                  "devicePlugin": {
                    "devicePluginImage": "rocm/k8s-device-plugin:rhubi-latest",
                    "nodeLabellerImage": "rocm/k8s-node-labeller:rhubi-latest"
                  },
                  "driver": {
                    "enable": "true",
                    "version": "6.4.2"
                  },
                  "selector": {
                    "feature.node.kubernetes.io/amd-gpu": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning,Monitoring
          containerImage: docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
          createdAt: "2025-08-01T01:55:48Z"
          description: |-
            Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
            For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
          deviceConfigManagerImage: docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
          devicePluginImage: docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          metricsExporterImage: docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
          nodelabellerImage: docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-amd-gpu
          operators.openshift.io/valid-subscription: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/ROCm/gpu-operator
          support: Advanced Micro Devices, Inc.
          testRunnerImage: docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
          utilsContainerImage: docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DeviceConfig describes how to enable AMD GPU device
            displayName: DeviceConfig
            kind: DeviceConfig
            name: deviceconfigs.amd.com
            version: v1alpha1
        description: |-
          Operator responsible for deploying AMD GPU kernel drivers, device plugin, device test runner and device metrics exporter
          For more information, visit [documentation](https://instinct.docs.amd.com/projects/gpu-operator/en/latest/)
        displayName: amd-gpu-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AMD
        - GPU
        - AI
        - Deep Learning
        - Hardware
        - Driver
        - Monitoring
        links:
        - name: AMD GPU Operator
          url: https://github.com/ROCm/gpu-operator
        maintainers:
        - email: Yan.Sun3@amd.com
          name: Yan Sun
        - email: farshad.ghodsian@amd.com
          name: Farshad Ghodsian
        - email: shrey.ajmera@amd.com
          name: Shrey Ajmera
        maturity: stable
        provider:
          name: Advanced Micro Devices, Inc.
        relatedImages:
        - docker.io/rocm/k8s-node-labeller@sha256:75f916c5a407e40e9ffdd26cdfc65b7da855710568937e740a969094e2991065
        - docker.io/rocm/test-runner@sha256:98e698263b4df26959341b14b3ddded9df51e504e7c6819d5c7c56c434855e9d
        - registry.connect.redhat.com/amd/gpu-operator-v1@sha256:8a92a0d18a5ae6a837ca6590c2beb65d889cd744a098c22d3f01d2fffdc4dc57
        - docker.io/rocm/device-config-manager@sha256:d2896271eacfc664ac7651425f46a85af4f4ef98d713988c2a4fee4a846548ee
        - docker.io/rocm/device-metrics-exporter@sha256:6c946c6ca10d5c0f40e67b5759759f12d51c45e3b934cb1af912821d5fd76e6e
        - docker.io/rocm/gpu-operator-utils@sha256:9cbe60dce5b3a2ffd93f7148a363ed5f072a320377d5e5bcbbfb1e05d4959287
        - docker.io/rocm/gpu-operator@sha256:1583d4ec172d50d4416f4030eb85910ec54633b7e8e061d4bd79b14dbf98d28c
        - docker.io/rocm/k8s-device-plugin@sha256:840c4ea1bf58df60ef3e80e3bf27bbf620a393854e42f58b2224b288767e3dd4
        version: 1.3.2
      entries:
      - name: amd-gpu-operator.v1.3.2
        version: 1.3.2
      - name: amd-gpu-operator.v1.3.1
        version: 1.3.1
      - name: amd-gpu-operator.v1.2.1
        version: 1.2.1
      - name: amd-gpu-operator.v1.1.1
        version: 1.1.1
      name: alpha
    defaultChannel: alpha
    packageName: amd-gpu-operator
    provider:
      name: Advanced Micro Devices, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/community-operator-pipeline-prod/mongodb-atlas-kubernetes@sha256:f12108323fc990f869394e09b4ff6d9751d0a07f3c152fbd45b0f738746a468b
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator:2.11.0
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: nfd
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "imagePullPolicy": "IfNotPresent",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
          createdAt: "2025-04-23T07:03:43Z"
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.9.0 <4.18.0-202509100344'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - kind: NodeFeatureGroup
            name: nodefeaturegroups.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - kind: NodeFeature
            name: nodefeatures.nfd.openshift.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.18/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: support@redhat.com
          name: Red Hat Support
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-nfd-operator-bundle@sha256:ffb2d53dfdc63bfc9320bc4e927d037244111ab0cdbe5721ad69ce06792d9753
        - registry.redhat.io/openshift4/ose-cluster-nfd-rhel9-operator@sha256:e53a29c3de1f68c5ae32713e6f8176b6baa076be8277d9fb0b55812b52ef6f80
        - registry.redhat.io/openshift4/ose-node-feature-discovery-rhel9@sha256:a1a60b53db81b32de0127dbb63f8c3a9ef7f70c95968aa53ffe8b49eced7dc43
        version: 4.18.0-202509100344
      entries:
      - name: nfd.4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: nfd.4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: nfd.4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: nfd.4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: nfd.4.18.0-202506230505
        version: 4.18.0-202506230505
      - name: nfd.4.18.0-202506092139
        version: 4.18.0-202506092139
      - name: nfd.4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: nfd.4.18.0-202505052133
        version: 4.18.0-202505052133
      - name: nfd.4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: nfd.4.18.0-202504232132
        version: 4.18.0-202504232132
      - name: nfd.4.18.0-202504151633
        version: 4.18.0-202504151633
      - name: nfd.4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: nfd.4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: nfd.4.18.0-202503210101
        version: 4.18.0-202503210101
      - name: nfd.4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: nfd.4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: nfd.4.18.0-202502101733
        version: 4.18.0-202502101733
      name: stable
    defaultChannel: stable
    packageName: nfd
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: submariner.io
      provider-url: ""
    name: submariner
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: submariner.v0.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
          createdAt: "2025-04-22 21:12:45"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.19.4'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.19 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:70008af49a9d1ee516204715f66bd415c9caa7462f6b178dc1dc91e513605fcd
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:5938d67a0a2653e92f0995440700512992ed82269820679f68483dcc1b2d98d9
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:f6de1136ced78ddf1ff48fdadc259c251919dd3a70819fe0e9247411999d51f3
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:4afaf6689e2225c5c7f6fbe002202a1cdf9444e23fd210cc2d86da71c9f78626
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:ef73b8964cd66ffb496191056594e0886afa9d17c03976b4159413c89ed569bf
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:1bbb624bbefbb75407071449d0144258732e219e3a7fc67ac709c3bcee560576
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:d9d4d03013d65d11a4e8ecda411015f359e4215a3a53059a37a810acf2e3ea6c
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5b83c5889aa0d8f7838dd5afb8c901dfb0ce0fbe01175e8468ec50ddedef9223
        version: 0.19.4
      entries:
      - name: submariner.v0.19.4
        version: 0.19.4
      - name: submariner.v0.19.3
        version: 0.19.3
      - name: submariner.v0.19.2
        version: 0.19.2
      - name: submariner.v0.19.0
        version: 0.19.0
      name: stable-0.19
    - currentCSV: submariner.v0.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "registry.redhat.io/rhacm2",
                  "version": "$(VERSION)"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "registry.redhat.io/rhacm2",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v$(VERSION)"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
          createdAt: "2025-05-15 07:58:52"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.18.0 <0.20.1'
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |-
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.20 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```

          ## Support & Troubleshooting

          Full product support, including Support Cases, Product Pages and Knowledgebase, comes through a
          [Red Hat Advanced Cluster Management](https://www.redhat.com/en/technologies/management/advanced-cluster-management) subscription.
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:0885c82cd09105a078d4308b4ec1047b96f3cbf1fd555a1c90a8b053befee27e
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:5783a00c386858d1eb45f224431d07d40bb9be746de965e7b129ee30969d4014
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:5e822b12b5aadaa08a2f95b879be911455607a9c04922a641625515f955dc9ca
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:2217d9128fe50d61d1dff48b588113b5267002f72d025b1b6fbf87a1c2e670d1
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:90505e5ffaee7af330306c4e045d3755e0c74d30b9f45e1e7739c205d945872d
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:9bd4cce195d28beee61f1adf9144b16c4de4c14bb5b5c4046921b730d758ae93
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:2e18b858f09e7595a89c4dc6ea5a85f8c5952f581b769093bfaab8156bab0813
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:6b4d34e7a40b21cf464e2f42cef21ef72a69ae8a3f4e7600a7bd85aa689117d9
        version: 0.20.1
      entries:
      - name: submariner.v0.20.1
        version: 0.20.1
      - name: submariner.v0.20.0
        version: 0.20.0
      name: stable-0.20
    - currentCSV: submariner.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Broker",
                "metadata": {
                  "name": "submariner-broker"
                },
                "spec": {
                  "components": [
                    "service-discovery",
                    "connectivity"
                  ],
                  "defaultGlobalnetClusterSize": 8192,
                  "globalnetEnabled": false
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "ServiceDiscovery",
                "metadata": {
                  "name": "servicediscovery"
                },
                "spec": {
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "clusterID": "cluster1",
                  "debug": false,
                  "globalnetEnabled": false,
                  "namespace": "submariner-operator",
                  "repository": "quay.io/submariner",
                  "version": "v0.21.0"
                }
              },
              {
                "apiVersion": "submariner.io/v1alpha1",
                "kind": "Submariner",
                "metadata": {
                  "name": "submariner"
                },
                "spec": {
                  "broker": "k8s",
                  "brokerK8sApiServer": "192.168.67.110:8443",
                  "brokerK8sApiServerToken": "$(BROKER_K8S_API_TOKEN)",
                  "brokerK8sCA": "$(BROKER_K8S_CA)",
                  "brokerK8sRemoteNamespace": "submariner-k8s-broker",
                  "cableDriver": "libreswan",
                  "ceIPSecDebug": false,
                  "ceIPSecIKEPort": 500,
                  "ceIPSecNATTPort": 4500,
                  "ceIPSecPSK": "$(IPSEC_PSK)",
                  "clusterCIDR": "192.168.67.0/24",
                  "clusterID": "cluster1",
                  "connectionHealthCheck": {
                    "enabled": false,
                    "intervalSeconds": 1,
                    "maxPacketLossCount": 5
                  },
                  "debug": false,
                  "globalCIDR": "",
                  "namespace": "submariner-operator",
                  "natEnabled": true,
                  "repository": "quay.io/submariner",
                  "serviceCIDR": "192.168.66.0/24",
                  "serviceDiscoveryEnabled": true,
                  "version": "v0.21.0"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: $(SUBMARINER_OPERATOR_IMAGE)
          createdAt: "2025-08-13 21:57:14"
          description: Creates and manages Submariner deployments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: submariner-operator
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Advanced Cluster Management for Kubernetes"]'
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/submariner-io/submariner-operator
          support: Submariner
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Broker is the Schema for the brokers API.
            displayName: Submariner Broker
            kind: Broker
            name: brokers.submariner.io
            version: v1alpha1
          - description: ServiceDiscovery is the Schema for the servicediscoveries
              API.
            displayName: Service Discovery
            kind: ServiceDiscovery
            name: servicediscoveries.submariner.io
            version: v1alpha1
          - description: Submariner is the Schema for the submariners API.
            displayName: Submariner
            kind: Submariner
            name: submariners.submariner.io
            version: v1alpha1
        description: |
          [Submariner](https://submariner.io) enables direct networking between Pods and Services in different Kubernetes
          clusters. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions.

          Submariner is designed to be cloud provider and network plugin (CNI) agnostic.

          Submariner consists of several main components that work in conjunction to securely connect workloads across
          multiple Kubernetes clusters, both on-premise and on public clouds:
          * [Gateway Engine](https://submariner.io/getting-started/architecture/gateway-engine/): manages the secure tunnels to other clusters.
          * [Route Agent](https://submariner.io/getting-started/architecture/route-agent/): routes cross-cluster traffic from nodes
          to the active Gateway Engine.
          * [Broker](https://submariner.io/getting-started/architecture/broker/): facilitates the exchange of metadata between Gateway
          Engines enabling them to discover one another.

          Submariner has optional components that provide additional functionality:
          * [Globalnet Controller](https://submariner.io/getting-started/architecture/globalnet/): handles overlapping CIDRs across clusters.
          * [Service Discovery](https://submariner.io/getting-started/architecture/service-discovery/): provides DNS discovery of services
          across clusters.

          ### Prerequisites
          Submariner has a few requirements to get started, all are documented [here](https://submariner.io/getting-started/quickstart/kind/#prerequisites).

          If running on an OpenShift cluster, perform the following steps to bind the Submariner service accounts to the **privileged** SCC.

          ```shell
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-gateway
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-routeagent
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-globalnet
          oc adm policy add-scc-to-user privileged system:serviceaccount:submariner-operator:submariner-lighthouse-coredns
          ```

          ### Deployment
          Submariner provides an [Operator](https://github.com/submariner-io/submariner-operator) for easy API-based
          installation and management.
          A command line utility, [subctl](https://github.com/submariner-io/submariner-operator/releases), wraps the
          Operator to aid users with manual deployments and easy experimentation.
          subctl greatly simplifies the deployment of Submariner, and is therefore the recommended deployment method.
          For complete information about subctl, please refer to [this page](https://submariner.io/operations/deployment/subctl).
          In addition to Operator and subctl, Submariner also provides [Helm Charts](https://submariner.io/operations/deployment/helm).

          On the broker cluster:
          1. Create the broker namespace called `submariner-k8s-broker`.
          2. Apply the broker RBAC definitions:

              ```shell
              git clone --depth 1 --single-branch --branch release-0.21 https://github.com/submariner-io/submariner-operator
              kubectl apply -k submariner-operator/config/broker -n submariner-k8s-broker
              ```

          3. Install the Submariner Operator from the catalog.
          4. Create a new `Broker` custom resource named **submariner-broker**.

          On connected clusters:
          1. Install the Submariner Operator from the catalog.
          2. Create a new `Submariner` custom resource named **submariner**.

          The following values can be specified for the **Broker** CR:
          * **defaultCustomDomains**: List of domains to use for multicluster service discovery.
          * **defaultGlobalnetClusterSize**: Default cluster size for GlobalCIDR allocated to each cluster (amount of global IPs).
          * **globalnetEnabled**: Enable support for Overlapping CIDRs in connecting clusters (default disabled).
          * **globalnetCIDRRange**: GlobalCIDR supernet range for allocating GlobalCIDRs to each cluster.
          * **components**: List of components to be installed [service-discovery, connectivity].

          The following values can be specified for the **Submariner** CR:
          * **namespace**: The namespace to install in.
          * **serviceCIDR**: The service CIDR.
          * **clusterCIDR**: The cluster CIDR.
          * **clusterID**: The cluster ID used to identify the tunnels.
          * **colorCodes**: The color code(s) for the cluster.
          * **debug**: *'true'* to enable operator debugging information - verbose logging.
          * **natEnabled**: *'true'* if there is NAT between clusters. Typically *'true'* in public cloud or mixed on-prem/public cloud environments and *'false'* when all clusters are on-prem.
          * **broker**: This must be *k8s* currently.
          * **brokerK8sApiServer**: The broker API server URL.
              To retrieve it run the following command on the broker cluster:

              `kubectl -n default get endpoints kubernetes -o
              jsonpath="{.subsets[0].addresses[0].ip}:{.subsets[0].ports[?(@.name=='https')].port}"`

          * **brokerK8sApiServerToken**: The broker API server token.
              To retrieve the token run the following command on the broker cluster, assuming the broker is running in the
              *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data.token != null)] | .[0].data.token' | base64 --decode`

          * **brokerK8sRemoteNamespace**: The broker namespace.
          * **brokerK8sCA**: The broker certificate authority.
              To retrieve the broker CA run the following command on the broker cluster, assuming the broker is running in
              the *submariner-k8s-broker* namespace:

              `kubectl -n submariner-k8s-broker get secrets -o json
              | jq -r -c '[.items[] | select(.metadata.annotations."kubernetes.io/service-account.name"=="submariner-k8s-broker-client")
              | select(.data."ca.crt" != null)] | .[0].data."ca.crt"' | base64 --decode`

          * **ceIPSecPSK**: The IPsec Pre-Shared Key which must be identical in all route agents across the cluster.
              To generate a unique PSK run the following command on your machine:

              `dd if=/dev/urandom count=64 bs=8 | LC_CTYPE=C tr -dc 'a-zA-Z0-9' | fold -w 64 | head -n 1`

          * **ceIPSecDebug**: *'true'* to enable logging IPsec debugging information - verbose logging from the IPsec daemon.
          * **ceIPSecIKEPort**: The IPsec IKE port (500 usually).
          * **ceIPSecNATTPort**: The IPsec NAT traversal port (4500 usually).
          * **repository**: The container repository to use.
          * **version**: The container version to use.
          * **imageOverrides**: A string map that overrides the images by defining a path for each component.
                                This field is meant to be used by CI, development or testing.

            **example:**

            ```yaml
            imageOverrides:
              submariner-route-agent=my_repo/my_custom_image:my_tag
              submariner-globalnet=my_repo/my_custom_image:my_tag
            ```
        displayName: Submariner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - multi-cluster
        - ipsec
        - tunnel
        - submariner-operator
        links:
        - name: Website
          url: https://submariner.io
        - name: GitHub
          url: https://github.com/submariner-io
        - name: Slack
          url: https://kubernetes.slack.com/archives/C010RJV694M
        - name: YouTube
          url: https://www.youtube.com/channel/UCZ3brSgl2v4boglZoeChClQ/videos
        - name: Twitter
          url: https://twitter.com/submarinerio
        maintainers:
        - email: submariner-dev@googlegroups.com
          name: Submariner project
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: submariner.io
        relatedImages:
        - registry.redhat.io/rhacm2/submariner-globalnet-rhel9@sha256:b307f748125fd459c3474fdab645688675a06515a9924b47cedba80e038458db
        - registry.redhat.io/rhacm2/submariner-rhel9-operator@sha256:65765c7f08c5454b557fbda91c49efbd2cc0d4d973cf8acc7ff40ec8a381839d
        - registry.redhat.io/rhacm2/submariner-route-agent-rhel9@sha256:f5909b1802986622ab169a82d5dd7877af72660903e961b3097cc29e30eb9799
        - registry.redhat.io/rhacm2/submariner-operator-bundle@sha256:e533406d832461cd8889f5e95569ea9f8dbbd57b51dcd08b56050ba1e3775ad0
        - registry.redhat.io/rhacm2/lighthouse-agent-rhel9@sha256:0b74f058400f24a63cf9b57a388914cc6dda38e03d639196871f3fd76fb7f2ae
        - registry.redhat.io/rhacm2/lighthouse-coredns-rhel9@sha256:cd851eab4a2214ea7a699bd01d2abac1322d01562f27b35f58f3c0b8716f88fb
        - registry.redhat.io/rhacm2/nettest-rhel9@sha256:bcdeaa67664568c5e2a6928236ff69b175ce65ddf7a58c1664d091b926000052
        - registry.redhat.io/rhacm2/submariner-gateway-rhel9@sha256:b0ad338aed82367048cf024767b0b09e6bb09bbf7dd5128e6711e1feec6f696e
        version: 0.21.0
      entries:
      - name: submariner.v0.21.0
        version: 0.21.0
      name: stable-0.21
    defaultChannel: stable-0.21
    packageName: submariner
    provider:
      name: submariner.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: fuse-apicurito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fuse-apicurito.v7.10.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
          createdAt: 2022-02-28 09:35:38 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.10.3'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:8d19ad579f874d6b9a08f127c8900906913a5a7c1e69cbe4b5cb0333b8d86152
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:8dbeeb8b04c81f41d896f71db90813de57e67493fa85389b63ac23f390785043
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:c02c218ccfbb385ea29af98409995a3d9c26cb0b4c76caf2e0779fa233b5ff8c
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:4530ed52522a74051aaedabd62b99d5c5274a0ab72eb28d4e75207d9e9d3391b
        version: 7.10.3
      entries:
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.10.x
    - currentCSV: fuse-apicurito.v7.11.1-0.1687357785.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.11.2'
          olm.substitutesFor: fuse-apicurito.v7.11.1-0.1682509045.p
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:df61aafc4023b9068a8636c35eabb0d5fda400f523b64cb1fc5d8c480272c435
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4c493a2d690d0d7c1684c53d15c06b47d691aeffdde6dd471b6e75fed2fec6a9
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:02e6baf8619cc2b782cbb1f3072e554a7638eaf3fd804abe1ba5f43f692e511b
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:2f8d4450187839f0e0c761e5ddcf61b03f5996acfe3d75b60120e76e5b0167c1
        version: 7.11.1+0.1687357785.p
      entries:
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.11.x
    - currentCSV: fuse-apicurito.v7.12.1-0.1706187670.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.12.1'
          olm.substitutesFor: fuse-apicurito.v7.12.1
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:df59ef548b82e9d72fecd06aa6b9c569f96485693c8a4ecb8ec278ae92df436f
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:7ccefee63165e7eaf9a609e8a9f978a06f27deff69a318fe399a29292ad99106
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:4cd612283e8136839f1f48a3a220dac7c6c38adf14b61bc03a981e514baefe9b
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:bee0eeaa7898af835f80c5217059d73361daa42197fa6cbbaf416e1ba373eb54
        version: 7.12.1+0.1706187670.p
      entries:
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.12.x
    - currentCSV: fuse-apicurito.v7.13.0-0.1753785079.p
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                    "apiVersion": "apicur.io/v1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
          createdAt: 2022-08-31 15:11:41 UTC
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=7.7.0 <7.13.0'
          olm.substitutesFor: fuse-apicurito.v7.13.0-0.1751291353.p
          operators.openshift.io/valid-subscription: '["Red Hat Integration"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Fuse Apicurito
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Apicurito is the Schema for the apicuritoes API
            displayName: Apicurito
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: Red Hat Integration - API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:f5bf9240b28b3664bdb11b1f57f611e6194ff8643ae939b6ab3b2b176705f487
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:4e64a5930a10b76147ac128a216cd7c0346da2ff3e9f31ed86311941ed193cc1
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:e68814481ca33a422ed5c2b55b8a369e8b48092fbc71b138cb938a119d39e689
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:d58d345d74059242d57d5c1a75f31517769230d3014a42243fda20aa93ebd5c8
        version: 7.13.0+0.1753785079.p
      entries:
      - name: fuse-apicurito.v7.13.0-0.1739885035.p
        version: 7.13.0+0.1739885035.p
      - name: fuse-apicurito.v7.13.0-0.1741884705.p
        version: 7.13.0+0.1741884705.p
      - name: fuse-apicurito.v7.13.0-0.1753785079.p
        version: 7.13.0+0.1753785079.p
      - name: fuse-apicurito.v7.13.0-0.1721224330.p
        version: 7.13.0+0.1721224330.p
      - name: fuse-apicurito.v7.13.0-0.1753778625.p
        version: 7.13.0+0.1753778625.p
      - name: fuse-apicurito.v7.13.0-0.1753702605.p
        version: 7.13.0+0.1753702605.p
      - name: fuse-apicurito.v7.13.0-0.1721607497.p
        version: 7.13.0+0.1721607497.p
      - name: fuse-apicurito.v7.13.0-0.1751291353.p
        version: 7.13.0+0.1751291353.p
      - name: fuse-apicurito.v7.13.0-0.1748960166.p
        version: 7.13.0+0.1748960166.p
      - name: fuse-apicurito.v7.13.0-0.1745523192.p
        version: 7.13.0+0.1745523192.p
      - name: fuse-apicurito.v7.13.0-0.1744737027.p
        version: 7.13.0+0.1744737027.p
      - name: fuse-apicurito.v7.13.0-0.1744280419.p
        version: 7.13.0+0.1744280419.p
      - name: fuse-apicurito.v7.13.0-0.1741875975.p
        version: 7.13.0+0.1741875975.p
      - name: fuse-apicurito.v7.13.0-0.1739473863.p
        version: 7.13.0+0.1739473863.p
      - name: fuse-apicurito.v7.13.0-0.1738056248.p
        version: 7.13.0+0.1738056248.p
      - name: fuse-apicurito.v7.13.0-0.1731460815.p
        version: 7.13.0+0.1731460815.p
      - name: fuse-apicurito.v7.13.0-0.1729189341.p
        version: 7.13.0+0.1729189341.p
      - name: fuse-apicurito.v7.13.0-0.1727957847.p
        version: 7.13.0+0.1727957847.p
      - name: fuse-apicurito.v7.13.0-0.1727193594.p
        version: 7.13.0+0.1727193594.p
      - name: fuse-apicurito.v7.13.0-0.1725022690.p
        version: 7.13.0+0.1725022690.p
      - name: fuse-apicurito.v7.13.0-0.1724659732.p
        version: 7.13.0+0.1724659732.p
      - name: fuse-apicurito.v7.13.0-0.1724053718.p
        version: 7.13.0+0.1724053718.p
      - name: fuse-apicurito.v7.13.0-0.1723709434.p
        version: 7.13.0+0.1723709434.p
      - name: fuse-apicurito.v7.13.0
        version: 7.13.0
      - name: fuse-apicurito.v7.13.0-0.1720707249.p
        version: 7.13.0+0.1720707249.p
      - name: fuse-apicurito.v7.13.0-0.1718285973.p
        version: 7.13.0+0.1718285973.p
      - name: fuse-apicurito.v7.13.0-0.1717672390.p
        version: 7.13.0+0.1717672390.p
      - name: fuse-apicurito.v7.13.0-0.1753778936.p
        version: 7.13.0+0.1753778936.p
      - name: fuse-apicurito.v7.13.0-0.1753701210.p
        version: 7.13.0+0.1753701210.p
      - name: fuse-apicurito.v7.12.1-0.1706187670.p
        version: 7.12.1+0.1706187670.p
      - name: fuse-apicurito.v7.12.1
        version: 7.12.1
      - name: fuse-apicurito.v7.12.0-0.1697419531.p
        version: 7.12.0+0.1697419531.p
      - name: fuse-apicurito.v7.12.0-0.1697748407.p
        version: 7.12.0+0.1697748407.p
      - name: fuse-apicurito.v7.12.0-0.1694087055.p
        version: 7.12.0+0.1694087055.p
      - name: fuse-apicurito.v7.12.0
        version: 7.12.0
      - name: fuse-apicurito.v7.11.1-0.1673364838.p
        version: 7.11.1+0.1673364838.p
      - name: fuse-apicurito.v7.11.1-0.1687357785.p
        version: 7.11.1+0.1687357785.p
      - name: fuse-apicurito.v7.11.1-0.1681224697.p
        version: 7.11.1+0.1681224697.p
      - name: fuse-apicurito.v7.11.1-0.1679272341.p
        version: 7.11.1+0.1679272341.p
      - name: fuse-apicurito.v7.11.1-0.1682509045.p
        version: 7.11.1+0.1682509045.p
      - name: fuse-apicurito.v7.11.1
        version: 7.11.1
      - name: fuse-apicurito.v7.11.1-0.1675972858.p
        version: 7.11.1+0.1675972858.p
      - name: fuse-apicurito.v7.11.0
        version: 7.11.0
      - name: fuse-apicurito.v7.11.0-0.1666805761.p
        version: 7.11.0+0.1666805761.p
      - name: fuse-apicurito.v7.11.0-0.1665994968.p
        version: 7.11.0+0.1665994968.p
      - name: fuse-apicurito.v7.11.0-0.1662457152.p
        version: 7.11.0+0.1662457152.p
      - name: fuse-apicurito.v7.11.0-0.1659517292.p
        version: 7.11.0+0.1659517292.p
      - name: fuse-apicurito.v7.10.3
        version: 7.10.3
      - name: fuse-apicurito.v7.10.2-0.1649640025.p
        version: 7.10.2+0.1649640025.p
      - name: fuse-apicurito.v7.10.2-0.1648491130.p
        version: 7.10.2+0.1648491130.p
      - name: fuse-apicurito.v7.10.1
        version: 7.10.1
      - name: fuse-apicurito.v7.10.0
        version: 7.10.0
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.13.x
    - currentCSV: fuse-apicurito.v7.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [{
                    "apiVersion": "apicur.io/v1alpha1",
                    "kind": "Apicurito",
                    "metadata": {
                      "name": "apicurito-service"
                    },
                    "spec": {
                      "size": 3
                    }
            }]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
          createdAt: "2020-10-06 12:48:47"
          description: Manages the installation and upgrades of apicurito, a small/minimal
            version of Apicurio
          olm.skipRange: '>=7.7.0 <7.9.3'
          repository: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
          support: Apicurito Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CRD for Apicurito
            displayName: Apicurito CRD
            kind: Apicurito
            name: apicuritoes.apicur.io
            version: v1alpha1
        description: |+
          The API Designer is a small/minimal version of Apicurio, a standalone API design studio that can be used to create new or edit existing API designs (using the OpenAPI specification).

          This operator supports the installation and upgrade of apicurito. Apicurito components are:
             - apicurito-ui (Apicurito API designer application)
             - apicurito-generator (Apicurito fuse project generator)
             - apicurito route (to access apicurito from outside openshift)

          ### How to install
          When the operator is installed (you have created a subscription and the operator is running in the selected namespace) create a new CR of Kind Apicurito (click the Create New button). The CR spec contains all defaults.

          At the moment, following fields are supported as part of the CR:
             - size: how many pods apicurito operands will have.

        displayName: API Designer
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - api
        - apicurio
        - apicurito
        links:
        - name: Apicurito source code
          url: https://github.com/Apicurio/apicurito
        - name: Apicurito operator source code
          url: https://github.com/Apicurio/apicurio-operators/tree/master/apicurito
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurito Project
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8@sha256:5535864ce393ae0ae5ee52ed0c7661e0351fb7df2f9235da70608245519530b2
        - registry.redhat.io/fuse7/fuse-apicurito-generator-rhel8@sha256:dc39b79c360511ff999eba5b2c299db3e08c9c244a0a1f3da8c90c02695e506d
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator-bundle@sha256:8ce258f36d3baab3175b29d5ed6df89588f128d249285bb1e7cb45539933c511
        - registry.redhat.io/fuse7/fuse-apicurito-rhel8-operator@sha256:05c9676aada2d34a985a59abcf55105a3765d75b0d3764089d2b32e9ca5d367c
        version: 7.9.3
      entries:
      - name: fuse-apicurito.v7.9.3
        version: 7.9.3
      name: fuse-apicurito-7.9.x
    defaultChannel: fuse-apicurito-7.13.x
    packageName: fuse-apicurito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Skupper Project
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: alpha
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1
    - currentCSV: skupper-operator.v1.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
          createdAt: "2024-02-26T18:22:07Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/flow-collector@sha256:3870c5bc89e623b7cebe21856bfefc22de76313705efa0ec551a5963789fff3b
        - quay.io/skupper/service-controller@sha256:345edf2d9d72b5a97bcce0e8cea8cda82d120f7bf83756d73f9ac114c2fcced8
        - quay.io/skupper/site-controller@sha256:e5f8b2ddf378cf3e40f4c7b7ec8482eb3321947cd46481c868d727932c449d78
        - quay.io/skupper/skupper-router@sha256:bf8d5955ce4e406275a4934278efe51c1af9a1a269a70e8fe642b7d94d80cb81
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:3a96ab2db995d24a9ca064b5566821828d6e73ea20df77c9932d8972cb8e5780
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:659fe28d1df7c99bba3bb854b1e9991c715ae4de99d91c88d5db5e4f9ef4c852
        version: 1.6.0
      entries:
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.6
    - currentCSV: skupper-operator.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
          createdAt: "2024-06-12T14:36:59Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/service-controller@sha256:624e88e695297df28e485db7e802945ad1978ac6b0c9e9edca8b5e6013af8991
        - quay.io/skupper/site-controller@sha256:4809db7263d77a95797061798e910f1bdbb5d5fe2db2bb3c743b97f5d0c7b3af
        - quay.io/skupper/skupper-router@sha256:5eb258a0927e7706e7c035f1def2eb5211c8c33dd371fafa7210517fd894f68b
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:c2b996e8dac2f4c81a3e48e4203499f48e4e15d0692e98d61dc8c97192e0a776
        - quay.io/openshift/origin-oauth-proxy@sha256:a7dff785d821acc4b972d94964947713fade45f007dedb5deee17a2869bf5d24
        - quay.io/prometheus/prometheus@sha256:d2ab0a27783fd4ad96a8853e2847b99a0be0043687b8a5d1ebfb2dd3fa4fd1b8
        - quay.io/skupper/config-sync@sha256:a884b26a32179af5a2bfa85118bf873852eec6e2930e92da778b57506c0a0880
        - quay.io/skupper/flow-collector@sha256:bf646b0cb75e1c29bbe0261aa9943e06d33118f5e6a5ccb7503c60c974167692
        version: 1.7.1
      entries:
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.7
    - currentCSV: skupper-operator.v1.8.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
          createdAt: "2025-03-27T20:11:53Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:9e96d6a585f037f60ca3d89b5efee27ec92444d353e8b289ca35b94aeed1e309
        - quay.io/skupper/flow-collector@sha256:252244ab6aaf6b80118bfc06b690d4f5e8d7829b999fca60d194a84452fee7f6
        - quay.io/skupper/service-controller@sha256:a45de211b8f8a8294a3588932a0a72766cf0d5ce64b01f4f744832c8d09aaa46
        - quay.io/skupper/site-controller@sha256:ffe3ea5067d350b64b69120bdf6d9e3ee9a30e8a2559b77170a10d66635623b8
        - quay.io/skupper/skupper-router@sha256:7633b143f1844ee559322e4642ca80e8248a08711c8e814d489c26c39ecf9bd0
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:68eb666168d3a41cc98f91a0f2d74656af73cf76ec2cbe2cc308ae2ffab84875
        - quay.io/openshift/origin-oauth-proxy@sha256:c15baf57dbd17dcc517cd2fa2299462d62124f689d6f5a48a3c9fd154fc77ccf
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.8.4
      entries:
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
          createdAt: "2025-08-28T19:54:31Z"
          description: Skupper Operator provides the ability to create a service network
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Skupper Project
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Skupper enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Skupper instance.

          The Skupper operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Skupper operator will create the Skupper instance and configure the site.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Skupper
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - service
        - mesh
        - van
        links:
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Skupper Project
          url: https://skupper.io
        relatedImages:
        - quay.io/skupper/config-sync@sha256:d0e6a601274bc98a28962fd0ea09aa0328a19c2a83eab5b8c54bc13e0b433aa7
        - quay.io/skupper/flow-collector@sha256:504d0d28d2a79a93bfbb572cc5bc9394469163605b9c087d21e58864ee4ae1c0
        - quay.io/skupper/service-controller@sha256:bf620cbd297824f94b9be28a5abd416d5a6b9f6a2de5af9c11a87d34bc223015
        - quay.io/skupper/site-controller@sha256:928209bc28667bef30acc2e3762d5ea60c55ddf189da74a7108ec650677d645e
        - quay.io/skupper/skupper-router@sha256:d87c4fbc6a1cbe40a5805cc8b886bf98b5481a56f71b80e950260e00bb58eff5
        - quay.io/community-operator-pipeline-prod/skupper-operator@sha256:b32411e6dff4d74ed34ba1985e2d912da8f86dd2e6730e1777f86ae24405cf4d
        - quay.io/openshift/origin-oauth-proxy@sha256:702ba441b48cddec62c2bfe4d3d58080792243faaad094ae4e10062b25ec5fbd
        - quay.io/prometheus/prometheus@sha256:b1935d181b6dd8e9c827705e89438815337e1b10ae35605126f05f44e5c6940f
        version: 1.9.4
      entries:
      - name: skupper-operator.v1.9.4
        version: 1.9.4
      - name: skupper-operator.v1.9.3
        version: 1.9.3
      - name: skupper-operator.v1.9.2
        version: 1.9.2
      - name: skupper-operator.v1.9.1
        version: 1.9.1
      - name: skupper-operator.v1.9.0
        version: 1.9.0
      - name: skupper-operator.v1.8.4
        version: 1.8.4
      - name: skupper-operator.v1.8.3
        version: 1.8.3
      - name: skupper-operator.v1.8.2
        version: 1.8.2
      - name: skupper-operator.v1.8.1
        version: 1.8.1
      - name: skupper-operator.v1.8.0
        version: 1.8.0
      - name: skupper-operator.v1.7.1
        version: 1.7.1
      - name: skupper-operator.v1.7.0
        version: 1.7.0
      - name: skupper-operator.v1.6.0
        version: 1.6.0
      - name: skupper-operator.v1.5.3
        version: 1.5.3
      name: stable-1.9
    defaultChannel: stable
    packageName: skupper-operator
    provider:
      name: Skupper Project
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: opentelemetry-product
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: opentelemetry-operator.v0.135.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "Instrumentation",
                "metadata": {
                  "name": "instrumentation"
                },
                "spec": {
                  "exporter": {
                    "endpoint": "http://otel-collector-headless:4317"
                  }
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpAMPBridge",
                "metadata": {
                  "name": "opampbridge-sample"
                },
                "spec": {
                  "capabilities": {
                    "AcceptsOpAMPConnectionSettings": true,
                    "AcceptsOtherConnectionSettings": true,
                    "AcceptsRemoteConfig": true,
                    "AcceptsRestartCommand": true,
                    "ReportsEffectiveConfig": true,
                    "ReportsHealth": true,
                    "ReportsOwnLogs": true,
                    "ReportsOwnMetrics": true,
                    "ReportsOwnTraces": true,
                    "ReportsRemoteConfig": true,
                    "ReportsStatus": true
                  },
                  "componentsAllowed": {
                    "exporters": [
                      "debug"
                    ],
                    "processors": [
                      "memory_limiter"
                    ],
                    "receivers": [
                      "otlp"
                    ]
                  },
                  "endpoint": "ws://opamp-server:4320/v1/opamp"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1alpha1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": "receivers:\n  otlp:\n    protocols: \n      grpc: {}\n      http: {}\n\nexporters:\n  debug: {}\n\nservice:\n  pipelines:\n    traces:\n      receivers: [otlp]\n      exporters: [debug]\n"
                }
              },
              {
                "apiVersion": "opentelemetry.io/v1beta1",
                "kind": "OpenTelemetryCollector",
                "metadata": {
                  "name": "otel"
                },
                "spec": {
                  "config": {
                    "exporters": {
                      "debug": {}
                    },
                    "receivers": {
                      "otlp": {
                        "protocols": {
                          "grpc": {},
                          "http": {}
                        }
                      }
                    },
                    "service": {
                      "pipelines": {
                        "traces": {
                          "exporters": [
                            "debug"
                          ],
                          "receivers": [
                            "otlp"
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Logging & Tracing,Monitoring
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
          createdAt: 24 Sep 2025, 10:19
          description: Provides the OpenTelemetry components, including the Collector
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.33.0 <0.135.0-1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-opentelemetry-operator
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/open-telemetry/opentelemetry-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instrumentation is the spec for OpenTelemetry instrumentation.
            displayName: OpenTelemetry Instrumentation
            kind: Instrumentation
            name: instrumentations.opentelemetry.io
            version: v1alpha1
          - description: OpAMPBridge is the Schema for the opampbridges API.
            displayName: OpAMP Bridge
            kind: OpAMPBridge
            name: opampbridges.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1alpha1
          - description: OpenTelemetryCollector is the Schema for the opentelemetrycollectors
              API.
            displayName: OpenTelemetry Collector
            kind: OpenTelemetryCollector
            name: opentelemetrycollectors.opentelemetry.io
            version: v1beta1
          - description: TargetAllocator is the Schema for the targetallocators API.
            displayName: Target Allocator
            kind: TargetAllocator
            name: targetallocators.opentelemetry.io
            version: v1alpha1
        description: |-
          Red Hat build of OpenTelemetry is a collection of tools, APIs, and SDKs. You use it to instrument, generate, collect, and export telemetry data (metrics, logs, and traces) for analysis in order to understand your software's performance and behavior.
          This operator was previously called Red Hat OpenShift distributed tracing data collection.

          ### Operator features

          * **Sidecar injection** - annotate your pods and let the operator inject a sidecar.
          * **Managed upgrades** - updating the operator will automatically update your OpenTelemetry collectors.
          * **Deployment modes** - your collector can be deployed as sidecar, daemon set, or regular deployment.
          * **Service port management** - the operator detects which ports need to be exposed based on the provided configuration.

          ## Examples

          There is a list of examples to help you deploy different configurations in [this GitHub repository](https://github.com/os-observability/redhat-rhosdt-samples)

          ### Support & Troubleshooting

          Red Hat build of OpenTelemetry is available as part of a Red Hat OpenShift subscription.
          Checking the [Red Hat Documentation](https://docs.redhat.com/en/documentation/openshift_container_platform/4.17/html/red_hat_build_of_opentelemetry/index) is recommended when installing, configuring, and managing the Operator and instances.
        displayName: Red Hat build of OpenTelemetry
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opentelemetry
        - tracing
        - logging
        - metrics
        - monitoring
        - troubleshooting
        links:
        - name: OpenTelemetry Operator
          url: https://github.com/open-telemetry/opentelemetry-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat support
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhosdt/opentelemetry-rhel8-operator@sha256:45c6c52744e9a17ffd883a2864cb7722ce3f9580e522c976d4f55623118f3504
        - registry.redhat.io/rhosdt/opentelemetry-target-allocator-rhel8@sha256:f77a26edafba06a4fca1a07083e06a908664852cf228b1814455ffa8155cc979
        - registry.redhat.io/rhosdt/opentelemetry-operator-bundle@sha256:1844430afea89707b70bba22cb9a2db98161cfc62b53b1086b63f109058028c9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5a6edfecf1636805e6f0bcad02df10bceec764f0580f3ac8fe62aba978d50006
        - registry.redhat.io/rhosdt/opentelemetry-collector-rhel8@sha256:5f31caf1f4a12cf2832ac3e84877f4fb0c856073820498c8a5be13698348c38a
        version: 0.135.0-1
      entries:
      - name: opentelemetry-operator.v0.135.0-1
        version: 0.135.0-1
      - name: opentelemetry-operator.v0.127.0-2
        version: 0.127.0-2
      - name: opentelemetry-operator.v0.127.0-1
        version: 0.127.0-1
      - name: opentelemetry-operator.v0.119.0-2
        version: 0.119.0-2
      - name: opentelemetry-operator.v0.119.0-1
        version: 0.119.0-1
      - name: opentelemetry-operator.v0.113.0-2
        version: 0.113.0-2
      - name: opentelemetry-operator.v0.113.0-1
        version: 0.113.0-1
      - name: opentelemetry-operator.v0.107.0-5
        version: 0.107.0-5
      - name: opentelemetry-operator.v0.107.0-4-0.1727168137.p
        version: 0.107.0-4+0.1727168137.p
      - name: opentelemetry-operator.v0.107.0-4
        version: 0.107.0-4
      - name: opentelemetry-operator.v0.102.0-3
        version: 0.102.0-3
      - name: opentelemetry-operator.v0.102.0-2
        version: 0.102.0-2
      - name: opentelemetry-operator.v0.100.1-3
        version: 0.100.1-3
      - name: opentelemetry-operator.v0.93.0-3
        version: 0.93.0-3
      - name: opentelemetry-operator.v0.93.0-2
        version: 0.93.0-2
      - name: opentelemetry-operator.v0.89.0-3
        version: 0.89.0-3
      - name: opentelemetry-operator.v0.81.1-5
        version: 0.81.1-5
      - name: opentelemetry-operator.v0.74.0-5-0.1687199949.p
        version: 0.74.0-5+0.1687199949.p
      - name: opentelemetry-operator.v0.74.0-5
        version: 0.74.0-5
      - name: opentelemetry-operator.v0.60.0-2
        version: 0.60.0-2
      - name: opentelemetry-operator.v0.56.0-1
        version: 0.56.0-1
      name: stable
    defaultChannel: stable
    packageName: opentelemetry-product
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Netscaler
      provider-url: https://www.netscaler.com
    name: netscaler-adm-agent-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: netscaler-adm-agent-operator.v141.37.44
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "netscaler.com/v1",
                "kind": "AdmAgent",
                "metadata": {
                  "name": "admagent"
                },
                "spec": {
                  "accessSecret": "",
                  "admURL": "adm.cloud.com",
                  "affinity": {},
                  "clusterPrefix": "",
                  "core": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 30,
                      "initialDelaySeconds": 60,
                      "periodSeconds": 30,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "2000Mi"
                      }
                    }
                  },
                  "fullnameOverride": "",
                  "helper": {
                    "image": "{{ .Values.helper.imageRegistry }}/{{ .Values.helper.imageRepository }}:{{ .Values.helper.imageTag }}",
                    "imageRegistry": "quay.io",
                    "imageRepository": "citrix/adm-agent-helper",
                    "imageTag": "2.0.1",
                    "pullPolicy": "IfNotPresent",
                    "required": false,
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "200Mi"
                      }
                    }
                  },
                  "hpa": {
                    "maxReplicas": 10,
                    "minReplicas": 1,
                    "required": false,
                    "targetCPUUtilization": "1000m",
                    "targetMemoryUtilization": "4000Mi"
                  },
                  "image": "{{ .Values.imageRegistry }}/{{ .Values.imageRepository }}:{{ .Values.imageTag }}",
                  "imagePullSecrets": [],
                  "imageRegistry": "quay.io",
                  "imageRepository": "citrix/adm-agent",
                  "imageTag": "14.1-37.44",
                  "kad": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "50m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "lic": {
                    "livenessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 600,
                      "periodSeconds": 20,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "100m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "loginSecret": "",
                  "nameOverride": "",
                  "nodeSelector": {},
                  "podAnnotations": {},
                  "pullPolicy": "IfNotPresent",
                  "redis": {
                    "livenessProbe": {
                      "failureThreshold": 5,
                      "initialDelaySeconds": 30,
                      "periodSeconds": 60,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "readinessProbe": {
                      "failureThreshold": 20,
                      "initialDelaySeconds": 20,
                      "periodSeconds": 10,
                      "required": true,
                      "successThreshold": 1,
                      "timeoutSeconds": 5
                    },
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "1000Mi"
                      }
                    }
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "create": true
                  },
                  "sharding": {
                    "resources": {
                      "limits": {},
                      "requests": {
                        "cpu": "10m",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
          createdAt: "2024-11-18T06:40:58Z"
          description: This is an operator to install ADM Agent
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/netscaler/netscaler-k8s-ingress-controller
          support: NetScaler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: To install NetScaler ADM Agent Operator
            displayName: NetScaler ADM Agent Operator
            kind: AdmAgent
            name: admagents.netscaler.com
            version: v1
        description: Netscaler ADM Agent Operator
        displayName: Netscaler ADM Agent Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Ingress
        - NetScaler
        - ADC
        - Kubernetes
        - OpenShift
        - Citrix
        - Controller
        - Observability
        - Metrices
        - adm
        - agent
        - monitoring
        - proxy
        links:
        - name: Netscaler Adm Agent Operator
          url: https://github.com/netscaler/netscaler-helm-charts/blob/master/adm-agent/README.md
        maintainers:
        - email: pavan.belani@cloud.com
          name: Pavan Belani
        - email: swapnil.khanderaokakade@cloud.com
          name: Swapnil Khanderao
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Netscaler
          url: https://www.netscaler.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6b8c101124579dd46086ec83715618dd9a6aebe66176d676242e6c0b0900a53c
        - quay.io/citrix/adm-agent-helper@sha256:f567a0637e8024f9d1674977f6c013741cc0fd12cf78703f8dbe01126561c3c4
        - quay.io/citrix/adm-agent@sha256:45cab90524687c76a6d53f014b55063fa25c8aaeec0261f0f4f0efa6bff42968
        - registry.connect.redhat.com/citrix/admagent-operator-bundle@sha256:2a32d9a8cc3bd67ba7ebfe3f96b59e42009121d630918d417717d8ac9f2e40f4
        - registry.connect.redhat.com/citrix/netscaler-adm-agent-operator@sha256:1c462c2d6161877f87d5eed4888f8d94c733501e67a94757ab0873d983fb3a4a
        version: 141.37.44
      entries:
      - name: netscaler-adm-agent-operator.v141.37.44
        version: 141.37.44
      - name: netscaler-adm-agent-operator.v141.31.38
        version: 141.31.38
      - name: netscaler-adm-agent-operator.v141.25.33
        version: 141.25.33
      - name: netscaler-adm-agent-operator.v141.23.31
        version: 141.23.31
      - name: netscaler-adm-agent-operator.v141.21.41
        version: 141.21.41
      - name: netscaler-adm-agent-operator.v141.16.39
        version: 141.16.39
      - name: netscaler-adm-agent-operator.v141.13.23
        version: 141.13.23
      name: stable
    defaultChannel: stable
    packageName: netscaler-adm-agent-operator
    provider:
      name: Netscaler
      url: https://www.netscaler.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://github.com/redhat-developer/app-services-operator/
    name: rhoas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhoas-operator.0.13.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "ServiceRegistryConnection",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken",
                  "serviceRegistryId": "exampleId",
                  "credentials": {
                    "serviceAccountSecretName": "service-account-secret"
                  }
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServiceAccountRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace"
                },
                "spec": {
                  "serviceAccountName": "rhoas-sa",
                  "serviceAccountDescription": "Operator created service account",
                  "serviceAccountSecretName": "service-account-credentials",
                  "accessTokenSecretName": "rh-managed-services-api-accesstoken"
                }
              },
              {
                "apiVersion": "rhoas.redhat.com/v1alpha1",
                "kind": "CloudServicesRequest",
                "metadata": {
                  "name": "example",
                  "namespace": "example-namespace",
                  "labels": {
                    "app.kubernetes.io/component": "external-service",
                    "app.kubernetes.io/managed-by": "rhoas"
                  }
                },
                "spec": {
                  "accessTokenSecretName": "rh-cloud-services-api-accesstoken"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, OpenShift Optional, Integration & Delivery
          certified: "false"
          containerImage: quay.io/rhoas/service-operator:0.13.0
          createdAt: "2023-01-20 16:09:37"
          description: Red Hat OpenShift Application Services Operator
          olm.skipRange: '>=0.0.1 <0.13.0'
          repository: https://github.com/redhat-developer/app-services-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Creates RHOAS Service Accounts
            displayName: Cloud Service Account Request
            kind: CloudServiceAccountRequest
            name: cloudserviceaccountrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Cloud Services Request
            displayName: Cloud Services Request
            kind: CloudServicesRequest
            name: cloudservicesrequests.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Kafka Instance
            displayName: Kafka Connection
            kind: KafkaConnection
            name: kafkaconnections.rhoas.redhat.com
            version: v1alpha1
          - description: Binds Service Registry Instance
            displayName: Service Registry Connection
            kind: ServiceRegistryConnection
            name: serviceregistryconnections.rhoas.redhat.com
            version: v1alpha1
        description: "**OpenShift Application Services Operator (RHOAS-Operator)**
          creates instances of Red Hat OpenShift Application Services in your cluster.
          Operator is currently available as limited Beta and it cannot be consumed
          without invitation.\n\nRed Hat OpenShift Application Services, when added
          to Red Hat OpenShift Dedicated,  provide a streamlined developer experience
          for building, deploying,  and scaling cloud-native applications in open
          hybrid-cloud environments.   \n\nRHOAS-Operator enables developers to connect
          with their Application Services directly in their own cluster by using [RHOAS
          cli](https://github.com/redhat-developer/app-services-cli) and OpenShift
          Console UI.  \n\nRHOAS-Operator requires developers to connect with their
          Red Hat OpenShift Streams for Apache Kafka (RHOSAK) instances.  "
        displayName: OpenShift Application Services (RHOAS)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - rhoas
        - kafka
        - serviceregistry
        - app-services
        - application
        - services
        links:
        - name: RHOAS Operator Documentation
          url: https://github.com/redhat-developer/app-services-operator/tree/main/docs
        maintainers:
        - email: rhosak-eval-support@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat Inc.
          url: https://github.com/redhat-developer/app-services-operator/
        relatedImages:
        - quay.io/rhoas/service-operator:0.13.0
        - quay.io/openshift-community-operators/rhoas-operator@sha256:5299cd4c88012692b31d560ef981f80f994a0468552f43e03623adbf89b49ff6
        version: 0.13.0
      entries:
      - name: rhoas-operator.0.13.0
        version: 0.13.0
      - name: rhoas-operator.0.12.0
        version: 0.12.0
      - name: rhoas-operator.0.11.0
        version: 0.11.0
      - name: rhoas-operator.0.9.6
        version: 0.9.6
      - name: rhoas-operator.0.9.5
        version: 0.9.5
      - name: rhoas-operator.0.9.3
        version: 0.9.3
      - name: rhoas-operator.0.9.2
        version: 0.9.2
      - name: rhoas-operator.0.9.1
        version: 0.9.1
      - name: rhoas-operator.0.9.0
        version: 0.9.0
      - name: rhoas-operator.0.8.2
        version: 0.8.2
      - name: rhoas-operator.0.7.10
        version: 0.7.10
      - name: rhoas-operator.0.7.9
        version: 0.7.9
      - name: rhoas-operator.0.7.8
        version: 0.7.8
      - name: rhoas-operator.0.7.6
        version: 0.7.6
      - name: rhoas-operator.0.7.2
        version: 0.7.2
      - name: rhoas-operator.0.7.1
        version: 0.7.1
      - name: rhoas-operator.0.6.8
        version: 0.6.8
      name: beta
    defaultChannel: beta
    packageName: rhoas-operator
    provider:
      name: Red Hat Inc.
      url: https://github.com/redhat-developer/app-services-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-networkfirewall-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-networkfirewall-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "Firewall",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "FirewallPolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "networkfirewall.services.k8s.aws/v1alpha1",
                "kind": "RuleGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
          createdAt: "2025-09-19T21:03:12Z"
          description: AWS Network Firewall controller is a service controller for
            managing Network Firewall resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FirewallPolicy represents the state of an AWS networkfirewall
              FirewallPolicy resource.
            displayName: FirewallPolicy
            kind: FirewallPolicy
            name: firewallpolicies.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: Firewall represents the state of an AWS networkfirewall Firewall
              resource.
            displayName: Firewall
            kind: Firewall
            name: firewalls.networkfirewall.services.k8s.aws
            version: v1alpha1
          - description: RuleGroup represents the state of an AWS networkfirewall
              RuleGroup resource.
            displayName: RuleGroup
            kind: RuleGroup
            name: rulegroups.networkfirewall.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Network Firewall resources in AWS from within your Kubernetes cluster.

          **About Amazon Network Firewall**

          AWS Network Firewall is a stateful, managed, network firewall and intrusion detection and prevention service for your virtual private cloud (VPC) that you create in Amazon Virtual Private Cloud (Amazon VPC).
          With Network Firewall, you can filter traffic at the perimeter of your VPC. This includes filtering traffic going to and coming from an internet gateway, NAT gateway, or over VPN or AWS Direct Connect. Network Firewall uses the open source intrusion prevention system (IPS), Suricata, for stateful inspection. Network Firewall supports Suricata compatible rules. For more information, see [Working with stateful rule groups in AWS Network Firewall](https://docs.aws.amazon.com/network-firewall/latest/developerguide/stateful-rule-groups-ips.html).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Network Firewall
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networkfirewall
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Network Firewall Developer Resources
          url: https://aws.amazon.com/network-firewall/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: network firewall maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/networkfirewall-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-networkfirewall-controller@sha256:96c35a47f58c48532fa3a5750e52a6d7bc2d530b54e807d57547f8780e9d7aea
        version: 1.1.1
      entries:
      - name: ack-networkfirewall-controller.v1.1.1
        version: 1.1.1
      - name: ack-networkfirewall-controller.v1.1.0
        version: 1.1.0
      - name: ack-networkfirewall-controller.v1.0.13
        version: 1.0.13
      - name: ack-networkfirewall-controller.v1.0.12
        version: 1.0.12
      - name: ack-networkfirewall-controller.v1.0.11
        version: 1.0.11
      - name: ack-networkfirewall-controller.v1.0.10
        version: 1.0.10
      - name: ack-networkfirewall-controller.v1.0.9
        version: 1.0.9
      - name: ack-networkfirewall-controller.v1.0.8
        version: 1.0.8
      - name: ack-networkfirewall-controller.v1.0.7
        version: 1.0.7
      - name: ack-networkfirewall-controller.v1.0.6
        version: 1.0.6
      - name: ack-networkfirewall-controller.v1.0.5
        version: 1.0.5
      - name: ack-networkfirewall-controller.v1.0.4
        version: 1.0.4
      - name: ack-networkfirewall-controller.v1.0.1
        version: 1.0.1
      - name: ack-networkfirewall-controller.v1.0.0
        version: 1.0.0
      - name: ack-networkfirewall-controller.v0.0.10
        version: 0.0.10
      - name: ack-networkfirewall-controller.v0.0.9
        version: 0.0.9
      - name: ack-networkfirewall-controller.v0.0.8
        version: 0.0.8
      - name: ack-networkfirewall-controller.v0.0.7
        version: 0.0.7
      - name: ack-networkfirewall-controller.v0.0.6
        version: 0.0.6
      - name: ack-networkfirewall-controller.v0.0.5
        version: 0.0.5
      - name: ack-networkfirewall-controller.v0.0.2
        version: 0.0.2
      - name: ack-networkfirewall-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-networkfirewall-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Korrel8r
      provider-url: https://github.com/korrel8r/korrel8r
    name: korrel8r
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: korrel8r.v0.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "korrel8r.openshift.io/v1alpha1",
                "kind": "Korrel8r",
                "metadata": {
                  "name": "korrel8r",
                  "namespace": "korrel8r"
                },
                "spec": {
                  "debug": {
                    "verbose": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional,Monitoring,Logging & Tracing
          containerImage: quay.io/korrel8r/operator:0.1.8
          createdAt: "2024-09-27T17:45:00Z"
          description: Correlate cluster resources and observability data.
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "korrel8r.openshift.io/v1alpha1",
              "kind": "Korrel8r",
              "metadata": {
                "name": "korrel8r",
                "namespace": "korrel8r"
              }
            }
          operatorframework.io/suggested-namespace: korrel8r
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/korrel8r/korrel8r
          support: Korrel8r Project
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Korrel8r is a service that correlates observabililty signals
              in the cluster.
            displayName: Korrel8r
            kind: Korrel8r
            name: korrel8rs.korrel8r.openshift.io
            version: v1alpha1
        description: |
          Korrel8r helps navigate from problem symptoms to related resources and signal data that can reveal the cause.
          It can follow relationships between disjointed observability "silos" (logs, metrics, alerts and more) to bring together all the data available to solve a problem.
          This operator allows Korrel8r to be deployed as a cluster service with a REST API.
          It can be used as a back-end for data correlation by the OpenShift console and other tools.
        displayName: Korrel8r
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - signal
        - correlation
        links:
        - name: Documentation
          url: https://korrel8r.github.io/korrel8r
        - name: Source Code and Issues
          url: https://github.com/korrel8r/korrel8r
        maintainers:
        - email: korrel8r@gmail.com
          name: Korrel8r Project
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Korrel8r
          url: https://github.com/korrel8r/korrel8r
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/korrel8r@sha256:f00f8a7e2011bcc97d87fc7f5bb40a19d6b5737d1f5d7f3b2322ec39009c0618
        - quay.io/korrel8r/operator:0.1.8
        version: 0.1.8
      entries:
      - name: korrel8r.v0.1.8
        version: 0.1.8
      - name: korrel8r.v0.1.7
        version: 0.1.7
      - name: korrel8r.v0.1.5
        version: 0.1.5
      - name: korrel8r.v0.1.3
        version: 0.1.3
      - name: korrel8r.v0.1.2
        version: 0.1.2
      - name: korrel8r.v0.0.8
        version: 0.0.8
      name: stable
    defaultChannel: stable
    packageName: korrel8r
    provider:
      name: Korrel8r
      url: https://github.com/korrel8r/korrel8r
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-route53resolver-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-route53resolver-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "route53resolver.services.k8s.aws/v1alpha1",
                "kind": "ResolverRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
          createdAt: "2025-09-20T01:21:08Z"
          description: AWS Route53Resolver controller is a service controller for
            managing Route53Resolver resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ResolverEndpoint represents the state of an AWS route53resolver
              ResolverEndpoint resource.
            displayName: ResolverEndpoint
            kind: ResolverEndpoint
            name: resolverendpoints.route53resolver.services.k8s.aws
            version: v1alpha1
          - description: ResolverRule represents the state of an AWS route53resolver
              ResolverRule resource.
            displayName: ResolverRule
            kind: ResolverRule
            name: resolverrules.route53resolver.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Route53Resolver resources in AWS from within your Kubernetes cluster.

          **About Amazon Route53Resolver**

          Amazon Route 53 Resolver responds recursively to DNS queries from AWS resources for public records, Amazon VPC-specific DNS names, and Amazon Route 53 private hosted zones, and is available by default in all VPCs.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Route53Resolver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - route53resolver
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Route53Resolver Developer Resources
          url: https://docs.aws.amazon.com/Route53/latest/DeveloperGuide/resolver.html
        maintainers:
        - email: ack-maintainers@amazon.com
          name: route53resolver maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/route53resolver-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-route53resolver-controller@sha256:542229539932458d76180d23f1d5c5795e2de89169d8054f7b95be6f3700881b
        version: 1.1.1
      entries:
      - name: ack-route53resolver-controller.v1.1.1
        version: 1.1.1
      - name: ack-route53resolver-controller.v1.0.15
        version: 1.0.15
      - name: ack-route53resolver-controller.v1.0.14
        version: 1.0.14
      - name: ack-route53resolver-controller.v1.0.13
        version: 1.0.13
      - name: ack-route53resolver-controller.v1.0.11
        version: 1.0.11
      - name: ack-route53resolver-controller.v1.0.10
        version: 1.0.10
      - name: ack-route53resolver-controller.v1.0.9
        version: 1.0.9
      - name: ack-route53resolver-controller.v1.0.7
        version: 1.0.7
      - name: ack-route53resolver-controller.v1.0.6
        version: 1.0.6
      - name: ack-route53resolver-controller.v1.0.5
        version: 1.0.5
      - name: ack-route53resolver-controller.v1.0.3
        version: 1.0.3
      - name: ack-route53resolver-controller.v1.0.2
        version: 1.0.2
      - name: ack-route53resolver-controller.v1.0.0
        version: 1.0.0
      - name: ack-route53resolver-controller.v0.0.11
        version: 0.0.11
      - name: ack-route53resolver-controller.v0.0.10
        version: 0.0.10
      - name: ack-route53resolver-controller.v0.0.9
        version: 0.0.9
      - name: ack-route53resolver-controller.v0.0.7
        version: 0.0.7
      - name: ack-route53resolver-controller.v0.0.4
        version: 0.0.4
      - name: ack-route53resolver-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-route53resolver-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: recipe
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: recipe.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-09T14:20:34Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:85ba83a60fece2cc1c39ff245193712f918fbd55e1e283e940425fbb1b1f0e29
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        version: 4.17.13-rhodf
      entries:
      - name: recipe.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: recipe.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: recipe.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: recipe.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: recipe.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: recipe.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: recipe.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: recipe.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: recipe.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: recipe.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: recipe.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: recipe.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: recipe.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: recipe.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: recipe.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "Recipe",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "recipe",
                    "app.kubernetes.io/instance": "recipe-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "recipe",
                    "app.kubernetes.io/part-of": "recipe"
                  },
                  "name": "recipe-sample"
                },
                "spec": null
              }
            ]
          capabilities: Deep Insights
          createdAt: "2025-09-11T15:27:55Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Recipe is the Schema for the recipes API
            displayName: Recipe
            kind: Recipe
            name: recipes.ramendr.openshift.io
            version: v1alpha1
        description: |
          Recipe describes a workflow used for capturing or recovering Kubernetes resources. This can be referred to by a Ramen DRPlacementControl object when kubeObjectProtection is enabled.
        displayName: Recipe
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - recipe
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/recipe
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/odf4/odr-recipe-operator-bundle@sha256:3c7033c68ff5e65533d2c3d0f2032cf34aecc153fd34f33ad3239331442d370e
        version: 4.18.11-rhodf
      entries:
      - name: recipe.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: recipe.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: recipe.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: recipe.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: recipe.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: recipe.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: recipe.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: recipe.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: recipe.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: recipe.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: recipe.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: recipe.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: recipe
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nxiq-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nxiq-operator-certified.v1.193.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusIQ",
                "metadata": {
                  "name": "example-nexusiq"
                },
                "spec": {
                  "deployment": {
                    "postStart": {
                      "command": null
                    }
                  },
                  "deploymentStrategy": {
                    "type": "Recreate"
                  },
                  "fullnameOverride": "",
                  "imagePullSecrets": [],
                  "ingress": {
                    "annotations": {},
                    "enabled": false,
                    "hosts": [],
                    "path": "/",
                    "tls": []
                  },
                  "iq": {
                    "adminPort": 8071,
                    "applicationPort": 8070,
                    "env": [
                      {
                        "name": "JAVA_OPTS",
                        "value": "-Djava.util.prefs.userRoot=$(SONATYPE_WORK)/javaprefs -Djavax.net.ssl.trustStore=/etc/nexus-iq-server/jks/keystore.jks"
                      }
                    ],
                    "configYaml": {
                      "sonatypeWork": "/sonatype-work",
                      "createSampleData": true,
                      "server": {
                        "adminConnectors": [
                          {
                            "port": 8071,
                            "type": "http"
                          }
                        ],
                        "applicationConnectors": [
                          {
                            "port": 8070,
                            "type": "http"
                          }
                        ],
                        "requestLog": {
                          "appenders": [
                            {
                              "type": "file",
                              "currentLogFilename": "/var/log/nexus-iq-server/request.log",
                              "logFormat": "%clientHost %l %user [%date] \"%requestURL\" %statusCode %bytesSent %elapsedTime \"%header{User-Agent}\"",
                              "archivedLogFilenamePattern": "/var/log/nexus-iq-server/request-%d.log.gz",
                              "archivedFileCount": 50
                            }
                          ]
                        }
                      },
                      "logging": {
                        "level": "DEBUG",
                        "loggers": {
                          "com.sonatype.insight.scan": "INFO",
                          "eu.medsea.mimeutil.MimeUtil2": "INFO",
                          "org.apache.http": "INFO",
                          "org.apache.http.wire": "ERROR",
                          "org.eclipse.birt.report.engine.layout.pdf.font.FontConfigReader": "WARN",
                          "org.eclipse.jetty": "INFO",
                          "org.apache.shiro.web.filter.authc.BasicHttpAuthenticationFilter": "INFO",
                          "com.sonatype.insight.audit": {
                            "appenders": [
                              {
                                "type": "file",
                                "currentLogFilename": "/var/log/nexus-iq-server/audit.log",
                                "archivedLogFilenamePattern": "/var/log/nexus-iq-server/audit-%d.log.gz",
                                "archivedFileCount": 50
                              }
                            ]
                          },
                          "com.sonatype.insight.policy.violation": {
                            "appenders": [
                              {
                                "type": "console",
                                "threshold": "INFO"
                              }
                            ]
                          }
                        },
                        "appenders": [
                          {
                            "type": "console",
                            "threshold": "INFO",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n"
                          },
                          {
                            "type": "file",
                            "threshold": "ALL",
                            "logFormat": "%d{'yyyy-MM-dd HH:mm:ss,SSSZ'} %level [%thread] %X{username} %logger - %msg%n",
                            "currentLogFilename": "/var/log/nexus-iq-server/clm-server.log",
                            "archivedLogFilenamePattern": "/var/log/nexus-iq-server/clm-server-%d.log.gz",
                            "archivedFileCount": 50
                          }
                        ]
                      }
                    },
                    "imageName": "registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0",
                    "imagePullPolicy": "IfNotPresent",
                    "imagePullSecret": "",
                    "licenseSecret": "",
                    "memory": "1Gi",
                    "name": "nxiq",
                    "certificates": {}
                  },
                  "nameOverride": "",
                  "persistence": {
                    "accessMode": "ReadWriteOnce",
                    "storageSize": "1Gi"
                  },
                  "service": {
                    "annotations": {},
                    "enabled": false,
                    "labels": {},
                    "ports": [
                      {
                        "name": "nexus-service",
                        "port": 80,
                        "targetPort": 80
                      }
                    ]
                  },
                  "serviceAccount": {
                    "create": true,
                    "name": null
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
          createdAt: "2020-07-20"
          description: |-
            Nexus Lifecycle is an open source governance platform that continuously identifies risk,
            automatically enforces policy, and provides visibility throughout the entire SDLC.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <1.193.0-1
          repository: https://github.com/sonatype/operator-iq
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus IQ Server
            displayName: NexusIQ
            kind: NexusIQ
            name: nexusiqs.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Lifecycle is an open source governance platform that continuously identifies risk,
          automatically enforces policy, and provides visibility throughout the entire SDLC.
          Developers leveraging open source require a solution that helps them make safer choices
          while still delivering at DevOps speed.
          Nexus Lifecycle empowers developers and security teams to choose the best components
          and govern open source usage, ensuring organizations continue to innovate with less risk.

          ## Core Capabilities

          * **Dependency Management**:
            Generate a precise software bill of materials for all applications
            to identify open source components and gather contextual direct
            and transitive dependency information for faster action.
          * **One-Click Remediation**:
            Upgrade to the best components and versions based on real-time intelligence.
            Available in your preferred IDE or a GitHub pull request.
          * **Automated Open Source Policy Enforcement**:
            Create custom security, license, and architectural policies based
            on application type or organization and contextually enforce those policies
            across every stage of the SDLC.
          * **Continuous Monitoring**:
            Real-time notifications of newly discovered defects,
            as well as vulnerabilities based on component,
            risk level or application/container affected.

          ## Extras

          * IDE Integrations: Eclipse, IntelliJ, and Visual Studio
          * Source Control Integrations: GitHub, GitLab, and Bitbucket
          * Chrome Browser Extension
          * Red Hat Clair for Container Scanning

          ## Limitations

          This operator will be released on a quarterly basis.

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus IQ Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage

          Once the server instance is created by the operator and running,
          you'll want to expose the service as you see fit:
          1. Create a Route to the new service for iq.applicationPort (8070).
          2. Visit the URL provided by the Route, login, and set new credentials.
            The default credentials are `admin`/`admin123`.

          The Nexus IQ Server can be further configured via the NexusIQ custom resource definition:

          | Parameter            | Description                                                  | Default           |
          | -------------------- | ------------------------------------------------------------ | ----------------- |
          | `iq.applicationPort` | Port of the application connector. Must match the value in the `configYaml` property | `8070`            |
          | `iq.adminPort`       | Port of the application connector. Must match the value in the `configYaml` property | `8071`            |
          | `iq.memory`          | The amount of RAM to allocate                                | `1Gi`             |
          | `iq.env`             | Customize the server environment, including `JAVA_OPTS`      | See example YAML show when ceating a NexusIQ. |
          | `iq.licenseSecret`   | The base-64 encoded license file to be installed at startup  | `""`              |
          | `iq.configYaml`      | A YAML block which will be used as a configuration block for IQ Server. | See example YAML shown when creating a NexusIQ. |
          | `iq.certificates`   | Add your own certificates for your own trusted Certificate Authorities with names and plain armored text. | `{}` |
          | `ingress.enabled`                           | Create an ingress for Nexus         | `true`                                  |
          | `ingress.annotations`                       | Annotations to enhance ingress configuration  | `{}`                          |
          | `ingress.tls.enabled`                       | Enable TLS                          | `true`                                 |
          | `ingress.tls.secretName`                    | Name of the secret storing TLS cert, `false` to use the Ingress' default certificate | `nexus-tls`                             |
          | `ingress.path`                              | Path for ingress rules. GCP users should set to `/*` | `/`                    |
          | `deploymentStrategy.type`                   | Deployment Strategy                 |  `Recreate` |
          | `initContainers`                            | Containers to run before starting the server | `nil` |
          | `deployment.postStart.command`              | Command to run after starting the IQ Server container  | `nil`                    |
          | `deployment.terminationGracePeriodSeconds` | Time to allow for clean shutdown                        | 120                      |
          | `persistence.storageClass` | The provisioner class                        | `-` (disables dynamic provisioning)            |
          | `persistence.storageSize` | The amount of drive space to allocate                        | `1Gi`             |
          | `persistence.accessMode` | Default access mode                        | `ReadWriteOnce`             |
          | `persistence.volumeConfiguration` | A YAML block to configure the persistent volume type. Defaults to `hostPath` which should not be used in production | `hostPath`             |

          ## Configuring IQ Server

          You can define the `config.yml` for IQ Server in your CRD on startup.
          It is the `iq.configYaml` property. For more details, see the [Configuring IQ Server](https://help.sonatype.com/iqserver/configuring) help page.

          ## Installing the License

          The license file can be installed via the UI when IQ server is running, or it can be done as a part of the deploy.
          If you leave the `licenseFile` field empty/commented, IQ Server will start and prompt you to manually install the license
          when you first enter the GUI.

          ### Installing the License Automatically
          To do it automatically, first encode your `.lic` file in Base 64 with no line breaks, eg:

          ```bash
          base64 --wrap=0 mylicense.lic > lic.base64
          ```

          Then add this value to your CRD file as `iq.licenseSecret`, eg:

          ```yaml
          iq:
            licenseSecret: bXkgc2FtcGxlIGxpY2Vuc2U=
          ```

          Specify the `licenseFile` path in `iq.configYaml` as:

          ```yaml
          iq:
            configYaml:
              sonatypeWork: /sonatype-work
              createSampleData: true
              # add this line and the `licenseSecret` above to autoconfigure licensing
              licenseFile: /etc/nexus-iq-license/license_lic
              server:
                applicationConnectors:
                  - type: http
                    port: 8070
                adminConnectors:
                  - type: http
                    port: 8071
          ```

          ## Adding Certificates for Your Own Trusted Certificate Authorities

          Specify a set of certificates to be imported into the Java trust store
          of the running IQ Server. More than one certificate can be provided.

          ```yaml
          iq:
            certificates:
              my-company-cert: |
                -----BEGIN CERTIFICATE-----
                ...
                -----END CERTIFICATE-----
          ```

          ## Running Commands Before Initialization of the Server

          If you need custom initialization, you can use standard Kubernetes
          [`initContainers`](https://kubernetes.io/docs/concepts/workloads/pods/init-containers/)
          to run any series of containers and commands you need.

          ```yaml
          initContainers:
            - name: an-init-container
              image: busybox:1.28
              command: ['/bin/sh', '-c', 'echo "initializing"']
          ```
        displayName: Nexus IQ Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - sonatype
        - licensing
        - policy
        - components
        links:
        - name: Nexus Lifecycle
          url: https://www.sonatype.com/product-nexus-lifecycle
        maintainers:
        - email: support@sonaype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-iq-server@sha256:d65df6bdff1f32bf32a87e5cb38f5e9f98b75b81e9acccbffbd967e1f71691f0
        - registry.connect.redhat.com/sonatype/nxiq-operator-bundle@sha256:48bccf9c8cc07127153af7974b949b8cd8429adf8714dcc866f026806e38fb08
        - registry.connect.redhat.com/sonatype/nxiq-operator-certified@sha256:59c52729189da87d9b5c3de8ef85326662d94c99835ab99d6e8d7fefdcf53ef0
        version: 1.193.0-1
      entries:
      - name: nxiq-operator-certified.v1.193.0-1
        version: 1.193.0-1
      name: stable
    defaultChannel: stable
    packageName: nxiq-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: CyberArk
      provider-url: https://www.cyberark.com/
    name: vcp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vcp-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
          createdAt: "2024-05-30T11:35:41Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          Venafi Control Plane Operator for Red Hat OpenShift simplifies deployment of Venafi's enterprise components and dependencies.

          The enterprise components include:
          - Enterprise cert-manager: An enhanced version of cert-manager designed for application certificate management within your clusters.
          - Venafi Enhanced Issuer: Enables seamless certificate enrollment for your clusters directly from the Venafi Control Plane.
          - Venafi Firefly: A high-performance, lightweight issuer for certificates, centrally managed by the Venafi Control Plane.
          - Venafi Kubernetes Agent: Provides real-time visibility into certificates and Kubernetes resources directly within the Venafi Control Plane.
        displayName: Venafi Control Plane Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - venafi
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Venafi
          url: https://venafi.com
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:9d24507aabd06856c7c35d73417b6ce065164f9c966053b494d68f14f0e260b7
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:9da79fd0941c21e433f22aaaccc43ebbd4646e685aea0050194954d8b4019c05
        version: 1.0.0
      entries:
      - name: vcp-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: vcp-operator.v1.8.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "installer.venafi.com/v1alpha1",
                "kind": "VenafiInstall",
                "metadata": {
                  "name": "install-venafi-enhanced-issuer"
                },
                "spec": {
                  "certManager": {
                    "install": true
                  },
                  "globals": {
                    "imagePullSecretNames": [
                      "venafi"
                    ]
                  },
                  "venafiConnection": {
                    "install": true
                  },
                  "venafiEnhancedIssuer": {
                    "install": true
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
          createdAt: "2025-07-22T06:44:30Z"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: unknown
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              VenafiInstall is an object for installing Venafi components onto a Kubernetes
              cluster
            displayName: VenafiInstall
            kind: VenafiInstall
            name: venafiinstalls.installer.venafi.com
            version: v1alpha1
        description: |
          The operator, formerly known as the Venafi Control Plane Operator, simplifies the deployment of enterprise components part of the CyberArk Certificate Manager for Kubernetes.

          The components include:
          - cert-manager: An enterprise distribution of cert-manager designed for application certificate management within your clusters.
          - Enterprise Issuer: Enables seamless certificate enrollment directly from the Certificate Manager.
          - Workload Identity Manager: A high-performance issuer for certificates, centrally managed by the Certificate Manager.
          - Agent: Provides real-time visibility into cluster resources directly within the Certificate Manager.
        displayName: CyberArk Certificate Manager Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cert-manager
        - cyberark
        - vcp-operator
        links:
        - name: Documentation
          url: https://docs.venafi.cloud/vaas/k8s-components/c-vcpo-overview/
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: CyberArk
          url: https://www.cyberark.com/
        relatedImages:
        - registry.connect.redhat.com/venafi/vcp-operator@sha256:00e6945419be8ac83aa7c888fe243c2fb67ee4ae5030cae117c111fdf80f5fe4
        - registry.venafi.cloud/public/venafi-images/vcp-operator@sha256:eb2b06d191dbfbb02332375a7bca6bc010d5410e53b3bc5fb8c320101292408c
        version: 1.8.1
      entries:
      - name: vcp-operator.v1.8.1
        version: 1.8.1
      - name: vcp-operator.v1.8.0
        version: 1.8.0
      - name: vcp-operator.v1.7.0
        version: 1.7.0
      - name: vcp-operator.v1.6.0
        version: 1.6.0
      - name: vcp-operator.v1.5.0
        version: 1.5.0
      - name: vcp-operator.v1.4.0
        version: 1.4.0
      - name: vcp-operator.v1.3.0
        version: 1.3.0
      - name: vcp-operator.v1.2.1
        version: 1.2.1
      - name: vcp-operator.v1.2.0
        version: 1.2.0
      - name: vcp-operator.v1.1.1
        version: 1.1.1
      name: stable
    defaultChannel: stable
    packageName: vcp-operator
    provider:
      name: CyberArk
      url: https://www.cyberark.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: ptp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ptp-operator.v4.18.0-202509100344
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "NodePtpDevice",
                "metadata": {
                  "name": "node.example.com"
                },
                "spec": {}
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpConfig",
                "metadata": {
                  "name": "example-ptpconfig"
                },
                "spec": {
                  "profile": [
                    {
                      "interface": "ens787f1",
                      "name": "profile1",
                      "phc2sysOpts": "-a -r",
                      "ptp4lOpts": "-s -2",
                      "ptpClockThreshold": {}
                    }
                  ],
                  "recommend": [
                    {
                      "match": [
                        {
                          "nodeLabel": "node-role.kubernetes.io/worker=",
                          "nodeName": "node.example.com"
                        }
                      ],
                      "priority": 4,
                      "profile": "profile1"
                    }
                  ]
                }
              },
              {
                "apiVersion": "ptp.openshift.io/v1",
                "kind": "PtpOperatorConfig",
                "metadata": {
                  "name": "default"
                },
                "spec": {
                  "daemonNodeSelector": {
                    "feature.node.kubernetes.io/ptp-capable": "yes",
                    "node-role.kubernetes.io/worker": ""
                  },
                  "ptpEventConfig": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
          createdAt: "2024-06-12"
          description: This software enables configuration of Precision Time Protocol(PTP)
            on Kubernetes. It detects hardware capable PTP devices on each node, and
            configures linuxptp processes such as ptp4l, phc2sys and timemaster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.3.0-0 <4.18.0-202509100344'
          operatorframework.io/suggested-namespace: openshift-ptp
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.22.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/ptp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodePtpDevice is the Schema for the nodeptpdevices API
            displayName: Node Ptp Device
            kind: NodePtpDevice
            name: nodeptpdevices.ptp.openshift.io
            version: v1
          - description: PtpConfig is the Schema for the ptpconfigs API
            displayName: Ptp Config
            kind: PtpConfig
            name: ptpconfigs.ptp.openshift.io
            version: v1
          - description: PtpOperatorConfig is the Schema for the ptpoperatorconfigs
              API
            displayName: Ptp Operator Config
            kind: PtpOperatorConfig
            name: ptpoperatorconfigs.ptp.openshift.io
            version: v1
        description: |
          # Precision Time Protocol (PTP) Operator for Openshift

          ## Introdution
          PTP Operator manages cluster wide PTP configuration.
          This operator has to run in namespace 'openshift-ptp'. An Operator Group is also required to install this operator:
          ```
          $ oc create -f - <<EOF
          apiVersion: v1
          kind: Namespace
          metadata:
            name: openshift-ptp
            labels:
              name: openshift-ptp
              openshift.io/cluster-monitoring: "true"
          EOF

          $ oc create -f - <<EOF
          apiVersion: operators.coreos.com/v1
          kind: OperatorGroup
          metadata:
            name: ptp-operators
            namespace: openshift-ptp
          spec:
            targetNamespaces:
            - openshift-ptp
          EOF
          ```
        displayName: PTP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ptp
        links:
        - name: Documentation operator
          url: https://github.com/openshift/ptp-operator
        - name: Documentation operand
          url: https://github.com/openshift/linuxptp-daemon
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-cloud-event-proxy-rhel9@sha256:87c29bad5a920ea676a5aa8f6e582492547e391ea166694f84cbb75044718a19
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:467557f453bde3feaa0129f38c52fba00c3b98acddbf8978bf21ef13a9890fab
        - registry.redhat.io/openshift4/ose-ptp-rhel9-operator@sha256:86db22fdfb52fb3cdce61c9e919026b1a70109373f26b7e54a20ab627e97448c
        - registry.redhat.io/openshift4/ose-ptp-rhel9@sha256:a29fc3bfed3add2768499e94435e8993bfd704a8a46268e6e63a8d20d2c4ca6c
        - registry.redhat.io/openshift4/ose-ptp-operator-bundle@sha256:7202697ea75c3503d4165b1d970802ac776c2977755a8b240094046f74c162c0
        version: 4.18.0-202509100344
      entries:
      - name: ptp-operator.v4.18.0-202509100344
        version: 4.18.0-202509100344
      - name: ptp-operator.v4.18.0-202508201347
        version: 4.18.0-202508201347
      - name: ptp-operator.v4.18.0-202507211933
        version: 4.18.0-202507211933
      - name: ptp-operator.v4.18.0-202507081733
        version: 4.18.0-202507081733
      - name: ptp-operator.v4.18.0-202506260833
        version: 4.18.0-202506260833
      - name: ptp-operator.v4.18.0-202506121732
        version: 4.18.0-202506121732
      - name: ptp-operator.v4.18.0-202505291136
        version: 4.18.0-202505291136
      - name: ptp-operator.v4.18.0-202505200035
        version: 4.18.0-202505200035
      - name: ptp-operator.v4.18.0-202505071704
        version: 4.18.0-202505071704
      - name: ptp-operator.v4.18.0-202504282205
        version: 4.18.0-202504282205
      - name: ptp-operator.v4.18.0-202504231304
        version: 4.18.0-202504231304
      - name: ptp-operator.v4.18.0-202504170406
        version: 4.18.0-202504170406
      - name: ptp-operator.v4.18.0-202504090803
        version: 4.18.0-202504090803
      - name: ptp-operator.v4.18.0-202504021503
        version: 4.18.0-202504021503
      - name: ptp-operator.v4.18.0-202503211332
        version: 4.18.0-202503211332
      - name: ptp-operator.v4.18.0-202503121135
        version: 4.18.0-202503121135
      - name: ptp-operator.v4.18.0-202502250302
        version: 4.18.0-202502250302
      - name: ptp-operator.v4.18.0-202501230001
        version: 4.18.0-202501230001
      name: stable
    defaultChannel: stable
    packageName: ptp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: cluster-kube-descheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: clusterkubedescheduleroperator.4.13.0-202309181427
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
          createdAt: 2021/10/13
          description: An operator to run descheduler in Openshift cluster.
          olm.skipRange: '>=4.10.0 <4.13.0-202309181427'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator-bundle@sha256:b473fba287414d3ccb09aaabc64f463af2c912c322ca2c41723020b216d98d14
        - registry.redhat.io/openshift4/ose-cluster-kube-descheduler-operator@sha256:ba0b71ff2a30a069b4a8a8f3c1e0898aaadc6db112e4cc12aff7c77ced7a0405
        - registry.redhat.io/openshift4/ose-descheduler@sha256:45dc69ad93ab50bdf9ce1bb79f6d98f849e320db68af30475b10b7f5497a1b13
        version: 4.13.0-202309181427
      entries:
      - name: clusterkubedescheduleroperator.4.13.0-202309181427
        version: 4.13.0-202309181427
      name: "4.13"
    - currentCSV: clusterkubedescheduleroperator.v5.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "KubeDescheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-kube-descheduler-operator"
                },
                "spec": {
                  "deschedulingIntervalSeconds": 3600,
                  "managementState": "Managed",
                  "mode": "Predictive",
                  "profiles": ["AffinityAndTaints"]
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
          createdAt: 2025/06/23
          description: An operator to run descheduler in Openshift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.0.0 <5.1.3'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-kube-descheduler-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/cluster-kube-descheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeDescheduler is the Schema for the deschedulers API
            displayName: Kube Descheduler
            kind: KubeDescheduler
            name: kubedeschedulers.operator.openshift.io
            version: v1
        description: |
          The Kube Descheduler Operator provides the ability to evict a running pod so that the pod can be rescheduled onto a more suitable node.

          There are several situations where descheduling can benefit your cluster:

          * Nodes are underutilized or overutilized.
          * Pod and node affinity requirements, such as taints or labels, have changed and the original scheduling decisions are no longer appropriate for certain nodes.
          * Node failure requires pods to be moved.
          * New nodes are added to clusters.

          ## Descheduler Profiles

          Once the operator is installed, you can configure one or more profiles to identify pods to evict. The scheduler will schedule the replacement of the evicted pods.

          The following profiles are available:

          * AffinityAndTaints
          * TopologyAndDuplicates
          * SoftTopologyAndDuplicates
          * LifecycleAndUtilization
          * LongLifecycle
          * CompactAndScale
          * EvictPodsWithPVC
          * EvictPodsWithLocalStorage

          These profiles are documented in detail in the [descheduler operator README](https://github.com/openshift/cluster-kube-descheduler-operator#profiles).

          ## Additional Parameters

          In addition to the profiles, the following parameters can be configured:

          * `deschedulingIntervalSeconds` - Set the number of seconds between descheduler runs. A value of `0` in this field runs the descheduler once and exits.
          * `mode` - Configures the descheduler to either evict pods or to simulate the eviction.
          * `profileCustomizations` - Allows certain profile parameters to be tweaked, such as `podLifetime` (see [README](https://github.com/openshift/cluster-kube-descheduler-operator#profile-customizations) for more info).
          * `evictionLimits` - Restricts the number of evictions during each descheduling run.
        displayName: Kube Descheduler Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - scheduling
        - descheduler
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/cluster-kube-descheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/kube-descheduler-operator/descheduler-rhel9@sha256:4c653bee02c0135c8625fea6a01e18a046013ee330c5253ff66a7cd705ba08c1
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-rhel9-operator@sha256:f139f223bddc4a5f4bdbc7b096b2e3a58e7f4102104dca411afe0b653c181bf5
        - registry.redhat.io/kube-descheduler-operator/kube-descheduler-operator-bundle@sha256:37afe091e3b9656c107196d7120ce171d8b0bf895264a9791c4559e97c81a00a
        version: 5.1.3
      entries:
      - name: clusterkubedescheduleroperator.v5.1.3
        version: 5.1.3
      - name: clusterkubedescheduleroperator.v5.1.2
        version: 5.1.2
      - name: clusterkubedescheduleroperator.v5.1.1
        version: 5.1.1
      - name: clusterkubedescheduleroperator.v5.1.0
        version: 5.1.0
      - name: clusterkubedescheduleroperator.v5.0.1
        version: 5.0.1
      - name: clusterkubedescheduleroperator.4.14.0-202311021650
        version: 4.14.0-202311021650
      - name: clusterkubedescheduleroperator.4.14.0-202310201027
        version: 4.14.0-202310201027
      name: stable
    defaultChannel: stable
    packageName: cluster-kube-descheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: t8c-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: t8c-operator.v42.2.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Xl","metadata":{"name":"xl-release"},"spec":{"global":{"customImageNames":
            false,"externalArangoDBName": "arangodb","repository": "registry.connect.redhat.com/turbonomic","tag":"8.3.4"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
          createdAt: "2019-07-01T00:00:00.000Z"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/t8c-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/turbonomic/t8c-install/tree/master/operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Turbonomic Platform Operator
            kind: Xl
            name: xls.charts.helm.k8s.io
            version: v1
        description: |-
          ### Realtime Decision Automation for Multicloud Applications
          Turbonomic Workload Automation for Multicloud simultaneously optimizes performance, compliance, and cost in real-time. Workloads are precisely resourced, automatically, to perform while satisfying business constraints:
          * Continuous placement of workload across multiple clouds both on-prem and public clouds providers.
          * Continuous scaling for applications and the underlying infrastructure.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a public APIs already exposed by application and infrastructure instrumentation to discover and monitor your environment.
          Turbonomic determines the right actions that drive continuous health, including continuous placement and continuous scaling for applications and the underlying cluster.
          Turbonomic leverages the built-on orchestration provided by the application and infrastructure deployment tools and automates the execution of these actions to continiously meet the respective service level objective of each application service.
        displayName: Turbonomic Platform Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Turbonomic Platform Operator
          url: https://github.com/turbonomic/t8c-install/tree/master/operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - registry.connect.redhat.com/turbonomic/t8c-operator-bundle@sha256:afbe40e3dd6894210741fa65a3f0bb6e4a3d00ca261eb6bc3adac6a22251a74e
        - registry.marketplace.redhat.com/rhm/turbonomic/t8c-operator@sha256:d5164e6dd24a0feff10cab62530cafff1bc92d0e8ec2d16c6f4d880c6e2f50bd
        version: 42.2.0
      entries:
      - name: t8c-operator.v42.2.0
        version: 42.2.0
      name: stable
    defaultChannel: stable
    packageName: t8c-certified-rhmp
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/instance: vfunction-operator
      app.kubernetes.io/managed-by: operator
      app.kubernetes.io/name: vfunction
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      name: vfunction-server-operator
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: vFunction
      provider-url: ""
    name: vfunction-server-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: vfunction-server-operator.v2.2.469
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "vfunction.com/v1",
                "kind": "VfunctionServer",
                "metadata": {
                  "name": "example-vfunction-server",
                  "namespace": "vfunction",
                  "labels": {
                      "name": "vfunction",
                      "app.kubernetes.io/name": "vfunction",
                      "app.kubernetes.io/instance": "vfunction-server",
                      "app.kubernetes.io/managed-by": "operator"
                  }
                },
                "spec": {
                  "admin": {
                    "email": "admin@mycompany.com",
                    "name": "Admin",
                    "password": "Password1!"
                  },
                  "host": "http://my.domain.com",
                  "org_name": "MyCompany",
                  "upgrade": "Daily",
                  "measurements": {
                    "auto_scaling": "No",
                    "min_number_of_services": "2",
                    "max_number_of_services": "10"
                  },
                  "smtp": {
                    "identity": "",
                    "password": "",
                    "url": "",
                    "user": ""
                  },
                  "tls": {
                    "crt": "",
                    "key": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools, AI/Machine Learning
          certified: Red Hat certified
          containerImage: registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
          createdAt: "2021-08-03T19:08:54Z"
          description: vFunction is a cutting-edge code analysis, machine learning,
            and automation to boost your Java modernization projects.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/vfunction-server-operator-rhmp/support?utm_source=openshift_console
          support: vFunction
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an instance of vFunction running on your cluster.
            displayName: vFunction
            kind: VfunctionServer
            name: vfunctionservers.vfunction.com
            version: v1
        description: |2-

          ## Introduction

          [vFunction](https://www.vfunction.com/) is a cloud-native modernization platform that combines dynamic and static code analysis, machine learning, and automation to automatically identify and extract services from existing applications. vFunction is the only platform purpose-built for modernization of Java applications.

          The vFunction modernization process starts by learning the running monolithic application, and surfacing the interdependencies within it. Using AI, the platform analyzes and identifies services that can be separated from the application. This decomposition can present a range of micro, mini, or even macro services, depending on your application environment, each being an independently deployable and scalable application component.

          vFunction automates the extraction of these services, enabling you to modernize your monolith, quickly and easily.

          ## The vFunction platform

          The platform consists of 3 basic components; the server, the controller package, and a tools package. The server runs as an operand on an OpenShift environment. The controller package is installed on the machine that runs the monolithic application which can be either a Linux or a Windows machine, and the tools are run on a development machine, with access to the code of the monolithic application.

          The controller package consists of three elements: the vFunction agent, that collects data during the dynamic analysis phase; the vFunction Viper application, that performs static analysis on the binaries of the application; and the vFunction controller that handles all the communication between the agent, Viper, and the vFunction server.

          The vFunction agent is a mix of a Java and native agent, and needs to run on the JVM that is currently running your application. [Refer to the vFunction OOB Support Matrix document](https://drive.google.com/file/d/1ccq8LFab1FrYAimDUxwgjiCOdk4QuVzs/view) for a list of supported application servers and JVMs.

          This operator installs a vFunction server instance to be connected later with one or more vFunction controllers installed on your application machine(s).

          ## Before You Start

          The vFunction operator requires a Red Hat OpenShift Kubernetes Application Platform 4.x.

          The cluster recommended configuration (which is the same as the minimal one) should consist of:

          - *1 master node*

          - *1 worker node*

          - *Storage capacity enough for dynamically provision 2 PVs of 50G each*

          - *A default StorageClass configured*

          vFunction doesn’t provide any encryption mechanism for data stored on any attached storage. If required, encryption can be achieved by the client by using its own external means on the storage itself.

          vFunction supports backup of all critical data as part of the application. Nevertheless, for complete backup of entire data it is recommended that the client apply its own storage based backup mechanism.

          ## Install Prerequisites

          ###Cluster:
          The operator was certified and tested on OCP 4.6 and 4.7.
          If you encounter any issue with other OCP versions or cloud providers please contact info@vfunction.com.

          ###Storage:
          The operator creates two new PersistentVolumeClaims (PVCs) during installation time. For their creation, it relies on the default StorageClass to dynamically provision two PersistentVolumes (PVs). The system must have a default StorageClass in place.

          Both PersistentVolumes are accessed with ReadWriteOnce mode.

          ###Project:
          The operator should be installed in a new and dedicated project (namespace). If you intend to install more than one vFunction operands in the same cluster, please use different projects for each one. Installation of more than one operand on the same cluster will allow scaling by load balancing multiple applications to different vFunction servers.

          ## Installation instructions

          1. Prepare the YAML file to use in the installation:

              a. Copy the following YAML template into a text editor:

              ```yaml
                  apiVersion: vfunction.com/v1
                  kind: VfunctionServer
                  metadata:
                    name: vfunction
                    namespace: vfunction
                    labels:
                      name: vfunction
                      app.kubernetes.io/name: vfunction
                      app.kubernetes.io/instance: vfunction-server
                      app.kubernetes.io/managed-by: operator
                  spec:
                    host: "http://my.domain.com"
                    org_name: "MyCompany"
                    upgrade: "Daily"
                    admin:
                      email: "admin@mycompany.com"
                      name: "Admin"
                      password: "Password1!"
                    smtp:
                      password: ""
                      url: ""
                      identity: ""
                      user: ""
                    measurements:
                      auto_scaling: "No"
                      min_number_of_services: "2"
                      max_number_of_services: "10"
                    tls:
                      crt: |
                        -----BEGIN CERTIFICATE-----
                        ...
                        -----END CERTIFICATE-----
                      key: |
                        -----BEGIN PRIVATE KEY-----
                        ...
                        -----END PRIVATE KEY-----
              ```
              b. Customize the template fields by replacing with your information:

              - **host**: Enter the FQDN for accessing the vFunction dashboard.
                          Make sure that you use a domain name and not an IP address. Writing "https://" will mandate TLS while writing "http://" will expose the server through HTTP.
              - **org_name**: Enter your organization name.
              - **upgrade**: Choose auto-upgrade mode
                  - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
                  - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
                  - **Never**: The automatic upgrading mechanism is off.
                  - **Maintenance**: Put the operator in a maintenance mode.

              - **admin.email**: Enter the email address of your administrator.
              - **admin.name**: Enter the name of your administrator.
              - **admin.password**: Enter the password you want to use for the vFunction administrator.
                                    The password should be at least 8 characters long, and consist of at least one lowercase letter, at least one uppercase letter, at least one number, and at least one special character.
              - **smtp.user**: Enter the email address for a designated user for the SMTP server (optional).
              - **smtp.password**: Enter this user’s password (optional).
              - **smtp.identity**: Enter this user’s password (optional).
              - **smtp.url**: Enter the SMTP server URL (optional).
              - **tls.crt**: If your host FQDN starts with “https”, paste in the certifications you have for using the TLS connection.
              - **tls.key**: If your host FQDN starts with “https”, enter the key you have for using the TLS connection.
              - **measurements.auto_scaling**: Indicates if the measurement services auto scalling is active.
              - **measurements.min_num_of_services**: Minimum Number of measurement service pods.
              - **measurements.max_num_of_services**: Maximum Number of measurement service pods.
              - **measurements.S3**: Save all measurements data to S3, instead of local PV (optional).
                  - **measurements.S3.bucket**: Measurements S3 bucket name.
                  - **measurements.S3.key**: Measurements S3 key.
                  - **measurements.S3.secret**: Measurements S3 secret.
                  - **measurements.S3.region**: Measurements S3 region.
              - **backup.mysql.S3**: Backup MySQL to S3 (optional).
                  - **backup.mysql.S3.bucket**: MySQL backup S3 bucket name.
                  - **backup.mysql.S3.key**: MySQL backup S3 key.
                  - **backup.mysql.S3.secret**: MySQL backup S3 secret.
                  - **backup.mysql.S3.region**: MySQL backup S3 region.

          1. In the Installed Operators window choose and click on vFunction Operator and then Create Instance.
          1. In the Create vFunctionServer window, paste the YAML file that you created in step 1.
          1. Click Create.
          1. You can now click on your new created vFunction operand in order to see its details and installation progress.

          ## Verifying the installation

          Check that the operand installed successfully:

          1. Check that the "Operand State" property shows “Working”.
          1. Check that the "Successfully Installed" property shows "Yes".

          If "Operand State" shows "Failed", it indicates the operand wasn’t installed correctly. Check the events and logs for all pods (operator and image containers) for any issues. In the event that you cannot troubleshoot, contact vFunction support.

          ## After installation

          The vFunction site is now accessible via the newly created vFunction application custom address, for example, my.domain.com. There are two ways you can access the vFunction dashboard.

          ### Access using the router canonical hostname

          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address, and to the ”vfunction” subdomain of the  OpenShift canonical router hostname as the alias.
          For example, ***my.domain.com.   CNAME   vfunction.apps.ocp4.my-openshift.com.***
          2. Find your cluster Router Canonical Hostname address in the newly created vfunction-route-xxx under your vFunction OpenShift project -> **Networking** -> **Routes** -> **vfunction-route-xxx** route -> **Router Canonical Hostname** field.
          You can now access the dashboard using your defined "host" spec property (as above).

          ### Access using the nginx service location

          You may use this access method if your OpenShift is installed on a provider that supports exposing LoadBalancer-type services.
          1. Update your DNS provider by creating a canonical name (CNAME) record.
          This record should point to your host address and to the vfunction-nginx-xxx service location.
          For example, ***my.domain.com.   CNAME   a05951ed7cdf-1394239323.us-east-1.elb.amazonaws.com.***
          1. Point your custom domain to the vfunction-nginx-xxx service's external IP location, which you can find in the OpenShift project > **Networking** -> **Services** -> **vfunction-nginx-xxx** service > **Service Address** > **Location** field.

          ## Upgrade and Rollback

          The vFunction operator includes a built-in auto-upgrade mechanism.

          You can choose one of 4 upgrading modes:
          - **Daily**: The operator will check for a new version every day at 3 AM and will install it automatically (default).
          - **Always**: The operator will check for a new version every 10 minutes and will install it automatically.
          - **Never**: The automatic upgrading mechanism is off.
          - **Maintenance**: Put the operator in a maintenance mode.

          For an on-demand upgrade,  you can change anytime the upgrading mode from “Never” to “Always”, wait for the operand to be upgraded, and change again to “Never”.

          During the upgrade, the "Operand State" property will change to “Upgrading...” and back again to “Working” after a successful upgrade.

          In case of a failure upgrading the operand, an automatic rollback to the last working version will occur. The failed version will be marked as defective and the operator will not try to upgrade to it again.

          Choosing the "Maintenance" option will tell the operator to take down all pods, so you can safely fix storage issues, return volumes from snapshots, etc.
          After finishing the maintenance, return the upgrade mode to your original desired policy.
        displayName: vFunction Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - vfunction
        - analyzing
        - services
        links:
        - name: vFunction
          url: https://vfunction.com
        - name: Terms
          url: https://www.vfunction.com/eula/
        maintainers:
        - email: raviv@vfunction.com
          name: Raviv Halivni
        maturity: alpha
        provider:
          name: vFunction
        relatedImages:
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-organizations@sha256:bcc2c3d6222af95e614318daea3108fffc6bc77aa17b23031cce296a3811b79f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-server-operator@sha256:60d1965460101d0fd90ecc9499d0a6d4c9684a54ada61a43a253d30661fab22f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-nginx@sha256:435d1489aaf29a811081ddb51f40ae692e5833f42c31dbced6bc189c0ddc40ab
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-measurements@sha256:35ffbee4b0eead58706b23cf5a702b1d55d2a7913cd0019cefbff94957d4e5d6
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-mysql@sha256:a6e06ea08fffa33b709d4d1240587bb3ab5b4cc6a28f493a227555e88e645798
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-idp@sha256:21ca95a0209f3c557a45658452d154777f928539a93129148470562c79e68d70
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-storage@sha256:b778cb9c724a59e51b4cb9088047be471cdf338f1138c31b81f79e50a491203f
        - registry.marketplace.redhat.com/rhm/vfunction/vfunction-vfapi-users@sha256:e47cc051953976a0e075e8cfb457cd4749e9b9e4394f2fbecab2458c2ade92ca
        - registry.connect.redhat.com/vfunction/vfunction-server-operator-bundle@sha256:f1784afe63b2e834a6b0bf6da254339ffe3597d02dc2d389f9162d806bedbe74
        version: 2.2.469
      entries:
      - name: vfunction-server-operator.v2.2.469
        version: 2.2.469
      name: stable
    defaultChannel: stable
    packageName: vfunction-server-operator-rhmp
    provider:
      name: vFunction
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-organizations-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-organizations-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "organizations.services.k8s.aws/v1alpha1",
                "kind": "OrganizationalUnit",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
          createdAt: "2025-09-19T19:35:20Z"
          description: AWS Organizations controller is a service controller for managing
            Organizations resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Account represents the state of an AWS organizations Account
              resource.
            displayName: Account
            kind: Account
            name: accounts.organizations.services.k8s.aws
            version: v1alpha1
          - description: OrganizationalUnit represents the state of an AWS organizations
              OrganizationalUnit resource.
            displayName: OrganizationalUnit
            kind: OrganizationalUnit
            name: organizationalunits.organizations.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Organizations resources in AWS from within your Kubernetes cluster.

          **About Amazon Organizations**

          With AWS Organizations you can perform account management activities at scale by consolidating multiple AWS accounts into a single organization. Consolidating accounts simplifies how you use other AWS services. You can leverage the multi-account management services available in AWS Organizations with select AWS services to perform tasks on all accounts that are members of your organization.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Organizations
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - organizations
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Organizations Developer Resources
          url: https://aws.amazon.com/organizations/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: organizations maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/organizations-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-organizations-controller@sha256:76ea247239dde4e52e5016944a41908de8fdcc4173cad8dfff04413a0e3cdf48
        version: 1.1.1
      entries:
      - name: ack-organizations-controller.v1.1.1
        version: 1.1.1
      - name: ack-organizations-controller.v1.0.13
        version: 1.0.13
      - name: ack-organizations-controller.v1.0.11
        version: 1.0.11
      - name: ack-organizations-controller.v1.0.10
        version: 1.0.10
      - name: ack-organizations-controller.v1.0.9
        version: 1.0.9
      - name: ack-organizations-controller.v1.0.8
        version: 1.0.8
      - name: ack-organizations-controller.v1.0.7
        version: 1.0.7
      - name: ack-organizations-controller.v1.0.6
        version: 1.0.6
      - name: ack-organizations-controller.v1.0.5
        version: 1.0.5
      - name: ack-organizations-controller.v1.0.4
        version: 1.0.4
      - name: ack-organizations-controller.v1.0.1
        version: 1.0.1
      - name: ack-organizations-controller.v1.0.0
        version: 1.0.0
      - name: ack-organizations-controller.v0.0.10
        version: 0.0.10
      - name: ack-organizations-controller.v0.0.9
        version: 0.0.9
      - name: ack-organizations-controller.v0.0.8
        version: 0.0.8
      - name: ack-organizations-controller.v0.0.7
        version: 0.0.7
      - name: ack-organizations-controller.v0.0.6
        version: 0.0.6
      name: alpha
    defaultChannel: alpha
    packageName: ack-organizations-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Upbound Inc.
      provider-url: https://upbound.io
    name: universal-crossplane
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: universal-crossplane.1.5.1-up.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Configuration",
                "metadata":{
                  "name":"xp-getting-started-with-aws"
                },
                "spec":{
                  "package":"registry.upbound.io/xp/getting-started-with-aws:v1.2.2"
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"CompositeResourceDefinition",
                "metadata":{
                  "name":"compositepostgresqlinstances.database.example.org"
                },
                "spec":{
                  "group":"database.example.org",
                  "names":{
                    "kind":"CompositePostgreSQLInstance",
                    "plural":"compositepostgresqlinstances"
                  },
                  "claimNames":{
                    "kind":"PostgreSQLInstance",
                    "plural":"postgresqlinstances"
                  },
                  "connectionSecretKeys":[
                      "username",
                      "password",
                      "endpoint",
                      "port"
                  ],
                  "versions":[
                    {
                      "name":"v1alpha1",
                      "served":true,
                      "referenceable":true,
                      "schema":{
                        "openAPIV3Schema":{
                          "type":"object",
                          "properties":{
                            "spec":{
                              "type":"object",
                              "properties":{
                                "parameters":{
                                  "type":"object",
                                  "properties":{
                                    "storageGB":{
                                      "type":"integer"
                                    }
                                  },
                                  "required":[
                                      "storageGB"
                                  ]
                                }
                              },
                              "required":[
                                  "parameters"
                              ]
                            }
                          }
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion":"apiextensions.crossplane.io/v1",
                "kind":"Composition",
                "metadata":{
                  "name":"compositepostgresqlinstances.aws.database.example.org",
                  "labels":{
                    "provider":"aws",
                    "guide":"quickstart",
                    "vpc":"default"
                  }
                },
                "spec":{
                  "writeConnectionSecretsToNamespace":"crossplane-system",
                  "compositeTypeRef":{
                    "apiVersion":"database.example.org/v1alpha1",
                    "kind":"CompositePostgreSQLInstance"
                  },
                  "resources":[
                    {
                      "name":"rdsinstance",
                      "base":{
                        "apiVersion":"database.aws.crossplane.io/v1beta1",
                        "kind":"RDSInstance",
                        "spec":{
                          "forProvider":{
                            "region":"us-east-1",
                            "dbInstanceClass":"db.t2.small",
                            "masterUsername":"masteruser",
                            "engine":"postgres",
                            "engineVersion":"9.6",
                            "skipFinalSnapshotBeforeDeletion":true,
                            "publiclyAccessible":true
                          },
                          "writeConnectionSecretToRef":{
                            "namespace":"crossplane-system"
                          }
                        }
                      },
                      "patches":[
                        {
                          "fromFieldPath":"metadata.uid",
                          "toFieldPath":"spec.writeConnectionSecretToRef.name",
                          "transforms":[
                            {
                              "type":"string",
                              "string":{
                                "fmt":"%s-postgresql"
                              }
                            }
                          ]
                        },
                        {
                          "fromFieldPath":"spec.parameters.storageGB",
                          "toFieldPath":"spec.forProvider.allocatedStorage"
                        }
                      ],
                      "connectionDetails":[
                        {
                          "fromConnectionSecretKey":"username"
                        },
                        {
                          "fromConnectionSecretKey":"password"
                        },
                        {
                          "fromConnectionSecretKey":"endpoint"
                        },
                        {
                          "fromConnectionSecretKey":"port"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion":"pkg.crossplane.io/v1",
                "kind":"Provider",
                "metadata":{
                  "name":"provider-aws"
                },
                "spec":{
                  "package":"crossplane/provider-aws:master"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Application Runtime, Cloud Provider, Database, Integration &
            Delivery, Networking, Storage
          com.redhat.component: universal-crossplane
          com.redhat.delivery.backport: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.6
          containerImage: upbound/uxp-bootstrapper
          createdAt: 2021-05-18T00:00:00+0000
          description: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          io.k8s.display-name: universal-crossplane
          io.openshift.tags: uxp,crossplane,upbound
          maintainer: Upbound Inc. <info@upbound.io>
          name: universal-crossplane
          ocs.tags: v4.6
          operatorframework.io/initialization-resource: |
            apiVersion: v1
            kind: Secret
            metadata:
              name: upbound-control-plane-token
              namespace: upbound-system
            type: Opaque
            data:
              token: "Output of `up cloud controlplane attach | base64` command"
          operatorframework.io/suggested-namespace: upbound-system
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io.bundle.channels.v1: stable
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: universal-crossplane
          operators.operatorframework.io/internal-objects: '["locks.pkg.crossplane.io"]'
          repository: https://github.com/upbound/universal-crossplane
          summary: Upbound Universal Crossplane (UXP) is Upbound's official enterprise-grade
            distribution of Crossplane.
          support: Upbound Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - displayName: CompositeResourceDefinition
            kind: CompositeResourceDefinition
            name: compositeresourcedefinitions.apiextensions.crossplane.io
            version: v1
          - displayName: CompositionRevision
            kind: CompositionRevision
            name: compositionrevisions.apiextensions.crossplane.io
            version: v1alpha1
          - displayName: Composition
            kind: Composition
            name: compositions.apiextensions.crossplane.io
            version: v1
          - displayName: ConfigurationRevision
            kind: ConfigurationRevision
            name: configurationrevisions.pkg.crossplane.io
            version: v1
          - displayName: Configuration
            kind: Configuration
            name: configurations.pkg.crossplane.io
            version: v1
          - displayName: ControllerConfig
            kind: ControllerConfig
            name: controllerconfigs.pkg.crossplane.io
            version: v1alpha1
          - displayName: Lock
            kind: Lock
            name: locks.pkg.crossplane.io
            version: v1alpha1
          - displayName: ProviderRevision
            kind: ProviderRevision
            name: providerrevisions.pkg.crossplane.io
            version: v1
          - displayName: Provider
            kind: Provider
            name: providers.pkg.crossplane.io
            version: v1
        description: |
          ## Overview

          Upbound Universal Crossplane (UXP) is [Upbound's][upbound] official
          enterprise-grade distribution of [Crossplane][crossplane]. It's fully compatible
          with upstream Crossplane, open source, capable of connecting to Upbound Cloud
          for real-time dashboard visibility, and maintained by Upbound. It's the easiest
          way for both individual community members and enterprises to build their
          production control planes.

          ## Quick Start

          1. Install the [Upbound CLI][upbound-cli].
              ```console
              curl -sL https://cli.upbound.io | sh
              # Install with Homebrew
              brew install upbound/tap/up
              ```

          2. [Create an Upbound account][create-account] for a free dashboard for UXP.

          3. Connect UXP to Upbound Cloud.
              ```console
              # The name of your new UXP control plane.
              UXP_NAME=mycrossplane

              up cloud login
              up cloud controlplane attach ${UXP_NAME} | up uxp connect -
              ```

          4. Manage your UXP control plane by [signing in][login] to your Upbound account.

          ## Additional Resources

          - The [UXP Documentation][uxp-documentation] provides additional information
            about UXP and resources for developers, like examples.
          - The [developer guide][developer-guide] describes how to build and run UXP
            locally from source.
          - [UXP Slack][uxp-slack] is where you can go to get all of your UXP questions
            answered.

          [upbound]: https://upbound.io
          [crossplane]: https://crossplane.io/
          [upbound-cli]: https://github.com/upbound/up
          [create-account]: https://cloud.upbound.io/register
          [login]: https://cloud.upbound.io/login
          [uxp-documentation]: https://cloud.upbound.io/docs/uxp
          [developer-guide]: https://github.com/upbound/universal-crossplane/blob/main/docs/developer-guide.md
          [uxp-slack]: https://crossplane.slack.com/archives/upbound
        displayName: Upbound Universal Crossplane (UXP)
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - infrastructure
        - services
        - application
        - database
        - cache
        - bucket
        - infra
        - app
        - ops
        - oam
        - gcp
        - azure
        - aws
        - alibaba
        - cloudsql
        - rds
        - s3
        - azuredatabase
        - asparadb
        - gke
        - aks
        - eks
        links:
        - name: GitHub
          url: https://github.com/upbound/universal-crossplane
        - name: Website
          url: https://upbound.io
        - name: Twitter
          url: https://twitter.com/upbound_io
        - name: Slack
          url: https://slack.crossplane.io/
        maintainers:
        - email: info@upbound.io
          name: Upbound Inc.
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: Upbound Inc.
          url: https://upbound.io
        relatedImages:
        - quay.io/openshift-community-operators/universal-crossplane@sha256:b3104adb471083de96d49fa53f663dcf0e3e7ff5e3573bb1422aa6045277bea0
        - upbound/crossplane:v1.5.1-up.1
        - upbound/uxp-bootstrapper:v1.5.1-up.1
        - upbound/xgql:v0.1.5
        version: 1.5.1-up.1
      entries:
      - name: universal-crossplane.1.5.1-up.1
        version: 1.5.1-up.1
      - name: universal-crossplane.1.4.3-up.1
        version: 1.4.3-up.1
      - name: universal-crossplane.1.3.3-up.1
        version: 1.3.3-up.1
      - name: universal-crossplane.1.3.1-up.1
        version: 1.3.1-up.1
      - name: universal-crossplane.1.2.3-up.1
        version: 1.2.3-up.1
      - name: universal-crossplane.1.2.2-up.1
        version: 1.2.2-up.1
      - name: universal-crossplane.1.2.1-up.4
        version: 1.2.1-up.4
      - name: universal-crossplane.1.2.1-up.3
        version: 1.2.1-up.3
      name: stable
    defaultChannel: stable
    packageName: universal-crossplane
    provider:
      name: Upbound Inc.
      url: https://upbound.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: hyperfoil-bundle
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hyperfoil-operator.v0.26.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "hyperfoil.io/v1alpha2",
                "kind": "Hyperfoil",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "hyperfoil-operator",
                    "app.kubernetes.io/instance": "hyperfoil-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "hyperfoil",
                    "app.kubernetes.io/part-of": "hyperfoil-operator"
                  },
                  "name": "hyperfoil-sample"
                },
                "spec": {
                  "version": "latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/hyperfoil/hyperfoil-operator:0.26.0
          createdAt: "2024-06-11T17:39:20Z"
          description: Microservice-oriented distributed benchmark framework.
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Hyperfoil/hyperfoil-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Hyperfoil is the Schema for the hyperfoils API
            displayName: Hyperfoil
            kind: Hyperfoil
            name: hyperfoils.hyperfoil.io
            version: v1alpha2
        description: |
          Hyperfoil is a modern benchmark framework oriented promising
          accurate results with more flexibility and distributed architecture.

          See more on [hyperfoil.io](http://hyperfoil.io).

          Hyperfoil is licensed under Apache License 2.0
        displayName: Hyperfoil
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hyperfoil
        - benchmark
        - test
        - load
        - driver
        links:
        - name: Website
          url: http://hyperfoil.io
        - name: GitHub
          url: http://github.com/Hyperfoil
        maintainers:
        - email: rvansa@redhat.com
          name: Radim Vansa
        - email: spederse@redhat.com
          name: Ståle W. Pedersen
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.15.0
        - quay.io/community-operator-pipeline-prod/hyperfoil-bundle@sha256:1d3625d12d95deb4f48660459fd0fc8b250b52b46d68136927d0f12465aaadf1
        - quay.io/hyperfoil/hyperfoil-operator:0.26.0
        version: 0.26.0
      entries:
      - name: hyperfoil-operator.v0.26.0
        version: 0.26.0
      - name: hyperfoil-operator.v0.24.2
        version: 0.24.2
      - name: hyperfoil-operator.v0.21.0
        version: 0.21.0
      - name: hyperfoil-operator.v0.15.1
        version: 0.15.1
      - name: hyperfoil-operator.v0.15.0
        version: 0.15.0
      name: alpha
    defaultChannel: alpha
    packageName: hyperfoil-bundle
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Weights & Biases
      provider-url: https://wandb.ai
    name: wandb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: wandb-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apps.wandb.com/v1",
                "kind": "WeightsAndBiases",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "wandb-operator",
                    "app.kubernetes.io/instance": "weightsandbiases-sample",
                    "app.kubernetes.io/managed-by": "olm",
                    "app.kubernetes.io/name": "weightsandbiases",
                    "app.kubernetes.io/part-of": "wandb-operator"
                  },
                  "name": "weightsandbiases-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning,Developer Tools
          certified: "false"
          containerImage: quay.io/wandb_tools/wandb-k8s-operator:1.0.0
          createdAt: "2024-10-10T14:46:54Z"
          description: Operator for Weights and Biases, enabling seamless integration
            and management of machine learning experiments in Kubernetes environments.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: wandb-system
          repository: https://github.com/wandb/operator
          support: WeightsAndBiases
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: WeightsAndBiases is a custom resource for managing W&B configurations.
            displayName: Weights And Biases
            kind: WeightsAndBiases
            name: weightsandbiases.apps.wandb.com
            version: v1
        description: |-
          The Weights and Biases Operator simplifies the management of machine learning experiments, tracking, and infrastructure. It integrates with Kubernetes to deploy and manage Weights and Biases instances.
          Features:
          - Experiment Tracking
          - Hyperparameter Tuning
          - Model Management
          - Integration with popular ML frameworks
        displayName: Weights & Biases Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - machine learning
        - experiment tracking
        - ML infrastructure
        - weights and biases
        links:
        - name: Weights & Biases Documentation
          url: https://docs.wandb.ai/
        maintainers:
        - email: abhinav.garg@domain.com
          name: Abhinav Garg
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Weights & Biases
          url: https://wandb.ai
        relatedImages:
        - quay.io/wandb_tools/wandb-k8s-operator:1.0.0
        - quay.io/community-operator-pipeline-prod/wandb-operator@sha256:9bebaf0e72f0b3ce20dabf7ace62cb5ecd1015243438248e8101a238988cd60e
        version: 1.0.0
      entries:
      - name: wandb-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: wandb-operator
    provider:
      name: Weights & Biases
      url: https://wandb.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: netobserv-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: network-observability-operator.v1.9.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowMetric",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "netobserv-operator",
                    "app.kubernetes.io/instance": "flowmetric-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "flowmetric",
                    "app.kubernetes.io/part-of": "netobserv-operator"
                  },
                  "name": "flowmetric-sample"
                },
                "spec": {
                  "charts": [
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "",
                          "promQL": "sum(rate($METRIC[2m]))"
                        }
                      ],
                      "title": "External ingress traffic",
                      "type": "SingleStat",
                      "unit": "Bps"
                    },
                    {
                      "dashboardName": "Main",
                      "queries": [
                        {
                          "legend": "{{DstK8S_Namespace}} / {{DstK8S_OwnerName}}",
                          "promQL": "sum(rate($METRIC{DstK8S_Namespace!=\"\"}[2m])) by (DstK8S_Namespace, DstK8S_OwnerName)"
                        }
                      ],
                      "sectionName": "External",
                      "title": "Top external ingress traffic per workload",
                      "type": "StackArea",
                      "unit": "Bps"
                    }
                  ],
                  "direction": "Ingress",
                  "filters": [
                    {
                      "field": "SrcSubnetLabel",
                      "matchType": "Absence"
                    }
                  ],
                  "labels": [
                    "DstK8S_HostName",
                    "DstK8S_Namespace",
                    "DstK8S_OwnerName",
                    "DstK8S_OwnerType"
                  ],
                  "metricName": "cluster_external_ingress_bytes_total",
                  "type": "Counter",
                  "valueField": "Bytes"
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "app"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "infra"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "enable": true,
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "includeList": [
                        "node_ingress_bytes_total",
                        "workload_ingress_bytes_total",
                        "namespace_flows_total"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta2",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 1048576,
                      "logLevel": "info",
                      "metrics": {
                        "server": {
                          "port": 9400
                        }
                      },
                      "privileged": false,
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "eBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "Disabled"
                    },
                    "enable": true,
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "flow_layer": "\"app\""
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "flow_layer": "\"infra\""
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "\"Pod\"",
                          "src_kind": "\"Pod\""
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "\"Service\""
                        },
                        "name": "Services network"
                      }
                    ],
                    "replicas": 1,
                    "resources": {
                      "limits": {
                        "memory": "100Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    }
                  },
                  "deploymentModel": "Direct",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "enable": true,
                    "lokiStack": {
                      "name": "loki"
                    },
                    "mode": "Monolithic",
                    "monolithic": {
                      "tenantID": "netobserv",
                      "tls": {
                        "caCert": {
                          "certFile": "service-ca.crt",
                          "name": "loki-gateway-ca-bundle",
                          "type": "configmap"
                        },
                        "enable": false
                      },
                      "url": "http://loki.netobserv.svc:3100/"
                    },
                    "readTimeout": "30s",
                    "writeBatchSize": 10485760,
                    "writeBatchWait": "1s",
                    "writeTimeout": "10s"
                  },
                  "namespace": "netobserv",
                  "networkPolicy": {
                    "additionalNamespaces": [],
                    "enable": false
                  },
                  "processor": {
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "Flows",
                    "metrics": {
                      "disableAlerts": [],
                      "server": {
                        "port": 9401
                      }
                    },
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "prometheus": {
                    "querier": {
                      "enable": true,
                      "mode": "Auto",
                      "timeout": "30s"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring, Networking, Observability, OpenShift Optional
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
          createdAt: 2025-09-08T21:48:09
          description: Network flows collector and monitoring solution
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: '{"apiVersion":"flows.netobserv.io/v1beta2",
            "kind":"FlowCollector","metadata":{"name":"cluster"},"spec": {}}'
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
          support: NetObserv team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
          - description: '`FlowCollector` is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.'
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta2
          - description: '`FlowMetric` is the schema for the custom metrics API, which
              allows to generate more metrics out of flow logs. You can find examples
              here: https://github.com/netobserv/network-observability-operator/tree/main/config/samples/flowmetrics'
            displayName: Flow Metric
            kind: FlowMetric
            name: flowmetrics.flows.netobserv.io
            version: v1alpha1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline consisting in:
          - an eBPF agent, that generates network flows from captured packets
          - flowlogs-pipeline, a component that collects, enriches and exports these flows
          - a Console plugin for flows visualization with powerful filtering options, a topology representation and more

          Flow data is then available in multiple ways, each optional:

          - As Cluster Monitoring metrics
          - As raw flow logs stored in Grafana Loki
          - As raw flow logs exported to a collector

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, can optionally be used as the backend to store all collected flows. The Network Observability operator does not install Loki directly, however we provide some guidance to help you there.

          - For a production or production-like environment usage, refer to [the operator documentation](https://docs.openshift.com/container-platform/latest/network_observability/installing-operators.html).

          - For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 10GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/5410e65b8e05aaabd1244a9524cfedd8ac8c56b5/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          If you prefer to not use Loki, you must set `spec.loki.enable` to `false` in `FlowCollector`.
          In that case, you can still get the Prometheus metrics or export raw flows to a custom collector. But be aware that some of the Console plugin features will be disabled. For instance, you will not be able to view raw flows there, and the metrics / topology will have a more limited level of details, missing information such as pods or IPs.

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for example [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/config/samples/flows_v1beta2_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide on every node, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you might have to configure differently if you used another installation method. Make sure to disable it (`spec.loki.enable`) if you don't want to use Loki.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`) an optional list of exporters to which to send enriched flows. KAFKA and IPFIX exporters are supported. This allows you to define any custom storage or processing that can read from Kafka or use the IPFIX standard.

          - To enable availability zones awareness, set `spec.processor.addZone` to `true`.

          ## Resource considerations

          The following table outlines examples of resource considerations for clusters with certain workload sizes.
          The examples outlined in the table demonstrate scenarios that are tailored to specific workloads. Consider each example only as a baseline from which adjustments can be made to accommodate your workload needs.


          | Resource recommendations                        | Extra small (10 nodes) | Small (25 nodes)       | Medium (65 nodes) **    | Large (120 nodes) **          |
          | ----------------------------------------------- | ---------------------- | ---------------------- | ----------------------- | ----------------------------- |
          | *Worker Node vCPU and memory*                   | 4 vCPUs\| 16GiB mem *  | 16 vCPUs\| 64GiB mem * | 16 vCPUs\| 64GiB mem  * |16 vCPUs\| 64GiB Mem *         |
          | *LokiStack size*                                | `1x.extra-small`       | `1x.small`             | `1x.small`              | `1x.medium`                   |
          | *Network Observability controller memory limit* | 400Mi (default)        | 400Mi (default)        | 400Mi (default)         | 800Mi                         |
          | *eBPF sampling rate*                            | 50 (default)           | 50 (default)           | 50 (default)            | 50 (default)                  |
          | *eBPF memory limit*                             | 800Mi (default)        | 800Mi (default)        | 2000Mi                  | 800Mi (default)               |
          | *FLP memory limit*                              | 800Mi (default)        | 800Mi (default)        | 800Mi (default)         | 800Mi (default)               |
          | *FLP Kafka partitions*                          | N/A                    | 48                     | 48                      | 48                            |
          | *Kafka consumer replicas*                       | N/A                    | 24                     | 24                      | 24                            |
          | *Kafka brokers*                                 | N/A                    | 3 (default)            | 3 (default)             | 3 (default)                   |

          *. Tested with AWS M6i instances.
          **. In addition to this worker and its controller, 3 infra nodes (size `M6i.12xlarge`) and 1 workload node (size `M6i.8xlarge`) were tested.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - More information on [configuring metrics](https://github.com/netobserv/network-observability-operator/blob/1.9.1-community/docs/Metrics.md).
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - ipfix
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:09454618709a5c83beb4252f1c14c149716d1c09b6567c0a571d5390e8a639a2
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:bf79fd57e1df791fcd45dd4745c923046dd7759ff5b3762cd7a329c5bf611316
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:13b6dd0ddf4bdd2836b0d3ec9c79119a902456dcee100cfac6898768e1f22aab
        - registry.redhat.io/network-observability/network-observability-console-plugin-compat-rhel9@sha256:d130f94cdd3cc863a31c59f395a6c4f663e8e0da23747304353c6674b20d81be
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:7d92eadb3cd08a08672a65a196a5d3181e764c68e8c20cab2554933cdb10fd0c
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:0b4337f4db1e74a2bea1021a617cb42690251036f583bb771cc5ec888f38c97f
        version: 1.9.3
      entries:
      - name: network-observability-operator.v1.9.3
        version: 1.9.3
      - name: network-observability-operator.v1.9.2
        version: 1.9.2
      - name: network-observability-operator.v1.9.1
        version: 1.9.1
      - name: network-observability-operator.v1.9.0
        version: 1.9.0
      - name: network-observability-operator.v1.8.1
        version: 1.8.1
      - name: network-observability-operator.v1.8.0
        version: 1.8.0
      - name: network-observability-operator.v1.7.0
        version: 1.7.0
      - name: network-observability-operator.v1.6.2
        version: 1.6.2
      - name: network-observability-operator.v1.6.1
        version: 1.6.1
      - name: network-observability-operator.v1.6.0
        version: 1.6.0
      - name: network-observability-operator.v1.5.0
        version: 1.5.0
      - name: network-observability-operator.v1.4.2
        version: 1.4.2
      - name: network-observability-operator.v1.4.1
        version: 1.4.1
      - name: network-observability-operator.v1.4.0
        version: 1.4.0
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: stable
    - currentCSV: network-observability-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "flows.netobserv.io/v1alpha1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "metrics": {
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "flows.netobserv.io/v1beta1",
                "kind": "FlowCollector",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "agent": {
                    "ebpf": {
                      "cacheActiveTimeout": "5s",
                      "cacheMaxFlows": 100000,
                      "excludeInterfaces": [
                        "lo"
                      ],
                      "imagePullPolicy": "IfNotPresent",
                      "interfaces": [],
                      "kafkaBatchSize": 10485760,
                      "logLevel": "info",
                      "resources": {
                        "limits": {
                          "memory": "800Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "50Mi"
                        }
                      },
                      "sampling": 50
                    },
                    "type": "EBPF"
                  },
                  "consolePlugin": {
                    "autoscaler": {
                      "maxReplicas": 3,
                      "metrics": [
                        {
                          "resource": {
                            "name": "cpu",
                            "target": {
                              "averageUtilization": 50,
                              "type": "Utilization"
                            }
                          },
                          "type": "Resource"
                        }
                      ],
                      "minReplicas": 1,
                      "status": "DISABLED"
                    },
                    "imagePullPolicy": "IfNotPresent",
                    "logLevel": "info",
                    "port": 9001,
                    "portNaming": {
                      "enable": true,
                      "portNames": {
                        "3100": "loki"
                      }
                    },
                    "quickFilters": [
                      {
                        "default": true,
                        "filter": {
                          "dst_namespace!": "openshift-,netobserv",
                          "src_namespace!": "openshift-,netobserv"
                        },
                        "name": "Applications"
                      },
                      {
                        "filter": {
                          "dst_namespace": "openshift-,netobserv",
                          "src_namespace": "openshift-,netobserv"
                        },
                        "name": "Infrastructure"
                      },
                      {
                        "default": true,
                        "filter": {
                          "dst_kind": "Pod",
                          "src_kind": "Pod"
                        },
                        "name": "Pods network"
                      },
                      {
                        "filter": {
                          "dst_kind": "Service"
                        },
                        "name": "Services network"
                      }
                    ],
                    "register": true
                  },
                  "deploymentModel": "DIRECT",
                  "exporters": [],
                  "kafka": {
                    "address": "kafka-cluster-kafka-bootstrap.netobserv",
                    "tls": {
                      "caCert": {
                        "certFile": "ca.crt",
                        "name": "kafka-cluster-cluster-ca-cert",
                        "type": "secret"
                      },
                      "enable": false,
                      "userCert": {
                        "certFile": "user.crt",
                        "certKey": "user.key",
                        "name": "flp-kafka",
                        "type": "secret"
                      }
                    },
                    "topic": "network-flows"
                  },
                  "loki": {
                    "batchSize": 10485760,
                    "batchWait": "1s",
                    "maxBackoff": "5s",
                    "maxRetries": 2,
                    "minBackoff": "1s",
                    "statusTls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false,
                      "userCert": {
                        "certFile": "tls.crt",
                        "certKey": "tls.key",
                        "name": "loki-query-frontend-http",
                        "type": "secret"
                      }
                    },
                    "tls": {
                      "caCert": {
                        "certFile": "service-ca.crt",
                        "name": "loki-gateway-ca-bundle",
                        "type": "configmap"
                      },
                      "enable": false,
                      "insecureSkipVerify": false
                    },
                    "url": "http://loki.netobserv.svc:3100/"
                  },
                  "namespace": "netobserv",
                  "processor": {
                    "conversationEndTimeout": "10s",
                    "conversationHeartbeatInterval": "30s",
                    "conversationTerminatingTimeout": "5s",
                    "dropUnusedFields": true,
                    "imagePullPolicy": "IfNotPresent",
                    "kafkaConsumerAutoscaler": null,
                    "kafkaConsumerBatchSize": 10485760,
                    "kafkaConsumerQueueCapacity": 1000,
                    "kafkaConsumerReplicas": 3,
                    "logLevel": "info",
                    "logTypes": "FLOWS",
                    "metrics": {
                      "disableAlerts": [],
                      "ignoreTags": [
                        "egress",
                        "packets"
                      ],
                      "server": {
                        "port": 9102
                      }
                    },
                    "port": 2055,
                    "profilePort": 6060,
                    "resources": {
                      "limits": {
                        "memory": "800Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "memory": "100Mi"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          console.openshift.io/plugins: '["netobserv-plugin"]'
          containerImage: registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
          createdAt: 2023-06-22T09:31:50
          description: Network flows collector and monitoring solution
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-netobserv-operator
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.25.3
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/netobserv/network-observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FlowCollector is the Schema for the flowcollectors API, which
              pilots and configures netflow collection.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1alpha1
          - description: FlowCollector is the schema for the network flows collection
              API, which pilots and configures the underlying deployments.
            displayName: Flow Collector
            kind: FlowCollector
            name: flowcollectors.flows.netobserv.io
            version: v1beta1
        description: |
          Network Observability is an OpenShift operator that deploys a monitoring pipeline to collect and enrich network flows that are produced by the Network Observability eBPF agent.

          The operator provides dashboards, metrics, and keeps flows accessible in a queryable log store, Grafana Loki. When a `FlowCollector` instance is created, new views are available in the Console.

          ## Dependencies

          ### Loki

          [Loki](https://grafana.com/oss/loki/), from GrafanaLabs, is the backend that is used to store all collected flows. The NetObserv Operator does not install Loki directly, however we provide some guidance to help you there.

          For normal usage, we recommend two options:

          - Installing the [Loki Operator](https://docs.openshift.com/container-platform/4.11//logging/cluster-logging-loki.html). We have written [a guide](https://github.com/netobserv/documents/blob/main/loki_operator.md) to help you through those steps. Please note that it requires configuring an object storage. Note also that the Loki Operator can also be used for [OpenShift cluster logging](https://docs.openshift.com/container-platform/4.11/logging/cluster-logging.html). If you do so, you should not share the same `LokiStack` for Logging and NetObserv.

          - Installing using [Grafana's official documentation](https://grafana.com/docs/loki/latest/). Here also we wrote a ["distributed Loki" step by step guide](https://github.com/netobserv/documents/blob/main/loki_distributed.md).

          For a quick try that is not suitable for production and not scalable (it deploys a single pod, configures a 1GB storage PVC, with 24 hours of retention), you can simply run the following commands:

          ```
          oc create namespace netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/1-storage.yaml) -n netobserv
          oc apply -f <(curl -L https://raw.githubusercontent.com/netobserv/documents/252bb624cf0425a1b66f59ce68fb939f246ef77b/examples/zero-click-loki/2-loki.yaml) -n netobserv
          ```

          ### Kafka

          [Apache Kafka](https://kafka.apache.org/) can optionally be used for a more resilient and scalable architecture. You can use for instance [Strimzi](https://strimzi.io/), an operator-based distribution of Kafka for Kubernetes and OpenShift.

          ### Grafana

          [Grafana](https://grafana.com/oss/grafana/) can optionally be installed for custom dashboards and query capabilities.

          ## Configuration

          The `FlowCollector` resource is used to configure the operator and its managed components. A comprehensive documentation is [available here](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/FlowCollector.md), and a full sample file [there](https://github.com/netobserv/network-observability-operator/blob/1.0.2/config/samples/flows_v1beta1_flowcollector.yaml).

          To edit configuration in cluster, run:

          ```bash
          oc edit flowcollector cluster
          ```

          As it operates cluster-wide, only a single `FlowCollector` is allowed, and it has to be named `cluster`.

          A couple of settings deserve special attention:

          - Sampling (`spec.agent.ebpf.sampling`): a value of `100` means: one flow every 100 is sampled. `1` means all flows are sampled. The lower it is, the more flows you get, and the more accurate are derived metrics, but the higher amount of resources are consumed. By default, sampling is set to 50 (ie. 1:50). Note that more sampled flows also means more storage needed. We recommend to start with default values and refine empirically, to figure out which setting your cluster can manage.

          - Loki (`spec.loki`): configure here how to reach Loki. The default values match the Loki quick install paths mentioned above, but you may have to configure differently if you used another installation method.

          - Quick filters (`spec.consolePlugin.quickFilters`): configure preset filters to be displayed in the Console plugin. They offer a way to quickly switch from filters to others, such as showing / hiding pods network, or infrastructure network, or application network, etc. They can be tuned to reflect the different workloads running on your cluster. For a list of available filters, [check this page](https://github.com/netobserv/network-observability-operator/blob/1.0.2/docs/QuickFilters.md).

          - Kafka (`spec.deploymentModel: KAFKA` and `spec.kafka`): when enabled, integrates the flow collection pipeline with Kafka, by splitting ingestion from transformation (kube enrichment, derived metrics, ...). Kafka can provide better scalability, resiliency and high availability ([view more details](https://www.redhat.com/en/topics/integration/what-is-apache-kafka)). Assumes Kafka is already deployed and a topic is created.

          - Exporters (`spec.exporters`, _experimental_) an optional list of exporters to which to send enriched flows. Currently only KAFKA is supported. This allows you to define any custom storage or processing that can read from Kafka. This feature is flagged as _experimental_ as it has not been thoroughly or stress tested yet, so use at your own risk.

          ## Further reading

          Please refer to the documentation on GitHub for more information.

          This documentation includes:

          - An [overview](https://github.com/netobserv/network-observability-operator#openshift-console) of the features, with screenshots
          - A [performance](https://github.com/netobserv/network-observability-operator#performance-fine-tuning) section, for fine-tuning
          - A [security](https://github.com/netobserv/network-observability-operator#securing-data-and-communications) section
          - An [F.A.Q.](https://github.com/netobserv/network-observability-operator#faq--troubleshooting) section
        displayName: Network Observability
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network observability
        - ebpf
        - flow tracing
        - flows
        - topology
        - network
        - observability
        links:
        - name: Project page
          url: https://github.com/netobserv/network-observability-operator
        - name: Issue tracker
          url: https://github.com/netobserv/network-observability-operator/issues
        - name: Discussion board
          url: https://github.com/netobserv/network-observability-operator/discussions
        maintainers:
        - email: jpinsonn@redhat.com
          name: Julien Pinsonneau
        - email: jtakvori@redhat.com
          name: Joel Takvorian
        - email: kmeth@redhat.com
          name: Kalman Meth
        - email: mmahmoud@redhat.com
          name: Mohamed S. Mahmoud
        - email: ocazade@redhat.com
          name: Olivier Cazade
        - email: rschaffe@redhat.com
          name: Ronen Schaffer
        - email: stlee@redhat.com
          name: Steven Lee
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:37b3c9b2eed351d9d1a39c99045bbc7993ccb819f67f7b390deaa34930f45946
        - registry.redhat.io/network-observability/network-observability-console-plugin-rhel9@sha256:880c0812a2cd4b7194f417e31fbdcee951ce76776fa06e33c5c9cd479a854daa
        - registry.redhat.io/network-observability/network-observability-ebpf-agent-rhel9@sha256:e31acc12d8690c339e13b5dd22d33e707701362bdd31086c9b4c0f48cad62cc3
        - registry.redhat.io/network-observability/network-observability-flowlogs-pipeline-rhel9@sha256:7c7c42fff63e0599bb943a9e9617b43181360d5a5254b27f368010af275cf0c3
        - registry.redhat.io/network-observability/network-observability-operator-bundle@sha256:b03f4e88ab1fed4744356a584d673fb2fdb15de7e1de7786bdd222b853690670
        - registry.redhat.io/network-observability/network-observability-rhel9-operator@sha256:fd943746840cba26996dabb47b1beb3445571fa43cb1b859fb073e4aa114b376
        version: 1.3.0
      entries:
      - name: network-observability-operator.v1.3.0
        version: 1.3.0
      - name: network-observability-operator.v1.2.0
        version: 1.2.0
      - name: network-observability-operator.v1.1.0-202302110050
        version: 1.1.0
      - name: network-observability-operator.v1.0.0-202301070345
        version: 1.0.0
      name: v1.0.x
    defaultChannel: stable
    packageName: netobserv-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://skupper.io
    name: skupper-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: alpha
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1
    - currentCSV: skupper-operator.v1.4.8-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:c665677c639699d6a5ef24aed5a1a3d9c0a5ee25f077a6db42a9fc6247d93132
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:eedde9f2019d4e919612bcc18f427c0d88b4846e773751af231ba0e560d48869
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:be37ebbe5a0c07497ccd903a9c627dd3a3fec4dcedae1e8737dd1ffed367cf99
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5391e9ec3748fec2b820db5fb0b8bd167a1efd607b8a0c090cce5ad2e94f4fcc
        - registry.redhat.io/openshift4/ose-prometheus@sha256:65602221cdff7f718d4149f4dead167d05dbbe0cea1eaa8bf4b98be588135583
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:1fd908e80ef21fe7d7051c400cb099f2f3ca51c8cdd698705ef9784d34688911
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:4c9cc9df83baa1ef37cfb429babe5b86a962b8bb096209b3fb74291c67c94000
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:0f6c6faf7fdf7b6d69bc5cc4b0266064d0035e295d0562b957d95c8c81699f2b
        version: 1.4.8-rh-1
      entries:
      - name: skupper-operator.v1.4.8-rh-1
        version: 1.4.8-rh-1
      - name: skupper-operator.v1.4.7-rh-3
        version: 1.4.7-rh-3
      - name: skupper-operator.v1.4.7-rh-2
        version: 1.4.7-rh-2
      - name: skupper-operator.v1.4.7-rh-1
        version: 1.4.7-rh-1
      - name: skupper-operator.v1.4.5-rh-4
        version: 1.4.5-rh-4
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.4
    - currentCSV: skupper-operator.v1.5.5-rh-6
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus@sha256:53afaf53975807765fc3777314edd218c789c17680c1f57a8e33a8f9a76f38bf
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:5f8621d0013febcecbed4d7c82db11a4dded99ad8ce6b6d83858b9bdbe79ee5a
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:a5d7c1b2b9db672548126c0873d062a55667d93291fd3e32809d3c9c900c548e
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:aa261c79b421bd49f7b46974a3bb5a790d5a1a73dd8dea4e1d8a4634fb68a51f
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:be9dbcd2a3a1e1e536921390dcfd283fc45dc4423d2705376ef37a9fb53f7cf1
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:1de071dabf620265252f3922c632f17f676ba716ce4c3df931a0ccb3871dbb6c
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:9e9e1f2288f0218e3ba2abf4bebac9731d7cf9f9c567e0f6de6767b0b66cbee4
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:3e78ec8cdb729a704d5a26e54697f285ba7c0b9afee6160b7ecc703876120dea
        version: 1.5.5-rh-6
      entries:
      - name: skupper-operator.v1.5.5-rh-6
        version: 1.5.5-rh-6
      - name: skupper-operator.v1.5.5-rh-5
        version: 1.5.5-rh-5
      - name: skupper-operator.v1.5.5-rh-4
        version: 1.5.5-rh-4
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.5
    - currentCSV: skupper-operator.v1.8.7-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.8.7-rh-2'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:56b8fde75a6e7df235c560a0f4b9114398da5ba8b20f75f1ea7338f48ce19506
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:d7d252cff08ba1589afd8707ed349f87d3b6214906a4fb49932bda0a4d19fee4
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:123488f0877c054a348f5b4135e68d77c17523bc4e5ea14dd73168372592f4ce
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:f798745e799515065af099a816057b9c98a741e951d30a0f45ed47ca50e6724e
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:61542c5f0989f51ac5d6c96f5d0e29aca75614b79e7626299c77c5a74636b564
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:b0d43628b779a101a24e4b271318a34c21f3b71bcba2bf20804f941a93d27455
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:3adeef827eca0382cd410f7c3e99a0fee9b50791cb0e0c6fc130b72a2550684f
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:5ea36e96f878577104ab7601e57e7a2a233f11dbdc533454f95e2e14a066d21f
        version: 1.8.7-rh-2
      entries:
      - name: skupper-operator.v1.8.7-rh-2
        version: 1.8.7-rh-2
      - name: skupper-operator.v1.8.7-rh-1
        version: 1.8.7-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.8
    - currentCSV: skupper-operator.v1.9.4-rh-1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
          createdAt: "2021-04-14T08:00:00Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>1.4.0-rh-1 <1.9.4-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/skupperproject/skupper-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The Red Hat Service Interconnect operator allows you create and manage sites.
          The operator processes any ConfigMap named skupper-site and configures the site for the service network accordingly.

          To get started, create the following ConfigMap in the desired namespace, the
          Red Hat Service Interconnect operator will create and configure the site instance.

                apiVersion: v1
                kind: ConfigMap
                metadata:
                  name: skupper-site

          The operator can be installed to a single namespace or at cluster level, watching for the skupper-site ConfigMap in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper
          url: https://skupper.io
        maintainers:
        - email: ansmith@redhat.com
          name: Andy Smith
        - email: fgiorget@redhat.com
          name: Fernando Giorgetti
        - email: gsim@redhat.com
          name: Gordon Sim
        - email: nluaces@redhat.com
          name: Noe Luaces
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://redhat.com
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:97f1a693c36b909b31a8b3141a5ea07b9c369acb98f9f89771de8db09d209fad
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:9abea7559799ec9315c646ac00688582044a2783973efb0caf01f8c401ca4c82
        - registry.redhat.io/service-interconnect/skupper-service-controller-rhel9@sha256:b981b42e7ba47169636ab14f3ec092aadfdc374399b5856d77a0cefd5295fca1
        - registry.redhat.io/service-interconnect/skupper-site-controller-rhel9@sha256:a48771488b13663283783c8d2350cf67baf64b638e0108139ee37a702194593b
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:8010bd9fa41fcd706a84d20bf54c27f59ab1f6c6e1a10881e42e1f71a1c7e237
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:039fcc42f7935dc3a5fea4da1b9b858ae0dabf6855a101543edd5fcea2b57392
        - registry.redhat.io/service-interconnect/skupper-config-sync-rhel9@sha256:a9926486d25ca1d9b5886a08a660fb11a08a6100adfa410ae9c353009bfbf4c5
        - registry.redhat.io/service-interconnect/skupper-flow-collector-rhel9@sha256:0bc6bb33f6bce7a48847f35656f8a8ec2910247150f8db1f1515a3487f1f6bba
        version: 1.9.4-rh-1
      entries:
      - name: skupper-operator.v1.9.4-rh-1
        version: 1.9.4-rh-1
      - name: skupper-operator.v1.9.2-rh-1
        version: 1.9.2-rh-1
      - name: skupper-operator.v1.9.1-rh-1
        version: 1.9.1-rh-1
      - name: skupper-operator.v1.8.5-rh-2
        version: 1.8.5-rh-2
      - name: skupper-operator.v1.8.3-rh-1
        version: 1.8.3-rh-1
      - name: skupper-operator.v1.8.2-rh-2
        version: 1.8.2-rh-2
      - name: skupper-operator.v1.8.1-rh-3
        version: 1.8.1-rh-3
      - name: skupper-operator.v1.5.5-rh-1
        version: 1.5.5-rh-1
      - name: skupper-operator.v1.5.4-rh-2
        version: 1.5.4-rh-2
      - name: skupper-operator.v1.5.3-rh-3
        version: 1.5.3-rh-3
      - name: skupper-operator.v1.4.4-rh-1
        version: 1.4.4-rh-1
      - name: skupper-operator.v1.4.3-rh-6
        version: 1.4.3-rh-6
      - name: skupper-operator.v1.4.2-rh-2
        version: 1.4.2-rh-2
      - name: skupper-operator.v1.4.1-rh-1
        version: 1.4.1-rh-1
      - name: skupper-operator.v1.2.2
        version: 1.2.2
      - name: skupper-operator.v1.2.1
        version: 1.2.1
      - name: skupper-operator.v1.2.0
        version: 1.2.0
      - name: skupper-operator.v1.0.8
        version: 1.0.8
      - name: skupper-operator.v1.0.7
        version: 1.0.7
      - name: skupper-operator.v1.0.6
        version: 1.0.6
      - name: skupper-operator.v1.0.5
        version: 1.0.5
      - name: skupper-operator.v1.0.4
        version: 1.0.4
      name: stable-1.9
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2
    - currentCSV: skupper-operator.v2.0.1-rh-2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
          createdAt: "2025-05-29T13:35:40Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.0.1-rh-1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:bd3f72544841c91c2d573a65ac3c988756f3523e17e0988027793b0f5f3ec3d1
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:95062ff1256135cfa17907e82b09b2a211371ea0e76ccc711b7c737e040d87e9
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:b5c73dc59a9184cd120c2ee35256cc097c790b8ef488af0c732d5efe3192b768
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:f2a41674ed196c089748d6386c18c14580e502de33235ef23c37eddeb6350f41
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:498b5bb648e5bbaad07663941600173abac7d60a10744305f619509ddf7563e0
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:e66501276e3a10ae6d08dad3c4658fdb8b0b635e4b2fe72fa943b0e5300d7f2d
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:e239c859a1f49816ae3068455e703b8a12919f11158b0f6ee0e6bfb31e8a8082
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:92f44cff160874efe57a985609e599b13c49c0c4cef846dccc80833af2ae74f3
        version: 2.0.1-rh-2
      entries:
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.0
    - currentCSV: skupper-operator.v2.1.1-rh-3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessGrant",
                "metadata": {
                  "name": "my-grant"
                },
                "spec": {
                  "expirationWindow": "1h",
                  "redemptionsAllowed": 10
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AccessToken",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "ca": "-----BEGIN CERTIFICATE-----\nMIIDNzCCAh+ … Q=\n-----END CERTIFICATE-----\n",
                  "code": "L1qCl2mZ7D81gbXp3ax4YT0K",
                  "url": "https://10.107.241.191:9090/86b80f6b-2b02-47ac-bab4-d285bb4f7b9d"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnector",
                "metadata": {
                  "name": "backend",
                  "namespace": "other"
                },
                "spec": {
                  "port": 8080,
                  "selector": "app=backend",
                  "siteNamespace": "west"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "AttachedConnectorBinding",
                "metadata": {
                  "name": "backend",
                  "namespace": "west"
                },
                "spec": {
                  "connectorNamespace": "other",
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Certificate",
                "metadata": {
                  "name": "west-certificate"
                },
                "spec": {
                  "client": true
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Connector",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "port": 8080,
                  "routingKey": "backend",
                  "selector": "app=backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Link",
                "metadata": {
                  "name": "my-token"
                },
                "spec": {
                  "endpoints": [
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "inter-router",
                      "port": "55671"
                    },
                    {
                      "group": "skupper-router-1",
                      "host": "10.102.82.46",
                      "name": "edge",
                      "port": "45671"
                    }
                  ],
                  "tlsCredentials": "my-token"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "backend"
                },
                "spec": {
                  "host": "backend",
                  "port": 8080,
                  "routingKey": "backend"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "RouterAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "SecuredAccess",
                "metadata": {
                  "name": "west-access"
                },
                "spec": {
                  "issuer": "default"
                }
              },
              {
                "apiVersion": "skupper.io/v2alpha1",
                "kind": "Site",
                "metadata": {
                  "name": "west"
                },
                "spec": {
                  "ha": true,
                  "linkAccess": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery, Networking, Streaming & Messaging
          certified: "false"
          containerImage: registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
          createdAt: "2025-08-19T20:43:19Z"
          description: Red Hat Service Interconnect Operator provides the ability
            to create a service network
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>2.0.0-rh-1 <2.1.1-rh-3'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["Red Hat Service Interconnect"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/skupperproject/skupper
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AccessGrant
            name: accessgrants.skupper.io
            version: v2alpha1
          - kind: AccessToken
            name: accesstokens.skupper.io
            version: v2alpha1
          - kind: AttachedConnectorBinding
            name: attachedconnectorbindings.skupper.io
            version: v2alpha1
          - kind: AttachedConnector
            name: attachedconnectors.skupper.io
            version: v2alpha1
          - kind: Certificate
            name: certificates.skupper.io
            version: v2alpha1
          - kind: Connector
            name: connectors.skupper.io
            version: v2alpha1
          - kind: Link
            name: links.skupper.io
            version: v2alpha1
          - kind: Listener
            name: listeners.skupper.io
            version: v2alpha1
          - kind: RouterAccess
            name: routeraccesses.skupper.io
            version: v2alpha1
          - kind: SecuredAccess
            name: securedaccesses.skupper.io
            version: v2alpha1
          - kind: Site
            name: sites.skupper.io
            version: v2alpha1
        description: |
          Red Hat Service Interconnect enables communication between services running in different network locations.
          It allows geographically distributed services to connect as if they were all running in the same site.

          Each site is a namespace with a running Red Hat Service Interconnect instance.

          The operator can only be installed at cluster level, watching for the skupper resources in all namespaces.
        displayName: Red Hat Service Interconnect
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - skupper
        - rhsi
        - service
        - interconnect
        - mesh
        - van
        links:
        - name: Red Hat Service Interconnect
          url: https://www.redhat.com/en/technologies/cloud-computing/service-interconnect
        - name: Skupper Operator
          url: https://github.com/skupperproject/skupper-operator
        maintainers:
        - email: skupper@googlegroups.com
          name: Skupper Community
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
          url: https://skupper.io
        relatedImages:
        - registry.redhat.io/service-interconnect/skupper-operator-bundle@sha256:9df21dc1f414c784c13c91dcdd40c0f4a708ea719473e613c4a941f1c9c6a213
        - registry.redhat.io/service-interconnect/skupper-router-rhel9@sha256:1a290a5ac8b9d1b082e740054675fe432aa424c2f55dd662caf67c740044f229
        - registry.redhat.io/openshift4/ose-oauth-proxy-rhel9@sha256:e8df53454cf12a14338790af6ed209798010990ca3320096e8b44acf901c4123
        - registry.redhat.io/openshift4/ose-prometheus-rhel9@sha256:4e4842a8dc452bc6fe84e8d76a4cf2c8e32d5de8f7cbbdb88841768db7ec03bd
        - registry.redhat.io/service-interconnect/skupper-cli-rhel9@sha256:c01d0eab0841b1a5fddb58a0ae110beb464b91390f02662590bebc7b2a7213b6
        - registry.redhat.io/service-interconnect/skupper-controller-rhel9@sha256:d25839c53b3da2341980df300431a9ddfefe0a26c90e46f1bdc2b1bba755dda7
        - registry.redhat.io/service-interconnect/skupper-kube-adaptor-rhel9@sha256:36872037066fae806cbf88cab673229f25e592295972aa18227827233362eec5
        - registry.redhat.io/service-interconnect/skupper-network-observer-rhel9@sha256:4f66c398084041f0625f9075a38bcc760c7154d8c0ec5b5e6dada7de2a0b3ca2
        version: 2.1.1-rh-3
      entries:
      - name: skupper-operator.v2.1.1-rh-3
        version: 2.1.1-rh-3
      - name: skupper-operator.v2.0.1-rh-2
        version: 2.0.1-rh-2
      - name: skupper-operator.v2.0.0-rh-11
        version: 2.0.0-rh-11
      name: stable-2.1
    defaultChannel: stable-2
    packageName: skupper-operator
    provider:
      name: Red Hat
      url: https://skupper.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle@sha256:47ad4efd37ad85bccf915a6d9448d404b7ce3164d7afb4c944648eeba9f0b04b
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Sonatype
      provider-url: ""
    name: nexus-repository-ha-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nexus-repository-ha-operator-certified.v3.84.0-1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonatype.com/v1alpha1",
                "kind": "NexusRepo",
                "metadata": {
                  "name": "nexusrepo-sample"
                },
                "spec": {
                  "ingress": {
                    "additionalRules": null,
                    "annotations": null,
                    "defaultRule": false,
                    "dockerIngress": {
                      "annotations": null,
                      "enabled": false,
                      "host": "example.com",
                      "name": "nexus-docker-ingress"
                    },
                    "dockerSubdomain": false,
                    "enabled": false,
                    "host": "example.com",
                    "name": "nexus-ingress",
                    "tls": {
                      "enabled": false,
                      "secretName": "tlsSecretName"
                    }
                  },
                  "license": {
                    "fileContentsBase64": "your_license_file_contents_in_base_64",
                    "secretName": "nexus-repo-license"
                  },
                  "secret": {
                    "nexusSecret": {
                      "enabled" : false,
                      "name": "nexus-secret.json",
                      "nexusSecretsKeyId": "super_secret_key_id",
                      "secretKeyfileContentsBase64": "secretKeyfileContentsBase64",
                      "mountPath": "/var/nexus-repo-secrets"
                    }
                  },
                  "nexusData": {
                    "pvc": {
                      "accessMode": "ReadWriteOnce",
                      "size": "2Gi"
                    },
                    "storageClass": {
                      "enabled": false,
                      "name": "nexusrepo-storage",
                      "parameters": {},
                      "provisioner": "provisioner",
                      "reclaimPolicy": "Retain",
                      "volumeBindingMode": "WaitForFirstConsumer"
                    },
                    "volumeClaimTemplate": {
                      "enabled": false
                    }
                  },
                  "service": {
                    "docker": {
                      "enabled": false,
                      "name": "nexus-repo-docker-service",
                      "port": 9090,
                      "protocol": "TCP",
                      "targetPort": 9090,
                      "type": "NodePort"
                    },
                    "nexus": {
                      "enabled": false,
                      "name": "nexus-repo-service",
                      "port": 80,
                      "protocol": "TCP",
                      "targetPort": 8081,
                      "type": "NodePort"
                    }
                  },
                  "statefulset": {
                    "container": {
                      "containerPort": 8081,
                      "env": {
                        "clustered": false,
                        "install4jAddVmParams": "-Xms2703m -Xmx2703m",
                        "jdbcUrl": null,
                        "nexusInitialPassword": "nexusabc",
                        "password": "nexus",
                        "user": "nexus",
                        "zeroDowntimeEnabled": false
                      },
                      "imageName": "registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f",
                      "pullPolicy": "IfNotPresent",
                      "resources": {
                        "limits": {
                          "cpu": 4,
                          "memory": "8Gi"
                        },
                        "requests": {
                          "cpu": 4,
                          "memory": "8Gi"
                        }
                      },
                      "terminationGracePeriod": 120
                    },
                    "imagePullSecrets": {},
                    "livenessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "name": "nexusrepo-statefulset",
                    "readinessProbe": {
                      "failureThreshold": 6,
                      "initialDelaySeconds": 240,
                      "path": "/",
                      "periodSeconds": 60
                    },
                    "replicaCount": 1
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
          createdAt: "2023-08-14"
          description: |-
            Nexus Repository is the central source of control to efficiently manage all binaries
            and build artifacts across your DevOps pipeline.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.replaces: 3.80.0-2
          olm.skipRange: <3.84.0-1
          repository: https://github.com/sonatype/operator-nexus-repository
          support: Sonatype
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Nexus Repository
            displayName: NexusRepo
            kind: NexusRepo
            name: nexusrepos.sonatype.com
            version: v1alpha1
        description: |-
          Nexus Repository is the central source of control to efficiently manage all binaries
          and build artifacts across your DevOps pipeline.
          The flow of open source and third-party components into and through an organization
          creates a complex software supply chain.
          Nexus Repository delivers speed, efficiency, and quality to the governance
          and management of all dependencies, libraries, and applications for your DevOps teams.

          ## Core Capabilities

          * **Dependency Management**:
            Improves reliability with repeatable, fast access to secure dependencies
          * **Developer Productivity**:
            Streamline developer workflows by enabling the sharing of components and applications across teams
          * **Supply Chain Performance**:
            Improve speed-to-market and reduced build times with release advanced staging and component tagging
          * **CI/CD Integrations**:
            Increase DevOps scalability with integrations to the most popular build and deployment tools

          Version control systems and package registries do not scale when managing proprietary,
          open source, and third-party components.
          Organizations need a central binary and build artifact repository to manage dependencies
          across the entire software supply chain.

          ## Limitations

          ## Controlling Automatic vs Manual Update

          If you use the default configuration for the Nexus Repository Operator installation,
          please notice that on any new operator release, the corresponding deployments are
          also updated without user intervention, resulting in unscheduled downtime.

          If you want to avoid this unscheduled downtime, we recommend installing the operator
          into **its own namespace** with **manual approval** for updates.

          ## Usage
          By default, the example Custom Resource starts up a single instance of Nexus Repository in non-Pro mode using an H2 database.

          ### Connect to a PostgreSQL database
          We strongly recommend using Nexus Repository with a PostgreSQL database. To connect a Nexus Repository instance to a PostgreSQL database, do as follows:
          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.

          ### High Availability (HA) mode
          To start Nexus Repository in HA mode, you must edit the Custom Resource YAML as follows:

          * Update the `license.fileContentsBase64` parameter with a base 64 representation of your Nexus Repository Pro license.
          * Set the `statefulset.container.env.jdbcUrl` parameter to your database url.
          * Set the `statefulset.container.env.user` parameter to your database user name.
          * Set the `statefulset.container.env.password` parameter to your database password.
          * Set the `statefulset.container.env.clustered` parameter to true.
          * Set the `statefulset.replicaCount` parameter to the number of Nexus Repository pods you require in your HA cluster
          * You must use a persistent volume as described in the 'Storage' section below.

          ### Storage
          The default configuration uses an emptyDir volume for storing Nexus Repository logs. However, we strongly recommend that
          you configure dynamic provisioning of persistent storage based on your deployment environment as explained below.

          #### Cloud deployments (AWS/Azure)
          * Ensure the appropriate Container Storage Interface (CSI) driver(s) are installed for your chosen OpenShift cloud deployment.
            Please, refer to OpenShift documentation for details on configuring CSI drivers.
          Note: if you're using Red Hat OpenShift on AWS (ROSA), CSI drivers for dynamically provisioning EBS volumes are installed
          by default; you should see associated storage classes for them in your cluster web console.

          #### On premise deployments
          1. Attach separate disks (i.e. separate from the root disk) to your worker nodes.
          2. Install the Local Storage Operator. Please refer to Open Shift persistent storage documentation:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          and the Local Storage Operator documentation: https://github.com/openshift/local-storage-operator.
          3. Use the Local Storage Operator to automatically create persistent volumes for your chosen storage class name as documented here:
             https://docs.openshift.com/container-platform/4.13/storage/persistent_storage/persistent_storage_local/persistent-storage-local.html
          4. Specify the storage class name from step 3 as the value of the `nexusData.storageClass.name` property and also set `nexusData.volumeClaimTemplate.enabled` to true.

          #### Configuring the operator for dynamic persistent volume provisioning
          * Set the `nexusData.storageClass.name` parameter to a storage class name. This could be one of the default storage classes
            automatically created in your OpenShift cluster (e.g., if you're using ROSA) or one that you would like the operator to create.
          * If you would like to create a dedicated storage class (i.e., you don't want to use the default), then in addition to
            specifying a value for `nexusData.storageClass.name`, you must also set `nexusData.storageClass.enabled` parameter to `true`.
          * Set the `nexusData.volumeClaimTemplate.enabled` parameter to `true`.

          ### Additional configuration
          * Once the operator has created the server instance and it is running, you can further configure Nexus Repository via the NexusRepo custom resource:

          | Parameter                                                   | Description                         | Default                                 |
          | ----------------------------------------------------------- | ----------------------------------  | ----------------------------------------|
          | `statefulset.name`                                          |  The suffix for the statefulset's service name    | nexusrepo-statefulset     |
          | `statefulset.replicaCount`                                  |  The desired number of Nexus Repository pods    | 3                           |
          | `statefulset.container.imageName`                           |  The Nexus repository image registry url    |
          | | | registry.connect.redhat.com/sonatype/nexus-repository-manager:<app version>-ubi-1|
          | `statefulset.container.resources.requests.cpu`              |  The minimum cpu the Nexus repository pod can request    | 4 |
          | `statefulset.container.resources.requests.memory`           |  The minimum memory the Nexus repository pod can request    | 8Gi|
          | `statefulset.container.resources.limits.cpu`                |  The maximum cpu the Nexus repository pod may get.    | 4|
          | `statefulset.container.resources.limits.memory`             |  The maximum memory the Nexus repository pod may get.  | 8Gi|
          | `statefulset.container.containerPort`                       |  The Nexus Repository container's HTTP port  | 8081|
          | `statefulset.container.pullPolicy`                          |  The Nexus Repository docker image pull policy  | IfNotPresent|
          | `statefulset.container.terminationGracePeriod`              |  The time given for the pod to gracefully shut down  | 120 seconds|
          | `statefulset.container.env.clustered`                       |  Controls whether Nexus Repository should start in clustered mode  | False|
          | `statefulset.container.env.install4jAddVmParams`            |  Xmx and Xms settings for JVM  | -Xms2703m -Xmx2703m|
          | `statefulset.container.env.user`                            |  The database user name | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.password`                        |  The database password  | nexus (can be overridden via Kubernetes secret. See above)|
          | `statefulset.container.env.jdbcUrl`                         |  The database url |
          | | | jdbc:postgresql://your_database_server:5432/nexus (must be overridden via Kubernetes |
          | | | secret object which feeds into env variable variables defined in the operator.yaml . See above)|
          | `statefulset.container.env.nexusInitialPassword`            |  The initial admin password for Nexus Repository UI. This is for first time |
          | | installs only. Subsequent installs/restarts on the same node ignores this setting | nexusabc|
          | `statefulset.livenessProbe.initialDelaySeconds`             |  The initial delay before the Nexus Repository application is pinged for a |
          | | response i.e. to determine whether the application is still alive | 240|
          | `statefulset.livenessProbe.path`                            |  The path to ping the application | |
          | `statefulset.livenessProbe.periodSeconds`                   |  The interval for checking that the application is still alive i.e. responsive, not deadlocked etc | 60|
          | `statefulset.livenessProbe.failureThreshold`                |  The maximum number of times the application ping can fail before the |
          | | application is considered unresponsive. Kubernetes will attempt to restart the application's pod in this case | 6|
          | `statefulset.readinessProbe.initialDelaySeconds`            |  The initial delay before the Nexus Repository application is checked |
          | | whether it is ready to receive traffic | 240|
          | `statefulset.readinessProbe.path`                           |  The path where the Nexus Repository application is checked for traffic readiness | |
          | `statefulset.readinessProbe.periodSeconds`                  |  The interval for checking the Nexus Repository application for traffic readiness | 60|
          | `statefulset.readinessProbe.failureThreshold`               |  The maximum number of times to check the Nexus Repository application for traffic readiness | 6|
          | `statefulset.imagePullSecrets`                              |  The pull secret for private image registries | {}|
          | `ingress.name`                                              |  The name of an Ingress object which listens for requests to Nexus Repository on port 80 | nexus-ingress|
          | `ingress.enabled`                                           |  Whether or not to create the Ingress | false|
          | `ingress.host`                                              |  Ingress host | "example.com"|
          | `ingress.defaultRule`                                       |  Whether or not to add a default rule for the Nexus Repository |
          | | Ingress which forwards traffic to a Service object  | false|
          | `ingress.additionalRules`                                   |  Additional rules to add to the ingress | null|
          | `ingress.tls.enabled`                                       |  Whether or not TLS should be enabled for the Ingress | false|
          | `ingress.tls.secretName`                                    | The name of a Secret object in which to store the TLS secret | tlsSecretName|
          | `ingress.dockerSubdomain`                                   |  Whether or not to add rules for docker subdomains | false|
          | `ingress.annotations`                                       |  Annotations for the Ingress object | null |
          | `license.fileContentsBase64`                                |  Your Base64 encoded Nexus Repository license | your_license_file_contents_in_base_64 |
          | `license.secretName`                                        |  The name of a secret object in which to store your |
          | | Base64-encoded Nexus Repository license | nexus-repo-license|
          | `ingress.dockerIngress.name`                                |  The name of an Ingress object which listens for requests to an |
          | | existing docker repository in Nexus Repository | nexus-docker-ingress|
          | `ingress.dockerIngress.enabled`                             |  Whether or not to create an Ingress for a docker repository | false|
          | `ingress.dockerIngress.host`                                |  The docker Ingress's host | "example.com"|
          | `ingress.dockerIngress.annotations`                         |  Annotations for the docker Ingress object | null |
          | `nexusData.pvc.size`                                        |  The volume size to request for storing Nexus logs | 2Gi |
          | `nexusData.pvc.accessMode`                                  |  The persistent volume claim access mode | ReadWriteOnce |
          | `nexusData.storageClass.name`                               |  The name of the storage class to use for dynamically requesting that |
          | | storage be provisioned when using dynamic provisioning or to randomly select a pre-created persisted volume when using the Local Storage Operator. | nexusrepo-storage |
          | `nexusData.storageClass.reclaimPolicy`                      |  The reclaim policy for persistent volume's created by the storage class | Retain |
          | `nexusData.storageClass.enabled`                            |  Whether or not a storage class object should be created. | false |
          | `nexusData.storageClass.volumeBindingMode`                  |  It controls when volume binding and dynamic provisioning should occur | WaitForFirstConsumer |
          | `nexusData.storageClass.provisioner`                        |  The name of the volume provisioner | provisioner |
          | `nexusData.storageClass.parameters`                         |  The additional parameters for the storage class | {} |
          | `nexusData.volumeClaimTemplate.enabled`                           |  Whether or not volume should be requested using a claim template. |
          | | You should set this property to true when using dynamic volume provisioning (cloud deployments) |
          | | or if you're using Local Storage Operator to automatically create persistent volumes for locally attached disks (i.e. not root disk) | false |
          | `license.secretName | A name for the Kubernetes secret object which stores the base64 representation of your |
          | | Nexus Repository license specified via `license.fileContentsBase64` is stored | nexus-repo-license |
          | `license.fileContentsBase64 | A base64 representation of your Nexus Repository license | your_license_file_contents_in_base_64 |
          | `service.nexus.name`                                              |  The name of a kubernetes Service Object which allows access to |
          | | Nexus Repository's main HTTP port | nexus-repo-service |
          | `service.nexus.enabled`                                           |  Whether or not to create the Service object | false|
          | `service.nexus.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.nexus.protocol`                                       |  The protocol  | TCP |
          | `service.nexus.port`                                   |  The port to listen for incoming requests | 80|
          | `service.nexus.targetPort`                                   |  The port to forward requests to | 8081 |
          | `service.docker.name`                                              |  The name of a kubernetes Service Object which allows access to a |
          | | docker repository within Nexus Repository | nexus-repo-docker-service |
          | `service.docker.enabled`                                           |  Whether or not to create a Kubernetes Service object for a given |
          | | docker repository within Nexus Repository | false |
          | `service.docker.type`                                              |  The type of the Kubernetes Service | "NodePort"|
          | `service.docker.protocol`                                       |  The protocol  | TCP |
          | `service.docker.port`                                   |  The port to listen for incoming requests | 9090 |
          | `service.docker.targetPort`                                   |  The port to forward requests to | 9090 |
        displayName: Nexus Repository HA Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - repository
        - sonatype
        links:
        - name: Nexus Repository Manager
          url: https://www.sonatype.com/product-nexus-repository
        maintainers:
        - email: support@sonatype.com
          name: Sonatype
        maturity: stable
        provider:
          name: Sonatype
        relatedImages:
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-certified@sha256:99f30aa5ba8469d96de3efe69d1b21e73ff022172fc79144bba496d1119ec8a2
        - registry.connect.redhat.com/sonatype/nexus-repository-manager@sha256:fbad488ff2e368845473a11b4955992fe826a94811da3abf5cc3c6a33ff3974f
        - registry.connect.redhat.com/sonatype/nexus-repository-ha-operator-bundle@sha256:3e7ab287f64398b2aa434d9cfd5d99e2e9cf2d40c2e9fa1b1418537807b7aeea
        version: 3.84.0-1
      entries:
      - name: nexus-repository-ha-operator-certified.v3.84.0-1
        version: 3.84.0-1
      name: stable
    defaultChannel: stable
    packageName: nexus-repository-ha-operator-certified
    provider:
      name: Sonatype
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: AppDynamics LLC
      provider-url: https://www.appdynamics.com/
    name: appdynamics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: appdynamics-operator.v24.7.0-861
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "Clusteragent",
                "metadata": {
                  "name": "k8s-cluster-agent",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "appName": "<app-name>",
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/cluster-agent:tag"
                }
              },
              {
                "apiVersion": "cluster.appdynamics.com/v1alpha1",
                "kind": "InfraViz",
                "metadata": {
                  "name": "appdynamics-infraviz",
                  "namespace": "appdynamics"
                },
                "spec": {
                  "controllerUrl": "<appdynamics-controller-url",
                  "account": "<appdynamics-account-name>",
                  "image": "<your-docker-registry>/appdynamics/machine-agent:latest"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring,Logging & Tracing
          certified: "false"
          containerImage: registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
          createdAt: "2023-02-06T14:26:07Z"
          description: AppDynamics Operator to install AppDynamics Cluster Agent &
            AppDynamics Machine Agent for K8s Monitoring for various Cloud Providers
            like EKS, AKS etc.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: appdynamics
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: support@appdynamics.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Clusteragent is the Schema for the clusteragents API
            displayName: Clusteragent
            kind: Clusteragent
            name: clusteragents.cluster.appdynamics.com
            version: v1alpha1
          - description: InfraViz is the Schema for the infravizs API
            displayName: Infra Viz
            kind: InfraViz
            name: infravizs.cluster.appdynamics.com
            version: v1alpha1
        description: |-
          AppDynamics Operator to install AppDynamics Cluster Agent & Machine Agent for K8s Monitoring on various Cloud Providers like EKS, AKS, GCP, Openshift etc.
          **About AppDynamics Operator**
          AppDynamics Operator simplifies the installation and lifecycle management of the AppDynamics ClusterAgent and the AppDynamics Machine Agent for K8s Monitoring on different Cloud Providers like EKS, AKS, GCP, OpenShift etc.
          **Installation Steps**
          Please refer below links for installation guide:
          [Install AppDynamics Cluster Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli)
          [Install AppDynamics Machine Agent](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli)
          [Install AppDynamics Cluster Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts)
          [Install Appdynamics Machine Agent Using Helm Charts](https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts)
        displayName: AppDynamics Operator for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - AppDynamics
        - Operator
        - AppDynamics Cluster Agent
        - AppDynamics Machine Agent
        - AppDynamics K8s Monitoring
        links:
        - name: AppDynamics Kubernetes Monitoring Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent
        - name: AppDynamics Cluster Agent Overview
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/overview-of-cluster-monitoring
        - name: AppDynamics Cluster Agent Requirements
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/cluster-agent-requirements-and-supported-environments
        - name: Install AppDynamics Cluster Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-the-kubernetes-cli
        - name: Install AppDynamics Cluster Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-the-cluster-agent-with-helm-charts
        - name: Install AppDynamics Machine Agent & NetViz Agent
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-the-kubernetes-cli
        - name: Install AppDynamics Machine Agent & NetViz Agent Using Helm Charts
          url: https://docs.appdynamics.com/appd/23.x/23.1/en/infrastructure-visibility/monitor-kubernetes-with-the-cluster-agent/install-the-cluster-agent/install-infrastructure-visibility-with-helm-charts
        maintainers:
        - email: support@appdynamics.com
          name: AppDynamics
        maturity: alpha
        provider:
          name: AppDynamics LLC
          url: https://www.appdynamics.com/
        relatedImages:
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator@sha256:3c420cce8b626d8f5cf424410527c6a7a17ca3348e903dd8cdfcd74819b8de41
        - registry.connect.redhat.com/appdynamics/cluster-agent-operator-bundle@sha256:b2bc3e161311f8f5d2f5fc967815259697ce30bf94e980214bc05aa0e1529d55
        version: 24.7.0-861
      entries:
      - name: appdynamics-operator.v24.7.0-861
        version: 24.7.0-861
      name: alpha
    defaultChannel: alpha
    packageName: appdynamics-operator
    provider:
      name: AppDynamics LLC
      url: https://www.appdynamics.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: www.redhat.com
    name: odf-node-recovery-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-node-recovery-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "NodeRecovery",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "odf-node-recovery-operator",
                    "app.kubernetes.io/instance": "noderecovery-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "noderecovery",
                    "app.kubernetes.io/part-of": "odf-node-recovery-operator"
                  },
                  "name": "noderecovery-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Storage,OpenShift Optional
          description: ODF Node Recovery is an operator that assist in the recovery
            of and ODF cluster that has a device or a node replaced
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/jordigilh/odf-node-recovery-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NodeRecovery is the Schema for the noderecoveries API
            displayName: Node Recovery
            kind: NodeRecovery
            name: noderecoveries.odf.openshift.io
            version: v1alpha1
        description: |
          ODF Node Recovery is an operator that assist in the recovery of an
          ODF cluster in these 2 scenarios:
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_devices/index"> Device replacement<a/>
          * <a href="https://docs.redhat.com/en/documentation/red_hat_openshift_data_foundation/4.14/html-single/replacing_nodes/index"> Node replacement</a>
        displayName: ODF Node Recovery Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - odf
        - node recovery
        links:
        - name: GitHub Repository
          url: https://github.com/jordigilh/odf-node-recovery-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat, Inc.
          url: www.redhat.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/odf-node-recovery-operator@sha256:77b0ef6110a71da56216719ceed0f448257dec9633c98d7dfabc1038b23b6e29
        - quay.io/jordigilh/odf-node-recovery-controller@sha256:4164fafa9bb1c4d6121f4c5fc700fa2827c3994ca98a65d641471a2a46d4f59c
        version: 1.1.0
      entries:
      - name: odf-node-recovery-operator.v1.1.0
        version: 1.1.0
      - name: odf-node-recovery-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: odf-node-recovery-operator
    provider:
      name: Red Hat, Inc.
      url: www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: crunchy-postgres-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:0e283616c2d32e53258d62d02702bc4e179ef8c03933592154babed1a70bf666
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-exporter@sha256:3b231902b2116cd38de05c7378202716c95813b9961fb8cee811b61d36cddf86
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:4bdd0588ffbb3f628e7df1c72670e7a5087269438343b68779012bc3a13dea31
        - registry.connect.redhat.com/crunchydata/crunchy-upgrade@sha256:68bc93cce61d9b1e44822ae3e3ac8f7a9c2b712eddfaea090eabefacd60c67f6
        - registry.connect.redhat.com/crunchydata/crunchy-pgbouncer@sha256:17cb58dec78447c2f70a414a2aeff6facbf921b42a0f366d0dbf7be95aa8987a
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:a7f62f367817fc08294a0b8b82ac9826a42c0ae668efdfacaa158a4287b8a688
        - registry.connect.redhat.com/crunchydata/postgres-operator-bundle@sha256:6c09391edf49d71b558f8ddbaf4e947dee17490202521539910d78b5ddcb3531
        - registry.connect.redhat.com/crunchydata/crunchy-pgadmin4@sha256:0d84acb1ce6a9b046e3f8b222a583ddcc91dda0a117c9f637ef173d0ac992f57
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:56a5e3a5758982e086c9f78201a7137e0d934316f058a31cf30df44639a569a2
        - registry.connect.redhat.com/crunchydata/crunchy-postgres-gis@sha256:b2c3c74fdcc0e08f555e50c527f0d322292333c5d3dce061a0dee3680a1fbc56
        - registry.connect.redhat.com/crunchydata/crunchy-postgres@sha256:c559c896e174674d151f93543ac92a1b5647eb4676313f7dc6cbb8faee786f18
        - registry.connect.redhat.com/crunchydata/postgres-operator@sha256:682ff2e1229c5afb1bf4bee7019353bdaad55a723218361a0e989443d3e7fbc3
        - registry.connect.redhat.com/crunchydata/crunchy-pgbackrest@sha256:b70da0035b74379f2507cfabdd9d1cc1e3c3eecb9066c22f292981c9a27916e1
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.4
        version: 5.4.4
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.4.0
        version: 5.4.0
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: crunchy-postgres-operator
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-prometheusservice-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-prometheusservice-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "RuleGroupsNamespace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "prometheusservice.services.k8s.aws/v1alpha1",
                "kind": "Workspace",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
          createdAt: "2025-09-19T21:11:38Z"
          description: AWS Prometheus controller is a service controller for managing
            Prometheus resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertManagerDefinition represents the state of an AWS prometheusservice
              AlertManagerDefinition resource.
            displayName: AlertManagerDefinition
            kind: AlertManagerDefinition
            name: alertmanagerdefinitions.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: LoggingConfiguration represents the state of an AWS prometheusservice
              LoggingConfiguration resource.
            displayName: LoggingConfiguration
            kind: LoggingConfiguration
            name: loggingconfigurations.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: RuleGroupsNamespace represents the state of an AWS prometheusservice
              RuleGroupsNamespace resource.
            displayName: RuleGroupsNamespace
            kind: RuleGroupsNamespace
            name: rulegroupsnamespaces.prometheusservice.services.k8s.aws
            version: v1alpha1
          - description: Workspace represents the state of an AWS prometheusservice
              Workspace resource.
            displayName: Workspace
            kind: Workspace
            name: workspaces.prometheusservice.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Prometheus resources in AWS from within your Kubernetes cluster.

          **About Amazon Managed Service for Prometheus**

          Amazon Managed Service for Prometheus is a serverless, Prometheus-compatible monitoring service for container metrics that makes it easier to securely monitor container environments at scale. With Amazon Managed Service for Prometheus, you can use the same open-source Prometheus data model and query language that you use today to monitor the performance of your containerized workloads, and also enjoy improved scalability, availability, and security without having to manage the underlying infrastructure.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Prometheus
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - prometheusservice
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Prometheus Developer Resources
          url: https://aws.amazon.com/prometheus/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: prometheus maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/prometheusservice-controller:1.3.1
        - quay.io/community-operator-pipeline-prod/ack-prometheusservice-controller@sha256:dd9c62f7493d43c6f28055e6f1aa7d586e25b2d6d359418782cf18de45da32e8
        version: 1.3.1
      entries:
      - name: ack-prometheusservice-controller.v1.3.1
        version: 1.3.1
      - name: ack-prometheusservice-controller.v1.3.0
        version: 1.3.0
      - name: ack-prometheusservice-controller.v1.2.27
        version: 1.2.27
      - name: ack-prometheusservice-controller.v1.2.26
        version: 1.2.26
      - name: ack-prometheusservice-controller.v1.2.24
        version: 1.2.24
      - name: ack-prometheusservice-controller.v1.2.23
        version: 1.2.23
      - name: ack-prometheusservice-controller.v1.2.22
        version: 1.2.22
      - name: ack-prometheusservice-controller.v1.2.20
        version: 1.2.20
      - name: ack-prometheusservice-controller.v1.2.19
        version: 1.2.19
      - name: ack-prometheusservice-controller.v1.2.17
        version: 1.2.17
      - name: ack-prometheusservice-controller.v1.2.16
        version: 1.2.16
      - name: ack-prometheusservice-controller.v1.2.15
        version: 1.2.15
      - name: ack-prometheusservice-controller.v1.2.14
        version: 1.2.14
      - name: ack-prometheusservice-controller.v1.2.13
        version: 1.2.13
      - name: ack-prometheusservice-controller.v1.2.12
        version: 1.2.12
      - name: ack-prometheusservice-controller.v1.2.11
        version: 1.2.11
      - name: ack-prometheusservice-controller.v1.2.10
        version: 1.2.10
      - name: ack-prometheusservice-controller.v1.2.9
        version: 1.2.9
      - name: ack-prometheusservice-controller.v1.2.6
        version: 1.2.6
      - name: ack-prometheusservice-controller.v1.2.5
        version: 1.2.5
      - name: ack-prometheusservice-controller.v1.2.4
        version: 1.2.4
      - name: ack-prometheusservice-controller.v1.2.3
        version: 1.2.3
      - name: ack-prometheusservice-controller.v1.2.2
        version: 1.2.2
      - name: ack-prometheusservice-controller.v1.2.1
        version: 1.2.1
      - name: ack-prometheusservice-controller.v1.2.0
        version: 1.2.0
      - name: ack-prometheusservice-controller.v1.1.0
        version: 1.1.0
      - name: ack-prometheusservice-controller.v1.0.0
        version: 1.0.0
      - name: ack-prometheusservice-controller.v0.1.3
        version: 0.1.3
      - name: ack-prometheusservice-controller.v0.1.2
        version: 0.1.2
      - name: ack-prometheusservice-controller.v0.1.1
        version: 0.1.1
      - name: ack-prometheusservice-controller.v0.1.0
        version: 0.1.0
      - name: ack-prometheusservice-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-prometheusservice-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: DataTrucker.IO
      provider-url: https://www.datatrucker.io/
    name: datatrucker-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: datatrucker-operator.v2.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerConfig",
                "metadata": {
                  "name": "datatruckerconfig-sample"
                },
                "spec": {
                  "API": {
                    "crypto": {
                      "CredentailEncryption": {
                        "Algorithm": "aes-256-cbc",
                        "IV": "eThWmZq4t7w!z$C\u0026",
                        "Key": "JaNdRgUkXp2s5u8x/A?D(G+KbPeShVmY"
                      },
                      "JWT": {
                        "authorization_keys": {
                          "private_key": "privateJWT.key",
                          "public_key": "publicJWT.key"
                        },
                        "signOptions": {
                          "algorithm": "RS256",
                          "expiresIn": "60m",
                          "issuer": "datatrucker"
                        },
                        "verify": {
                          "issuer": "datatrucker"
                        }
                      },
                      "keycloak": {
                        "client_id": "nodejs",
                        "client_secret": "7c5542ae-7264-420d-bf5f-766f4ab39ad3",
                        "jwt_publickey_verifier": "keycloak.key",
                        "realm": "TestRealm",
                        "url": "http://localhost:30015"
                      },
                      "passwordHash": {
                        "Algorithm": "sha256",
                        "SaltLength": 16
                      }
                    },
                    "cryptokeys": {
                      "keycloakkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0KTUlJQklqQU5CZ2txaGtpRzl3MEJBUUVGQUFPQ0FROEFNSUlCQ2dLQ0FRRUF4ZnhGY1gxb3RmTUN6SWlkZ09tMndZakZWZXZHN1BQT1h5dTNrTEFuTWtTVkUxUXl4TzJ4SFFvSGcydDFVcUU4SVNXeTN3ZkZTbERzdU1VK0J1VVkvaWI5L04vSlpBV3k5MGRET0YxREdIeGVwM21HMVBTUllMQWFGQ0I3NTVsbWZjL1dna3VzZFdLdmxOTmZuNkd6S1J6OXBmV3RNUnhWWkFrejhmNlVUeG0zWE02Q0tOMUludThubEcxU3ZzaFc5cmt3Lyt6UmVvOFVCV2pPcTh5RjdPV0tnS0ZZZlhENFREa090OWFxN3RaaSsxaExYOE5iV2xxRnJZYmhJcFZINXlLSko3ZTA5cUVIRkVEelg5QXJzVVNHamNjbFFFYTl1dmJxU2d4Rzc3RUZVbk82aDZxUkNnU3lNNnVYdk9qcWlIMHp4RTNmRHB4R2NhY2pvTW1EMlFJREFRQUIKLS0tLS1FTkQgUFVCTElDIEtFWS0tLS0t",
                      "privateJWTkey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQ0KTUlJRW9RSUJBQUtDQVFCVWd1ZDN1OFlXcUJCWWtYZ3M2K25pMS84bnM5YjRGNkZudzA4RDRKdGRlalg5L1ZaMg0KWFdOamJ1ekMxUkRHMnVseFM4QjZIdTZYeEhYWU9XU2ZzOEY0UHZ2eC9TUWdmRWtWUlNDL1pQN3pzM1ZETllaMA0Kc1JScDNSZzUza29qUG80T3o2azFaeGprbFRFZEpMWUtqYlZud3U1M3A3MTFsVTJkV3B4R3VEeGYxVTJuUWgrWA0KaUhKZkhyUmxTZVA1aVhYQkpxd21VTDFkK25lTU5BOXlGakR0MnNZMG5iSDRySDVxbk1VRk4xMEpsZ3V4NFdiUw0KaGFUNklNN3h1dUVHbDRDckVENmdTejBNNGF1UVk4ekZxQWxJNlFWUTlML28zQnBCRmxYUUMwWmowb3lwQVFKLw0KQ1VLaU9FcEhLRHgram02RU9KOE1aejlMQ01yTmdLME5KSWlaQWdNQkFBRUNnZ0VBRTBMMnNLSjZob1Z0bkZJag0KZXRHV01CNmN5QmJFSXkxM215ZTNwYUdXcDVxSUhjaVU0VXR0c0tjQ2pYRjBibTVNa0liOGFwa01Oa2tQdHVWMQ0KMnJSa254eHozZmFkT2s0NWRnQ1ljeWI0dU9DcEZGVnB2TVR6YXlGNk9RNUNCWUJuYmxiRU1TaTdVclh5U0xzcQ0KUm9OLzJoTml1S0RobktBWDNZcGVYZnYxdnIyaWIrMGtiRDVvMVZLcnB6SlJ5emlsWHFwQWJwNDQ4RExFZHVGeA0KTnJKc0tWb1dXMVhmS1FvR2VDMXZqNFJudzdhWkNyU1BjMEhvMUdKb29xSXlrNjRXZnhGWjhpVElsWDJMU1gzbQ0KQm1pN1ZXZnVJK0QvUXRtWEVpN05oTE1jUmExVXB4QkdiK1ZPZXlucEVyMEpxM3VZSmlva1VIZzloOGIzSDM1Yg0KK0lJaEFRS0JnUUNtbzUrRUNpOEJUcFRHdlpqQXBtdmxGZ055NkpWd0RvVHJLSGdCcHJHVHpsVGt6Q0tLYzJhTw0KTTMvQmVIVUQ1NVFqcTdUM094eGgzQUk3cG9kT0tCWS84eXgzenlYUExQUmVCMlR0eTNVQjE5YU5CclhNUGQ5RQ0KTG5hbUQ0SmZQeTV0WHpWWEloWVB0UzBiNjdGbjZnZFl4MlNBbGFYYWFjejU4eURLZnlZL3VRS0JnUUNCMUxNYg0KTFZNTU1Pd1U4R2xsVVB1V0Q5VnBsMnlURytCN2Nnd0ZiYTlrWVlVSFQ4TmlvaUdxQzlLZWxqeklKNnlUdyt3aA0KQisvNGNJU1o5Y0JDNHB0VnRlTWJwUkg3UTQwK20xU3M1SmQzU1VFcDBMOW5VbmdVNDhrMzlpSTV5aEgxU3JTRw0KREtBamVka2Rjb3hNVWFwUHB6cVprOVNKdGVHckVsTHY5WmcvNFFLQmdHcExQTmt3MWNMenFSc01uWDlscUxVOA0KVVg3ZXRWUUJMdkFCcEdzSk5ZU2V2TEFsTm5NeUlWemJwTVo0Vzh2ZDBEVGd0bEp0dGpFRjhiYXpCdVBuMWlsOA0KcUFUVWNiMnhZejhIWUczK3RlZ3pFWHhycDFOY3RDbjlxVmxMMDFHR0w4czhTLzRFbENYNDM0bEwwcUlDT2lWNw0KS3JTS3dIeU51RjVIbzhjSFRCS2hBb0dBVjhVRVpENEs0N1g3RTlqSFY3QWN0UUQxUkpJUzI0ekFBSlJFSGpOZA0KbkIzMVFkRWhoSCtaQkkxRGtPUitLTTN4ZEYvQ05UNVpjM29WbXNPeGUwWVl2eVBjYWViNU5nVWY2cEExK1QvOA0KU2E4ak4zS2d3dWZVdUVLWjJKUXo1UjkvYm5WWGZ6dnYwcXZnSkhFNCthdEgyTnR6MlRsQU9PelZwOElXZ1pSbA0KbmVFQ2dZQXpVS0RLS1ZKaXBYdC9TRC8xd2ZzWElCazFxYWpmdENHOFVTcFlGWGp5VzVYMlZ5bTlVTldmVEdJcA0KQlJkVldXamdFbWhyOU5Da1UrVEU4WEVjTXJNRnhWbXBEeTQ4YmxsQWltWFJtOEhNMlhKODQ4TlArTG5GdEtHdA0KTWswT3JuUHdPYitualZGcXdWUllNQ0QzeFN0Z2RnZUppUHFYcWd3SDhjR0M0OGVmZ0E9PQ0KLS0tLS1FTkQgUlNBIFBSSVZBVEUgS0VZLS0tLS0=",
                      "publicJWTkey": "LS0tLS1CRUdJTiBQVUJMSUMgS0VZLS0tLS0NCk1JSUJJVEFOQmdrcWhraUc5dzBCQVFFRkFBT0NBUTRBTUlJQkNRS0NBUUJVZ3VkM3U4WVdxQkJZa1hnczYrbmkNCjEvOG5zOWI0RjZGbncwOEQ0SnRkZWpYOS9WWjJYV05qYnV6QzFSREcydWx4UzhCNkh1Nlh4SFhZT1dTZnM4RjQNClB2dngvU1FnZkVrVlJTQy9aUDd6czNWRE5ZWjBzUlJwM1JnNTNrb2pQbzRPejZrMVp4amtsVEVkSkxZS2piVm4NCnd1NTNwNzExbFUyZFdweEd1RHhmMVUyblFoK1hpSEpmSHJSbFNlUDVpWFhCSnF3bVVMMWQrbmVNTkE5eUZqRHQNCjJzWTBuYkg0ckg1cW5NVUZOMTBKbGd1eDRXYlNoYVQ2SU03eHV1RUdsNENyRUQ2Z1N6ME00YXVRWTh6RnFBbEkNCjZRVlE5TC9vM0JwQkZsWFFDMFpqMG95cEFRSi9DVUtpT0VwSEtEeCtqbTZFT0o4TVp6OUxDTXJOZ0swTkpJaVoNCkFnTUJBQUU9DQotLS0tLUVORCBQVUJMSUMgS0VZLS0tLS0=",
                      "truckerapicrt": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUQxVENDQXIyZ0F3SUJBZ0lVUFVzZ2NYVXZRYWtuQldjeUFDeEw2dmt3b204d0RRWUpLb1pJaHZjTkFRRUwKQlFBd2RqRUxNQWtHQTFVRUJoTUNRMEV4Q3pBSkJnTlZCQWdNQWs5T01ROHdEUVlEVlFRSERBWk5hV3gwYjI0eApGekFWQmdOVkJBb01Ea1JoZEdGVWNuVmphMlZ5TGtsUE1SUXdFZ1lEVlFRTERBdEZibWRwYm1WbGNtbHVaekVhCk1CZ0dBMVVFQXd3UlJHRjBZVlJ5ZFdOclpYSlNiMjkwUTBFd0hoY05NakF4TURFeU1UWXlOalV5V2hjTk1qTXcKTVRFMU1UWXlOalV5V2pCM01Rc3dDUVlEVlFRR0V3SkRRVEVMTUFrR0ExVUVDQXdDVDA0eER6QU5CZ05WQkFjTQpCazFwYkhSdmJqRVhNQlVHQTFVRUNnd09SR0YwWVZSeWRXTnJaWEl1U1U4eEZEQVNCZ05WQkFzTUMwVnVaMmx1ClpXVnlhVzVuTVJzd0dRWURWUVFEREJKa1lYUmhkSEoxWTJ0bGNpMWhjR2t1YVc4d2dnRWlNQTBHQ1NxR1NJYjMKRFFFQkFRVUFBNElCRHdBd2dnRUtBb0lCQVFEakM4ZmVCVEg4L3NnejFQcTJQYU9QQjc5YWtyeUZvWVpGV292bQpWOXNvMmJNRHp4NVBZSFJydVFGYmJQRzlVVlc2QWZpSHdEb1p2Z3NTVXpyS3ZyR1RlYjluNUhSVlVOMTl0NExTCjdtUFppNzFRbGE4bVR3QmdvTnE4TFppcnZUdzV0S2I4K3NmcFJxQUlUQkVsRmxVUEZISzllcDl2eFdvaHhwdGgKS1dPVkwvV2ZtWldhWXlRdjlLNDBkWUJpRWpGcTFKZVB4MDlnUFdHM2pOT1VYTTZQUVlMaHJDUXFFcy9VME9GWQpwVFRHUmNXZEx4TFVFZVNkcHJKK1ZEc3lEK1lVeUFRR2phYk14VmVTVlRUN1FGbFBjLzc3OVZsRkdXMC8xWVVaCmpqakNsSkFLUWZOTHg5bG85cWJERVBZL3grZmhnRnYyWUVaSTNTdGVyZWUvVDJoWEFnTUJBQUdqV2pCWU1COEcKQTFVZEl3UVlNQmFBRkNJVy9JYkZIUUNBeEdxZC9FVFpXbVhGeGMvQ01Ba0dBMVVkRXdRQ01BQXdDd1lEVlIwUApCQVFEQWdUd01CMEdBMVVkRVFRV01CU0NFbVJoZEdGMGNuVmphMlZ5TFdGd2FTNXBiekFOQmdrcWhraUc5dzBCCkFRc0ZBQU9DQVFFQUxrMlJDbnVweDA5Wmx0SmN3M1pUaFIveUZGNS9KVlVsWkdHTnc3MlAwQmR6SEloQncrVk4KRkNucGFPMjJ0WGF1NlJVeTVNUWhwc3JrbFNKYVdJT25jOUg0Uk0vbGh3YWsyRXN0MkFhdUFwNUdYSVFJMDVBeAo3NmtDZXEyMnUyalNXRWpFS1hGRnRiZXVDZnFpWlpqNWFLcGF1YlZWRG4ydEZTWjJRMFRib2RRN2dpTFdNUTBLCk5ycVgwQWNSU2J0ZnRSNjBhdGQ1YW41aFNXY0dyZUg5ZHdFUVRvVVNFblg2RFRZSURoeElsZStjQ2V5VjVSWk8KVS9hck5RZ0lFY3ZrZUM1d1lZaFFFL1FUN3lNb1lNajFnckxmOTZSZkRhbEVFNE1seUtDWm9wYWNJWUdzSmhuKwpLMWFISEQvbGZxTC9zbnNtY3pXdHlLUncxTDF1WkxPcjJnPT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=",
                      "truckerapikey": "LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBNHd2SDNnVXgvUDdJTTlUNnRqMmpqd2UvV3BLOGhhR0dSVnFMNWxmYktObXpBODhlClQyQjBhN2tCVzJ6eHZWRlZ1Z0g0aDhBNkdiNExFbE02eXI2eGszbS9aK1IwVlZEZGZiZUMwdTVqMll1OVVKV3YKSms4QVlLRGF2QzJZcTcwOE9iU20vUHJINlVhZ0NFd1JKUlpWRHhSeXZYcWZiOFZxSWNhYllTbGpsUy8xbjVtVgptbU1rTC9TdU5IV0FZaEl4YXRTWGo4ZFBZRDFodDR6VGxGek9qMEdDNGF3a0toTFAxTkRoV0tVMHhrWEZuUzhTCjFCSGtuYWF5ZmxRN01nL21GTWdFQm8ybXpNVlhrbFUwKzBCWlQzUCsrL1ZaUlJsdFA5V0ZHWTQ0d3BTUUNrSHoKUzhmWmFQYW13eEQyUDhmbjRZQmI5bUJHU04wclhxM252MDlvVndJREFRQUJBb0lCQUZSbFo2aiswMk1RcmdpZQpUWWQ2aUJhOXNZS1FjdVZWY2x6TFdKK0ozdHhqZCtpZGo4Mi9zK2JyQVRjajJ3UWRmTGxFekVaaHpQY09ybnprClZVRlIreDR2cmxlSEV6czNjTzJ2VGtEci9rKzZZdWZXakNHNFVoS1pHRlhFdVRCMWlqS0ZGV0dLTXM5aXdrL3oKUHBoRG9lQ1dJRVNRbFQxdDY1S2NpZHRuRUM0K3oxaURxR21ZZEVEa2xEZFFIVHorRGpseS80VzV2MFpVQWN2Vwo1Unkvd0JldmNURlBXQy9uZGFrN3hGUnVrU2xNZUlhTWhvTlllOU5uVCt2Nmh5VVh1MXp6cG45UEFJRTVpQWpOCnBsZkFEZitwakwrZFJDbW9sWW1LR3NQVjEzSmtCY1p2cEVoUWgrUmJnaWFwczJLc2plQ0NmTmtST0hndHc0MU0KeGxOTU1Ha0NnWUVBOHM1akpuZ2FMblIvTzZtakZEb0wvd1p2UWdnT1h3Z1RBN2tsQ0trbjBYM0xJS0hhaTZESgpyOVRGQktMdkhUazVFK1Y4NkhIZ29VMUFQQnYyMWJpaHhuQXhUYm8wSHZMdG5QOXU3YnZJejFNOWMyd2ZlWUgzCmliMG1FNkV0Y2wvNnVIUHdMVlZtRGh5bllneEZpeTRRRXp1ZHZwaFIvSnlPazQvUytQUzh2ZXNDZ1lFQTcySW8KWDlkY0xvUmYzSk0vRTF1SXhrRGRLZm1VakN6Mk1NN3pyRkpNUERobTFSUlJhQjdidHNxUzl4Mm5KeTI5amtoSQpndTR0Y2s5STdncm9WRXY0cEtkU0lOVzVNVUg0MXJPckxTbFBobCtMaTM5QXNrTERFVXVQbkFmYko3YWZibFFHCjNFWE1Ka1d2WFQxRG90aE11MUpUMTFPMjhmYlBmb3R3YVRucXFFVUNnWUE1RnArblh0dkJDbkVUNVJuTDJxNFIKWTBEcmhKdGthckNWMG9WTG1RSWN0bmY1RHpVa1dVU3lpOU5hK0xoQjViOXUxeWxjT2RJMytzTGNuektWSTVrZAo0Q2tuZmwxYzNYOVN4NWdjVzJlbW40cTJSYmpNeCtzY3ZwVmlUaVl5V3pESEI3Tno1VDFlbWd3YytJZXRac0l1CjAzaU9aSi82UmhFYnp0Z3E0dGd4eXdLQmdCanhwSDFnSm9YZVcxTFRzUlhtU3RsRnR3WUJzTDN1alhUR2k5T2YKKzd1bXNPaHhKbzV2OExzYVhyZ0xTZmltSm9ObCt0QlIyMEIzRElWcGM0NGRJSHFmeU1vTkd6QUx6TThxaDZLOApGRUpQUUc0M3cxa1A1ZDJ1ZnZkbVJpcmNONzY2V01DcE1UMm5zejlHKy81REJiUGJQMHRIdU45N3RXRmRURWhhCnRaMzFBb0dCQU5xZnVycEV4ZHozdXRNUm91RnJjTmVkV2s4VlVpVVZIV1hZMjhNNWQvRjcxMExIOTZuOW12V08Kc1cwcDFjcXBJMEsyQko2WlYwNVpybnlyRHBsVVpVdVpSQWxGR3hYR1dLU2trQzlmWm90SkFOWDBkR29KSitqTwpEYUNLNGxlcXdOMHNPRE1YZFErMlFPaUtlSUxkNU5KcDNUSW1YV1Z0RFlXUkRwOW9iSUdlCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="
                    },
                    "jobsServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": true,
                        "FileSFTP": true,
                        "IOTKafka": true,
                        "IOTProxy": false,
                        "IOTRedis": true,
                        "Login": false,
                        "ScriptJS": true,
                        "ScriptSSH": true,
                        "ScriptShell": true,
                        "UtilEcho": true,
                        "UtilFuzzy": true,
                        "UtilSentiment": true,
                        "credentials": false,
                        "jobs": true,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "loginServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "info"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": false,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": false,
                        "usermanagement": false
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "managementServer": {
                      "ajv": {
                        "customOptions": {
                          "allErrors": true,
                          "coerceTypes": true,
                          "jsPropertySyntax": true,
                          "keepErrors": false,
                          "singleError": true
                        }
                      },
                      "bind": "0.0.0.0",
                      "compression": {
                        "threshold": 2048
                      },
                      "fastify": {
                        "ignoreTrailingSlash": true,
                        "logger": {
                          "level": "debug"
                        }
                      },
                      "helmet": {
                        "framegaurd": {
                          "action": "deny"
                        }
                      },
                      "jobResponseErrors": true,
                      "keycloak": true,
                      "pluginsEnable": {
                        "DB": false,
                        "FileSFTP": false,
                        "IOTKafka": false,
                        "IOTProxy": false,
                        "IOTRedis": false,
                        "Login": true,
                        "ScriptJS": false,
                        "ScriptSSH": false,
                        "ScriptShell": false,
                        "UtilEcho": false,
                        "UtilFuzzy": false,
                        "UtilSentiment": false,
                        "credentials": true,
                        "jobs": false,
                        "resources": false,
                        "statuschecks": true,
                        "ui": true,
                        "usermanagement": true
                      },
                      "port": "8080",
                      "protocol": {
                        "http2": false,
                        "httpsconfig": {
                          "allowHTTP1": true,
                          "cert": "truckerapi.crt",
                          "key": "truckerapi.key"
                        },
                        "type": "http"
                      },
                      "requestLog": false,
                      "serverID": "ServerHandler",
                      "writableResource": false
                    },
                    "name": "API",
                    "resource": {
                      "Cache": {
                        "Creds": false,
                        "CredsFlush": 10,
                        "Resource": false,
                        "ResourceFlush": 10
                      },
                      "Templates": {
                        "Scripts": "scripts",
                        "Shell": "shell",
                        "keys": "keys",
                        "resourcedefinitions": "resourcedefinitions"
                      }
                    }
                  },
                  "DB": {
                    "databasename": "userdb",
                    "hostname": "db",
                    "password": "password",
                    "port": 5432,
                    "type": "pg",
                    "user": "testuser"
                  },
                  "TempDB": {
                    "enabled": false,
                    "pvc": "samplepvc"
                  }
                }
              },
              {
                "apiVersion": "datatrucker.datatrucker.io/v1",
                "kind": "DatatruckerFlow",
                "metadata": {
                  "name": "datatruckerflow-sample"
                },
                "spec": {
                  "API": {
                    "Image": {
                      "imageName": "datatrucker-server/datatrucker-api",
                      "imageNameUI": "datatruckerio/datatrucker-ui",
                      "repository": "registry.gitlab.com/datatruckerio",
                      "tagName": "latest"
                    },
                    "name": "api"
                  },
                  "DatatruckerConfig": "datatruckerconfig-sample",
                  "JobDefinitions": [
                    {
                      "name": "echoapi",
                      "restmethod": "GET",
                      "tenant": "Admin",
                      "type": "Util-Echo",
                      "validations": {
                        "properties": {
                          "voila": {
                            "enum": [
                              "true",
                              "false"
                            ],
                            "type": "string"
                          }
                        },
                        "type": "object"
                      }
                    }
                  ],
                  "Keys": {
                    "configmap": "placeholder"
                  },
                  "Replicas": 1,
                  "Resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "256Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "256Mi"
                    }
                  },
                  "Scripts": {
                    "configmap": "placeholder"
                  },
                  "Type": "Job"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          certified: "false"
          containerImage: docker.io/datatruckerio/datatrucker-operator:2.0.0
          createdAt: "2021-09-08 01:01:01"
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: hhttps://gitlab.com/datatruckerio/datatrucker-server.git
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              - creates secrets to hold server configs
              - create tempDB for protoypting
              - create configs for seperating Login, management and Jobs endpoints
              - Create crypto configs
            displayName: Datatrucker Config
            kind: DatatruckerConfig
            name: datatruckerconfigs.datatrucker.datatrucker.io
            version: v1
          - description: '- creates an API as per Job Definitions'
            displayName: Datatrucker Flow
            kind: DatatruckerFlow
            name: datatruckerflows.datatrucker.datatrucker.io
            version: v1
        description: "## DataTrucker.IO\n\nDataTrucker.IO is a simple no-code / less-code
          API Backend\n\n### What does it do ?\n\nDatatrucker.IO is a product capable
          of reading simple json/yaml configs and building th code necessary for converting
          it into an API.\nAlong with building the code, it will also host the code
          base on a nodejs server , i.e. it will immediately make it available for
          consumption.\n\n### What does it offload from a developer ?\n\nDataTrucker
          is capable of removing the most common activities which a developer needs
          to do on every new project A few of such common activities are\n\n- Creating
          an API endpoint with a specified business logic (using simple plugins)\n-
          Applying standard RBAC\n- Applying Authorization logic\n- Applying hardening
          on endpoints\n- Log management\n- Connecting to variety of systems\n- Modularizing
          business logic\n- The best of doing it with little to no code\n\n### watch
          the quick elevator pitch to understand datatrucker.io\n\n https://youtu.be/DKLFDqhjs5M\n "
        displayName: DataTrucker.IO
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - datatrucker
        - operator
        - nocode
        - apibuilder
        links:
        - name: Website
          url: https://www.datatrucker.io/
        - name: Git Repo
          url: https://gitlab.com/datatruckerio/datatrucker-server.git
        maintainers:
        - email: gauravshankar.can@gmail.com
          name: Gaurav Shankar
        maturity: alpha
        provider:
          name: DataTrucker.IO
          url: https://www.datatrucker.io/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/datatrucker-operator@sha256:4ab1d683d95d3eeba944b977594aff2e3b3e7203f56598b2fb2f7cff75354be3
        - docker.io/datatruckerio/datatrucker-operator:2.0.0
        version: 2.0.0
      entries:
      - name: datatrucker-operator.v2.0.0
        version: 2.0.0
      - name: datatrucker-operator.v1.3.2
        version: 1.3.2
      - name: datatrucker-operator.v1.3.1
        version: 1.3.1
      - name: datatrucker-operator.v1.3.0
        version: 1.3.0
      name: alpha
    defaultChannel: alpha
    packageName: datatrucker-operator
    provider:
      name: DataTrucker.IO
      url: https://www.datatrucker.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: NeuVector
      provider-url: ""
    name: neuvector-community-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: neuvector-operator.v1.5.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "apm.neuvector.com/v1alpha1",
                "kind": "Neuvector",
                "metadata": {
                  "name": "neuvector"
                },
                "spec": {
                  "openshift": true,
                  "registry": "docker.io",
                  "tag": "5.4.6",
                  "oem": "",
                  "imagePullSecrets": "",
                  "psp": false,
                  "rbac": true,
                  "serviceAccount": "default",
                  "leastPrivilege": false,
                  "global": {
                    "cattle": {
                      "url": ""
                    },
                    "azure": {
                      "enabled": false,
                      "identity": {
                        "clientId": "DONOTMODIFY"
                      },
                      "marketplace": {
                        "planId": "DONOTMODIFY"
                      },
                      "extension": {
                        "resourceId": "DONOTMODIFY"
                      },
                      "serviceAccount": "csp",
                      "imagePullSecrets": "",
                      "images": {
                        "neuvector_csp_pod": {
                          "tag": "latest",
                          "image": "neuvector-billing-azure-by-suse-llc",
                          "registry": "registry.suse.de/suse/sle-15-sp5/update/pubclouds/images",
                          "imagePullPolicy": "IfNotPresent"
                        },
                        "controller": {
                          "tag": "5.2.4",
                          "image": "controller",
                          "registry": "docker.io/neuvector"
                        },
                        "manager": {
                          "tag": "5.2.4",
                          "image": "manager",
                          "registry": "docker.io/neuvector"
                        },
                        "enforcer": {
                          "tag": "5.2.4",
                          "image": "enforcer",
                          "registry": "docker.io/neuvector"
                        }
                      }
                    },
                    "aws": {
                      "enabled": false,
                      "accountNumber": "",
                      "roleName": "",
                      "serviceAccount": "csp",
                      "annotations": {},
                      "imagePullSecrets": "",
                      "image": {
                        "digest": "",
                        "repository": "neuvector/neuvector-csp-adapter",
                        "tag": "latest",
                        "imagePullPolicy": "IfNotPresent"
                      }
                    }
                  },
                  "bootstrapPassword": "",
                  "autoGenerateCert": true,
                  "defaultValidityPeriod": 365,
                  "internal": {
                    "certmanager": {
                      "enabled": false,
                      "secretname": "neuvector-internal"
                    },
                    "autoGenerateCert": true,
                    "autoRotateCert": true
                  },
                  "controller": {
                    "enabled": true,
                    "annotations": {},
                    "strategy": {
                      "type": "RollingUpdate",
                      "rollingUpdate": {
                        "maxSurge": 1,
                        "maxUnavailable": 0
                      }
                    },
                    "image": {
                      "repository": "neuvector/controller",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "replicas": 3,
                    "disruptionbudget": 0,
                    "schedulerName": "",
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "searchRegistries": "",
                    "env": [],
                    "affinity": {
                      "podAntiAffinity": {
                        "preferredDuringSchedulingIgnoredDuringExecution": [
                          {
                            "weight": 100,
                            "podAffinityTerm": {
                              "labelSelector": {
                                "matchExpressions": [
                                  {
                                    "key": "app",
                                    "operator": "In",
                                    "values": [
                                      "neuvector-controller-pod"
                                    ]
                                  }
                                ]
                              },
                              "topologyKey": "kubernetes.io/hostname"
                            }
                          }
                        ]
                      }
                    },
                    "tolerations": [],
                    "topologySpreadConstraints": [],
                    "nodeSelector": {},
                    "apisvc": {
                      "type": "",
                      "annotations": {},
                      "nodePort": "",
                      "route": {
                        "enabled": false,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      }
                    },
                    "ranchersso": {
                      "enabled": false
                    },
                    "pvc": {
                      "enabled": false,
                      "existingClaim": false,
                      "accessModes": [
                        "ReadWriteMany"
                      ],
                      "storageClass": "",
                      "capacity": ""
                    },
                    "azureFileShare": {
                      "enabled": false,
                      "secretName": "",
                      "shareName": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "federation": {
                      "mastersvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      },
                      "managedsvc": {
                        "type": "",
                        "loadBalancerIP": "",
                        "clusterIP": "",
                        "nodePort": "",
                        "externalTrafficPolicy": "",
                        "internalTrafficPolicy": "",
                        "ingress": {
                          "enabled": false,
                          "host": "",
                          "ingressClassName": "",
                          "path": "/",
                          "annotations": {
                            "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                          },
                          "tls": false,
                          "secretName": ""
                        },
                        "annotations": {},
                        "route": {
                          "enabled": false,
                          "termination": "passthrough",
                          "host": "",
                          "tls": ""
                        }
                      }
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "configmap": {
                      "enabled": false,
                      "data": ""
                    },
                    "secret": {
                      "enabled": false,
                      "data": {
                        "userinitcfg.yaml": {
                          "users": [
                            {
                              "Fullname": "admin",
                              "Password": "",
                              "Role": "admin"
                            }
                          ]
                        }
                      }
                    },
                    "certupgrader": {
                      "env": [],
                      "schedule": "",
                      "imagePullPolicy": "IfNotPresent",
                      "timeout": 3600,
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "prime": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/compliance-config",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "1.0.7",
                        "hash": ""
                      }
                    }
                  },
                  "enforcer": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/enforcer",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "updateStrategy": {
                      "type": "RollingUpdate"
                    },
                    "priorityClassName": "",
                    "podLabels": {},
                    "podAnnotations": {},
                    "env": [],
                    "tolerations": [
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/master"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/control-plane"
                      },
                      {
                        "effect": "NoSchedule",
                        "key": "node-role.kubernetes.io/etcd"
                      }
                    ],
                    "resources": {},
                    "internal": {
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt",
                        "caFile": "ca.crt"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "manager": {
                    "enabled": true,
                    "image": {
                      "repository": "neuvector/manager",
                      "imagePullPolicy": "IfNotPresent",
                      "hash": ""
                    },
                    "priorityClassName": "",
                    "env": {
                      "ssl": true,
                      "envs": []
                    },
                    "svc": {
                      "type": "ClusterIP",
                      "nodePort": "",
                      "loadBalancerIP": "",
                      "annotations": {}
                    },
                    "route": {
                      "enabled": true,
                      "termination": "passthrough",
                      "host": "",
                      "tls": ""
                    },
                    "certificate": {
                      "secret": "",
                      "keyFile": "tls.key",
                      "pemFile": "tls.pem"
                    },
                    "ingress": {
                      "enabled": false,
                      "host": "",
                      "ingressClassName": "",
                      "path": "/",
                      "annotations": {
                        "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                      },
                      "tls": false,
                      "secretName": ""
                    },
                    "resources": {},
                    "topologySpreadConstraints": [],
                    "affinity": {},
                    "podLabels": {},
                    "podAnnotations": {},
                    "tolerations": [],
                    "nodeSelector": {},
                    "runAsUser": "",
                    "probes": {
                      "enabled": false,
                      "timeout": 1,
                      "periodSeconds": 10,
                      "startupFailureThreshold": 30
                    }
                  },
                  "cve": {
                    "adapter": {
                      "enabled": false,
                      "image": {
                        "repository": "neuvector/registry-adapter",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.1.9",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "certificate": {
                        "secret": "",
                        "keyFile": "tls.key",
                        "pemFile": "tls.crt"
                      },
                      "harbor": {
                        "protocol": "https",
                        "secretName": ""
                      },
                      "svc": {
                        "type": "ClusterIP",
                        "loadBalancerIP": "",
                        "annotations": {}
                      },
                      "route": {
                        "enabled": true,
                        "termination": "passthrough",
                        "host": "",
                        "tls": ""
                      },
                      "ingress": {
                        "enabled": false,
                        "host": "",
                        "ingressClassName": "",
                        "path": "/",
                        "annotations": {
                          "nginx.ingress.kubernetes.io/backend-protocol": "HTTPS"
                        },
                        "tls": false,
                        "secretName": ""
                      },
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      }
                    },
                    "updater": {
                      "enabled": true,
                      "secure": false,
                      "cacert": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                      "image": {
                        "registry": "",
                        "repository": "neuvector/updater",
                        "imagePullPolicy": "IfNotPresent",
                        "tag": "0.0.5",
                        "hash": ""
                      },
                      "schedule": "0 0 * * *",
                      "priorityClassName": "",
                      "resources": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": ""
                    },
                    "scanner": {
                      "enabled": true,
                      "replicas": 3,
                      "dockerPath": "",
                      "strategy": {
                        "type": "RollingUpdate",
                        "rollingUpdate": {
                          "maxSurge": 1,
                          "maxUnavailable": 0
                        }
                      },
                      "image": {
                        "registry": "",
                        "repository": "neuvector/scanner",
                        "imagePullPolicy": "Always",
                        "tag": "6",
                        "hash": ""
                      },
                      "priorityClassName": "",
                      "resources": {},
                      "topologySpreadConstraints": [],
                      "affinity": {},
                      "podLabels": {},
                      "podAnnotations": {},
                      "env": [],
                      "tolerations": [],
                      "nodeSelector": {},
                      "runAsUser": "",
                      "internal": {
                        "certificate": {
                          "secret": "",
                          "keyFile": "tls.key",
                          "pemFile": "tls.crt",
                          "caFile": "ca.crt"
                        }
                      },
                      "volumes": "",
                      "volumeMounts": ""
                    }
                  },
                  "resources": {},
                  "runtimePath": "",
                  "docker": {
                    "enabled": false,
                    "path": "/var/run/docker.sock"
                  },
                  "k3s": {
                    "enabled": false,
                    "runtimePath": "/run/k3s/containerd/containerd.sock"
                  },
                  "bottlerocket": {
                    "enabled": false,
                    "runtimePath": "/run/dockershim.sock"
                  },
                  "containerd": {
                    "enabled": false,
                    "path": "/var/run/containerd/containerd.sock"
                  },
                  "crio": {
                    "enabled": true,
                    "path": "/var/run/crio/crio.sock"
                  },
                  "admissionwebhook": {
                    "type": "ClusterIP"
                  },
                  "crdwebhooksvc": {
                    "enabled": true
                  },
                  "crdwebhook": {
                    "enabled": true,
                    "type": "ClusterIP"
                  },
                  "lease": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking, Security
          certified: "false"
          containerImage: docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
          createdAt: "2019-11-18T02:09:59Z"
          description: NeuVector delivers the only cloud-native Kubernetes security
            platform with uncompromising end-to-end protection from DevOps vulnerability
            protection to automated run-time security, and featuring a true Layer
            7 container firewall.
          repository: https://github.com/neuvector/neuvector-operator
          support: support@neuvector.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A Full LifeCycle Container Security Platform
            displayName: NeuVector
            kind: Neuvector
            name: neuvectors.apm.neuvector.com
            version: v1alpha1
        description: |+
          NeuVector delivers the only cloud-native Kubernetes security platform with uncompromising end-to-end protection from DevOps vulnerability protection to automated run-time security, and featuring a true Layer 7 container firewall.

          The NeuVector Operator runs  in the openshift container platform to deploy and manage the NeuVector Security cluster components. The NeuVector operator contains all necessary information to deploy NeuVector using helm charts. You simply need to install the NeuVector operator from the OpenShift embeded operator hub and create NeuVector instance. You can modify the NeuVector installation configuration by modifying yaml while creating the NeuVector instance such as imagePullSecrets, tag version, etc. Please refer to [github link](https://github.com/neuvector/neuvector-helm) for the values that can be modifed during installation. To upgrade to a newer version of NeuVector, just reapply the NeuVector instance with desired tag , which in turn pulls the specified NeuVector image tags and upgrades as per upgrade plan configured on the helm chart. NeuVector Operator version 1.5.5 deploys NeuVector app version 5.4.6.

          **Complete below steps to create namespace and Grant Service Account Access to the Privileged SCC before installation.**

          Create the NeuVector namespace

                   oc new-project  neuvector
          Login as system:admin account

                   oc login -u system:admin

          Privileged SCC is added to Service Account specified in the deploment yaml by Operator version 1.3.4 and above in new deployment. In case of upgrading NeuVector Operator from previous version to 1.3.4, please delete Privileged SCC before upgrading.

                   oc delete rolebinding -nneuvector system:openshift:scc:privileged

          The following info will be added in the Privileged SCC users:

                   - system:serviceaccount:neuvector:default

          In OpenShift 4.6+ use the following to check:

                   oc get rolebinding system:openshift:scc:privileged -n neuvector -o wide
                   system:openshift:scc:privileged   ClusterRole/system:openshift:scc:privileged   9m22s  neuvector/default


        displayName: NeuVector Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - wavefront
        - wavefront collector
        - wavefront proxy
        links:
        - name: NeuVector Operator
          url: https://github.com/neuvector/neuvector-operator
        maintainers:
        - email: support@neuvector.com
          name: A Full LifeCycle Container Security Platform
        maturity: beta
        provider:
          name: NeuVector
        relatedImages:
        - docker.io/neuvector/neuvector-operator@sha256:6815df6435eafaf7b2bc9d5e3f0c9b65f0a50ad4e05d2862e56b227e19aeca27
        - quay.io/community-operator-pipeline-prod/neuvector-community-operator@sha256:788f742c9f2def80d0b768eb96c5138f2cebdad767ea4ebb8cc26451771e23f5
        version: 1.5.5
      entries:
      - name: neuvector-operator.v1.5.5
        version: 1.5.5
      - name: neuvector-operator.v1.5.4
        version: 1.5.4
      - name: neuvector-operator.v1.5.3
        version: 1.5.3
      - name: neuvector-operator.v1.5.2
        version: 1.5.2
      - name: neuvector-operator.v1.5.1
        version: 1.5.1
      - name: neuvector-operator.v1.5.0
        version: 1.5.0
      - name: neuvector-operator.v1.4.9
        version: 1.4.9
      - name: neuvector-operator.v1.4.8
        version: 1.4.8
      - name: neuvector-operator.v1.4.7
        version: 1.4.7
      - name: neuvector-operator.v1.4.6
        version: 1.4.6
      - name: neuvector-operator.v1.4.5
        version: 1.4.5
      - name: neuvector-operator.v1.4.4
        version: 1.4.4
      - name: neuvector-operator.v1.4.3
        version: 1.4.3
      - name: neuvector-operator.v1.4.2
        version: 1.4.2
      - name: neuvector-operator.v1.4.1
        version: 1.4.1
      - name: neuvector-operator.v1.4.0
        version: 1.4.0
      - name: neuvector-operator.v1.3.9
        version: 1.3.9
      - name: neuvector-operator.v1.3.8
        version: 1.3.8
      - name: neuvector-operator.v1.3.7
        version: 1.3.7
      - name: neuvector-operator.v1.3.6
        version: 1.3.6
      - name: neuvector-operator.v1.3.5
        version: 1.3.5
      - name: neuvector-operator.v1.3.4
        version: 1.3.4
      - name: neuvector-operator.v1.3.3
        version: 1.3.3
      - name: neuvector-operator.v1.3.2
        version: 1.3.2
      - name: neuvector-operator.v1.3.1
        version: 1.3.1
      - name: neuvector-operator.v1.3.0
        version: 1.3.0
      - name: neuvector-operator.v1.2.9
        version: 1.2.9
      name: beta
    defaultChannel: beta
    packageName: neuvector-community-operator
    provider:
      name: NeuVector
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: TrendMicro
      provider-url: trendmicro.com
    name: visionone-containersecurity
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      - name: visionone-containersecurity.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: visionone-containersecurity.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              },
              {
                "apiVersion": "container-security.trendmicro.com/v1alpha1",
                "kind": "VisionOneContainerSecurity",
                "metadata": {
                  "name": "visionone-containersecurity-sample",
                  "namespace": "trendmicro-system"
                },
                "spec": {
                  "visionOne": {
                    "bootstrapToken": "",
                    "clusterRegistrationKey": false,
                    "endpoint": "https://api.xdr.trendmicro.com/external/v2/direct/vcs/external/vcs",
                    "exclusion": {
                      "namespaces": [
                        "kube-system",
                        "openshift",
                        "openshift-apiserver",
                        "openshift-apiserver-operator",
                        "openshift-aqua",
                        "openshift-authentication",
                        "openshift-authentication-operator",
                        "openshift-backplane",
                        "openshift-backplane-cee",
                        "openshift-backplane-srep",
                        "openshift-build-test",
                        "openshift-cloud-credential-operator",
                        "openshift-cloud-ingress-operator",
                        "openshift-cluster-csi-drivers",
                        "openshift-cluster-machine-approver",
                        "openshift-cluster-node-tuning-operator",
                        "openshift-cluster-samples-operator",
                        "openshift-cluster-storage-operator",
                        "openshift-cluster-version",
                        "openshift-codeready-workspaces",
                        "openshift-config",
                        "openshift-config-managed",
                        "openshift-config-operator",
                        "openshift-console",
                        "openshift-console-operator",
                        "openshift-console-user-settings",
                        "openshift-controller-manager",
                        "openshift-controller-manager-operator",
                        "openshift-custom-domains-operator",
                        "openshift-customer-monitoring",
                        "openshift-dns",
                        "openshift-dns-operator",
                        "openshift-etcd",
                        "openshift-etcd-operator",
                        "openshift-host-network",
                        "openshift-image-registry",
                        "openshift-infra",
                        "openshift-ingress",
                        "openshift-ingress-canary",
                        "openshift-ingress-operator",
                        "openshift-insights",
                        "openshift-kni-infra",
                        "openshift-kube-apiserver",
                        "openshift-kube-apiserver-operator",
                        "openshift-kube-controller-manager",
                        "openshift-kube-controller-manager-operator",
                        "openshift-kube-scheduler",
                        "openshift-kube-scheduler-operator",
                        "openshift-kube-storage-version-migrator",
                        "openshift-kube-storage-version-migrator-operator",
                        "openshift-kubevirt-infra",
                        "openshift-logging",
                        "openshift-machine-api",
                        "openshift-machine-config-operator",
                        "openshift-managed-upgrade-operator",
                        "openshift-marketplace",
                        "openshift-monitoring",
                        "openshift-multus",
                        "openshift-must-gather-operator",
                        "openshift-network-diagnostics",
                        "openshift-network-operator",
                        "openshift-node",
                        "openshift-oauth-apiserver",
                        "openshift-openstack-infra",
                        "openshift-operator-lifecycle-manager",
                        "openshift-operators",
                        "openshift-operators-redhat",
                        "openshift-osd-metrics",
                        "openshift-ovirt-infra",
                        "openshift-rbac-permissions",
                        "openshift-route-monitor-operator",
                        "openshift-sdn",
                        "openshift-security",
                        "openshift-service-ca",
                        "openshift-service-ca-operator",
                        "openshift-splunk-forwarder-operator",
                        "openshift-sre-pruning",
                        "openshift-sre-sshd",
                        "openshift-strimzi",
                        "openshift-user-workload-monitoring",
                        "openshift-validation-webhook",
                        "openshift-velero",
                        "openshift-vsphere-infra",
                        "openshift-addon-operator",
                        "openshift-backplane-managed-scripts",
                        "openshift-cloud-controller-manager-operator",
                        "openshift-cloud-network-config-controller",
                        "openshift-deployment-validation-operator",
                        "openshift-managed-node-metadata-operator",
                        "openshift-ocm-agent-operator",
                        "openshift-ovn-kubernetes",
                        "openshift-azure-operator",
                        "openshift-azure-loggi",
                        "openshift-cloud-controller-manager",
                        "openshift-nutanix-infra",
                        "openshift-route-controller-manager",
                        "dedicated-admin",
                        "default",
                        "openshift-aws-vpce-operator",
                        "openshift-backplane-csa",
                        "openshift-backplane-cse",
                        "openshift-backplane-csm",
                        "openshift-backplane-mobb",
                        "openshift-backplane-tam",
                        "openshift-cluster-api",
                        "openshift-compliance",
                        "openshift-compliance-monkey",
                        "openshift-container-security",
                        "openshift-file-integrity",
                        "openshift-observability-operator",
                        "openshift-platform-operators",
                        "openshift-scanning",
                        "openshift-suricata"
                      ]
                    },
                    "inventoryCollection": {
                      "enabled": true
                    },
                    "malwareScanning": {
                      "enabled": true
                    },
                    "runtimeSecurity": {
                      "enabled": true
                    },
                    "secretScanning": {
                      "enabled": true
                    },
                    "vulnerabilityScanning": {
                      "enabled": true
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          com.redhat.openshift.versions: v4.12-v4.20
          containerImage: public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
          createdAt: "2025-09-18T22:43:41Z"
          description: Trend Vision One™ Container Security delivers full lifecycle
            protection for containers with real-time threat detection, policy enforcement,
            compliance assurance, and rapid incident response.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: trendmicro-system
          operatorhub.io/ui-metadata-max-k8s-version: "1.26"
          operators.openshift.io/valid-subscription: VisionOneContainerSecurity Subscription
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/trendmicro/visionone-container-security-helm
          support: TrendMicro Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Trend Vision One™ Container Security delivers full lifecycle
              protection for containers with real-time threat detection, policy enforcement,
              compliance assurance, and rapid incident response.
            displayName: VisionOneContainerSecurity
            kind: VisionOneContainerSecurity
            name: visiononecontainersecurities.container-security.trendmicro.com
            version: v1alpha1
        description: "[Trend Vision One™ Container Security](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-container-security)
          delivers comprehensive, end-to-end\n protection for containerized workloads
          across OpenShift and Kubernetes environments.\n It continuously monitors
          for threats, vulnerabilities, policy violations, and\n exposed secrets ensuring
          compliance and enabling rapid, automated response. Purpose-built\n for hybrid
          cloud, it instantly isolates compromised containers to prevent lateral\n
          movement and minimize impact. Seamlessly integrating into OpenShift DevSecOps\n
          pipelines, it empowers security teams to enforce policies while enabling
          developers\n to innovate and deploy with confidence without disrupting workflows
          or compromising\n agility.\n\n **Installation**\n \n - Use the recommended
          `trendmicro-system` namespace.\n - Refer to the [Trend Vision One™ Documentation](https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator)
          for registration and installation steps.\n"
        displayName: Trend Vision One™ Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trendmicro
        - visionone
        - container-security
        links:
        - name: Documentation
          url: https://docs.trendmicro.com/en-us/documentation/article/trend-vision-one-red-hat-openshift-operator
        maintainers:
        - email: support@trendmicro.com
          name: TrendMicro VisionOne Support Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: TrendMicro
          url: trendmicro.com
        relatedImages:
        - public.ecr.aws/trendmicro/container-security/visionone-containersecurity-operator:v0.0.2
        - quay.io/community-operator-pipeline-prod/visionone-containersecurity:0.0.2
        version: 0.0.2
      entries:
      - name: visionone-containersecurity.v0.0.2
        version: 0.0.2
      name: stable
    defaultChannel: stable
    packageName: visionone-containersecurity
    provider:
      name: TrendMicro
      url: trendmicro.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SusQL Operator Contributors
      provider-url: https://github.com/sustainable-computing-io/susql-operator
    name: susql-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: susql-operator.v0.0.35
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "susql.ibm.com/v1",
                "kind": "LabelGroup",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "susql-operator",
                    "app.kubernetes.io/instance": "labelgroup-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "susql-operator",
                    "app.kubernetes.io/part-of": "susql-operator"
                  },
                  "name": "labelgroup-sample"
                },
                "spec": {
                  "labels": [
                    "labelgroup-sample-1",
                    "labelgroup-sample-2"
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/susql_operator:0.0.34
          createdAt: "2025-07-10T01:20:37Z"
          description: 'Aggregates energy and CO2 emission data for pods tagged with
            SusQL labels '
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/susql-operator
          support: https://github.com/sustainable-computing-io/susql-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LabelGroup is the Schema for the LabelGroups API
            displayName: Label Group
            kind: LabelGroup
            name: labelgroups.susql.ibm.com
            version: v1
        description: |-
          ### About this Operator

          SusQL is a Kubernetes operator that aggregates energy and estimated
          carbon dioxide emission data for pods tagged with SusQL specific
          labels. The energy measurements are taken from Kepler which should
          be deployed in the cluster before using SusQL.

          ### Prerequisites

          1. Deployment of [Kepler](https://sustainable-computing.io/) on the cluster
          2. Ensure that [User Project Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          is enabled to monitor energy consumed in user projects.
          3. Creation of a SusQL Service Monitor:
                `oc apply -n <NAMESPACE> -f https://raw.githubusercontent.com/sustainable-computing-io/susql-operator/main/hack/susql-servicemonitor.yaml`

          ### API Backward Compatibility

          Until Operator supports **seamless upgrades**, a new version of the operator may
          introduce a change that is **NOT** backwards compatible with the previous version.
          Thus, to upgrade the operator, uninstall the already installed version completely
          (including CRDs) and install the new version. For example to upgrade from 0.1.0
          to 0.2.0, you must first uninstall 0.1.0.

          ### LabelGroup spec Example

          ```
          apiVersion: susql.ibm.com/v1
          kind: LabelGroup
          metadata:
            name: labelgroup-name
            namespace: susql-operator
          spec:
            labels:
              - my-label-1
              - my-label-2
          ```

          ### Documentation

          Documentation can be found at [SusQL Operator](https://github.com/sustainable-computing-io/susql-operator#readme).

          ### Getting help

          If you encounter any issues while using the SusQL Operator, you can create an
          [issue](https://github.com/sustainable-computing-io/susql-operator/issues)
          on our GitHub repo for bugs, enhancements, or other requests.

          ### Contributing
          You can contribute by raising [issues](https://github.com/sustainable-computing-io/susql-operator/issues)
          related to susql-operator.

          ### License
          susql-operator is licensed under the Apache 2.0 license
        displayName: SusQL
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - aggregation
        - sustainability
        - energy
        - kepler
        - susql
        - carbon dioxide (CO2) emissions
        - green computing
        links:
        - name: SusQL Operator
          url: https://github.com/sustainable-computing-io/susql-operator
        - name: Sustainable Computing
          url: https://github.com/sustainable-computing-io
        maintainers:
        - email: trent@jp.ibm.com
          name: Scott Trent
        - email: mohammed.munir.abdi@ibm.com
          name: Mohammed Abdi
        maturity: alpha
        provider:
          name: SusQL Operator Contributors
          url: https://github.com/sustainable-computing-io/susql-operator
        relatedImages:
        - quay.io/community-operator-pipeline-prod/susql-operator@sha256:4a47a8448610e54b8dc9700828c9250b84ee19126d6d4edbb62b35597e40bf37
        - quay.io/sustainable_computing_io/susql_operator:0.0.34
        version: 0.0.35
      entries:
      - name: susql-operator.v0.0.35
        version: 0.0.35
      - name: susql-operator.v0.0.34
        version: 0.0.34
      name: alpha
    defaultChannel: alpha
    packageName: susql-operator
    provider:
      name: SusQL Operator Contributors
      url: https://github.com/sustainable-computing-io/susql-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Labs
      provider-url: https://grafana.com
    name: grafana-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: grafana-operator.v5.19.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "Grafana",
                "metadata": {
                  "labels": {
                    "dashboards": "grafana-a",
                    "folders": "grafana-a"
                  },
                  "name": "grafana-a"
                },
                "spec": {
                  "config": {
                    "auth": {
                      "disable_login_form": "false"
                    },
                    "log": {
                      "mode": "console"
                    },
                    "security": {
                      "admin_password": "start",
                      "admin_user": "root"
                    }
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaAlertRuleGroup",
                "metadata": {
                  "name": "grafanaalertrulegroup-sample"
                },
                "spec": {
                  "folderRef": "test-folder-from-operator",
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "interval": "5m",
                  "rules": [
                    {
                      "condition": "B",
                      "data": [
                        {
                          "datasourceUid": "grafanacloud-demoinfra-prom",
                          "model": {
                            "datasource": {
                              "type": "prometheus",
                              "uid": "grafanacloud-demoinfra-prom"
                            },
                            "editorMode": "code",
                            "expr": "weather_temp_c{}",
                            "instant": true,
                            "intervalMs": 1000,
                            "legendFormat": "__auto",
                            "maxDataPoints": 43200,
                            "range": false,
                            "refId": "A"
                          },
                          "refId": "A",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        },
                        {
                          "datasourceUid": "__expr__",
                          "model": {
                            "conditions": [
                              {
                                "evaluator": {
                                  "params": [
                                    0
                                  ],
                                  "type": "gt"
                                },
                                "operator": {
                                  "type": "and"
                                },
                                "query": {
                                  "params": [
                                    "C"
                                  ]
                                },
                                "reducer": {
                                  "params": [],
                                  "type": "last"
                                },
                                "type": "query"
                              }
                            ],
                            "datasource": {
                              "type": "__expr__",
                              "uid": "__expr__"
                            },
                            "expression": "A",
                            "intervalMs": 1000,
                            "maxDataPoints": 43200,
                            "refId": "B",
                            "type": "threshold"
                          },
                          "refId": "B",
                          "relativeTimeRange": {
                            "from": 600
                          }
                        }
                      ],
                      "execErrState": "Error",
                      "for": "5m0s",
                      "noDataState": "NoData",
                      "title": "Temperature below freezing",
                      "uid": "4843de5c-4f8a-4af0-9509-23526a04faf8"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaContactPoint",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafanacontactpoint-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafanacontactpoint",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafanacontactpoint-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "name": "grafanacontactpoint-sample",
                  "settings": {
                    "email": null
                  },
                  "type": "email"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDashboard",
                "metadata": {
                  "name": "grafanadashboard-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "json": "{\n\n  \"id\": null,\n  \"title\": \"Simple Dashboard\",\n  \"tags\": [],\n  \"style\": \"dark\",\n  \"timezone\": \"browser\",\n  \"editable\": true,\n  \"hideControls\": false,\n  \"graphTooltip\": 1,\n  \"panels\": [],\n  \"time\": {\n    \"from\": \"now-6h\",\n    \"to\": \"now\"\n  },\n  \"timepicker\": {\n    \"time_options\": [],\n    \"refresh_intervals\": []\n  },\n  \"templating\": {\n    \"list\": []\n  },\n  \"annotations\": {\n    \"list\": []\n  },\n  \"refresh\": \"5s\",\n  \"schemaVersion\": 17,\n  \"version\": 0,\n  \"links\": []\n}\n"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaDatasource",
                "metadata": {
                  "name": "grafanadatasource-sample"
                },
                "spec": {
                  "datasource": {
                    "access": "proxy",
                    "isDefault": true,
                    "jsonData": {
                      "timeInterval": "5s",
                      "tlsSkipVerify": true
                    },
                    "name": "prometheus",
                    "type": "prometheus",
                    "url": "http://prometheus-service:9090"
                  },
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "plugins": [
                    {
                      "name": "grafana-clock-panel",
                      "version": "1.3.0"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaFolder",
                "metadata": {
                  "name": "grafanafolder-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana-a"
                    }
                  },
                  "title": "Example Folder"
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaLibraryPanel",
                "metadata": {
                  "name": "grafana-library-panel-inline-envs"
                },
                "spec": {
                  "envs": [
                    {
                      "name": "CUSTOM_RANGE_ENV",
                      "value": "now - 12h"
                    }
                  ],
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "jsonnet": "local myRange = std.extVar('CUSTOM_RANGE_ENV'); {\n\n  model: {}\n}\n",
                  "plugins": [
                    {
                      "name": "grafana-piechart-panel",
                      "version": "1.3.9"
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaMuteTiming",
                "metadata": {
                  "name": "mutetiming-sample"
                },
                "spec": {
                  "editable": false,
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "mutetiming-sample",
                  "time_intervals": [
                    {
                      "days_of_month": [
                        "1",
                        "15"
                      ],
                      "location": "Asia/Shanghai",
                      "times": [
                        {
                          "end_time": "06:00",
                          "start_time": "00:00"
                        }
                      ],
                      "weekdays": [
                        "saturday"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicy",
                "metadata": {
                  "name": "grafananotificationpolicy-sample"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "route": {
                    "group_by": [
                      "grafana_folder",
                      "alertname"
                    ],
                    "receiver": "Grafana Cloud OnCall",
                    "routes": [
                      {
                        "object_matchers": [
                          [
                            "foo",
                            "=",
                            "bar"
                          ]
                        ],
                        "receiver": "grafana-default-email",
                        "routes": [
                          {
                            "object_matchers": [
                              [
                                "severity",
                                "=",
                                "critical"
                              ]
                            ],
                            "receiver": "Grafana Cloud OnCall"
                          }
                        ]
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationPolicyRoute",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "grafana-operator",
                    "app.kubernetes.io/instance": "grafananotificationpolicyroute-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "grafananotificationpolicyroute",
                    "app.kubernetes.io/part-of": "grafana-operator"
                  },
                  "name": "grafananotificationpolicyroute-sample"
                },
                "spec": null
              },
              {
                "apiVersion": "grafana.integreatly.org/v1beta1",
                "kind": "GrafanaNotificationTemplate",
                "metadata": {
                  "name": "test"
                },
                "spec": {
                  "instanceSelector": {
                    "matchLabels": {
                      "dashboards": "grafana"
                    }
                  },
                  "name": "test",
                  "template": "{{ define \"SlackAlert\" }}\n  [{{.Status}}] {{ .Labels.alertname }}\n  {{ .Annotations.AlertValues }}\n{{ end }}\n\n{{ define \"SlackAlertMessage\" }}\n  {{ if gt (len .Alerts.Firing) 0 }}\n    {{ len .Alerts.Firing }} firing:\n    {{ range .Alerts.Firing }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n  {{ if gt (len .Alerts.Resolved) 0 }}\n    {{ len .Alerts.Resolved }} resolved:\n    {{ range .Alerts.Resolved }} {{ template \"SlackAlert\" . }} {{ end }}\n  {{ end }}\n{{ end }}\n\n{{ template \"SlackAlertMessage\" . }}\n"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          containerImage: ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
          createdAt: "2025-08-21T08:38:00Z"
          description: Deploys and manages Grafana instances, dashboards and data
            sources
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/grafana/grafana-operator
          support: Grafana Labs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GrafanaAlertRuleGroup is the Schema for the grafanaalertrulegroups
              API
            displayName: Grafana Alert Rule Group
            kind: GrafanaAlertRuleGroup
            name: grafanaalertrulegroups.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaContactPoint is the Schema for the grafanacontactpoints
              API
            displayName: Grafana Contact Point
            kind: GrafanaContactPoint
            name: grafanacontactpoints.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDashboard is the Schema for the grafanadashboards
              API
            displayName: Grafana Dashboard
            kind: GrafanaDashboard
            name: grafanadashboards.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaDatasource is the Schema for the grafanadatasources
              API
            displayName: Grafana Datasource
            kind: GrafanaDatasource
            name: grafanadatasources.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaFolder is the Schema for the grafanafolders API
            displayName: Grafana Folder
            kind: GrafanaFolder
            name: grafanafolders.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaLibraryPanel
            name: grafanalibrarypanels.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaMuteTiming
            name: grafanamutetimings.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicy is the Schema for the GrafanaNotificationPolicy
              API
            displayName: Grafana Notification policy
            kind: GrafanaNotificationPolicy
            name: grafananotificationpolicies.grafana.integreatly.org
            version: v1beta1
          - description: GrafanaNotificationPolicyRoute is the Schema for the grafananotificationpolicyroutes
              API
            displayName: Grafana Notification Policy Route
            kind: GrafanaNotificationPolicyRoute
            name: grafananotificationpolicyroutes.grafana.integreatly.org
            version: v1beta1
          - kind: GrafanaNotificationTemplate
            name: grafananotificationtemplates.grafana.integreatly.org
            version: v1beta1
          - description: Grafana is the Schema for the grafanas API
            displayName: Grafana
            kind: Grafana
            name: grafanas.grafana.integreatly.org
            version: v1beta1
        description: Deploys and manages Grafana instances, dashboards and data sources
        displayName: Grafana Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Grafana
        - Metrics
        - Observability
        links:
        - name: Grafana Operator
          url: https://grafana.github.io/grafana-operator
        maintainers:
        - email: grafana-operator@grafana.com
          name: Grafana-operator maintainers
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Grafana Labs
          url: https://grafana.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/grafana-operator@sha256:ae9aa1d05ab9dadcc9961ca4293d0878fd7656be79c869c7fdbee718f259ad86
        - docker.io/grafana/grafana@sha256:6ac590e7cabc2fbe8d7b8fc1ce9c9f0582177b334e0df9c927ebd9670469440f
        - ghcr.io/grafana/grafana-operator@sha256:b678bc3faad11f4ef53c240e2a650ce939b9e8920dd72a47681b2efcadc7b837
        version: 5.19.4
      entries:
      - name: grafana-operator.v5.19.4
        version: 5.19.4
      - name: grafana-operator.v5.18.0
        version: 5.18.0
      - name: grafana-operator.v5.17.1
        version: 5.17.1
      - name: grafana-operator.v5.15.1
        version: 5.15.1
      name: v5
    defaultChannel: v5
    packageName: grafana-operator
    provider:
      name: Grafana Labs
      url: https://grafana.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: orchestrator-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: orchestrator-operator.v1.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhdh.redhat.com/v1alpha3",
                "kind": "Orchestrator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "orchestrator-sample"
                  },
                  "name": "orchestrator-sample"
                },
                "spec": {
                  "argocd": {
                    "enabled": false,
                    "namespace": "orchestrator-gitops"
                  },
                  "platform": {
                    "eventing": {
                      "broker": {}
                    },
                    "monitoring": {
                      "enabled": false
                    },
                    "namespace": "sonataflow-infra",
                    "resources": {
                      "limits": {
                        "cpu": "500m",
                        "memory": "1Gi"
                      },
                      "requests": {
                        "cpu": "250m",
                        "memory": "64Mi"
                      }
                    }
                  },
                  "postgres": {
                    "authSecret": {
                      "name": "sonataflow-psql-postgresql",
                      "passwordKey": "postgres-password",
                      "userKey": "postgres-username"
                    },
                    "database": "sonataflow",
                    "name": "sonataflow-psql-postgresql",
                    "namespace": "sonataflow-infra"
                  },
                  "rhdh": {
                    "devMode": true,
                    "installOperator": true,
                    "name": "my-rhdh",
                    "namespace": "rhdh",
                    "plugins": {
                      "notificationsEmail": {
                        "enabled": false,
                        "port": 587,
                        "replyTo": "",
                        "sender": ""
                      }
                    }
                  },
                  "serverless": {
                    "installOperator": true
                  },
                  "serverlessLogic": {
                    "installOperator": true
                  },
                  "tekton": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-08-15T00:51:04Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Developer Hub"]'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/rhdhorchestrator/orchestrator-go-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Orchestrator
            name: orchestrators.rhdh.redhat.com
            version: v1alpha3
        description: |
          Red Hat Developer Hub Orchestrator is a plugin that enables serverless asynchronous workflows to Backstage.

          ## More Information

          * [Red Hat Developer Hub Product Page](https://www.redhat.com/en/technologies/cloud-computing/developer-hub)
          * [Product Documentation](https://rhdhorchestrator.io)  displayName: Orchestrator Operator
        displayName: Orchestrator Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Orchestrator
        - RHDH
        links:
        - name: Orchestrator Operator
          url: https://github.com/rhdhorchestrator/orchestrator-go-operator
        maintainers:
        - email: jubah@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/controller-rhel9-operator@sha256:9f5423829ac55faf2395e2d0d3683b238339dc03285696e1b350c5c5a1c4df3a
        - registry.redhat.io/rhdh-orchestrator-dev-preview-beta/orchestrator-operator-bundle@sha256:b82fd43d2dad2e525464177081869c923e239dcb09319458162f33a96370a6e6
        version: 1.6.1
      entries:
      - name: orchestrator-operator.v1.6.1
        version: 1.6.1
      - name: orchestrator-operator.v1.6.0
        version: 1.6.0
      - name: orchestrator-operator.v1.5.2
        version: 1.5.2
      - name: orchestrator-operator.v1.5.1
        version: 1.5.1
      - name: orchestrator-operator.v1.5.0
        version: 1.5.0
      name: stable
    defaultChannel: stable
    packageName: orchestrator-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Dynatrace LLC
      provider-url: www.dynatrace.com
    name: dynatrace-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dynatrace-operator-rhmp.v1.7.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Deep Insights
          categories: Monitoring,Logging & Tracing
          containerImage: registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
          createdAt: 2025-09-22T09:15:43.437654
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/dynatrace-operator-rhmp/support?utm_source=openshift_console
          olm.skipRange: <1.7.1
          operatorframework.io/suggested-namespace: dynatrace
          operators.openshift.io/valid-subscription: '["Dynatrace Platform Subscription
            (DPS)","Dynatrace Classic License"]'
          operators.operatorframework.io/builder: operator-sdk-v1.36.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Dynatrace/dynatrace-operator
          support: Dynatrace
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta3
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta4
          - kind: DynaKube
            name: dynakubes.dynatrace.com
            version: v1beta5
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha1
          - kind: EdgeConnect
            name: edgeconnects.dynatrace.com
            version: v1alpha2
        description: |
          The Dynatrace Operator supports rollout and lifecycle management of various Dynatrace components in Kubernetes and OpenShift.

          Currently the Dynatrace Operator supports the following capabilities:

          ### OneAgent
            * `classicFullStack` rolls out one OneAgent pod per node to monitor its pods and the node itself
            * `applicationMonitoring` is a webhook based injection mechanism for automatic app-only injection
            * `hostMonitoring` monitors only the hosts, i.e., the nodes, in the cluster without app-only injection
          ### ActiveGate
            * `routing` routes OneAgent traffic through the ActiveGate
            * `kubernetes-monitoring` allows monitoring the Kubernetes API
            * `metrics-ingest` routes enriched metrics through an ActiveGate

          For more information please refer to [our DynaKube Custom Resource examples](https://dt-url.net/dynakube-samples).

          ### Installation
          Once you've installed the Dynatrace Operator, you can create a DynaKube custom resource.

          First, please add a Secret within the Project you've deployed the Dynatrace Operator to, which would contain your API and PaaS tokens. Create tokens of type *Dynatrace API* (`API_TOKEN`) and *Platform as a Service* (`PAAS_TOKEN`) and use their values in the following commands respectively.

          For assistance please refer to [Create user-generated access tokens](https://www.dynatrace.com/support/help/shortlink/token#create-user-generated-access-tokens).

          ``` $ oc -n <project> create secret generic dynakube --from-literal="apiToken=API_TOKEN" --from-literal="paasToken=PAAS_TOKEN" ```

          You may update this Secret at any time to rotate the tokens.

          After creation of the secret add the DynaKube object in the project where the Dynatrace Operator has been deployed, configured to your needs.

          ### Required Parameters
          * `apiUrl` - provide the URL to the API of your Dynatrace environment. In Dynatrace SaaS it will look like `https://<ENVIRONMENTID>.live.dynatrace.com/api` . In Dynatrace Managed like `https://<YourDynatraceServerURL>/e/<ENVIRONMENTID>/api` .

          ### Advanced Options
          * **Disable Certificate Checking** - disable any certificate validation that may interact poorly with proxies with in your cluster
          * **Image Override** - use a copy of the ActiveGate container image from a registry other than Docker's or Red Hat's

          For a complete list of supported parameters please consult the [Operator Deploy Guide](https://www.dynatrace.com/support/help/shortlink/openshift-deploy).

          ### Help
          You can find more about our instructions in our [documentation](https://www.dynatrace.com/support/help/shortlink/openshift-deploy#install-dynatrace-operator).
        displayName: Dynatrace Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - dynatrace
        - oneagent
        - activegate
        - observability
        links:
        - name: Dynatrace Operator
          url: https://github.com/Dynatrace/dynatrace-operator
        - name: OpenShift Monitoring Info
          url: https://www.dynatrace.com/technologies/openshift-monitoring
        maintainers:
        - email: support@dynatrace.com
          name: Dynatrace LLC
        maturity: alpha
        provider:
          name: Dynatrace LLC
          url: www.dynatrace.com
        relatedImages:
        - registry.connect.redhat.com/dynatrace/dynatrace-operator-bundle-marketplace@sha256:b7f67aec4e911d89b3e6217e8718c349c20a859149b829a0390807c113e27196
        - registry.connect.redhat.com/dynatrace/dynatrace-operator@sha256:dfe8db1706e1f709892cf53b02328f05612e6e18a38110c7835bf347ceae2e62
        version: 1.7.1
      entries:
      - name: dynatrace-operator-rhmp.v1.7.1
        version: 1.7.1
      - name: dynatrace-operator-rhmp.v1.7.0
        version: 1.7.0
      - name: dynatrace-operator-rhmp.v1.6.2
        version: 1.6.2
      - name: dynatrace-operator-rhmp.v1.6.1
        version: 1.6.1
      - name: dynatrace-operator-rhmp.v1.6.0
        version: 1.6.0
      - name: dynatrace-operator-rhmp.v1.5.1
        version: 1.5.1
      - name: dynatrace-operator-rhmp.v1.5.0
        version: 1.5.0
      - name: dynatrace-operator-rhmp.v1.4.2
        version: 1.4.2
      - name: dynatrace-operator-rhmp.v1.4.1
        version: 1.4.1
      - name: dynatrace-operator-rhmp.v1.4.0
        version: 1.4.0
      name: alpha
    defaultChannel: alpha
    packageName: dynatrace-operator-rhmp
    provider:
      name: Dynatrace LLC
      url: www.dynatrace.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: LINBIT
      provider-url: ""
    name: linstor-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: alpha
    - currentCSV: linstor-operator.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorController",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerImage": "",
                  "dbConnectionURL": "etcd://linstor-etcd:2379",
                  "drbdRepoCred": "",
                  "priorityClassName": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorCSIDriver",
                "metadata": {
                  "name": "linstor"
                },
                "spec": {
                  "controllerReplicas": 1,
                  "controllerEndpoint": "http://linstor:3370",
                  "csiControllerServiceAccountName": "csi-controller",
                  "csiNodeServiceAccountName": "csi-node",
                  "linstorPluginImage": "",
                  "imagePullSecret": ""
                },
                "status": {
                  "errors": []
                }
              },
              {
                "apiVersion": "linstor.linbit.com/v1",
                "kind": "LinstorSatelliteSet",
                "metadata": {
                  "name": "linstor-satellites"
                },
                "spec": {
                  "automaticStorageType": "None",
                  "controllerEndpoint": "http://linstor:3370",
                  "drbdRepoCred": "",
                  "kernelModuleInjectionMode": "ShippedModules",
                  "priorityClassName": "",
                  "satelliteImage": ""
                },
                "status": {
                  "errors": []
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          createdAt: "2021-06-22T07:57:53+00:00"
          description: LINSTOR Kubernetes Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/linstor-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/piraeusdatastore/piraeus-operator
          support: LINBIT
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: LinstorController is the Schema for the linstorcontrollersets
              API
            kind: LinstorController
            name: linstorcontrollers.linstor.linbit.com
            version: v1
          - description: LinstorCSIDriver is the Schema for the linstorcsidrivers
              API
            kind: LinstorCSIDriver
            name: linstorcsidrivers.linstor.linbit.com
            version: v1
          - description: LinstorSatelliteSet is the Schema for the linstornodesets
              API
            kind: LinstorSatelliteSet
            name: linstorsatellitesets.linstor.linbit.com
            version: v1
        description: |-
          LINSTOR is a configuration management system for storage on Linux systems.
          It manages LVM logical volumes and/or ZFS ZVOLs on a cluster of nodes.
          It leverages DRBD for replication between different nodes and to provide block
          storage devices to users and applications. It manages snapshots, encryption and
          caching of HDD backed data in SSDs via bcache.

          LINBIT provides a certified LINSTOR operator to ease deployment of LINSTOR
          on Openshift by installing DRBD, managing Satellite and Controller pods,
          and integrating with Openshift to provision persistent storage for your workloads.

          For detailed instructions and more configuration options, see our [user guide].

          [user guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift

          ## Install

          Unlike deployment via the helm chart, the certified Openshift
          operator does not deploy the needed etcd cluster. You must deploy this
          yourself ahead of time. We do this via the etcd operator available in the
          OperatorHub.

          IMPORTANT: It is advised that the etcd deployment uses persistent
          storage of some type. Either use an existing storage provisioner with
          a default `StorageClass` or simply use `hostPath` volumes.

          ### Installing the operator

          Hit "Install", select the stable update channel and a namespace for the
          operator. Use of a new namespace is recommended.

          Hit "Install" again. At this point you should have just one pod, the
          operator pod, running. Next we needs to configure the remaining provided APIs.

          #### A note on operator namespaces
          The LINSTOR operator can only watch for events and manage
          custom resources that are within the same namespace it is deployed
          within (OwnNamsespace). This means the LINSTOR Controller, LINSTOR
          Satellites, and LINSTOR CSI Driver pods all need to be deployed in the
          same namsepace as the LINSTOR Operator pod.

          ### Deploying the LINSTOR Controller

          Navigate to the left-hand control pane of the Openshift Web
          Console. Expand the "Operators" section, selecting "Installed Operators".
          Find the entry for the "Linstor Operator", then select the "LinstorController"
          from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorController". Click the "Create LinstorController" button.

          Here you will be presented with options to configure the LINSTOR
          Controller. Either via the web-form view or the YAML View. Regardless
          of which view you select, make sure that the `dbConnectionURL` matches
          the endpoint provided from your etcd deployment. Otherwise, the
          defaults are usually fine for most purposes.

          Lastly hit "Create", you should now see a linstor-controller pod
          running.

          ### Deploying the LINSTOR Satellites

          Next we need to deploy the Satellites Set. Just as before navigate
          to the left-hand control pane of the Openshift Web Console. Expand the
          "Operators" section, but this time select "Installed Operators". Find
          the entry for the "Linstor Operator", then select the
          "LinstorSatelliteSet" from the "Provided APIs" column on the right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorSatelliteSet". Click the "Create LinstorSatelliteSet" button.

          Here you will be presented with the options to configure the LINSTOR
          Satellites. The defaults should be enough to get you started.
          Make sure the `controllerEndpoint` matches what is available in the
          openshift endpoints. The default is usually correct here.

          You can edit the `storagePools` section to configure LINSTOR storage pools,
          including preparing the backing devices. See our [storage guide].

          [storage guide]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-storage

          Below is an example manifest:

          ```yaml
          apiVersion: linstor.linbit.com/v1
          kind: LinstorSatelliteSet
          metadata:
            name: linstor-satellites
          spec:
            satelliteImage: ''
            automaticStorageType: None
            storagePools:
              lvmThinPools:
              - name: openshift-pool
                volumeGroup: ""
                thinVolume: openshift
                devicePaths:
                - /dev/vdb
            drbdRepoCred: ''
            kernelModuleInjectionMode: ShippedModules
            controllerEndpoint: 'http://linstor:3370'
            priorityClassName: ''
          ```

          Lastly hit "Create", you should now see a linstor-node pod
          running on every worker node.

          ### Deploying the LINSTOR CSI driver

          Last bit left is the CSI pods to bridge the layer between the CSI and
          LINSTOR. Just as before navigate to the left-hand control pane of the
          Openshift Web Console. Expand the "Operators" section, but this time
          select "Installed Operators". Find the entry for the "Linstor Operator",
          then select the "LinstorCSIDriver" from the "Provided APIs" column on the
          right.

          From here you should see a page that says "No Operands Found" and will
          feature a large button on the right which says "Create
          LinstorCSIDriver". Click the "Create LinstorCSIDriver" button.

          Again, you will be presented with the options. Make sure that the
          `controllerEndpoint` is correct. Otherwise the defaults are fine for
          most use cases.

          Lastly hit "Create". You will now see a single "linstor-csi-controller" pod,
          as well as a "linstor-csi-node" pod on all worker nodes.

          ## Interacting with LINSTOR in Openshift.

          The Controller pod includes a LINSTOR Client, making it easy to interact directly with LINSTOR.
          For instance:

          ```
          oc exec deployment/linstor-cs-controller -- linstor storage-pool list
          ```

          This should only be necessary for investigating problems and accessing advanced functionality.
          Regular operation such as creating volumes should be achieved via the [Openshift/Kubernetes integration].

          [Openshift/Kubernetes integration]: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#s-kubernetes-basic-configuration-and-deployment
        displayName: Linstor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Storage
        - DRBD
        links:
        - name: LINBIT
          url: https://www.linbit.com/
        - name: User Guide
          url: https://www.linbit.com/drbd-user-guide/linstor-guide-1_0-en/#ch-openshift
        maintainers:
        - email: sales@linbit.com
          name: LINBIT
        maturity: beta
        provider:
          name: LINBIT
        relatedImages:
        - registry.marketplace.redhat.com/rhm/linbit/drbd-9@sha256:6337b3adfa3de87e74adcc974edafa09b618cb8e0315658c1cbdf8dd1c6c4c32
        - registry.marketplace.redhat.com/rhm/linbit/linstor-operator@sha256:75e3ec9a9331072dc88c13979f0cacdf6232f8b3a6bfaf55c95b2ee3a03a5d5b
        - registry.redhat.io/openshift4/ose-csi-external-resizer@sha256:ba217e7e51f0748b54f477486263ff4983b0029635fd1f7ccc67160a09f81608
        - registry.redhat.io/openshift4/ose-csi-external-snapshotter@sha256:bba872052e42fe0d6cbb2eea6541bf57f035575cbc7d838a3260b8582b8fe572
        - registry.redhat.io/openshift4/ose-csi-livenessprobe@sha256:8e52a6b31f9f3834d709dbca3614c2110b8ebb4af382c76b7056921221c87600
        - registry.marketplace.redhat.com/rhm/linbit/linstor-controller@sha256:1084ba14eaec7511682dd87c562398ce5e6690ded700c964373188b73bfe3b17
        - registry.marketplace.redhat.com/rhm/linbit/linstor-satellite@sha256:4a3503092c8cb9435ccfb4837ee6a9183b680eb18a35c7d40227fa5c8f4da996
        - registry.connect.redhat.com/linbit/linstor-operator-bundle@sha256:2a41233da53d1ba9812c5ada5d1a2ed606f76a8473556f38cb8f0fd0bc6d0226
        - registry.marketplace.redhat.com/rhm/linbit/linstor-csi@sha256:034e3641376aa7ad47ed309fb389083c3ff82974f246f4d71b8789be8f8f618d
        - registry.redhat.io/openshift4/ose-csi-external-attacher@sha256:53b0f48eeecaa3ff1e211a2a413cd49266b7be85a5aa3a0431ab72d32e3a4e5c
        - registry.redhat.io/openshift4/ose-csi-node-driver-registrar@sha256:91cfe63acc12c91c3053ba7cc12d72cce404d33aab4afcfac285311f5ddfc7fe
        - registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:395a5a4aa4cfe3a0093d2225ce2e67acdcec0fd894e4b61e30a750f22931448d
        version: 1.5.1
      entries:
      - name: linstor-operator.v1.5.1
        version: 1.5.1
      name: stable
    defaultChannel: stable
    packageName: linstor-operator-rhmp
    provider:
      name: LINBIT
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kedify
      provider-url: kedify.io
    name: kedify-agent-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kedify-agent-helm-operator.v0.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.kedify.io/v1alpha1",
                "kind": "KedifyAgent",
                "metadata": {
                  "name": "kedifyagent-sample"
                },
                "spec": {
                  "agent": {
                    "agentId": "",
                    "apiKey": "kfy_xxxxxxxxxxxxxxxxxxxxx",
                    "createCrds": true,
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a"
                    },
                    "kedifyProxy": {
                      "globalValues": {
                        "openshift": true
                      },
                      "image": {
                        "tag": "v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e"
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "orgId": "00000000-0000-0000-0000-000000000000",
                    "securityContext": {
                      "readOnlyRootFilesystem": true
                    }
                  },
                  "keda": {
                    "enabled": true,
                    "image": {
                      "keda": {
                        "tag": "v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52"
                      },
                      "metricsApiServer": {
                        "tag": "v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389"
                      },
                      "pullPolicy": "IfNotPresent",
                      "webhooks": {
                        "tag": "v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41"
                      }
                    }
                  },
                  "keda-add-ons-http": {
                    "enabled": true,
                    "interceptor": {
                      "pullPolicy": "IfNotPresent",
                      "scaledObject": {
                        "waitForCrd": true
                      }
                    },
                    "kubectlImage": {
                      "pullPolicy": "IfNotPresent",
                      "tag": "v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3"
                    },
                    "pinnedImages": {
                      "interceptor": {
                        "tag": "v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb"
                      },
                      "scaler": {
                        "tag": "v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2"
                      }
                    },
                    "scaler": {
                      "pullPolicy": "IfNotPresent"
                    }
                  }
                }
              }
            ]
          alm-examples-metadata: '{"kedify-openshift-default":{"description":"Default
            Kedify Agent CR for OpenShift. Register at https://dashboard.kedify.io
            to get started."}}'
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
          createdAt: "2025-08-07T14:04:11Z"
          description: Kedify Autoscaler, powered by proven open source technologies
            and KEDA, is an enterprise-grade service that allows teams to autoscale
            any workload using various event sources including HTTP events
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: keda
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: http://github.com/kedify/kedify-agent-helm-operator
          support: support@kedify.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: KedifyAgent
            name: kedifyagents.charts.kedify.io
            version: v1alpha1
        description: Kedify Autoscaler, powered by proven open source technologies
          and KEDA, is an enterprise-grade service that allows teams to autoscale
          any workload using various event sources including HTTP events
        displayName: kedify-agent-helm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - autoscaling
        - keda
        - kedify
        links:
        - name: Kedify Homepage
          url: https://kedify.io
        maintainers:
        - email: support@kedify.io
          name: Kedify
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Kedify
          url: kedify.io
        relatedImages:
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:720b90a4f112a91a94b1bdafe445df1681657ddaa4fb20aa6d2904a0eb664a2a
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:d61332d50587ce2c75aff29033fdd5cb1bf0ec49d199ab72323a27d9d19ff402
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:9300a5382373b179f7f541a663cd30e50284e54f352993e8cca34460acbd890d
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:4b63409c022ba137760e7a534b6308fc1e3639f06fdb53ee1f6df32787b68400
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:2ea25099363a3f16a491fc2c5a30fff304a8bbd939c782f0f80b399e821084c3
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:ea6b22bb12c8ff7cbab4633c2f4a0e49eaf3883b07ab1f47dbebff2a3e07f029
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:73e180bd5793dec3785fcb9ecd7e7184e645f7c37d561a5728970e5303c19a41
        - registry.connect.redhat.com/kedify/kedify-agent-helm-operator@sha256:d176136a44e790298158ea3c4232e2aa7d0ad88ff18b6a5b39238162ed41226a
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:c1c093a6fda7efc9bc30f110c55f8bb075250db929dbdb2534b9071efbfc3169
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:fc91d1f7f8b632614f31dea14c0c8d4a5f8311a83fdef82303857aa8b43b5782
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:8e3a458912ec6540cb414446425ad7b1546528f04a5d84fcd64d8ee5424392e6
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:4b7277ad53dfb01bf5f2201dd31d94f24519da00d008638c3c36026b96ea78b4
        - ghcr.io/kedify/agent:v0.2.16@sha256:9a57d5eafe368aab1d94f0eec3101e02f0ff2e356d8499984030ab89161e42ae
        - ghcr.io/kedify/kubectl:v1.33.1@sha256:850922994153104a8e382416561c251b46a95ff2b2a216aacbd7fcb4be7f6382
        - envoyproxy/envoy:v1.34.1@sha256:683e2f32809657c8d4a1508648545481130ac5409cd1ce63ccc795ad8c71202e
        - ghcr.io/kedify/kedify-agent-helm-operator:v0.2.16@sha256:a38fdaf3a49eb4c71c4f3ab1211befa22fa5de56391f88efc11335f0bb3b1d29
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:33cd5a8c6dd5e90614341dc1863e1f041f2c2475bd61edf5176a2a0d98ea54f9
        - ghcr.io/kedify/agent:v0.2.16@sha256:c59d765091a31659236b8d80f4033912f5cf47d86c2e2da96bcfa133b9f6fe25
        - ghcr.io/kedify/keda-operator:v2.17.2-0@sha256:802592b691a7b81a249055fa117994927c2fd376b5023a3888bcc7973c65ba52
        - ghcr.io/kedify/agent:v0.2.16@sha256:bfcc0b24ff75de17464ec213406f48bc211a932308e888c450ed23ee1786d37a
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:82b041027dc900fdab717e8db6a522139067801c04cd1ebfe950572e714938ab
        - ghcr.io/kedify/http-add-on-interceptor:v0.10.0-17@sha256:30ccf20e79c6a62c14781872a73ad0a811063add7136b77df986bf403d6ca6bb
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:5863ee0ca09efeca225966c2eb8781f90f2dfb4aed872d84a6075b711ab042b2
        - ghcr.io/kedify/keda-metrics-apiserver:v2.17.2-0@sha256:df44bc1de75324158017d623a9d0adf3e3e9f1d64db6afa7133f777adc7b6389
        - envoyproxy/envoy:v1.34.1@sha256:007da57c2c328a90bd4e6d99b70bd899132f1b4a9426ccafe25437cf84a60c14
        - ghcr.io/kedify/keda-admission-webhooks:v2.17.2-0@sha256:cb40290842976fd09f2008d8e2e64ebf90a215f0584656d4886122d322006873
        - ghcr.io/kedify/http-add-on-scaler:v0.10.0-17@sha256:9fcc76f983035eb0ceff7beb941bf32fe7fc28b341ec1aac5191ffc1471f401b
        - envoyproxy/envoy:v1.34.1@sha256:75545b91d050aa51b52cb0673be172f2f263c971dae3d2a82a70d128d92e87c2
        version: 0.2.16
      entries:
      - name: kedify-agent-helm-operator.v0.2.16
        version: 0.2.16
      name: stable
    defaultChannel: stable
    packageName: kedify-agent-helm-operator
    provider:
      name: Kedify
      url: kedify.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CodeFlare
      provider-url: https://github.com/project-codeflare
    name: codeflare-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: codeflare-operator.v1.16.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: AI/Machine Learning, Big Data
          containerImage: quay.io/project-codeflare/codeflare-operator:v1.16.0
          createdAt: "2025-06-16T10:53:51Z"
          operatorframework.io/suggested-namespace: openshift-operators
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/project-codeflare/codeflare-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AppWrapper
            name: appwrappers.workload.codeflare.dev
            version: v1beta2
        description: CodeFlare allows you to scale complex pipelines anywhere
        displayName: CodeFlare Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Pipelines
        - Scaling
        - App
        - Jobs
        links:
        - name: Codeflare Operator
          url: https://codeflare.dev/
        maintainers:
        - email: aasthana@redhat.com
          name: Anish Asthana
        - email: kpostlet@redhat.com
          name: Kevin Postlethwait
        - email: meyceoz@redhat.com
          name: Mustafa Eyceoz
        maturity: alpha
        provider:
          name: CodeFlare
          url: https://github.com/project-codeflare
        relatedImages:
        - quay.io/community-operator-pipeline-prod/codeflare-operator@sha256:3dbd2b65d7841cc97c37db59ae8aff9991c4d0f0324c4391afd41c4de20c232a
        - quay.io/project-codeflare/codeflare-operator:v1.16.0
        version: 1.16.0
      entries:
      - name: codeflare-operator.v1.16.0
        version: 1.16.0
      - name: codeflare-operator.v1.15.0
        version: 1.15.0
      - name: codeflare-operator.v1.14.0
        version: 1.14.0
      - name: codeflare-operator.v1.13.0
        version: 1.13.0
      - name: codeflare-operator.v1.12.0
        version: 1.12.0
      - name: codeflare-operator.v1.11.0
        version: 1.11.0
      - name: codeflare-operator.v1.10.0
        version: 1.10.0
      - name: codeflare-operator.v1.9.0
        version: 1.9.0
      - name: codeflare-operator.v1.8.0
        version: 1.8.0
      - name: codeflare-operator.v1.7.0
        version: 1.7.0
      - name: codeflare-operator.v1.6.0
        version: 1.6.0
      - name: codeflare-operator.v1.5.0
        version: 1.5.0
      - name: codeflare-operator.v1.4.4
        version: 1.4.4
      - name: codeflare-operator.v1.4.3
        version: 1.4.3
      - name: codeflare-operator.v1.4.2
        version: 1.4.2
      - name: codeflare-operator.v1.4.1
        version: 1.4.1
      - name: codeflare-operator.v1.4.0
        version: 1.4.0
      - name: codeflare-operator.v1.3.1
        version: 1.3.1
      - name: codeflare-operator.v1.3.0
        version: 1.3.0
      - name: codeflare-operator.v1.2.0
        version: 1.2.0
      - name: codeflare-operator.v1.1.0
        version: 1.1.0
      - name: codeflare-operator.v1.0.1
        version: 1.0.1
      - name: codeflare-operator.v1.0.0
        version: 1.0.0
      - name: codeflare-operator.v1.0.0-rc.4
        version: 1.0.0-rc.4
      - name: codeflare-operator.v1.0.0-rc.3
        version: 1.0.0-rc.3
      - name: codeflare-operator.v1.0.0-rc.2
        version: 1.0.0-rc.2
      - name: codeflare-operator.v1.0.0-rc.1
        version: 1.0.0-rc.1
      - name: codeflare-operator.v0.2.3
        version: 0.2.3
      - name: codeflare-operator.v0.2.2
        version: 0.2.2
      - name: codeflare-operator.v0.2.1
        version: 0.2.1
      - name: codeflare-operator.v0.2.0
        version: 0.2.0
      - name: codeflare-operator.v0.1.0
        version: 0.1.0
      - name: codeflare-operator.v0.0.6
        version: 0.0.6
      - name: codeflare-operator.v0.0.5
        version: 0.0.5
      - name: codeflare-operator.v0.0.4
        version: 0.0.4
      - name: codeflare-operator.v0.0.3
        version: 0.0.3
      - name: codeflare-operator.v0.0.2
        version: 0.0.2
      - name: codeflare-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: codeflare-operator
    provider:
      name: CodeFlare
      url: https://github.com/project-codeflare
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Trustification
      provider-url: ""
    name: trustify-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: trustify-operator.v0.1.0-alpha.9
      currentCSVDesc:
        annotations:
          alm-examples: |
            [{
              "apiVersion": "org.trustify/v1alpha1",
              "kind": "Trustify",
              "metadata": {
                "name": "myapp"
              },
              "spec": { }
            }]
          capabilities: Basic Install
          categories: Application Runtime
          certified: "false"
          containerImage: ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
          createdAt: "2025-03-07T11:57:13+00:00"
          description: An Operator for installing and managing Trustify
          repository: https://github.com/trustification/trustify-operator
          support: https://github.com/trustification/trustify-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Trustify instance
            displayName: Trustify
            kind: Trustify
            name: trustifies.org.trustify
            version: v1alpha1
        description: |+
          Trustify is vendor-neutral, thought-leadering, mostly informational collection of resources devoted to making Software Supply Chains easier to create, manage, consume and ultimately... to trust!


          You can use Trustify for:

          - Store and Manage all the SBOM (Software Bill of Materials) files of your company.
          - Understand which are the Vulnerabilities that affect each of your SBOM files
          - Understand exactly which Packages are included/shipped within each SBOM

          Trustify is a project within the [Trustification community](https://trustification.io/).


          ### Install
          Once you have successfully installed the Operator, proceed to deploy components by creating the required CR.

          By default, the Operator installs the following components on a target cluster:

          * Server
          * UI

          ### Documentation
          Documentation can be found on our [website](https://trustification.io/).

          ### Getting help
          If you encounter any issues while using Trustify, you can create an issue on our [Github repo](https://github.com/trustification/trustify/issues), for bugs, enhancements or other requests.

          ### Contributing
          You can contribute by:

          * Raising any issues you find
          * Fixing issues by opening [Pull Requests](https://github.com/trustification/trustify/pulls)

        displayName: Trustify Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - trust
        links:
        - name: Website
          url: https://trustification.io/
        - name: Github
          url: https://github.com/trustification/trustify
        maintainers:
        - email: trustification@googlegroups.com
          name: Trustification
        maturity: alpha
        minKubeVersion: 1.23.0
        nativeApis:
        - group: ""
          kind: ConfigMap
          version: v1
        - group: ""
          kind: PersistentVolumeClaim
          version: v1
        - group: ""
          kind: Secret
          version: v1
        - group: ""
          kind: Service
          version: v1
        - group: apps
          kind: Deployment
          version: v1
        - group: apps
          kind: StatefulSet
          version: v1
        - group: networking.k8s.io
          kind: Ingress
          version: v1
        provider:
          name: Trustification
        relatedImages:
        - ghcr.io/trustification/trustify-operator:v0.1.0-alpha.9
        - quay.io/community-operator-pipeline-prod/trustify-operator@sha256:49def1c9612621f90d2398cbe7385488c167cc8ccdcd951f098433eb42cb9ac5
        version: 0.1.0-alpha.9
      entries:
      - name: trustify-operator.v0.1.0-alpha.9
        version: 0.1.0-alpha.9
      - name: trustify-operator.v0.1.0-alpha.8
        version: 0.1.0-alpha.8
      - name: trustify-operator.v0.1.0-alpha.7
        version: 0.1.0-alpha.7
      - name: trustify-operator.v0.1.0-alpha.6
        version: 0.1.0-alpha.6
      - name: trustify-operator.v0.1.0-alpha.5
        version: 0.1.0-alpha.5
      - name: trustify-operator.v0.1.0-alpha.4
        version: 0.1.0-alpha.4
      - name: trustify-operator.v0.1.0-alpha.3
        version: 0.1.0-alpha.3
      - name: trustify-operator.v0.1.0-alpha.2
        version: 0.1.0-alpha.2
      name: alpha
    defaultChannel: alpha
    packageName: trustify-operator
    provider:
      name: Trustification
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: PingCAP
      provider-url: ""
    name: tidb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: tidb-operator.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pingcap.com/v1alpha1",
                "kind": "TidbCluster",
                "metadata": {
                  "name": "basic"
                },
                "spec": {
                  "version": "v5.4.0",
                  "timezone": "UTC",
                  "pvReclaimPolicy": "Delete",
                  "discovery": {},
                  "pd": {
                    "baseImage": "pingcap/pd",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {}
                  },
                  "tikv": {
                    "baseImage": "pingcap/tikv",
                    "replicas": 1,
                    "requests": {
                      "storage": "1Gi"
                    },
                    "config": {
                      "storage": {
                        "reserve-space": "0MB"
                      }
                    }
                  },
                  "tidb": {
                    "baseImage": "pingcap/tidb",
                    "replicas": 1,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "config": {}
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "false"
          containerImage: pingcap/tidb-operator:v1.3.1
          createdAt: 2021-2-15T07-18-39Z
          description: TiDB Operator manages TiDB clusters on Kubernetes and automates
            tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native
            database.
          repository: https://github.com/pingcap/tidb-operator
          support: KanShiori
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup
            displayName: Backup
            kind: Backup
            name: backups.pingcap.com
            version: v1alpha1
          - description: Backup Schedule
            displayName: BackupSchedule
            kind: BackupSchedule
            name: backupschedules.pingcap.com
            version: v1alpha1
          - description: DM Cluster
            displayName: DMCluster
            kind: DMCluster
            name: dmclusters.pingcap.com
            version: v1alpha1
          - description: Restore
            displayName: Restore
            kind: Restore
            name: restores.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster Auto Scaler
            displayName: TidbClusterAutoScaler
            kind: TidbClusterAutoScaler
            name: tidbclusterautoscalers.pingcap.com
            version: v1alpha1
          - description: Tidb Cluster
            displayName: TidbCluster
            kind: TidbCluster
            name: tidbclusters.pingcap.com
            version: v1alpha1
          - description: Tidb Initializer
            displayName: TidbInitializer
            kind: TidbInitializer
            name: tidbinitializers.pingcap.com
            version: v1alpha1
          - description: Tidb Monitor
            displayName: TidbMonitor
            kind: TidbMonitor
            name: tidbmonitors.pingcap.com
            version: v1alpha1
          - description: Tidb NG Monitoring
            displayName: TidbNGMonitoring
            kind: TidbNGMonitoring
            name: tidbngmonitorings.pingcap.com
            version: v1alpha1
        description: |
          ## About the managed application
          TiDB ("Ti" stands for Titanium) is an open-source, distributed, NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloads. It is MySQL compatible and features horizontal scalability, strong consistency, and high availability. TiDB can be deployed on-premise or in-cloud.
          ## About this Operator

          TiDB Operator manages TiDB clusters on Kubernetes and automates tasks related to operating a TiDB cluster. It makes TiDB a truly cloud-native database.

          ## Features

          * Safely scaling the TiDB cluster

          TiDB Operator empowers TiDB with horizontal scalability on the cloud.

          * Rolling update of the TiDB cluster

          Gracefully perform rolling updates for the TiDB cluster in order, achieving zero-downtime of the TiDB cluster.

          * Multi-tenant support

          Users can deploy and manage multiple TiDB clusters on a single Kubernetes cluster easily.

          * Automatic failover

          TiDB Operator automatically performs failover for your TiDB cluster when node failures occur.

          * Kubernetes package manager support

          By embracing Kubernetes package manager Helm, users can easily deploy TiDB clusters with only one command.

          * Automatically monitoring TiDB cluster at creating

          Automatically deploy Prometheus, Grafana for TiDB cluster monitoring.

          ## Quick Start
          You can follow our [Get Started](https://docs.pingcap.com/tidb-in-kubernetes/stable/get-started) guide to quickly start a testing Kubernetes cluster and play with TiDB Operator on your own machine.

          ## Documentation

          You can see our documentation at PingCAP website for more in-depth installation and instructions for production:

          - [English](https://docs.pingcap.com/tidb-in-kubernetes/stable)
          - [简体中文](https://docs.pingcap.com/zh/tidb-in-kubernetes/stable)

          All the TiDB Operator documentation is maintained in the [docs-tidb-operator repository](https://github.com/pingcap/docs-tidb-operator).

          ## Community

          Feel free to reach out if you have any questions. The maintainers of this project are reachable via:

          - [TiDB Community Slack](https://pingcap.com/tidbslack/) in the [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel
          - [Filing an issue](https://github.com/pingcap/tidb-operator/issue) against this repo

          Pull Requests are welcome! Check the [issue tracker](https://github.com/pingcap/tidb-operator/issue) for `status/help-wanted` issues if you're unsure where to start.

          If you're planning a new feature, please file an issue or join [#sig-k8s](https://tidbcommunity.slack.com/archives/CHD0HA3LZ) channel to discuss first.
          ## Prerequisites for enabling this Operator
        displayName: TiDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - TiDB
        - SQL
        - Database
        links:
        - name: GitHub
          url: https://github.com/pingcap/tidb-operator
        - name: Documentation
          url: https://docs.pingcap.com/tidb-in-kubernetes/stable
        maintainers:
        - email: zhanghailong@pingcap.com
          name: Daniel Zhang
        maturity: stable
        minKubeVersion: 1.16.0
        provider:
          name: PingCAP
        relatedImages:
        - pingcap/tidb-operator:v1.3.1
        - quay.io/openshift-community-operators/tidb-operator@sha256:4f73b39de6129b906b5eb7bb0b7485f3da406cfad8616f81e375cc8ffcb156ed
        version: 1.3.1
      entries:
      - name: tidb-operator.v1.3.1
        version: 1.3.1
      name: stable
    defaultChannel: stable
    packageName: tidb-operator
    provider:
      name: PingCAP
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.2022.4.829
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "service": "ClusterIP"
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "replicas": 1
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080"
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer"
                  }
                },
                "runAsNonRoot": false
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaLightning",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "global": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "Default-cluster-name"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "kubeEnforcer": {
                    "infra": {
                      "serviceAccount": "aqua-kube-enforcer-sa"
                    },
                    "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                    "deploy": {
                      "service": "ClusterIP"
                    },
                    "starboard": {
                      "infra": {
                        "serviceAccount": "aqua-kube-enforcer-sa"
                      },
                      "config": {
                        "imagePullSecret": "aqua-registry"
                      },
                      "deploy": {
                        "replicas": 1
                      }
                    },
                    "env": []
                  },
                  "enforcer": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "deploy": null,
                    "token": "<<ENFORCER_GROUP_TOKEN>>",
                    "env": [
                    ],
                    "secret": {
                      "name": null,
                      "key": null
                    },
                    "runAsNonRoot": null,
                    "rhcosVersion": null
                  }
                }
              },
            {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
          createdAt: ""
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.2 <2022.4.829'
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnectors.operator.aquasec.com
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer) - Do not deploy using Certified Operator.
          * AquaCloudConnector.

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          The Aqua Operator provides a few [Custom Resources](https://github.com/aquasecurity/aqua-operator/tree/master/config/crd) for managing the Aqua CSP platform.

          ## Prerequisites
          There are only 4 prerequisites:
          1. Verify that the OpenShift version is officially supported [Verify Here](https://docs.aquasec.com/saas/workload-protection/product-information/system-requirements/)
          2. Make sure you have a license and to obtain one, please contact Aqua Security at [Contact Us](mailto:cloudsales@aquasec.com).
          3. Create a new project for aqua

              ```oc new-project aqua```

          4. Create the secret for Aqua Database password. If you are using an External Database, make sure you use those credentials.
          NOTE: This step is optional and you can specify the Database password when creating the CRs.

              ```oc create secret generic aqua-database-password --from-literal=db-password=<password> -n aqua```

          Please note that for the certified operator, the Docker registry secret is NOT needed. For the Red Hat certified operator, the Aqua application images are stored in the Red Hat Connect registry and will be automatically pulled by the Aqua operator.

          ## Choosing the right channel

          Channels are a flexible publishing mechanism that offers you different Aqua versions to deploy, at any point. These channels map to their respective Aqua versions, describing the maturity based on where they are in their lifecycle:
          * 4.6: This maps to the older version of Aqua 4.6.0
          * 5.0: This maps to the Aqua version 5.0.0
          * 5.3: This maps to the Aqua version 5.3.0
          * 6.0: This maps to the Aqua version 6.0.0
          * 6.2: This maps to the Aqua version 6.2.0
          * 6.5: This maps to the Aqua version 6.5.0
          * 2022.4: This is the latest and greatest version of Aqua (default)

          ## Deploying the Aqua Operator

          Aqua Operator follows an easy push button deployment that installs the operator in its own aqua namespace.
          Please refer to the instructions [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md).

          ## Configuring the Aqua Operator

          Operators leverage CRDs as a configuration mechanism for the application. Aqua offers a number of CRDs that can be tweaked and configured as per your needs. Please refer to this [link](https://github.com/aquasecurity/aqua-operator/blob/6.2.0/docs/DeployOpenShiftOperator.md#aquacsp-crds) for Custom Resource examples for popular scenarios.

          ## Support

          For support please contact support@aquasec.com.
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - registry.connect.redhat.com/aquasec/aquasec-operator-bundle@sha256:59e62e5cbd479eed9ddd083db881469056c35253d76aef6bda9cbdb578e05b3f
        - registry.connect.redhat.com/aquasec/kube-bench@sha256:bbb4e8a61033b303525af6764f685e62a2f441453c3f0eaaf4cbab6997bf93ee
        - registry.connect.redhat.com/aquasec/console@sha256:05bd172efdb4729027ea6a491c874a1a3368156a506b7b5d228837b12282ee1d
        - registry.connect.redhat.com/aquasec/database@sha256:c91ae82b0179c3ec4e156835a9f23b66767ba15514ff33fb268b35d8f04d6fde
        - registry.connect.redhat.com/aquasec/gateway@sha256:cc8a3e950ae45fb40866b7369aa9393e92d1ab91dc0795ee682217fc421e9ec2
        - registry.connect.redhat.com/aquasec/enforcer@sha256:b48f026d7876656c0b365a90bcaaca7f4544f38b85400c110cf20e57c67ac8f6
        - registry.connect.redhat.com/aquasec/aquasec@sha256:67b78f927f02c88c31a2f7a28aa2f90621eaaac1e3f13982513b895fa98e6118
        - registry.connect.redhat.com/aquasec/cloud-connector@sha256:f9da1936ac5e4b8e4537dca69f920c682fb9fac98d091680e01d2122f2b0f85f
        - registry.connect.redhat.com/aquasec/kube-enforcer@sha256:ce827db4a2b01b054e33f1091ee7046a14757b8781444bc9868783fe1efbf641
        - registry.connect.redhat.com/aquasec/scanner@sha256:d0ab5e0dad4e042324874de02ee36585a93b53ad1c3e905416af2fcb98f15e2f
        - registry.connect.redhat.com/aquasec/starboard@sha256:61673f7ba1d983ff4bb4bd76036ac55e5b81696c9a3e3aa4ab1a49a27a7a2106
        version: 2022.4.829
      entries:
      - name: aqua-operator.2022.4.829
        version: 2022.4.829
      - name: aqua-operator.2022.4.818
        version: 2022.4.818
      - name: aqua-operator.2022.4.805
        version: 2022.4.805
      - name: aqua-operator.2022.4.760
        version: 2022.4.760
      - name: aqua-operator.2022.4.759
        version: 2022.4.759
      - name: aqua-operator.2022.4.720
        version: 2022.4.720
      - name: aqua-operator.2022.4.683
        version: 2022.4.683
      - name: aqua-operator.2022.4.662
        version: 2022.4.662
      - name: aqua-operator.2022.4.588
        version: 2022.4.588
      - name: aqua-operator.2022.4.587
        version: 2022.4.587
      - name: aqua-operator.2022.4.519
        version: 2022.4.519
      - name: aqua-operator.2022.4.518
        version: 2022.4.518
      name: "2022.4"
    defaultChannel: "2022.4"
    packageName: aqua-operator-certified
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzo-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzo-operator.v3.1.200
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzo.cambridgesemantics.com/v2",
                "kind": "Anzo",
                "metadata": {
                  "name": "no-persistence"
                },
                "spec": {
                  "jsonActivation": true,
                  "jvmMemory": "4096m",
                  "nodeConfig": {
                    "spec": {
                      "replicas": 1,
                      "selector": {
                        "matchLabels": {
                          "app": "anzo"
                        }
                      },
                      "serviceName": "anzo-no-persistence",
                      "template": {
                        "metadata": {
                          "annotations": {
                            "cluster-autoscaler.kubernetes.io/safe-to-evict": "false"
                          },
                          "labels": {
                            "app": "anzo"
                          }
                        },
                        "spec": {
                          "containers": [
                            {
                              "image": "registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d",
                              "imagePullPolicy": "Always",
                              "name": "anzo",
                              "resources": {
                                "limits": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                },
                                "requests": {
                                  "cpu": "2.0",
                                  "memory": "8.0Gi"
                                }
                              }
                            }
                          ],
                          "initContainers": [
                            {
                              "name": "anzo-init",
                              "resources": {
                                "limits": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                },
                                "requests": {
                                  "cpu": "0.5",
                                  "memory": "2Gi"
                                }
                              }
                            }
                          ],
                          "securityContext": {
                            "fsGroup": 1001
                          },
                          "serviceAccountName": "anzo-operator"
                        }
                      }
                    }
                  },
                  "role": "AnzoServer"
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
          createdAt: "2025-07-09T06:52:12Z"
          description: kubernetes operator for Altair Graph Studio
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Customer Support
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoBackup is the Schema for the anzobackups API
            displayName: Altair Graph Studio Backup
            kind: AnzoBackup
            name: anzobackups.anzobackup.cambridgesemantics.com
            version: v1beta1
          - description: Altair Graph Studio is the Schema for the anzos API
            displayName: Altair Graph Studio Operator
            kind: Anzo
            name: anzos.anzo.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Studio Operator provides the way to install and configure an altair graph studio server and agent setup on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzo/userdoc/deploy-k8s.htm )

          ### Support

          We offer Support to our customers [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Studio Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        links:
        - name: Altair Graph Studio Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzo/blob/v3.1.2/README_openshift.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator-bundle@sha256:ad6b8e10abfff1cdac1ea35b44c456e3d59b4dd1372cd47c2e4221ed93e1f71f
        - registry.connect.redhat.com/cambridgesemantics/anzo-operator@sha256:4c0c8bf7acbfbb96f15459a20364608e5d0a0933c6326a89d219666ee0d88ca2
        - registry.connect.redhat.com/cambridgesemantics/anzo@sha256:5b192604338f602fdabe55fecefd9eb82400a77c5aa7f04bd8e54aead8644c6d
        version: 3.1.200
      entries:
      - name: anzo-operator.v3.1.200
        version: 3.1.200
      - name: anzo-operator.v3.1.106
        version: 3.1.106
      - name: anzo-operator.v3.1.105
        version: 3.1.105
      name: stable
    defaultChannel: stable
    packageName: anzo-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: EDB
      provider-url: ""
    name: cloud-native-postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cloud-native-postgresql.v1.27.0-rc1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
          createdAt: "2025-07-30T09:28:05Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:167b9063a2dec7f9a5f1c5f3c02daf481c45c5e6a2f1b25353102e85921b6fd7
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:5f5bebbf4d8021dafd0fb7485d629b30aaa362e899e69eb52fa5e74757af6e36
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.27.0-rc1
      entries:
      - name: cloud-native-postgresql.v1.27.0-rc1
        version: 1.27.0-rc1
      - name: cloud-native-postgresql.v1.26.0-rc3
        version: 1.26.0-rc3
      - name: cloud-native-postgresql.v1.26.0-rc2
        version: 1.26.0-rc2
      - name: cloud-native-postgresql.v1.26.0-rc1
        version: 1.26.0-rc1
      name: candidate
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.21.1
        version: 1.21.1
      - name: cloud-native-postgresql.v1.21.0
        version: 1.21.0
      - name: cloud-native-postgresql.v1.20.2
        version: 1.20.2
      - name: cloud-native-postgresql.v1.20.1
        version: 1.20.1
      - name: cloud-native-postgresql.v1.20.0
        version: 1.20.0
      - name: cloud-native-postgresql.v1.19.1
        version: 1.19.1
      - name: cloud-native-postgresql.v1.19.0
        version: 1.19.0
      name: fast
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      - name: cloud-native-postgresql.v1.18.13
        version: 1.18.13
      - name: cloud-native-postgresql.v1.18.12
        version: 1.18.12
      - name: cloud-native-postgresql.v1.18.11
        version: 1.18.11
      - name: cloud-native-postgresql.v1.18.10
        version: 1.18.10
      - name: cloud-native-postgresql.v1.18.9
        version: 1.18.9
      - name: cloud-native-postgresql.v1.18.8
        version: 1.18.8
      - name: cloud-native-postgresql.v1.18.7
        version: 1.18.7
      name: stable
    - currentCSV: cloud-native-postgresql.v1.22.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
          createdAt: "2025-07-25T19:02:56Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images. Only
              available in 1.25 or later
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster.
              Only available in 1.25 or later
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster. Only available in 1.25 or later
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication.
              Only available in 1.25 or later
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:83c5e966addca5197f25c5f9c81a52648db44c2fd9e4969bf30b82c4880f0a40
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:4d55ad54cc7e7807bc99f5a6f96bc92e88b4060f424ffd2cb6552be9505f38f4
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.22.11
      entries:
      - name: cloud-native-postgresql.v1.22.11
        version: 1.22.11
      - name: cloud-native-postgresql.v1.22.10
        version: 1.22.10
      - name: cloud-native-postgresql.v1.22.9
        version: 1.22.9
      - name: cloud-native-postgresql.v1.22.8
        version: 1.22.8
      - name: cloud-native-postgresql.v1.22.7
        version: 1.22.7
      - name: cloud-native-postgresql.v1.22.6
        version: 1.22.6
      - name: cloud-native-postgresql.v1.22.5
        version: 1.22.5
      - name: cloud-native-postgresql.v1.22.4
        version: 1.22.4
      - name: cloud-native-postgresql.v1.22.3
        version: 1.22.3
      - name: cloud-native-postgresql.v1.22.2
        version: 1.22.2
      - name: cloud-native-postgresql.v1.22.1
        version: 1.22.1
      - name: cloud-native-postgresql.v1.22.0
        version: 1.22.0
      name: stable-v1.22
    - currentCSV: cloud-native-postgresql.v1.23.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
          createdAt: "2024-12-23T16:37:54Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>= 1.18.0 < 1.23.6'
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:934e01ed5cd35cf681414b9136a337514a1b73c0110871f1c78ba98447badb0f
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:bfbb9d0d5556eb067543f4e139a935000319b439f700240adaf393c68bd8244b
        version: 1.23.6
      entries:
      - name: cloud-native-postgresql.v1.23.6
        version: 1.23.6
      - name: cloud-native-postgresql.v1.23.5
        version: 1.23.5
      - name: cloud-native-postgresql.v1.23.4
        version: 1.23.4
      - name: cloud-native-postgresql.v1.23.3
        version: 1.23.3
      - name: cloud-native-postgresql.v1.23.2
        version: 1.23.2
      - name: cloud-native-postgresql.v1.23.1
        version: 1.23.1
      name: stable-v1.23
    - currentCSV: cloud-native-postgresql.v1.24.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
          createdAt: "2025-05-23T15:20:37Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:a1acaaf5a413db0f33f3824a15aefe598974d1593172bcbccd0fc3d39d7b46d0
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:e0b1cff317bd4e35334429d1cc7c548dbda54065716aea96ef30610f0b91901b
        version: 1.24.4
      entries:
      - name: cloud-native-postgresql.v1.24.4
        version: 1.24.4
      - name: cloud-native-postgresql.v1.24.3
        version: 1.24.3
      - name: cloud-native-postgresql.v1.24.2
        version: 1.24.2
      - name: cloud-native-postgresql.v1.24.1
        version: 1.24.1
      - name: cloud-native-postgresql.v1.24.0
        version: 1.24.0
      name: stable-v1.24
    - currentCSV: cloud-native-postgresql.v1.25.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
          createdAt: "2025-07-26T06:26:53Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          * Direct integration with Kubernetes API server for High Availability,
            without requiring an external tool
          * Self-Healing capability, through:
              * failover of the primary instance by promoting the most aligned replica
              * automated recreation of a replica
          * Planned switchover of the primary instance by promoting a selected replica
          * Scale up/down capabilities
          * Definition of an arbitrary number of instances (minimum 1 - one primary server)
          * Definition of the *read-write* service, to connect your applications to the only primary server of the cluster
          * Definition of the *read-only* service, to connect your applications to any of the instances for reading workloads
          * Declarative management of PostgreSQL configuration
          * Declarative management of Postgres roles, users and groups
          * Support for Local Persistent Volumes with PVC templates
          * Reuse of Persistent Volumes storage in Pods
          * Separate volume for WAL files
          * Rolling updates for PostgreSQL minor versions
          * In-place or rolling updates for operator upgrades
          * TLS connections and client certificate authentication
          * Support for custom TLS certificates (including integration with cert-manager)
          * Continuous WAL archiving to an object store (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Backups on volume snapshots (where supported by the underlying storage classes)
          * Backups on object stores (AWS S3 and S3-compatible, Azure Blob Storage, and Google Cloud Storage)
          * Full recovery and Point-In-Time recovery from an existing backup on volume snapshots or object stores
          * Offline import of existing PostgreSQL databases, including major upgrades of PostgreSQL
          * Fencing of an entire PostgreSQL cluster, or a subset of the instances in a declarative way
          * Hibernation of a PostgreSQL cluster in a declarative way
          * Support for Synchronous Replicas
          * Support for HA physical replication slots at cluster level
          * Backup from a standby
          * Backup retention policies (based on recovery window, only on object stores)
          * Parallel WAL archiving and restore to allow the database to keep up with WAL
            generation on high write systems
          * Support tagging backup files uploaded to an object store to enable optional
            retention management at the object store layer Replica clusters for
          * PostgreSQL deployments across multiple Kubernetes
            clusters, enabling private, public, hybrid, and multi-cloud architectures
          * Connection pooling with PgBouncer
          * Support for node affinity via `nodeSelector`
          * Native customizable exporter of user defined metrics for Prometheus through the `metrics` port (9187)
          * Standard output logging of PostgreSQL error messages in JSON format
          * Automatically set `readOnlyRootFilesystem` security context for pods
          * `cnp` plugin for `kubectl`
          * Simple bind and search+bind LDAP client authentication
          * Multi-arch format container images
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:cad1dd788daaa47dc3ee70a61137dab56f4e4264e1dfbd59e87940a699642028
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:58001b8492f1be05e033d5e76ba03837c27d0b793d92ddb9e8846eb1b2b6eae2
        version: 1.25.3
      entries:
      - name: cloud-native-postgresql.v1.25.3
        version: 1.25.3
      - name: cloud-native-postgresql.v1.25.2
        version: 1.25.2
      - name: cloud-native-postgresql.v1.25.1
        version: 1.25.1
      - name: cloud-native-postgresql.v1.25.0
        version: 1.25.0
      name: stable-v1.25
    - currentCSV: cloud-native-postgresql.v1.26.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
          createdAt: "2025-07-26T07:59:14Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:eef6e7cf8f76ccd5e9ed16a5e4bdeb4398859121a6b0e86284b9766408cc2b30
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:47d70233908ea855ddab3ff79c7a226719c3f2be5b8a5018c7934cd44bf0f546
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        version: 1.26.1
      entries:
      - name: cloud-native-postgresql.v1.26.1
        version: 1.26.1
      - name: cloud-native-postgresql.v1.26.0
        version: 1.26.0
      name: stable-v1.26
    - currentCSV: cloud-native-postgresql.v1.27.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  }
                },
                "status": {
                  "serverName": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Cluster",
                "metadata": {
                  "name": "cluster-sample"
                },
                "spec": {
                  "instances": 3,
                  "logLevel": "info",
                  "primaryUpdateStrategy": "unsupervised",
                  "storage": {
                    "size": "1Gi"
                  },
                  "walStorage": {
                    "size": "1Gi"
                  }
                },
                "status": {
                  "instances": 3
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ClusterImageCatalog",
                "metadata": {
                  "name": "postgresql"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Database",
                "metadata": {
                  "name": "database-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "name": "database-sample",
                  "owner": "app"
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ImageCatalog",
                "metadata": {
                  "name": "postgresql",
                  "namespace": "default"
                },
                "spec": {
                  "images": [
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3",
                      "major": 15
                    },
                    {
                      "image": "quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9",
                      "major": 16
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Pooler",
                "metadata": {
                  "name": "pooler-sample-rw"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "instances": 1,
                  "pgbouncer": {
                    "poolMode": "session"
                  },
                  "type": "rw"
                },
                "status": {
                  "instances": 1
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Publication",
                "metadata": {
                  "name": "publication-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "dbname": "app",
                  "name": "pub",
                  "target": {
                    "allTables": true
                  }
                },
                "status": {
                  "applied": false
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "ScheduledBackup",
                "metadata": {
                  "name": "scheduledbackup-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample"
                  },
                  "schedule": "0 0 0 * * *"
                },
                "status": {
                  "lastCheckTime": null
                }
              },
              {
                "apiVersion": "postgresql.k8s.enterprisedb.io/v1",
                "kind": "Subscription",
                "metadata": {
                  "name": "subscription-sample"
                },
                "spec": {
                  "cluster": {
                    "name": "cluster-sample-dest"
                  },
                  "dbname": "app",
                  "externalClusterName": "cluster-sample",
                  "name": "sub",
                  "publicationName": "pub"
                },
                "status": {
                  "applied": false
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          certified: "true"
          containerImage: quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
          createdAt: "2025-08-19T12:00:36Z"
          description: EDB Postgres for Kubernetes is an operator designed to manage
            highly available PostgreSQL databases with a primary/standby architecture
            on any supported Kubernetes cluster.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/EnterpriseDB/cloud-native-postgres
          support: EDB
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PostgreSQL backup (physical base backup)
            displayName: Backups
            kind: Backup
            name: backups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A cluster-wide catalog of PostgreSQL operand images
            displayName: Cluster Image Catalog
            kind: ClusterImageCatalog
            name: clusterimagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: PostgreSQL cluster (primary/standby architecture)
            displayName: Cluster
            kind: Cluster
            name: clusters.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a database on a Cluster
            displayName: Postgres Database
            kind: Database
            name: databases.postgresql.k8s.enterprisedb.io
            version: v1
          - description: FailoverQuorum contains the information about the current
              failover quorum status of a PG cluster
            displayName: Failover Quorum
            kind: FailoverQuorum
            name: failoverquorums.postgresql.k8s.enterprisedb.io
            version: v1
          - description: A catalog of PostgreSQL operand images
            displayName: Image Catalog
            kind: ImageCatalog
            name: imagecatalogs.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Pooler for a Postgres Cluster (with PgBouncer)
            displayName: Pooler
            kind: Pooler
            name: poolers.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Publication in a PostgreSQL Cluster
            displayName: Postgres Publication
            kind: Publication
            name: publications.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Backup scheduler for a given Postgres cluster
            displayName: Scheduled Backups
            kind: ScheduledBackup
            name: scheduledbackups.postgresql.k8s.enterprisedb.io
            version: v1
          - description: Declarative creation and management of a Logical Replication
              Subscription in a PostgreSQL Cluster to a previously defined Publication
            displayName: Postgres Subscription
            kind: Subscription
            name: subscriptions.postgresql.k8s.enterprisedb.io
            version: v1
        description: |
          Main features:

          - Direct integration with the Kubernetes API server for High Availability,
            eliminating the need for external tools.
          - Self-healing capabilities, including:
              - Automated failover by promoting the most aligned replica.
              - Automatic recreation of failed replicas.
          - Planned switchover of the primary instance by promoting a selected replica.
          - Declarative management of key PostgreSQL configurations, including:
              - PostgreSQL settings.
              - Roles, users, and groups.
              - Databases, extensions, and schemas.
              - Tablespaces (including temporary tablespaces).
          - Flexible instance definition, supporting any number of instances (minimum 1
            primary server).
          - Scale-up/down capabilities to dynamically adjust cluster size.
          - Read-Write and Read-Only Services, ensuring applications connect correctly:
              - *Read-Write Service*: Routes connections to the primary server.
              - *Read-Only Service*: Distributes connections among replicas for read workloads.
          - Support for quorum-based and priority-based PostgreSQL Synchronous
            Replication.
          - Replica clusters enabling PostgreSQL distributed topologies across multiple
            Kubernetes clusters (private, public, hybrid, and multi-cloud).
          - Delayed Replica clusters for point-in-time access to historical data.
          - Persistent volume management, including:
              - Support for Local Persistent Volumes with PVC templates.
              - Reuse of Persistent Volumes storage in Pods.
              - Separate volumes for WAL files and tablespaces.
          - Backup and recovery options, including:
              - Integration with the [Barman Cloud plugin](https://github.com/cloudnative-pg/plugin-barman-cloud)
                for continuous online backup via WAL archiving to AWS S3, S3-compatible
                services, Azure Blob Storage, and Google Cloud Storage, with support for
                retention policies based on a configurable recovery window.
              - Backups using volume snapshots (where supported by storage classes).
              - Full and Point-In-Time recovery from volume snapshots or object stores (via Barman Cloud plugin).
              - Backup from standby replicas to reduce primary workload impact.
          - Offline and online import of PostgreSQL databases, including major upgrades:
              - *Offline Import*: Direct restore from existing databases.
              - *Online Import*: PostgreSQL native logical replication via the `Subscription` resource.
          - Offline In-Place Major Upgrades of PostgreSQL
          - High Availability physical replication slots, including synchronization of
            user-defined replication slots.
          - Parallel WAL archiving and restore, ensuring high-performance data
            synchronization in high-write environments.
          - TLS support, including:
              - Secure connections and client certificate authentication.
              - Custom TLS certificates (integrated with `cert-manager`).
          - Startup and readiness probes, including replica probes based on desired lag
            from the primary.
          - Declarative rolling updates for:
              - PostgreSQL minor versions.
              - Operator upgrades (in-place or rolling updates).
          - Standard output logging of PostgreSQL error messages in JSON format for
            easier integration with log aggregation tools.
          - Prometheus-compatible metrics exporter (`metrics` port 9187) for custom
            monitoring.
          - `cnp` plugin for `kubectl` to simplify cluster operations.
          - Cluster hibernation for resource efficiency in inactive states.
          - Fencing of PostgreSQL clusters (full cluster or subset) to isolate instances
            when needed.
          - Connection pooling with PgBouncer for improved database efficiency.
          - OLM (Operator Lifecycle Manager) installation support for streamlined
            deployments.
          - Multi-arch container images, including Software Bill of Materials (SBOM) and
            provenance attestations for security compliance.
        displayName: EDB Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgresql
        - postgres
        - database
        - sql
        - cloudnativepg
        - cloudnative-pg
        - cnp
        links:
        - name: EDB Postgres for Kubernetes
          url: https://www.enterprisedb.com/products/postgresql-on-kubernetes-ha-clusters-k8s-containers-scalable
        - name: Documentation
          url: https://www.enterprisedb.com/docs/postgres_for_kubernetes/latest/
        maintainers:
        - email: jonathan.gonzalez@enterprisedb.com
          name: Jonathan Gonzalez V.
        - email: jonathan.battiato@enterprisedb.com
          name: Jonathan Battiato
        - email: niccolo.fei@enterprisedb.com
          name: Niccolo Fei
        - email: gabriele.bartolini@enterprisedb.com
          name: Gabriele Bartolini
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: EDB
        relatedImages:
        - registry.connect.redhat.com/enterprisedb/cloud-native-postgresql@sha256:92ff276ec982a16f8bd84d451bc3def511a2757fa1c960d3537918ed1093358c
        - quay.io/enterprisedb/cloud-native-postgresql@sha256:75e712db8a7005b180b399c1c3bf991bc1a5e723bcc98066667a3ea27980b83c
        - quay.io/enterprisedb/postgresql@sha256:8f68afde9b1217ad13682bd2adfc0a7cac691cbdb496cc22915f6783012067f9
        - quay.io/enterprisedb/postgresql@sha256:d2e7f433904cf4f5193d044434b1d6294222047398692697fc5c3f6311997ed3
        version: 1.27.0
      entries:
      - name: cloud-native-postgresql.v1.27.0
        version: 1.27.0
      name: stable-v1.27
    defaultChannel: fast
    packageName: cloud-native-postgresql
    provider:
      name: EDB
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: OpenShift Database Access
      provider-url: ""
    name: dbaas-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dbaas-operator.v0.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSConnection",
                "metadata": {
                  "name": "dbaasconnection-example"
                },
                "spec": {
                  "databaseServiceID": "instance-example",
                  "databaseServiceType": "instance",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInstance",
                "metadata": {
                  "name": "dbaasinstance-sample"
                },
                "spec": {
                  "cloudProvider": "AWS",
                  "cloudRegion": "US_EAST_1",
                  "inventoryRef": {
                    "name": "inventory-example",
                    "namespace": "openshift-dbaas-operator"
                  },
                  "name": "my-cluster-free"
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSInventory",
                "metadata": {
                  "name": "inventory-example"
                },
                "spec": {
                  "credentialsRef": {
                    "name": "dbaas-vendor-credentials"
                  },
                  "providerRef": {
                    "name": "provider-example-registration"
                  }
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSPolicy",
                "metadata": {
                  "name": "dbaaspolicy-sample"
                },
                "spec": {
                  "disableProvisions": false
                }
              },
              {
                "apiVersion": "dbaas.redhat.com/v1beta1",
                "kind": "DBaaSProvider",
                "metadata": {
                  "labels": {
                    "related-to": "dbaas-operator",
                    "type": "dbaas-provider-registration"
                  },
                  "name": "provider-example-registration"
                },
                "spec": {
                  "allowsFreeTrial": true,
                  "connectionKind": "ProviderConnection",
                  "credentialFields": [
                    {
                      "displayName": "ProviderAPIKeyD",
                      "helpText": "This is the API Key for the example provider to show on UI",
                      "key": "APIKey",
                      "required": true,
                      "type": "maskedstring"
                    }
                  ],
                  "externalProvisionDescription": "Follow the guide to start a free Provider Serverless (beta) cluster",
                  "externalProvisionURL": "https://www.exmample.com/docs/provider/quickstart.html",
                  "groupVersion": "dbaas.redhat.com/v1beta1",
                  "instanceKind": "ProviderInstance",
                  "inventoryKind": "ProviderInventory",
                  "provider": {
                    "displayDescription": "This is an example for providers on how to implement their operator to integrate with DBaaS.",
                    "displayName": "DBaaS Provider Example ",
                    "icon": {
                      "base64data": "SGVsbG8sIHdvcmxkLg==",
                      "mediatype": "image/png"
                    },
                    "name": "Provider Example"
                  },
                  "provisioningParameters": {
                    "cloudProvider": {
                      "conditionalData": [
                        {
                          "defaultValue": "GCP",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "FREETRIAL"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        },
                        {
                          "defaultValue": "AWS",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "Amazon Web Services",
                              "value": "AWS"
                            },
                            {
                              "displayValue": "Google Cloud Platform",
                              "value": "GCP"
                            }
                          ]
                        }
                      ],
                      "displayName": "Cloud Provider"
                    },
                    "dedicatedLocationLabel": {
                      "displayName": "Select regions \u0026 nodes",
                      "helpText": "Select the geographical region where you want the database instance to run, and set the number of nodes you want running in this dedicated cluster."
                    },
                    "hardwareLabel": {
                      "displayName": "Hardware per node",
                      "helpText": "Select the compute and storage requirements for this database instance."
                    },
                    "machineType": {
                      "conditionalData": [
                        {
                          "defaultValue": "m5.large",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 8 GiB RAM",
                              "value": "m5.large"
                            },
                            {
                              "displayValue": "4 vCPU, 16 GiB RAM",
                              "value": "m5.xlarge"
                            },
                            {
                              "displayValue": "8 vCPU, 32 GiB RAM",
                              "value": "m5.2xlarge"
                            },
                            {
                              "displayValue": "16 vCPU, 64 GiB RAM",
                              "value": "m5.4xlarge"
                            },
                            {
                              "displayValue": "32 vCPU, 128 GiB RAM",
                              "value": "m5.8xlarge"
                            }
                          ]
                        },
                        {
                          "defaultValue": "n1-standard-2",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "2 vCPU, 7.5 GiB RAM",
                              "value": "n1-standard-2"
                            },
                            {
                              "displayValue": "4 vCPU, 15 GiB RAM",
                              "value": "n1-standard-4"
                            },
                            {
                              "displayValue": "8 vCPU, 30 GiB RAM",
                              "value": "n1-standard-8"
                            },
                            {
                              "displayValue": "16 vCPU, 60 GiB RAM",
                              "value": "n1-standard-16"
                            },
                            {
                              "displayValue": "32 vCPU, 120 GiB RAM",
                              "value": "n1-standard-32"
                            }
                          ]
                        }
                      ],
                      "displayName": "Compute"
                    },
                    "name": {
                      "displayName": "Cluster name"
                    },
                    "nodes": {
                      "displayName": "nodes"
                    },
                    "plan": {
                      "conditionalData": [
                        {
                          "defaultValue": "SERVERLESS",
                          "options": [
                            {
                              "displayValue": "Free trial",
                              "value": "FREETRIAL"
                            },
                            {
                              "displayValue": "Serverless",
                              "value": "SERVERLESS"
                            },
                            {
                              "displayValue": "Dedicated",
                              "value": "DEDICATED"
                            }
                          ]
                        }
                      ],
                      "displayName": "Hosting plan"
                    },
                    "planLabel": {
                      "displayName": "Select a plan"
                    },
                    "regions": {
                      "displayName": "us-west-2"
                    },
                    "serverlessLocationLabel": {
                      "displayName": "Select regions",
                      "helpText": "Select the geographical region where you want the database instance to run."
                    },
                    "spendLimit": {
                      "conditionalData": [
                        {
                          "defaultValue": "0",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "SERVERLESS"
                            }
                          ]
                        }
                      ],
                      "displayName": "Spend limit"
                    },
                    "spendLimitLabel": {
                      "displayName": "Spend limit",
                      "helpText": "Set a spending limit on resources for this database instance."
                    },
                    "storageGib": {
                      "conditionalData": [
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "AWS"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        },
                        {
                          "defaultValue": "15",
                          "dependencies": [
                            {
                              "field": "plan",
                              "value": "DEDICATED"
                            },
                            {
                              "field": "cloudProvider",
                              "value": "GCP"
                            }
                          ],
                          "options": [
                            {
                              "displayValue": "15 GiB",
                              "value": "15"
                            },
                            {
                              "displayValue": "35 GiB",
                              "value": "35"
                            },
                            {
                              "displayValue": "75 GiB",
                              "value": "75"
                            },
                            {
                              "displayValue": "150 GiB",
                              "value": "150"
                            },
                            {
                              "displayValue": "300 GiB",
                              "value": "300"
                            },
                            {
                              "displayValue": "600 GiB",
                              "value": "600"
                            }
                          ]
                        }
                      ],
                      "displayName": "Storage"
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Database
          containerImage: quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
          description: Discover & connect with database instances hosted on 3rd-party
            ISV cloud platforms.
          operatorframework.io/suggested-namespace: openshift-dbaas-operator
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/internal-objects: '["dbaasconnections.dbaas.redhat.com","dbaasproviders.dbaas.redhat.com","dbaasplatforms.dbaas.redhat.com","dbaasinstances.dbaas.redhat.com"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/RHEcosystemAppEng/dbaas-operator
          support: rhoda-requests@redhat.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSConnection defines the schema for the DBaaSConnection
              API.
            displayName: DBaaSConnection
            kind: DBaaSConnection
            name: dbaasconnections.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInstance defines the schema for the DBaaSInstance API.
            displayName: DBaaSInstance
            kind: DBaaSInstance
            name: dbaasinstances.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSInventory defines the schema for the DBaaSInventory
              API. Inventory objects must be created in a valid namespace, determined
              by the existence of a DBaaSPolicy object.
            displayName: Provider Account
            kind: DBaaSInventory
            name: dbaasinventories.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPlatform defines the schema for the DBaaSPlatform API.
            displayName: DBaaSPlatform
            kind: DBaaSPlatform
            name: dbaasplatforms.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSPolicy enables administrative capabilities within a
              namespace, and sets a default inventory policy. Policy defaults can
              be overridden on a per-inventory basis.
            displayName: Provider Account Policy
            kind: DBaaSPolicy
            name: dbaaspolicies.dbaas.redhat.com
            version: v1beta1
          - kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1alpha1
          - description: DBaaSProvider defines the schema for the DBaaSProvider API.
            displayName: DBaaSProvider
            kind: DBaaSProvider
            name: dbaasproviders.dbaas.redhat.com
            version: v1beta1
        description: |
          The OpenShift Database Access Operator enables OpenShift users to discover & connect with database instances
          hosted on 3rd-party ISV cloud platforms such as CrunchyData Bridge & CockroachCloud.

          ## Feedback & Support

          Please contact us at [rhoda-requests@redhat.com](mailto:rhoda-requests@redhat.com)

          For more information on how to use our service, please refer to our [Quick Starts](quickstart?keyword=database).
        displayName: OpenShift Database Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - rhoda
        - oda
        - dbaas
        - database
        - db
        - data
        - access
        - crunchy
        - bridge
        - cockroach
        - cockroachdb
        - cloud
        - ack
        - amazon
        - rds
        - relational
        links:
        - name: GitHub
          url: https://github.com/RHEcosystemAppEng/dbaas-operator
        maintainers:
        - email: ecosystem-appeng@redhat.com
          name: Red Hat Ecosystem App Engineering
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: OpenShift Database Access
        relatedImages:
        - quay.io/ecosystem-appeng/dbaas-dynamic-plugin:0.5.0
        - quay.io/ecosystem-appeng/dbaas-operator:v0.5.0
        - quay.io/ecosystem-appeng/rds-dbaas-operator-catalog:v0.4.0
        - quay.io/openshift-community-operators/dbaas-operator@sha256:b2474b569406fb0f47a7042042429bbc5cd2edbbd119e701c6f440a139e1b3ac
        - quay.io/openshift/origin-kube-rbac-proxy:4.10.0
        - registry.developers.crunchydata.com/crunchydata/crunchy-bridge-operator-catalog:v0.0.6
        - gcr.io/cockroach-shared/ccapi-k8s-operator-catalog:v0.0.5
        version: 0.5.0
      entries:
      - name: dbaas-operator.v0.5.0
        version: 0.5.0
      name: stable
    defaultChannel: stable
    packageName: dbaas-operator
    provider:
      name: OpenShift Database Access
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Inc.
      provider-url: https://www.redhat.com/
    name: analytics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: alpha
    - currentCSV: analytics-operator.v0.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyData",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalydata-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalydata",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalydata-sample"
                },
                "spec": {
                  "anomalyname": "etcd_object_namespaces",
                  "config": {
                    "max": 100,
                    "min": 60,
                    "query": "max(apiserver_storage_objects{resource=~\"namespaces\"}) by (resource)"
                  },
                  "method": "min_max",
                  "metricdata": {
                    "groupeddata": "{ \"resource\": \"namespaces\" }",
                    "latestvalue": 121,
                    "timestamp": 1695804362
                  }
                }
              },
              {
                "apiVersion": "observability-analytics.redhat.com/v1alpha1",
                "kind": "AnomalyEngine",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "analytics-operator",
                    "app.kubernetes.io/instance": "anomalyengine-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "anomalyengine",
                    "app.kubernetes.io/part-of": "analytics-operator"
                  },
                  "name": "anomalyengine-sample"
                },
                "spec": {
                  "anomalyqueryconfiguration": "etcd_objects_all:\n  method: percentage_change\n  query: max(apiserver_storage_objects{resource!~\"certificatesigningrequests.certificates.k8s.io|installplans.operators.coreos.com|operators.operators.coreos.com|subscriptions.operators.coreos.com|anomalydata.observability-analytics.redhat.com|anomalyengines.observability-analytics.redhat.com\"}) by (resource)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: True\nfailed_pods:\n  method: min_max\n  query: count(kube_pod_status_phase{phase!~\"Succeeded|Pending|Running|Unknown\"}) by (phase, namespace)\n  min: 0\n  max: 10\n  have_multi_result_data: True\nworkload_memory_usage_bytes:\n  method: percentage_change\n  query: sum(container_memory_working_set_bytes{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nworkload_cpu_usage_cores:\n  method: percentage_change\n  query: sum(container_cpu_usage_seconds_total{namespace!~\"openshift-.+\",pod!=\"\",container=\"\"})\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\nrest_client_requests_total:\n  method: percentage_change\n  query: sum(rest_client_requests_total)\n  step: 5  # minutes\n  percentage_change: 60  # percentage\n  period_range: 120  # minutes\n  have_multi_result_data: False\n",
                  "cronjobconfig": {
                    "anomalyqueries": "etcd_objects_all,failed_pods,workload_memory_usage_bytes,workload_cpu_usage_cores,rest_client_requests_total",
                    "name": "osa-anomaly-detection",
                    "resource": {
                      "cpulimit": "256m",
                      "cpurequest": "128m",
                      "memorylimit": "512Mi",
                      "memoryrequest": "256Mi"
                    },
                    "schedule": "*/5 * * * *"
                  },
                  "namespace": "osa-anomaly-detection",
                  "serviceaccountrolebinding": {
                    "clusterrolename": "cluster-monitoring-view",
                    "name": "osa-in-cluster-thanos-querier"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
          createdAt: "2024-03-12T08:02:25Z"
          description: Deploys and Manages Observability Analytics Operator
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4-alpha
          repository: https://github.com/openshift/analytics-operator
          support: RedHat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnomalyData is the Schema for the anomalydata API
            displayName: Anomaly Data
            kind: AnomalyData
            name: anomalydata.observability-analytics.redhat.com
            version: v1alpha1
          - description: AnomalyEngine is the Schema for the anomalyengines API
            displayName: Anomaly Engine
            kind: AnomalyEngine
            name: anomalyengines.observability-analytics.redhat.com
            version: v1alpha1
        description: "### Introduction\nObservability Analytics for Red Hat OpenShift
          runs OpenShift analytics processes like anomaly detection.\n\n### API Backward
          Compatibility\n**NOTE:** Until the Operator supports **seamless upgrades**,
          a new version of the operator may introduce a change that is **NOT** backward-compatible
          with the previous version. Thus, to upgrade the operator, uninstall the
          already installed version completely (including CRDs) and install the new
          version.\n\n### Installtion Steps for the Anomaly Engine\n Create namespace
          if you would like to install this operator into separate namespace (You
          can utilize existing namespace as well). ```bash kubectl create ns <Namespace>
          ``` Install the Observability Analytics operator to specific `<Namespace>`,
          wait until you see the operator in the **Installed Operator** list. Click
          on the Operator name and go to Operator details.\n\nHere you should see
          a tab called **Anomaly Engine**. Click on it, and then click on the **Create
          AnomalyEngine** button, which will redirect you to a form where you need
          to provide the required inputs for the Anomaly Engine. Once successfully
          created, you should see a cronjob in the given namespace. This cronjob will
          identify anomalies and store them into **anomalydata**.\n\n### How to See
          Anomaly Data Captured by Anomaly Engine?\nWe are storing anomaly data into
          CR.\n ####Option 1:Through `kubectl`` command\nYou can check detected anomalies
          using the following command from your terminal. \n\n```bash\nkubectl get
          anomalydata -n <Namespace>\nkubectl describe anomalydata <NAME of the anomalydata>
          -n <Namespace> \n```\n #### Option 2: From Installed Operator UI Go to installed
          operator and select **Observability Analytics** operator, now go to **Anomaly
          Data** tab where you can able to see the detected anomaly, click on specific
          anomaly and you should be able to see details of the same. ### Documentation\nDocumentation
          around how to test Operator codebase as a developer can be found [here](https://github.com/openshift/analytics-operator#readme)\n\n###
          Contributing\nYou can contribute by:\n  * Raising [issues](https://github.com/openshift/analytics-operator/issues)\n
          \   related to observability-analytics-operator\n  * Fixing issues by opening
          [Pull Requests](https://github.com/openshift/analytics-operator/pulls)\n
          \ * Discussing observability-analytics-operator\n\n### License\nobservability-analytics-operator
          is licensed under the Apache 2.0 license.\n"
        displayName: Observability Analytics
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - observability-analytics-operator
        - analytics-operator
        - analytics
        - anomaly
        links:
        - name: Analytics Operator
          url: https://analytics-operator.domain
        maintainers:
        - email: rzalavad@redhat.com
          name: Raj
        maturity: alpha
        provider:
          name: Red Hat Inc.
          url: https://www.redhat.com/
        relatedImages:
        - quay.io/openshiftanalytics/observability-analytics-operator:0.1.1
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/analytics-operator@sha256:6622e769acf5af500d901409575016013d97e88ff53f0b6db0c9f503e6928b8f
        version: 0.1.1
      entries:
      - name: analytics-operator.v0.1.1
        version: 0.1.1
      - name: analytics-operator.v0.1.0
        version: 0.1.0
      name: dev-preview
    defaultChannel: alpha
    packageName: analytics-operator
    provider:
      name: Red Hat Inc.
      url: https://www.redhat.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: odf-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
          createdAt: "2025-09-09T14:18:23Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.18"}]'
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:5189547a19ea81eb18142ad41e6412577eb9ac64d3987ba15b6665a1117a2e55
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:63d8b19d73e5cddcdce67c94e677666d3385ab8dfa1a534b9860fe1db33a2f91
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:a3080bbc8c8633a364a8c9c09951e13f1052be24e2bde3e44c9ec4f3c195c690
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 4.17.13-rhodf
      entries:
      - name: odf-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-flashsystemcluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "flashsystemcluster.odf.ibm.com/v1alpha1",
                  "name": "my-flashsystemcluster",
                  "namespace": "openshift-storage"
                }
              },
              {
                "apiVersion": "odf.openshift.io/v1alpha1",
                "kind": "StorageSystem",
                "metadata": {
                  "name": "my-storagecluster-storagesystem",
                  "namespace": "openshift-storage"
                },
                "spec": {
                  "kind": "storagecluster.ocs.openshift.io/v1",
                  "name": "my-storagecluster",
                  "namespace": "openshift-storage"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Storage
          console.openshift.io/plugins: '["odf-console"]'
          containerImage: registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
          createdAt: "2025-09-11T15:26:46Z"
          description: OpenShift Data Foundation provides a common control plane for
            storage solutions on OpenShift Container Platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.properties: '[{"type": "olm.maxOpenShiftVersion", "value": "4.19"}]'
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operator.openshift.io/uninstall-message: 'Warning: deleting the operand
            (StorageSystem) will result in data loss'
          operatorframework.io/initialization-resource: |-
            {
              "apiVersion": "odf.openshift.io/v1alpha1",
              "kind": "StorageSystem",
              "metadata": {
                "name": "ocs-storagecluster-storagesystem",
                "namespace": "openshift-storage"
              },
              "spec": {
                "kind": "storagecluster.ocs.openshift.io/v1",
                "name": "ocs-storagecluster",
                "namespace": "openshift-storage"
              }
            }
          operatorframework.io/suggested-namespace: openshift-storage
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/operator-type: standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/red-hat-storage/odf-operator
          support: Red Hat
          vendors.odf.openshift.io/kind: '["storagecluster.ocs.openshift.io/v1", "flashsystemcluster.odf.ibm.com/v1alpha1"]'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StorageSystem is the Schema for the storagesystems API
            displayName: Storage System
            kind: StorageSystem
            name: storagesystems.odf.openshift.io
            version: v1alpha1
        description: |
          **Red Hat OpenShift Data Foundation** deploys three operators.

          ### OpenShift Data Foundation
          The OpenShift Data Foundation operator is the primary operator for
          OpenShift Data Foundation. It serves to facilitate the other operators in
          OpenShift Data Foundation by performing administrative tasks outside their
          scope as well as watching and configuring their CustomResources.

          ### Rook
          [Rook][1] deploys and manages Ceph on OpenShift, which provides block,
          file and object storage.

          ### NooBaa operator
          The NooBaa operator deploys and manages the [NooBaa][2] Multi-Cloud Gateway
          on OpenShift, which provides object storage.

          # Core Capabilities
          * **Self-managing service:** No matter which supported storage technologies
          you choose, OpenShift Data Foundation ensures that resources can be
          deployed and managed automatically.

          * **Hyper-scale or hyper-converged:** With OpenShift Data Foundation you
          can either build dedicated storage clusters or hyper-converged clusters
          where your apps run alongside storage.

          * **File, Block, and Object provided by OpenShift Data Foundation:**
          OpenShift Data Foundation integrates Ceph with multiple storage
          presentations including object storage (compatible with S3), block storage,
          and POSIX-compliant shared file system.

          * **Your data, protected:** OpenShift Data Foundation efficiently
          distributes and replicates your data across your cluster to minimize the
          risk of data loss. With snapshots, cloning, and versioning, no more losing
          sleep over your data.

          * **Elastic storage in your datacenter:** Scale is now possible in your
          datacenter. Get started with a few terabytes, and easily scale up.

          * **Simplified data management:** Easily create hybrid and multi-cloud data
          storage for your workloads, using a single namespace.

          # Subscriptions
          OpenShift Data Foundation requires a matching 1:1 ratio with the OpenShift
          subscription used. The Essential subscription is included in the OpenShift
          Platform Plus subscription. Disaster Recovery features, external mode, and
          KMS integration for encryption at rest require a bundle of OpenShift
          Platform Plus with OpenShift Data Foundation Advanced subscription.

          [1]: https://rook.io
          [2]: https://www.noobaa.io
        displayName: OpenShift Data Foundation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - operator
        - data
        - storage
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/odf-operator
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/odf4/odf-operator-bundle@sha256:d45bf7fc74c58bc192a0c9d30157bd745f7d48b2a47820e3b768e8cf29b3c6a0
        - registry.redhat.io/odf4/odf-console-rhel9@sha256:e76c59039b39b3caeabca73353a91501e8374e832c45eaf4005b1c08244e00b1
        - registry.redhat.io/odf4/odf-rhel9-operator@sha256:95cd73d5eba4aaef8c305dc531b069ca927162cd6c1cd25934264294f3be1e79
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odf-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cluster-logging
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cluster-logging.v6.1.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.15-v4.18
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
          createdAt: "2025-05-14T12:38:55Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=5.9.0-0 <6.1.8'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.1
          operators.operatorframework.io.bundle.channels.v1: stable-6.1
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:b4e4d262801b8ff616710075ff05689a8956e975e8ac563e50204a26d1725458
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f7101c962de4e5643bd7f760104a41cc48f020889159a97e2d6711703fbbec64
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:76037c744b34b1c4fccf18f17888813d9670bb6c061c2c3de843aa43c0497170
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:96e438dd5c080cfa47080410f375dc71da24f6d046cf609db41c8b873d85404f
        version: 6.1.8
      entries:
      - name: cluster-logging.v6.1.8
        version: 6.1.8
      - name: cluster-logging.v6.1.7
        version: 6.1.7
      - name: cluster-logging.v6.1.6
        version: 6.1.6
      - name: cluster-logging.v6.1.5
        version: 6.1.5
      - name: cluster-logging.v6.1.4
        version: 6.1.4
      - name: cluster-logging.v6.1.3
        version: 6.1.3
      - name: cluster-logging.v6.1.2
        version: 6.1.2
      - name: cluster-logging.v6.1.1
        version: 6.1.1
      - name: cluster-logging.v6.1.0
        version: 6.1.0
      name: stable-6.1
    - currentCSV: cluster-logging.v6.2.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.16-v4.19
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          createdAt: "2025-07-29T15:25:15Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.0.0-0 <6.2.5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
          operators.operatorframework.io.bundle.channels.v1: stable-6.2
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.


              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Installing logging](https://docs.redhat.com/en/documentation/openshift_container_platform/latest/html-single/logging/index#installing-logging-6-2)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5ac679fe15f6b5ae3c6e52f5864043eca607241fca6b0684b29fe3272d45f816
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:502a5d2891049bbc32115fda28086841351cd9e6a3110b4c3c412f683f2b2491
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:c76c853565fc93da88ba0116425839cfb86cab63cbde9856d7cca78278d7758a
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:afc1e5bdade3ff0e3918fffb8fafb2700dfba7c473f99a2f347cfc7f889c8678
        version: 6.2.5
      entries:
      - name: cluster-logging.v6.2.5
        version: 6.2.5
      - name: cluster-logging.v6.2.4
        version: 6.2.4
      - name: cluster-logging.v6.2.3
        version: 6.2.3
      - name: cluster-logging.v6.2.2
        version: 6.2.2
      - name: cluster-logging.v6.2.1
        version: 6.2.1
      - name: cluster-logging.v6.2.0
        version: 6.2.0
      name: stable-6.2
    - currentCSV: cluster-logging.v6.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "logging.openshift.io/v1alpha1",
                "kind": "LogFileMetricExporter",
                "metadata": {
                  "name": "instance",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "resources": {
                    "limits": {
                      "cpu": "500m"
                    },
                    "requests": {
                      "cpu": "200m",
                      "memory": "128Mi"
                    }
                  },
                  "tolerations": [
                    {
                      "effect": "NoSchedule",
                      "key": "node-role.kubernetes.io/master",
                      "operator": "Exists"
                    }
                  ]
                }
              },
              {
                "apiVersion": "observability.openshift.io/v1",
                "kind": "ClusterLogForwarder",
                "metadata": {
                  "name": "logging",
                  "namespace": "openshift-logging"
                },
                "spec": {
                  "outputs": [
                    {
                      "lokiStack": {
                        "authentication": {
                          "token": {
                            "from": "serviceAccount"
                          }
                        },
                        "target": {
                          "name": "logging-loki"
                        }
                      },
                      "name": "logging-loki",
                      "tls": {
                        "ca": {
                          "configMapName": "openshift-service-ca.crt",
                          "key": "service-ca.crt"
                        }
                      },
                      "type": "lokiStack"
                    }
                  ],
                  "pipelines": [
                    {
                      "inputRefs": [
                        "application",
                        "infrastructure"
                      ],
                      "name": "logs-to-loki",
                      "outputRefs": [
                        "logging-loki"
                      ]
                    }
                  ],
                  "serviceAccount": {
                    "name": "log-collector"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional, Logging & Tracing, Observability
          certified: "false"
          com.redhat.delivery.appregistry: "false"
          com.redhat.delivery.operator.bundle: "true"
          com.redhat.openshift.versions: v4.17-v4.20
          containerImage: registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
          createdAt: "2025-05-14T12:47:05Z"
          description: The Red Hat OpenShift Logging Operator for OCP provides a means
            for configuring and managing log collection and forwarding.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=6.1.0-0 <6.3.1'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/initialization-resource: |
            {
              "apiVersion": "observability.openshift.io/v1",
              "kind": "ClusterLogForwarder",
              "metadata": {
                "name": "logging"
              }
            }
          operatorframework.io/suggested-namespace: openshift-logging
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io.bundle.channel.default.v1: stable-6.3
          operators.operatorframework.io.bundle.channels.v1: stable-6.3
          operators.operatorframework.io.bundle.manifests.v1: manifests/
          operators.operatorframework.io.bundle.mediatype.v1: registry+v1
          operators.operatorframework.io.bundle.metadata.v1: metadata/
          operators.operatorframework.io.bundle.package.v1: cluster-logging
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: AOS Logging (aos-logging@redhat.com)
          target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}'
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |-
              ClusterLogForwarder is an API to configure forwarding logs.

              You configure forwarding by specifying a list of `pipelines`,
              which forward from a set of named inputs to a set of named outputs.
            displayName: Cluster Log Forwarder
            kind: ClusterLogForwarder
            name: clusterlogforwarders.observability.openshift.io
            version: v1
          - description: A Log File Metric Exporter instance. LogFileMetricExporter
              is the Schema for the logFileMetricExporters API
            displayName: Log File Metric Exporter
            kind: LogFileMetricExporter
            name: logfilemetricexporters.logging.openshift.io
            version: v1alpha1
        description: |-
          # Red Hat OpenShift Logging
          The Red Hat OpenShift Logging Operator orchestrates log collection and forwarding to Red Had managed log stores and
          other third-party receivers.

          ##Features
          * **Create/Destroy**: Deploy log collectors and forwarders to support observability of OCP cluster.
          * **Simplified Configuration**: Spec collectors using a simplified API to configure log collection from opinionated sources to one or more third-party receivers.

          ## Prerequisites and Requirements
          ### Red Hat OpenShift Logging Namespace
          It is recommended to deploy the Red Hat OpenShift Logging Operator to the **openshift-logging** namespace. This namespace
          must be explicitly created by a cluster administrator (e.g. `oc create ns openshift-logging`). To enable metrics
          service discovery add namespace label `openshift.io/cluster-monitoring: "true"`.

          For additional installation documentation see [Deploying cluster logging](https://docs.openshift.com/container-platform/latest/logging/cluster-logging-deploying.html)
          in the OpenShift product documentation.
        displayName: Red Hat OpenShift Logging
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - observability
        - loki
        - vector
        - otlp
        links:
        - name: Documentation
          url: https://github.com/openshift/cluster-logging-operator/blob/master/README.adoc
        - name: Red Hat OpenShift Logging Operator
          url: https://github.com/openshift/cluster-logging-operator
        - name: Vector
          url: https://vector.dev/
        minKubeVersion: 1.18.3
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:255a73d34b85f25b3655fb2b63954aa6b635b89080a8c249fc41568fd8365b10
        - registry.redhat.io/openshift-logging/vector-rhel9@sha256:2495de935f503808bf8700d9d66b8412d0f15fc754cf569d02abfbd74b5a6ee8
        - registry.redhat.io/openshift-logging/cluster-logging-operator-bundle@sha256:8c6decb25972dffb37917fb5f5a91ea05fb21b87823eb75807d8c601c12fc1bb
        - registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:5f96aaba248207054bdf1dd8dd8e43a2550a210c62367535fb31fa9907c5ee7a
        version: 6.3.1
      entries:
      - name: cluster-logging.v6.3.1
        version: 6.3.1
      - name: cluster-logging.v6.3.0
        version: 6.3.0
      name: stable-6.3
    defaultChannel: stable-6.3
    packageName: cluster-logging
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NGINX Inc
      provider-url: ""
    name: nginx-ingress-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nginx-ingress-operator.v3.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.nginx.org/v1alpha1",
                "kind": "NginxIngress",
                "metadata": {
                  "name": "nginxingress-sample"
                },
                "spec": {
                  "controller": {
                    "affinity": {},
                    "annotations": {},
                    "appprotect": {
                      "enable": false
                    },
                    "appprotectdos": {
                      "debug": false,
                      "enable": false,
                      "maxDaemons": 0,
                      "maxWorkers": 0,
                      "memory": 0
                    },
                    "autoscaling": {
                      "annotations": {},
                      "behavior": {},
                      "enabled": false,
                      "maxReplicas": 3,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 50,
                      "targetMemoryUtilizationPercentage": 50
                    },
                    "config": {
                      "annotations": {},
                      "entries": {}
                    },
                    "containerPort": {
                      "http": 80,
                      "https": 443
                    },
                    "customConfigMap": "",
                    "customPorts": [],
                    "defaultHTTPListenerPort": 80,
                    "defaultHTTPSListenerPort": 443,
                    "defaultTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    },
                    "disableIPV6": false,
                    "dnsPolicy": "ClusterFirst",
                    "enableCertManager": false,
                    "enableCustomResources": true,
                    "enableExternalDNS": false,
                    "enableLatencyMetrics": false,
                    "enableOIDC": false,
                    "enableSSLDynamicReload": true,
                    "enableSnippets": false,
                    "enableTLSPassthrough": false,
                    "env": [],
                    "extraContainers": [],
                    "globalConfiguration": {
                      "create": false,
                      "spec": {}
                    },
                    "healthStatus": false,
                    "healthStatusURI": "/nginx-health",
                    "hostNetwork": false,
                    "hostPort": {
                      "enable": false,
                      "http": 80,
                      "https": 443
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "nginx/nginx-ingress",
                      "tag": "5.2.0-ubi"
                    },
                    "ingressClass": {
                      "create": true,
                      "name": "nginx",
                      "setAsDefaultIngress": false
                    },
                    "initContainerResources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "initContainers": [],
                    "kind": "deployment",
                    "lifecycle": {},
                    "logFormat": "glog",
                    "logLevel": "info",
                    "mgmt": {
                      "licenseTokenSecretName": "license-token"
                    },
                    "minReadySeconds": 0,
                    "name": "controller",
                    "nginxDebug": false,
                    "nginxReloadTimeout": 60000,
                    "nginxStatus": {
                      "allowCidrs": "127.0.0.1",
                      "enable": true,
                      "port": 8080
                    },
                    "nginxplus": false,
                    "pod": {
                      "annotations": {},
                      "extraLabels": {}
                    },
                    "podDisruptionBudget": {
                      "annotations": {},
                      "enabled": false
                    },
                    "readOnlyRootFilesystem": false,
                    "readyStatus": {
                      "enable": true,
                      "initialDelaySeconds": 0,
                      "port": 8081
                    },
                    "replicaCount": 1,
                    "reportIngressStatus": {
                      "annotations": {},
                      "enable": true,
                      "enableLeaderElection": true,
                      "ingressLink": "",
                      "leaderElectionLockName": "nginx-ingress-leader"
                    },
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "128Mi"
                      }
                    },
                    "selectorLabels": {},
                    "service": {
                      "annotations": {},
                      "clusterIP": "",
                      "create": true,
                      "customPorts": [],
                      "externalIPs": [],
                      "externalTrafficPolicy": "Local",
                      "extraLabels": {},
                      "httpPort": {
                        "enable": true,
                        "port": 80,
                        "targetPort": 80
                      },
                      "httpsPort": {
                        "enable": true,
                        "port": 443,
                        "targetPort": 443
                      },
                      "loadBalancerIP": "",
                      "loadBalancerSourceRanges": [],
                      "type": "LoadBalancer"
                    },
                    "serviceAccount": {
                      "annotations": {},
                      "imagePullSecretName": "",
                      "imagePullSecretsNames": []
                    },
                    "shareProcessNamespace": false,
                    "strategy": {},
                    "terminationGracePeriodSeconds": 30,
                    "tlsPassthroughPort": 443,
                    "tolerations": [],
                    "volumeMounts": [],
                    "volumes": [],
                    "watchNamespace": "",
                    "watchNamespaceLabel": "",
                    "watchSecretNamespace": "",
                    "wildcardTLS": {
                      "cert": "",
                      "key": "",
                      "secret": ""
                    }
                  },
                  "nginxServiceMesh": {
                    "enable": false,
                    "enableEgress": false
                  },
                  "prometheus": {
                    "create": true,
                    "port": 9113,
                    "scheme": "http",
                    "secret": "",
                    "service": {
                      "create": false,
                      "labels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    },
                    "serviceMonitor": {
                      "create": false,
                      "endpoints": [
                        {
                          "port": "prometheus"
                        }
                      ],
                      "labels": {},
                      "selectorMatchLabels": {
                        "service": "nginx-ingress-prometheus-service"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "serviceInsight": {
                    "create": false,
                    "port": 9114,
                    "scheme": "http",
                    "secret": ""
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring, Networking
          certified: "true"
          containerImage: quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
          createdAt: "2025-09-15T16:08:18Z"
          description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
            which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nginx-ingress
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/nginx/nginx-ingress-helm-operator
          support: NGINX Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The `NginxIngress` Custom Resource is the definition of a
              deployment of the Ingress Controller.
            displayName: Nginx Ingress Controller
            kind: NginxIngress
            name: nginxingresses.charts.nginx.org
            version: v1alpha1
        description: The NGINX Ingress Operator is a Kubernetes/OpenShift component
          which deploys and manages one or more NGINX/NGINX Plus Ingress Controllers
        displayName: Nginx Ingress Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nginx
        - ingress-controller
        - ingress
        - controller
        - kubernetes
        - openshift
        links:
        - name: Nginx Ingress Operator
          url: https://github.com/nginx/nginx-ingress-helm-operator
        maintainers:
        - email: kubernetes@nginx.com
          name: NGINX Inc
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: NGINX Inc
        relatedImages:
        - quay.io/nginx/nginx-ingress-operator@sha256:fb372438634b4a667218e725091bd448a253e7a9c40e7dda1044a318780d995f
        - registry.connect.redhat.com/nginx/nginx-ingress-operator-bundle@sha256:4e032549f2c36c5a660994b5ec460ef214bb875a53f6f7e9b814d57049683a7a
        - quay.io/brancz/kube-rbac-proxy@sha256:754ab2a723c8fc7ea8e09ae1224f8af5be91849b8b2f2a3735c6237f1c534187
        version: 3.3.0
      entries:
      - name: nginx-ingress-operator.v3.3.0
        version: 3.3.0
      name: alpha
    defaultChannel: alpha
    packageName: nginx-ingress-operator
    provider:
      name: NGINX Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM Edge
      provider-url: ""
    name: infrastructure-asset-orchestrator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infrastructure-asset-orchestrator.v1.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Generatekey",
                "metadata": {
                  "name": "modelbuilder-token"
                },
                "spec": {}
              },
              {
                "apiVersion": "infrastructure.asset.orchestrator.com/v1",
                "kind": "Mb-broker-service",
                "metadata": {
                  "name": "mb-broker-service-sample"
                },
                "spec": {
                  "env_type": "prod",
                  "cloud_type": "IBM",
                  "external_db": {
                    "enabled": false
                  },
                  "postgres": {
                    "enable_pg_tls": true,
                    "storage_class": "",
                    "storage_size": "20G",
                    "backup_type": "incremental",
                    "backup_frequency": "@daily"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
          createdAt: "2021-08-19T12:00:00Z"
          olm.skipRange: <1.0.2
          operators.operatorframework.io/builder: operator-sdk-v1.12.0+git
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/infrastructure-asset-orchestrator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Generate key CR
            displayName: GenerateKey
            kind: Generatekey
            name: generatekeys.infrastructure.asset.orchestrator.com
            version: v1
          - description: Deploy MB broker deployment
            displayName: MBBrokerService
            kind: Mb-broker-service
            name: mb-broker-services.infrastructure.asset.orchestrator.com
            version: v1
        description: Infrastructure Asset Orchestrator provides users access to cloud
          resources and services. Leverages the dynamic nature of the cloud platforms
          to provide significant cost savings and efficiencies. Using the Infrastructure
          Asset Orchestrator user can provision, configure, utilize and deprovision
          variety of services on the IBM Cloud Catalog.
        displayName: Infrastructure Asset Orchestrator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - cloud-service
        - mbbroker-service
        links:
        - name: Infrastructure Asset Orchestrator
          url: https://infrastructure-asset-orchestrator.domain
        maintainers:
        - email: edge@us.ibm.com
          name: IBM-Edge
        maturity: alpha
        provider:
          name: IBM Edge
        relatedImages:
        - registry.connect.redhat.com/ibm/iao-mb-broker-service@sha256:6531b4196c51246958027453c2d9c7dbcdb4a1f55c826d1da238d256cdbc8a10
        - registry.connect.redhat.com/ibm/iao-operator@sha256:5cc395f22d86641e175339deb134af6c74de92b81470a65a5c736c8005d645fd
        - registry.connect.redhat.com/ibm/iao-orchestrator-service@sha256:145928929fd0c2b7d9ed5fb374d7cc0e6b48269847e810c5472580a1538ad44c
        - registry.connect.redhat.com/ibm/iao-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:609c19d1718947ee814b829334c37e1d720acf5fb642b0712b8ab1b14d57f53f
        - registry.connect.redhat.com/ibm/iao-bundle@sha256:31fd95c5232105d3354515102f099b9b662a8659f08da5d67d4e147d4f934fc7
        version: 1.0.2
      entries:
      - name: infrastructure-asset-orchestrator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: infrastructure-asset-orchestrator-certified
    provider:
      name: IBM Edge
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel® Corporation
      provider-url: ""
    name: pmem-csi-operator-os
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: pmem-csi-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "pmem-csi.intel.com/v1beta1",
                "kind": "PmemCSIDeployment",
                "metadata": {
                  "name": "pmem-csi.intel.com"
                },
                "spec": {
                  "deviceMode": "lvm",
                  "nodeSelector": {
                    "feature.node.kubernetes.io/memory-nv.dax": "true"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
          createdAt: "2021-08-26T10:45:23Z"
          description: An operator for deploying and managing the PMEM-CSI driver
          operators.operatorframework.io/builder: operator-sdk-v1.6.1+git
          operators.operatorframework.io/project_layout: ""
          repository: https://github.com/intel/pmem-csi/tree/release-1.0
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a PMEM-CSI driver deployment in the cluster.
            displayName: PmemCSIDeployment
            kind: PmemCSIDeployment
            name: pmemcsideployments.pmem-csi.intel.com
            version: v1beta1
        description: |2

          Intel PMEM-CSI is a [CSI](https://github.com/container-storage-interface/spec)
          storage driver for container orchestrators like Kubernetes. It makes local
          persistent memory ([PMEM](https://pmem.io/)) available as a filesystem volume to
          container applications. This is the operator to deploy and manage the
          [PMEM-CSI](https://intel.github.io/pmem-csi/1.0/README.html)
          driver on a Kubernetes cluster.

          Refer to the [PMEM-CSI documentation](https://intel.github.io/pmem-csi/1.0/README.html)
          for more details on deploying and using the PMEM-CSI driver.
        displayName: Operator for PMEM-CSI driver
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - PMEM-CSI
        - PMEM-CSI operator
        - Persistent Memory
        - PMEM
        links:
        - name: PMEM-CSI
          url: https://intel.github.io/pmem-csi/1.0
        maintainers:
        - email: olev.kartau@intel.com
          name: Olev Kartau
        - email: patrick.ohly@intel.com
          name: Patrick Ohly
        - email: amarnath.valluri@intel.com
          name: Amarnath Valluri
        maturity: alpha
        minKubeVersion: 1.18.0
        provider:
          name: Intel® Corporation
        relatedImages:
        - registry.connect.redhat.com/intel/pmem-csi-driver-os@sha256:09c98a6588f3d9227f8ae8fc175b533b38590048b79cbdf367eeaa140926010a
        - registry.connect.redhat.com/intel/pmem-csi-operator-os@sha256:72c1ec043ba8bba57e710d1cabac54374612a18cbce05caca9b14310942faa8d
        version: 1.0.0
      entries:
      - name: pmem-csi-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: pmem-csi-operator-os
    provider:
      name: Intel® Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-spark-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: spark-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
          description: Stackable Operator for Apache Spark
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/spark-k8s-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for SparkApplicationSpec via
              `CustomResource`
            displayName: sparkapplications.spark.stackable.tech
            kind: SparkApplication
            name: sparkapplications.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkHistoryServerSpec via
              `CustomResource`
            displayName: sparkhistoryservers.spark.stackable.tech
            kind: SparkHistoryServer
            name: sparkhistoryservers.spark.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for SparkConnectServerSpec via
              `CustomResource`
            displayName: sparkconnectservers.spark.stackable.tech
            kind: SparkConnectServer
            name: sparkconnectservers.spark.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Spark
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - spark-k8s
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/spark-k8s-operator@sha256:27d2cadecc60de35ce855638792c30ab4553447ff6a69b2885001ca141e499ff
        - registry.connect.redhat.com/stackable/stackable-apache-spark-k8s-operator@sha256:916d6051452282647b3eefa7ffea525a8de69fdc2c72c32d87cc56720c7b16e4
        version: 25.7.0
      entries:
      - name: spark-operator.v25.7.0
        version: 25.7.0
      - name: spark-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-spark-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: CSI Addons Community
      provider-url: ""
    name: odf-csi-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odf-csi-addons-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:17:12Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dd20cf6cd0b69404d2ddf5554e7f89870429c117740d2df51f3be5a8163b8db9
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:a4d9350112eaf713a092c2f5c121bf72cf5703f5db3285392d2f653baba3a788
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:26c47d0498f8b8a61338cda3fffb10724c7e0f78d74d14537580763614130256
        version: 4.17.13-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odf-csi-addons-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odf-csi-addons-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odf-csi-addons-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odf-csi-addons-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odf-csi-addons-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odf-csi-addons-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odf-csi-addons-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odf-csi-addons-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odf-csi-addons-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odf-csi-addons-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odf-csi-addons-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odf-csi-addons-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odf-csi-addons-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odf-csi-addons-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:25:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CSIAddonsNode is the Schema for the csiaddonsnode API
            displayName: CSIAddons Node
            kind: CSIAddonsNode
            name: csiaddonsnodes.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationCronJob is the Schema for the encryptionkeyrotationcronjobs
              API
            displayName: EncryptionKeyRotation CronJob
            kind: EncryptionKeyRotationCronJob
            name: encryptionkeyrotationcronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: EncryptionKeyRotationJob is the Schema for the encryptionkeyrotationjobs
              API
            displayName: EncryptionKeyRotation Job
            kind: EncryptionKeyRotationJob
            name: encryptionkeyrotationjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFenceClass is the Schema for the networkfenceclasses
              API
            displayName: Network Fence Class
            kind: NetworkFenceClass
            name: networkfenceclasses.csiaddons.openshift.io
            version: v1alpha1
          - description: NetworkFence is the Schema for the networkfences API
            displayName: Network Fence
            kind: NetworkFence
            name: networkfences.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceCronJob is the Schema for the reclaimspacecronjobs
              API
            displayName: Reclaim Space CronJob
            kind: ReclaimSpaceCronJob
            name: reclaimspacecronjobs.csiaddons.openshift.io
            version: v1alpha1
          - description: ReclaimSpaceJob is the Schema for the reclaimspacejobs API
            displayName: Reclaim Space Job
            kind: ReclaimSpaceJob
            name: reclaimspacejobs.csiaddons.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationClass is the Schema for the volumegroupreplicationclasses
              API
            displayName: Volume Group Replication Class
            kind: VolumeGroupReplicationClass
            name: volumegroupreplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeGroupReplicationContent is the Schema for the volumegroupreplicationcontents
              API
            displayName: Volume Group Replication Content
            kind: VolumeGroupReplicationContent
            name: volumegroupreplicationcontents.replication.storage.openshift.io
            version: v1alpha1
          - kind: VolumeGroupReplication
            name: volumegroupreplications.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplicationClass is the Schema for the volumereplicationclasses
              API
            displayName: Volume Replication Class
            kind: VolumeReplicationClass
            name: volumereplicationclasses.replication.storage.openshift.io
            version: v1alpha1
          - description: VolumeReplication is the Schema for the volumereplications
              API
            displayName: Volume Replication
            kind: VolumeReplication
            name: volumereplications.replication.storage.openshift.io
            version: v1alpha1
        description: CSI Addons provides the CSI Addons Controller that enables advanced
          storage operations for CSI-drivers.
        displayName: CSI Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - storage
        - csi
        links:
        - name: CSI Addons
          url: https://csi-addons.github.io
        maintainers:
        - email: csi-addons@redhat.com
          name: CSI Addons Community
        maturity: alpha
        provider:
          name: CSI Addons Community
        relatedImages:
        - registry.redhat.io/odf4/odf-csi-addons-rhel9-operator@sha256:dc3cbef1d890b1be5cd59d4e2fcb91f4f88e2aa53bf59c78a2cbe9af136ec3cd
        - registry.redhat.io/odf4/odf-csi-addons-sidecar-rhel9@sha256:e61bb9f32f380ae4b65e3855011dbd6d5f8fa6c0f53d0ef4fb91b2ea0710f2ba
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        - registry.redhat.io/odf4/odf-csi-addons-operator-bundle@sha256:c3e9d2deb1454fbd423ff9eeca34225df86f92b536c962300a75158505341678
        version: 4.18.11-rhodf
      entries:
      - name: odf-csi-addons-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odf-csi-addons-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odf-csi-addons-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odf-csi-addons-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odf-csi-addons-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odf-csi-addons-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odf-csi-addons-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odf-csi-addons-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odf-csi-addons-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odf-csi-addons-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odf-csi-addons-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odf-csi-addons-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odf-csi-addons-operator
    provider:
      name: CSI Addons Community
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Altair Engineering Inc.
      provider-url: ""
    name: anzograph-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anzograph-operator.v3.2.500
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "anzograph.clusters.cambridgesemantics.com/v2",
                "kind": "AnzoGraph",
                "metadata": {
                  "name": "azg01"
                },
                "spec": {
                  "db": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_data": "anzograph-data-grpc",
                            "app_mgmt": "anzograph-mgmt-grpc"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_data": "anzograph-data-grpc",
                              "app_mgmt": "anzograph-mgmt-grpc"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e",
                                "name": "db",
                                "resources": {
                                  "limits": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  },
                                  "requests": {
                                    "cpu": "8000m",
                                    "memory": "8Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  },
                  "deployFrontend": false,
                  "frontend": {
                    "nodeConfig": {
                      "spec": {
                        "replicas": 1,
                        "selector": {
                          "matchLabels": {
                            "app_mgmt": "anzograph-frontend"
                          }
                        },
                        "serviceName": "anzograph-azg01",
                        "template": {
                          "metadata": {
                            "labels": {
                              "app_mgmt": "anzograph-frontend"
                            }
                          },
                          "spec": {
                            "containers": [
                              {
                                "image": "registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f",
                                "name": "frontend",
                                "resources": {
                                  "limits": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  },
                                  "requests": {
                                    "cpu": "2000m",
                                    "memory": "4Gi"
                                  }
                                }
                              }
                            ],
                            "securityContext": {
                              "fsGroup": 1000,
                              "runAsNonRoot": true,
                              "runAsUser": 1000,
                              "seccompProfile": {
                                "type": "RuntimeDefault"
                              }
                            },
                            "serviceAccountName": "anzograph-operator"
                          }
                        }
                      }
                    }
                  }
                },
                "status": {}
              }
            ]
          capabilities: Basic Install
          categories: Database, Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
          createdAt: "2025-09-22T12:03:18Z"
          description: kubernetes operator for Altair Graph Lakehouse DB
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.38.0
          operators.operatorframework.io/project_layout: unknown
          repository: ""
          support: Altair Engineering Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AnzoGraph is the Schema for the anzographs API
            displayName: Altair Graph Lakehouse Operator
            kind: AnzoGraph
            name: anzographs.anzograph.clusters.cambridgesemantics.com
            version: v2
        description: |-
          The Altair Graph Lakehouse Operator provides the way to install and configure an Altair Graph Lakehouse
          cluster on Red Hat K8S environment.

          ### Installation
           Refer [installation instructions]( https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md )

          ### Documentation

          You can find our documentation [here.]( https://docs.cambridgesemantics.com/anzograph/userdoc/ )

          ### Support

          We offer Support to our customers with the Altair Graph Lakehouse db Enterprise Edition License [here]( https://altair.com/customer-support ) or by emailing the Data Analytics Support team at dasupport@altair.com.
        displayName: Altair Graph Lakehouse Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Big Data
        - Database
        links:
        - name: Anzograph Operator
          url: https://github.com/cambridgesemantics/csi-k8s-operator-anzograph/blob/v3.2.5/README_openshift_marketplace.md
        maintainers:
        - email: fschwichtenberg@altair.com
          name: Frank Schwichtenberg
        maturity: stable
        provider:
          name: Altair Engineering Inc.
        relatedImages:
        - registry.connect.redhat.com/cambridgesemantics/anzograph-db@sha256:30ed2ec337135701ab8b54e75bb48ad8928a2c48818643c23b9dba4484255a2e
        - registry.connect.redhat.com/cambridgesemantics/anzograph-frontend@sha256:e5446b3c04889f2189e955702c4165eb243eb89f47ac797eb417b60b1459647f
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator-bundle@sha256:a7b60f1c65a6fb4ee0109c7b2b0bd12fb9487e2f6edb059bd18cfd2337ea9116
        - registry.connect.redhat.com/cambridgesemantics/anzograph-operator@sha256:2333154a14ab7f3a03a601f00c017feeafb9d9814aff5e59c95cef615ab2abd6
        version: 3.2.500
      entries:
      - name: anzograph-operator.v3.2.500
        version: 3.2.500
      - name: anzograph-operator.v3.2.300
        version: 3.2.300
      - name: anzograph-operator.v3.2.200
        version: 3.2.200
      - name: anzograph-operator.v3.2.104
        version: 3.2.104
      name: stable
    defaultChannel: stable
    packageName: anzograph-operator
    provider:
      name: Altair Engineering Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apicurio
      provider-url: ""
    name: apicurio-api-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: apicurio-api-controller.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "registry.apicur.io/v1",
                "kind": "ApicurioRegistry3",
                "metadata": {
                  "name": "example-registry-with-studio"
                },
                "spec": {
                  "app": {
                    "host": "example-registry-with-studio-app.apps.yourcluster.example"
                  },
                  "studioUi": {
                    "enabled": true,
                    "host": "example-registry-with-studio-studio-ui.apps.yourcluster.example"
                  },
                  "ui": {
                    "host": "example-registry-with-studio-ui.apps.yourcluster.example"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/apicurio/apicurio-registry-operator:3.0.5
          createdAt: "2024-12-03T11:30:52Z"
          description: Deploy and manage Apicurio API Controller on Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '[]'
          operators.operatorframework.io/builder: operator-sdk-v1.37.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/Apicurio/apicurio-registry/tree/main/operator
          support: Apicurio
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ApicurioRegistry3 represents an instance of Apicurio API
              Controller.
            displayName: Apicurio Registry 3
            kind: ApicurioRegistry3
            name: apicurioregistries3.registry.apicur.io
            version: v1
        description: |
          ## Apicurio API Controller

          Apicurio API Controller consists of 2 components:
           - Apicurio Registry 3
           - Apicurio Studio

          Storage options: PostgreSQL, Kafka, MySQL, SQL Server, in-memory.

          ## Apicurio API Controller Operator

          Provides a quick and easy way to deploy and manage Apicurio API Controller on Kubernetes.

          **Features**
          - Supports installation and configuration of Apicurio API Controller
          - Perform a rolling upgrade of Apicurio API Controller

          ## Prerequisites

          This operator does not deploy storage for Apicurio API Controller. Therefore, some storage options require that the chosen persistence service is already set up.

          ## License

          Apicurio API Controller and Apicurio API Controller Operator are licensed under the [Apache 2.0 license](https://github.com/Apicurio/apicurio-registry/blob/main/LICENSE)
        displayName: Apicurio API Controller
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - API Management
        - Apicurio Registry
        - Apicurio Studio
        links:
        - name: Website
          url: https://www.apicur.io/
        - name: GitHub
          url: https://github.com/Apicurio/apicurio-registry/
        - name: Issues
          url: https://github.com/Apicurio/apicurio-registry/issues
        maintainers:
        - email: apicurio@lists.jboss.org
          name: Apicurio
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Apicurio
        relatedImages:
        - quay.io/apicurio/apicurio-registry-operator:3.0.5
        - quay.io/community-operator-pipeline-prod/apicurio-api-controller@sha256:c783cd6db1ae5a6152848f5e4258e89b37077a177199be905d0cfe451a48e5ef
        version: 0.0.1
      entries:
      - name: apicurio-api-controller.v0.0.1
        version: 0.0.1
      name: 0.x
    defaultChannel: 0.x
    packageName: apicurio-api-controller
    provider:
      name: Apicurio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Fossul Framework
      provider-url: https://github.com/fossul/fossul
    name: fossul-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fossul-operator.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Backup\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n    \"kind\":
            \"Restore\",\n    \"metadata\": {\n      \"name\": \"mariadb-sample\"\n
            \   },\n    \"spec\": {\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\",\n      \"workflow_id\": \"xxxx\"\n    }\n  },        \n  {\n
            \   \"apiVersion\": \"fossul.io/v1\",\n    \"kind\": \"BackupConfig\",\n
            \   \"metadata\": {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\":
            {\n      \"container_platform\": \"openshift\",\n      \"operator_controlled\":
            true,\n      \"auto_discovery\": false,\n      \"deployment_name\": \"mariadb\",\n
            \     \"deployment_type\": \"DeploymentConfig\",\n      \"job_retention\":
            50,\n      \"overwrite_pcv_on_restore\": true,\n      \"policies\": [\n
            \       {\n          \"policy\": \"hourly\",\n          \"retentionNumber\":
            3\n        },\n        {\n          \"policy\": \"daily\",\n          \"retentionNumber\":
            10\n        }\n      ],\n      \"pvc_deletion_timeout\": 300,\n      \"restore_to_new_pvc\":
            false,\n      \"snapshot_timeout\": 180,\n      \"storage_plugin\": \"csi.so\",\n
            \     \"app_plugin\": \"mariadb.so\"\n    }\n  },\n  {\n    \"apiVersion\":
            \"fossul.io/v1\",\n    \"kind\": \"BackupSchedule\",\n    \"metadata\":
            {\n      \"name\": \"mariadb-sample\"\n    },\n    \"spec\": {\n      \"cron_schedule\":
            \"59 23 * * *\",\n      \"deployment_name\": \"mariadb\",\n      \"policy\":
            \"daily\"\n\n    }\n  },\n  {\n    \"apiVersion\": \"fossul.io/v1\",\n
            \   \"kind\": \"Fossul\",\n    \"metadata\": {\n      \"name\": \"fossul-sample\"\n
            \   },\n    \"spec\": {\n      \"enable_routes\": false,\n      \"enable_virtualmachines\":
            false,\n      \"container_platform\": \"openshift\"\n    }\n  }\n]"
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/fossul/fossul-operator:v0.6.0
          createdAt: "2022-01-25T20:09:00Z"
          description: Fossul is a application centric container native backup and
            restore framework for databases or applications
          operators.operatorframework.io/builder: operator-sdk-v1.10.1-ocp
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/fossul/fossul
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Manages application backup configurations
            kind: BackupConfig
            name: backupconfigs.fossul.io
            version: v1
          - description: Manages application backups
            kind: Backup
            name: backups.fossul.io
            version: v1
          - description: Manages application restores
            kind: Restore
            name: restores.fossul.io
            version: v1
          - description: Manages application backup scheduling
            kind: BackupSchedule
            name: backupschedules.fossul.io
            version: v1
          - description: Manages deployment of the Fossul Framework
            kind: Fossul
            name: fossuls.fossul.io
            version: v1
        description: "## Fossul Framework\nApplication backup and restore framework
          for container native applications on Kubernetes and OpenShift. Fossul provides
          application centric backup and recovery enabling integration with the application
          creating an application consistent backup. Since there are many applications
          and providers for storage or archive Fossul enables plugins that allow proper
          integration. Fossul provides plugins for postgresql, mariadb, mysql and
          mongo as well as plugin to integrate with CSI snapshots. Plugins can be
          written in any language. The Fossul operator deploys Fossul via the Fossul
          custom resource and additionally provides BackupConfig, Backup, BackupSchedule
          and Restore custom resources. Each user can configure and manage their own
          backups with simple custom resources that can be part of CI/CD or source
          control. Users can also opt to manage backups through Fossul CLI/API instead
          of the custom resources provided by the operator.\n\n## Installation\nIt
          is recommended to install the operator into the openshift-operators namespace
          which is the default location for operators. Once the operator is installed
          create a new project called ```fossul``` and create the Fossul custom resource
          under the fossul namespace. This will deploy the Fossul Framework. You can
          choose another namespace for the Fossul Framework but then also need to
          provide the optional 'fossul_namespace' parameter with the spec for all
          the additional custom resources. This is required so the Fossul operator
          can communicate with Fossul.\n\nOnce Fossul is deployed you can manage backups
          either through custom resources provided by the operator or the Fossul CLI/API.
          If using the customer resources, first create a BackupConfig custom resource
          for every application you want to backup. Once that is done, you can optionally
          create a backup by simply creating a Backup custom resource. You can also
          create a BackupSchedule custom resource using cron syntax which will schedule
          your backup to occur on a regular interval or do so via a Kubernetes job.
          Finally if a backup has been created you can perform a restore through a
          custom resource as well, providing the workflow id from the completed backup.
          If you are creating backups through the custom resource the workflow id
          will be appended to the spec once the backup completes. If you are using
          the Fossul CLI/API to create backups you will need to get the workflow id
          through the Fossul CLI/API directly using the backupList action.\n\nFor
          users of the cli, a cli pod is deployed with credentials and if desired
          you can create or manage backups from within that pod. To use the API you
          need the credentials stored in the fossul secret. The cli has already been
          pre-configured with these credentials. The API can also be optionally exposed
          via routes and API documentation is under ```/api/v1/index.html```.\n\n##
          Quick Setup\nWhen creating Fossul custom resources generally you will only
          need to update the ```deployment_name``` and ```deployment_type``` (DeploymentConfig,
          Deployment or VirtualMachine). All custom resources must be created in the
          namespace where the database or application exists. \n\n### Deploy Fossul
          in fossul namespace\n```$ oc new-project fossul```\n<pre>\n$ vi fossul.yaml\nkind:
          Fossul\napiVersion: fossul.io/v1\nmetadata:\n  name: fossul-sample\n  namespace:
          fossul\nspec: {}\n</pre>\n```$ oc create -f fossul.yaml```\n\n### Deploy
          Mariadb Database\nMake sure you create all Fossul custom resources in the
          same namespace as the database, application or virtual machine\n\n### Create
          MariaDB Fossul BackupConfig\nCreates a backup configuration for MariaDB
          or MySQL databases\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true\n  auto_discovery:
          false\n  deployment_name: mariadb\n  deployment_type: DeploymentConfig\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: mariadb.so \n</pre>\n```$ oc
          create -f backupconfig.yaml```\n\n### Create PostgreSQL Fossul BackupConfig\nCreates
          a backup config for PostgreSQL databases, you need to ensure the user defined
          in secret has backup permissions\n\n```$ vi backupconfig.yaml```\n<pre>\nkind:
          BackupConfig\napiVersion: fossul.io/v1\nmetadata:\n  name: postgresql-sample\n
          \ namespace: databases\nspec:\n  container_platform: openshift\n  operator_controlled:
          true    \n  auto_discovery: false\n  deployment_name: postgresql \n  deployment_type:
          DeploymentConfig\n  job_retention: 50\n  overwrite_pcv_on_restore: true\n
          \ policies:\n    - policy: hourly\n      retentionNumber: 3\n    - policy:
          daily\n      retentionNumber: 10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc:
          false\n  snapshot_timeout: 180\n  storage_plugin: csi.so\n  app_plugin:
          postgres.so \n</pre>\n```$ oc create -f backupconfig.yaml```\n\n### Create
          Kubevirt Fossul BackupConfig\nCreates a backup configuration for virtual
          machines\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: rhel8-vm-sample\n  namespace: virtualmachines\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: rhel8-vm\n  deployment_type: VirtualMachine\n
          \ job_retention: 50\n  overwrite_pcv_on_restore: true\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  pvc_deletion_timeout: 300\n  restore_to_new_pvc: false\n  snapshot_timeout:
          180\n  storage_plugin: csi.so\n  app_plugin: kubevirt.so \n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create Tar Fossul BackupConfig\nCreates
          a backup config using tar to backup folders and files where csi snapshots
          are not desired\n\n```$ vi backupconfig.yaml```\n<pre>\nkind: BackupConfig\napiVersion:
          fossul.io/v1\nmetadata:\n  name: tar-sample\n  namespace: databases\nspec:\n
          \ container_platform: openshift\n  operator_controlled: true    \n  auto_discovery:
          false\n  deployment_name: anyname\n  job_retention: 50\n  policies:\n    -
          policy: hourly\n      retentionNumber: 3\n    - policy: daily\n      retentionNumber:
          10\n  backup_source_paths: /opt/app-root/src/test\n  storage_plugin: tar.so\n</pre>\n```$
          oc create -f backupconfig.yaml```\n\n### Create BackupSchedule\nA backup
          can be scheduled per policy, defined by backup configuration using cron
          syntax\n\n```$ vi backupschedule.yaml```\n<pre>\nkind: BackupSchedule\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ cron_schedule: '59 23 * * *'\n  deployment_name: mariadb\n  policy: daily\n</pre>\n```$
          oc create -f backupschedule.yaml```\n\n### Create Backup\nA backup will
          be created according to policy and deleted either manually or as defined
          in retention policy\n\n```$ vi backup.yaml```\n<pre>\nkind: Backup\napiVersion:
          fossul.io/v1\nmetadata:\n  name: mariadb-sample\n  namespace: databases\nspec:\n
          \ deployment_name: mariadb\n  policy: daily\n</pre>\n```$ oc create -f backup.yaml```\n\n###
          Perform Restore\nA restore requires a workflow_id located in the backup
          spec, upon a successful restore the custom resource will deleted automatically\n\n```$
          vi restore.yaml```\n<pre>\nkind: Restore\napiVersion: fossul.io/v1\nmetadata:\n
          \ name: mariadb-sample\n  namespace: databases\nspec:\n  deployment_name:
          mariadb\n  policy: daily\n  workflow_id: xxxx\n</pre>\n```$ oc create -f
          restore.yaml```\n"
        displayName: Fossul Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - application
        - consistent
        links:
        - name: Fossul Framework Community
          url: https://github.com/fossul/fossul
        maintainers:
        - email: ktenzer@redhat.com
          name: Keith Tenzer
        maturity: alpha
        provider:
          name: Fossul Framework
          url: https://github.com/fossul/fossul
        relatedImages:
        - quay.io/fossul/fossul-operator:v0.6.0
        - quay.io/openshift-community-operators/fossul-operator@sha256:95c043ddc818f44a303e27508e0acdcbbcb3d9b1329712c85a8e3fd4cb647e9a
        version: 0.6.0
      entries:
      - name: fossul-operator.v0.6.0
        version: 0.6.0
      - name: fossul-operator.v0.5.0
        version: 0.5.0
      - name: fossul-operator.v0.4.0
        version: 0.4.0
      - name: fossul-operator.v0.3.0
        version: 0.3.0
      - name: fossul-operator.v0.2.0
        version: 0.2.0
      - name: fossul-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: fossul-operator
    provider:
      name: Fossul Framework
      url: https://github.com/fossul/fossul
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: The Flux project
      provider-url: ""
    name: flux
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: flux.v2.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "helm.toolkit.fluxcd.io/v2beta1",
                "kind": "HelmRelease",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "5m",
                    "chart": {
                      "spec": {
                          "chart": "podinfo",
                          "version": "4.0.x",
                          "sourceRef": {
                            "kind": "HelmRepository",
                            "name": "podinfo",
                            "namespace": "flux-system"
                          },
                          "interval": "1m"
                      }
                    },
                    "values": {
                      "replicaCount": 2
                    }
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "GitRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "1m",
                    "ref": {
                      "branch": "master"
                    },
                    "url": "https://github.com/stefanprodan/podinfo"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmRepository",
                "metadata": {
                    "name": "stable",
                    "namespace": "default"
                },
                "spec": {
                    "url": "https://kubernetes-charts.storage.googleapis.com/",
                    "interval": "10m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "HelmChart",
                "metadata": {
                    "name": "redis",
                    "namespace": "default"
                },
                "spec": {
                    "chart": "redis",
                    "version": "10.5.7",
                    "sourceRef": {
                      "name": "stable",
                      "kind": "HelmRepository"
                    },
                    "interval": "5m"
                }
              },
              {
                "apiVersion": "source.toolkit.fluxcd.io/v1beta1",
                "kind": "Bucket",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "interval": "1m",
                    "provider": "generic",
                    "bucketName": "podinfo",
                    "endpoint": "minio.minio.svc.cluster.local:9000",
                    "insecure": true,
                    "secretRef": {
                      "name": "minio-credentials"
                    }
                }
              },
              {
                "apiVersion": "kustomize.toolkit.fluxcd.io/v1beta2",
                "kind": "Kustomization",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "dev"
                },
                "spec": {
                    "interval": "5m",
                    "path": "./deploy/overlays/dev",
                    "prune": true,
                    "sourceRef": {
                      "kind": "GitRepository",
                      "name": "podinfo"
                    },
                    "timeout": "2m",
                    "wait": true
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Provider",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "address": "https://github.com/stefanprodan/podinfo",
                    "secretRef": {
                      "name": "api-token"
                    }
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Alert",
                "metadata": {
                    "name": "webapp",
                    "namespace": "default"
                },
                "spec": {
                    "providerRef": {
                      "name": "on-call-slack"
                    },
                    "eventSeverity": "info",
                    "eventSources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "Bucket",
                          "name": "secrets"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-backend"
                      },
                      {
                          "kind": "Kustomization",
                          "name": "webapp-frontend"
                      }
                    ]
                }
              },
              {
                "apiVersion": "notification.toolkit.fluxcd.io/v1beta1",
                "kind": "Receiver",
                "metadata": {
                    "name": "github-receiver",
                    "namespace": "default"
                },
                "spec": {
                    "type": "github",
                    "events": [
                      "ping",
                      "push"
                    ],
                    "secretRef": {
                      "name": "webhook-token"
                    },
                    "resources": [
                      {
                          "kind": "GitRepository",
                          "name": "webapp"
                      },
                      {
                          "kind": "HelmRepository",
                          "name": "webapp"
                      }
                    ]
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageRepository",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "image": "ghcr.io/stefanprodan/podinfo",
                    "interval": "1m0s"
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImagePolicy",
                "metadata": {
                    "name": "podinfo",
                    "namespace": "flux-system"
                },
                "spec": {
                    "imageRepositoryRef": {
                      "name": "podinfo"
                    },
                    "policy": {
                      "semver": {
                          "range": "5.0.x"
                      }
                    }
                }
              },
              {
                "apiVersion": "image.toolkit.fluxcd.io/v1beta1",
                "kind": "ImageUpdateAutomation",
                "metadata": {
                  "name": "flux-system",
                  "namespace": "flux-system"
                },
                "spec": {
                  "interval": "1m0s",
                  "sourceRef": {
                    "kind": "GitRepository",
                    "name": "flux-system"
                  },
                  "git": {
                    "checkout": {
                      "ref": {
                        "branch": "main"
                      }
                    },
                    "commit": {
                      "author": {
                        "email": "fluxcdbot@users.noreply.github.com",
                        "name": "fluxcdbot"
                      },
                      "messageTemplate": "{{range .Updated.Images}}{{println .}}{{end}}"
                    },
                    "push": {
                      "branch": "main"
                    }
                  },
                  "update": {
                    "path": "./clusters/my-cluster",
                    "strategy": "Setters"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
          description: Flux is a Continuous Delivery solution for Kubernetes.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/flux/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/flux/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: flux-system
          support: The Flux community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Alert
            displayName: Alert
            kind: Alert
            name: alerts.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta1
          - description: Bucket
            displayName: Bucket
            kind: Bucket
            name: buckets.source.toolkit.fluxcd.io
            version: v1beta2
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: GitRepository
            displayName: GitRepository
            kind: GitRepository
            name: gitrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmChart
            displayName: HelmChart
            kind: HelmChart
            name: helmcharts.source.toolkit.fluxcd.io
            version: v1beta2
          - description: HelmRelease
            displayName: HelmRelease
            kind: HelmRelease
            name: helmreleases.helm.toolkit.fluxcd.io
            version: v2beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta1
          - description: HelmRepository
            displayName: HelmRepository
            kind: HelmRepository
            name: helmrepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImagePolicy
            displayName: ImagePolicy
            kind: ImagePolicy
            name: imagepolicies.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta1
          - description: ImageRepository
            displayName: ImageRepository
            kind: ImageRepository
            name: imagerepositories.image.toolkit.fluxcd.io
            version: v1beta2
          - description: ImageUpdateAutomation
            displayName: ImageUpdateAutomation
            kind: ImageUpdateAutomation
            name: imageupdateautomations.image.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta1
          - description: Kustomization
            displayName: Kustomization
            kind: Kustomization
            name: kustomizations.kustomize.toolkit.fluxcd.io
            version: v1beta2
          - description: OCIRepository
            displayName: OCIRepository
            kind: OCIRepository
            name: ocirepositories.source.toolkit.fluxcd.io
            version: v1beta2
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Provider
            displayName: Provider
            kind: Provider
            name: providers.notification.toolkit.fluxcd.io
            version: v1beta2
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta1
          - description: Receiver
            displayName: Receiver
            kind: Receiver
            name: receivers.notification.toolkit.fluxcd.io
            version: v1beta2
        description: |-
          [Flux](https://fluxcd.io/) is a Continuous Delivery solution for Kubernetes.
          Flux is constructed with the [GitOps Toolkit](https://toolkit.fluxcd.io/components/), a set of composable APIs and specialized tools for keeping Kubernetes clusters in sync with sources of configuration (like Git & Helm repositories), and automating updates to configuration when there is new code to deploy.

          Flux is a CNCF incubation project made for:

          - **cluster operators** who automate provision and configuration of clusters
          - **platform engineers** who build continuous delivery for developer teams
          - **app developers** who rely on continuous delivery to get their code live

          Features:

          - Source configuration from Git and Helm repositories, and S3-compatible buckets (e.g., Minio)
          - Kustomize and Helm support
          - Event-triggered and periodic reconciliation
          - Integration with Kubernetes RBAC
          - Health assessment (clusters and workloads)
          - Dependency management (infrastructure and workloads)
          - Secrets management (builtin support for Mozilla SOPS and Cloud KMS)
          - Alerting to external systems (webhook senders)
          - External events handling (webhook receivers)
          - Automated container image updates to Git (image scanning and patching)
          - Policy-driven validation (OPA, Kyverno, admission controllers)
          - Seamless integration with Git providers (GitHub, GitLab, Bitbucket)
          - Interoperability with Cluster API (CAPI) providers

          ### Steps to support bootstraping with Flux CLI

          SecComp Profile should be removed as it requires a special privilege to get the access.
          Also, it requires to enforce `runUserAs` to the same UID provided by the images to prevent OpenShift
          to alter the user id expected by our controllers.

          To do so, please specifiy a patch using the following Kustomization file, for example.

          ```yaml
          apiVersion: kustomize.config.k8s.io/v1beta1
          kind: Kustomization
          resources:
            - gotk-components.yaml
            - gotk-sync.yaml
          patches:
            - patch: |
                apiVersion: apps/v1
                kind: Deployment
                metadata:
                  name: all
                spec:
                  template:
                    spec:
                      containers:
                        - name: manager
                          securityContext:
                            runAsUser: 65534
                            seccompProfile:
                              $patch: delete
              target:
                kind: Deployment
                labelSelector: app.kubernetes.io/part-of=flux
          ```

          Then before installing Flux, please also perform the following steps to bind Flux to the `nonroot` SecurityContextConstraint (SCC). This applies to installation via OC's OLM or via Flux CLI.

          ```sh
          # replace default "flux-system" with the project/namespace name for your flux installation
          namespace="flux-system"
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:source-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:kustomize-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:helm-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:notification-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-automation-controller
          oc adm policy add-scc-to-user nonroot system:serviceaccount:$namespace:image-reflector-controller

          ```

          To see complete steps please refer to this [documentation](https://github.com/fluxcd/website/blob/main/content/en/docs/use-cases/openshift.md).

          If you've already installed Flux, you might need to uninstall it after applying the above pre-requisite steps.

          ### Network Policies

          This version of Flux is shipped without the normally provided network Policies.
          By default all traffic is enabled, and anyone can access the source-controller
          artifacts.

          It is not recommended to run it in production as is.

          Please visit [Flux's policies](https://github.com/fluxcd/flux2/tree/main/manifests/policies)
          to deploy the needed network policies and secure your Flux setup.

          ### Breaking Changes

          - Flux 0.18.x ships with v1beta2 of the kustomize.toolkit.fluxcd.io/Kustomization CRD.
          - Flux 0.25.x is the last Flux release where Kubernetes 1.19 is supported.

          ### Flux on OpenShift Use Cases

          - [The GitOps with Flux + OpenShift](https://www.youtube.com/watch?v=8ybCtNVP5To) talk by Andrew Block (Red Hat) showing how to setup multi-tenacy on OpenShift using Flux.
          - [Flux OpenShift Use Case](https://fluxcd.io/docs/use-cases/openshift/) on FluxCD.io
        displayName: Flux
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - flux
        - gitops
        - cd
        - continuous delivery
        - fluxcd
        - helm
        - kustomize
        links:
        - name: Homepage
          url: https://fluxcd.io/
        - name: Documentation
          url: https://fluxcd.io/docs/
        - name: Slack
          url: https://cloud-native.slack.com/messages/flux
        - name: Mailing List
          url: https://lists.cncf.io/g/cncf-flux-dev
        - name: Source Code
          url: https://github.com/fluxcd/flux2
        maintainers:
        - email: support@weave.works
          name: The Flux maintainers
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: The Flux project
        relatedImages:
        - registry.connect.redhat.com/flux/image-automation-controller@sha256:0c51595307e14acc15f22d50b414a87513ebb03530ef4a8f7d6cd9f8b757243e
        - registry.connect.redhat.com/flux/image-reflector-controller@sha256:3bef7fde22be641304e5ca69f77150e416eba5141f8aa13f68be7f810e55aa59
        - registry.connect.redhat.com/flux/kustomize-controller@sha256:246290908d899b9f6feb920f88913bd2d4673cf852a3020714fd4152c7f78237
        - registry.connect.redhat.com/flux/notification-controller@sha256:b3d81234ba3312e0106f2f7cabc73fbe1d7bf683eda36b9ed5bbdaa43789be40
        - registry.connect.redhat.com/flux/source-controller@sha256:b576e3f010e3ff07e716064ab0dad042aba93f0ea5c250535dd18bd678c45337
        - registry.connect.redhat.com/flux/flux-operator@sha256:13c2f3b8092fc7b84548dd6383edbb8e41b7a8f3acefe63620f07f7a6454d8b9
        - registry.connect.redhat.com/flux/helm-controller@sha256:cc24a7b73b023ad86a6c3a845088d153ea290dda7cab1ec03eb85aaf814e6049
        version: 2.0.1
      entries:
      - name: flux.v2.0.1
        version: 2.0.1
      name: stable
    defaultChannel: stable
    packageName: flux
    provider:
      name: The Flux project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: model-builder-for-vision-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-modelbuilder-for-vision.v1.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "Modelbuilder",
                "metadata": {
                  "name": "modelbuilder-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "backup_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "metadata_storage": {
                    "storage_class": "ibmc-block-bronze",
                    "storage_size": "20G"
                  },
                  "photo_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "500G"
                  },
                  "in_memory_storage": {
                    "storage_class": "ibmc-file-gold-gid",
                    "storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              },
              {
                "apiVersion": "modelbuilder.com/v1alpha1",
                "kind": "ModelbuilderAWS",
                "metadata": {
                  "name": "modelbuilder-aws-sample"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "postgres": {
                    "storage_size": "20G",
                    "storage_class": "gp2",
                    "backup_storage_size": "20G"
                  },
                  "vm_request_method": "dynamic",
                  "env_type": "prod",
                  "email_notification": {
                    "accept": true,
                    "vm_mail_recipient": ""
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
          createdAt: "2021-08-28T12:00:00Z"
          olm.skipRange: <1.0.3
          operators.operatorframework.io/builder: operator-sdk-v1.11.0
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          support: Refer Readme - https://github.com/rhm-samples/ibm-modelbuilder-for-vision/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy Model Builder for Vision on AWS
            displayName: Deploy-Modelbuilder on AWS
            kind: ModelbuilderAWS
            name: modelbuildersaws.modelbuilder.com
            version: v1alpha1
          - description: Deploy Model Builder for Vision on IBM
            displayName: Deploy-Modelbuilder on IBM
            kind: Modelbuilder
            name: modelbuilders.modelbuilder.com
            version: v1alpha1
        description: IBM Services Software Model Builder for Vision is an AI training
          platform that can use the capabilities of IBM Cloud and Amazon Web Services
          and their GPUs to quickly train computer vision models for compatible mobile
          apps. IBM Services Software Inspection Workbench iOS and iPadOS app will
          be used as the exclusive UI for labeling and training. Specifically, IBM
          Model Builder for Vision will train computer vision models that can then
          be deployed for use on the IBM Services Software Inspector Portable , IBM
          Services Software Inspector Wearable and the IBM Maximo Visual Inspection
          Mobile apps.
        displayName: IBM Model Builder for Vision
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Model Builder
        - AI
        - Computer Vision
        - Inspection
        - Mobile
        - Core ML
        - TensorFlow Lite
        links:
        - name: IBM Model Builder for Vision
          url: http://ibm.biz/model_builder
        maintainers:
        - email: isshelp@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-12@sha256:82d171ab0ce78a0157408662155b53d4f637947a303bfecb684f6132f5f468be
        - registry.redhat.io/rhel8/redis-6@sha256:2f65f46b25d84a3cb41a6a10f93786bcab47b7354ca9ac13a8e602ea833eba60
        - registry.connect.redhat.com/ibm/modelbuilder-augmentor@sha256:ef5d0dde3eb334d919c2773af63007f1ef5e735ae040ef77fffe3440f854f068
        - registry.connect.redhat.com/ibm/modelbuilder-auth-service@sha256:f96b444fdcf7781775c9eba942bcc5e5ebdcf5f50820447f5eb26061c4c674cd
        - registry.connect.redhat.com/ibm/modelbuilder-cv-training-service@sha256:c3c9080b08b87022c784fd98ac3e2774ec89ce29a2dd19d91dee039e3fdb2b36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:d4c60531f59953f399c6a2b0f8239128cdc2640f72f2d1e9ac4d6add618f9f01
        - registry.connect.redhat.com/ibm/modelbuilder-bundle@sha256:4940d853ba477aeeec99550c35c132b69bab8bd7b6fbc1a2fe118c828c7ba088
        - registry.connect.redhat.com/ibm/modelbuilder-util@sha256:0ec55c7f25c24f786978e045e8a67facb965a08dd519d06c82de05dd15e0f4a1
        - registry.connect.redhat.com/ibm/modelbuilder-image-server@sha256:ecdb665af651417c50335e7259b89ef8096521abec431976bd7abe4c1edc49d6
        - registry.connect.redhat.com/ibm/modelbuilder-operator@sha256:96f4c45f82152ebbdd2b067a6dd8749c5572e00106142725a16ae30a3e09d2d0
        version: 1.0.3
      entries:
      - name: ibm-modelbuilder-for-vision.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: model-builder-for-vision-certified
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-elbv2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-elbv2-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Listener",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "LoadBalancer",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "Rule",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "elbv2.services.k8s.aws/v1alpha1",
                "kind": "TargetGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
          createdAt: "2025-09-19T22:16:38Z"
          description: AWS ELB controller is a service controller for managing ELB
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Listener represents the state of an AWS elbv2 Listener resource.
            displayName: Listener
            kind: Listener
            name: listeners.elbv2.services.k8s.aws
            version: v1alpha1
          - description: LoadBalancer represents the state of an AWS elbv2 LoadBalancer
              resource.
            displayName: LoadBalancer
            kind: LoadBalancer
            name: loadbalancers.elbv2.services.k8s.aws
            version: v1alpha1
          - description: Rule represents the state of an AWS elbv2 Rule resource.
            displayName: Rule
            kind: Rule
            name: rules.elbv2.services.k8s.aws
            version: v1alpha1
          - description: TargetGroup represents the state of an AWS elbv2 TargetGroup
              resource.
            displayName: TargetGroup
            kind: TargetGroup
            name: targetgroups.elbv2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Load Balancing (ELB) resources in AWS from within your Kubernetes cluster.

          **About Amazon ELB**

          Elastic Load Balancing automatically distributes your incoming traffic across multiple targets, such as EC2 instances, containers, and IP addresses, in one or more Availability Zones. It monitors the health of its registered targets, and routes traffic only to the healthy targets. Elastic Load Balancing scales your load balancer capacity automatically in response to changes in incoming traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ELB
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - elbv2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ELB Developer Resources
          url: https://aws.amazon.com/elasticloadbalancing
        maintainers:
        - email: ack-maintainers@amazon.com
          name: elb maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/elbv2-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-elbv2-controller@sha256:f40f12efb734af5930c225925e99f31eb8db5fecece704a27ea50bd8661658c2
        version: 1.1.1
      entries:
      - name: ack-elbv2-controller.v1.1.1
        version: 1.1.1
      - name: ack-elbv2-controller.v1.1.0
        version: 1.1.0
      - name: ack-elbv2-controller.v1.0.13
        version: 1.0.13
      - name: ack-elbv2-controller.v1.0.12
        version: 1.0.12
      - name: ack-elbv2-controller.v1.0.10
        version: 1.0.10
      - name: ack-elbv2-controller.v1.0.9
        version: 1.0.9
      - name: ack-elbv2-controller.v1.0.8
        version: 1.0.8
      - name: ack-elbv2-controller.v1.0.7
        version: 1.0.7
      - name: ack-elbv2-controller.v1.0.6
        version: 1.0.6
      - name: ack-elbv2-controller.v1.0.5
        version: 1.0.5
      - name: ack-elbv2-controller.v1.0.4
        version: 1.0.4
      - name: ack-elbv2-controller.v1.0.3
        version: 1.0.3
      - name: ack-elbv2-controller.v1.0.2
        version: 1.0.2
      - name: ack-elbv2-controller.v1.0.1
        version: 1.0.1
      - name: ack-elbv2-controller.v1.0.0
        version: 1.0.0
      - name: ack-elbv2-controller.v0.0.11
        version: 0.0.11
      - name: ack-elbv2-controller.v0.0.10
        version: 0.0.10
      - name: ack-elbv2-controller.v0.0.9
        version: 0.0.9
      - name: ack-elbv2-controller.v0.0.8
        version: 0.0.8
      - name: ack-elbv2-controller.v0.0.6
        version: 0.0.6
      - name: ack-elbv2-controller.v0.0.5
        version: 0.0.5
      - name: ack-elbv2-controller.v0.0.4
        version: 0.0.4
      - name: ack-elbv2-controller.v0.0.3
        version: 0.0.3
      - name: ack-elbv2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-elbv2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Antline-Soft
      provider-url: ""
    name: cxl-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cxl-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cxl.antline.com/v1",
                "kind": "CXL",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "cxl-operator",
                    "app.kubernetes.io/instance": "cxl-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cxl",
                    "app.kubernetes.io/part-of": "cxl-operator"
                  },
                  "name": "cxl-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          createdAt: "2024-05-08T08:36:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CXL is the Schema for the cxls API
            displayName: CXL
            kind: CXL
            name: cxls.cxl.antline.com
            version: v1
        description: memory extend operator
        displayName: extend-community-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - memory
        - extend
        - community
        - operator
        links:
        - name: Cxl Operator
          url: https://cxl-operator.domain
        maintainers:
        - email: tomboy@antlinesoft.com
          name: Antline-Soft
        maturity: alpha
        provider:
          name: Antline-Soft
        relatedImages:
        - rasen708/cxl-operator:latest
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy:v4.14
        - quay.io/community-operator-pipeline-prod/ncn-community-operator@sha256:15edc1898fe9f8f56708da7d6de3548d06f1df3bdbb3f66a9cf4f7990fe9fad6
        version: 0.0.1
      entries:
      - name: cxl-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: cxl-operator
    provider:
      name: Antline-Soft
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: VictoriaMetrics
      provider-url: ""
    name: victoriametrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: victoriametrics-operator.v0.63.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLAgent",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "persistentVolumeClaimRetentionPolicy": {
                    "whenDeleted": "Delete"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vlsingle-example-0.default.svc:9428/internal/insert"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxBlockSize": "30MB"
                  },
                  "replicaCount": 2,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "50Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLCluster",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "example"
                },
                "spec": {
                  "requestsLoadBalancer": {
                    "enabled": true,
                    "spec": {}
                  },
                  "vlinsert": {
                    "replicaCount": 1
                  },
                  "vlselect": {
                    "replicaCount": 1
                  },
                  "vlstorage": {
                    "replicaCount": 2,
                    "retentionPeriod": "1y",
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "50Gi"
                            }
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1",
                "kind": "VLSingle",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "victoriametrics-operator"
                  },
                  "name": "sample"
                },
                "spec": {
                  "retentionMaxDiskSpaceUsageBytes": "20GB",
                  "retentionPeriod": "1y",
                  "storage": {
                    "resources": {
                      "requests": {
                        "storage": "50Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VLogs",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "2"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAgent",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "extraArgs": {
                    "memory.allowedPercent": "40"
                  },
                  "remoteWrite": [
                    {
                      "maxDiskUsage": "10GB",
                      "url": "http://vmsingle-dc-1.default.svc:8429/api/v1/write"
                    },
                    {
                      "maxDiskUsage": "15GB",
                      "url": "http://vmsingle-dc-2.default.svc:8429/api/v1/write"
                    },
                    {
                      "url": "http://vminsert-main.default.svc:8480/insert/1:1/prometheus/api/v1/write"
                    }
                  ],
                  "remoteWriteSettings": {
                    "maxDiskUsagePerURL": "5GB"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "50m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlert",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "datasource": {
                    "url": "http://vmsingle-example-pvc.default.svc:8429"
                  },
                  "evaluationInterval": "30s",
                  "notifier": {
                    "url": "http://vmalertmanager-example.default.svc:9093"
                  },
                  "replicaCount": 1,
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanager",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "configSecret": "am-global-config",
                  "replicaCount": 1,
                  "selectAllByDefault": true,
                  "storage": {
                    "volumeClaimTemplate": {
                      "spec": {
                        "resources": {
                          "requests": {
                            "storage": "2Gi"
                          }
                        }
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAlertmanagerConfig",
                "metadata": {
                  "name": "example-email-web"
                },
                "spec": {
                  "receivers": [
                    {
                      "email_configs": [
                        {
                          "from": "alerting@example.com",
                          "smarthost": "example.com:25",
                          "text": "ALARM",
                          "to": "some-email@example.com"
                        }
                      ],
                      "name": "email",
                      "webhook_configs": [
                        {
                          "url": "http://some-other-wh"
                        }
                      ]
                    }
                  ],
                  "route": {
                    "group_interval": "1m",
                    "receiver": "email",
                    "routes": [
                      {
                        "receiver": "email"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMAuth",
                "metadata": {
                  "name": "vmauth-sample"
                },
                "spec": {
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "victoriametrics/vmauth"
                  },
                  "replicaCount": 1,
                  "resources": {
                    "limits": {
                      "cpu": "500m",
                      "memory": "850Mi"
                    },
                    "requests": {
                      "cpu": "250m",
                      "memory": "350Mi"
                    }
                  },
                  "selectAllByDefault": true
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMCluster",
                "metadata": {
                  "name": "example-persistent"
                },
                "spec": {
                  "replicationFactor": 2,
                  "retentionPeriod": "4",
                  "vminsert": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    }
                  },
                  "vmselect": {
                    "cacheMountPath": "/select-cache",
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1000Mi"
                      },
                      "requests": {
                        "cpu": "0.5",
                        "memory": "500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "2Gi"
                            }
                          }
                        }
                      }
                    }
                  },
                  "vmstorage": {
                    "replicaCount": 2,
                    "resources": {
                      "limits": {
                        "cpu": "1",
                        "memory": "1500Mi"
                      }
                    },
                    "storage": {
                      "volumeClaimTemplate": {
                        "spec": {
                          "resources": {
                            "requests": {
                              "storage": "10Gi"
                            }
                          }
                        }
                      }
                    },
                    "storageDataPath": "/vm-data"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMNodeScrape",
                "metadata": {
                  "name": "cadvisor-metrics"
                },
                "spec": {
                  "bearerTokenFile": "/var/run/secrets/kubernetes.io/serviceaccount/token",
                  "interval": "10s",
                  "relabelConfigs": [
                    {
                      "action": "labelmap",
                      "regex": "__meta_kubernetes_node_label_(.+)"
                    },
                    {
                      "replacement": "kubernetes.default.svc:443",
                      "targetLabel": "__address__"
                    },
                    {
                      "regex": "(.+)",
                      "replacement": "/api/v1/nodes/$1/proxy/metrics/cadvisor",
                      "sourceLabels": [
                        "__meta_kubernetes_node_name"
                      ],
                      "targetLabel": "__metrics_path__"
                    }
                  ],
                  "scheme": "https",
                  "scrapeTimeout": "2s",
                  "selector": {},
                  "tlsConfig": {
                    "caFile": "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt",
                    "insecureSkipVerify": true
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMPodScrape",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "podMetricsEndpoints": [
                    {
                      "port": "metrics",
                      "scheme": "http"
                    }
                  ],
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMProbe",
                "metadata": {
                  "name": "agent"
                },
                "spec": {
                  "interval": "2s",
                  "jobName": "static-probe",
                  "module": "http_2xx",
                  "targets": {
                    "staticConfig": {
                      "targets": [
                        "vmagent-example.default.svc:8429/health"
                      ]
                    }
                  },
                  "vmProberSpec": {
                    "url": "prometheus-blackbox-exporter.default.svc:9115"
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMRule",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "groups": [
                    {
                      "name": "kafka",
                      "rules": [
                        {
                          "alert": "coordinator down",
                          "annotations": {
                            "description": "kafka coordinator is down",
                            "value": "{{ $value }}"
                          },
                          "expr": "ml_app_gauge{exec_context=\"consumer_group_state\"} == 0",
                          "for": "60s",
                          "labels": {
                            "job": "{{ $labels.job }}",
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMScrapeConfig",
                "metadata": {
                  "name": "mongodb"
                },
                "spec": {
                  "consulSDConfigs": [
                    {
                      "server": "https://consul-dns:8500",
                      "services": [
                        "mongodb"
                      ]
                    }
                  ],
                  "relabelConfigs": [
                    {
                      "action": "replace",
                      "sourceLabels": [
                        "__meta_consul_service"
                      ],
                      "targetLabel": "job"
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMServiceScrape",
                "metadata": {
                  "name": "example-scrape"
                },
                "spec": {
                  "endpoints": [
                    {
                      "port": "http"
                    }
                  ],
                  "namespaceSelector": {},
                  "selector": {
                    "matchLabels": {
                      "app.kubernetes.io/name": "example-application",
                      "monitored-by": "vm-operator"
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "retentionPeriod": "1"
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMSingle",
                "metadata": {
                  "name": "example-pvc"
                },
                "spec": {
                  "removePvcAfterDelete": true,
                  "retentionPeriod": "1",
                  "storage": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  }
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMStaticScrape",
                "metadata": {
                  "name": "vmstaticscrape-sample"
                },
                "spec": {
                  "jobName": "static",
                  "targetEndpoints": [
                    {
                      "labels": {
                        "env": "dev",
                        "project": "operator"
                      },
                      "targets": [
                        "192.168.0.1:9100",
                        "196.168.0.50:9100"
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "operator.victoriametrics.com/v1beta1",
                "kind": "VMUser",
                "metadata": {
                  "name": "vmuser-tenant-1"
                },
                "spec": {
                  "bearerToken": "some-token",
                  "targetRefs": [
                    {
                      "crd": {
                        "kind": "VMCluster/vminsert",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/insert/1"
                    },
                    {
                      "crd": {
                        "kind": "VMCluster/vmselect",
                        "name": "test-persistent",
                        "namespace": "vm"
                      },
                      "target_path_suffix": "/select/1"
                    },
                    {
                      "paths": [
                        "/internal/resetRollupResultCache"
                      ],
                      "static": {
                        "url": "http://vmselect-test-persistent.default.svc:8481/"
                      }
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring
          certified: "false"
          containerImage: docker.io/victoriametrics/operator:v0.63.0
          createdAt: "2025-09-12T12:58:29Z"
          description: Provides monitoring capabilities for kubernetes clusters and
            applications
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorhub.io/ui-metadata-max-k8s-version: "1.30"
          operators.openshift.io/infrastructure-features: '[fips]'
          operators.operatorframework.io.bundle.channel.default.v1: beta
          operators.operatorframework.io.bundle.channels.v1: beta
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/VictoriaMetrics/operator
          support: VictoriaMetrics
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: VLAgent - is a tiny but brave agent, which helps you collect
              logs from various sources and stores them in VictoriaLogs.
            displayName: VLAgent
            kind: VLAgent
            name: vlagents.operator.victoriametrics.com
            version: v1
          - description: VLCluster is fast, cost-effective and scalable logs database.
            displayName: VLCluster
            kind: VLCluster
            name: vlclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VLogs is fast, cost-effective and scalable logs database.
              VLogs is the Schema for the vlogs API
            displayName: VLogs
            kind: VLogs
            name: vlogs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VLSingle is fast, cost-effective and scalable logs database.
              VLSingle is the Schema for the API
            displayName: VLSingle
            kind: VLSingle
            name: vlsingles.operator.victoriametrics.com
            version: v1
          - description: |-
              VMAgent - is a tiny but brave agent, which helps you collect metrics from various sources and stores them in VictoriaMetrics
              or any other Prometheus-compatible storage system that supports the remote_write protocol.
            displayName: VMAgent
            kind: VMAgent
            name: vmagents.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanagerConfig is the Schema for the vmalertmanagerconfigs
              API
            displayName: VMAlertmanager Config
            kind: VMAlertmanagerConfig
            name: vmalertmanagerconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlertmanager represents Victoria-Metrics deployment for
              Alertmanager.
            displayName: VMAlertmanager
            kind: VMAlertmanager
            name: vmalertmanagers.operator.victoriametrics.com
            version: v1beta1
          - description: VMAlert  executes a list of given alerting or recording rules
              against configured address.
            displayName: VMAlert
            kind: VMAlert
            name: vmalerts.operator.victoriametrics.com
            version: v1beta1
          - description: VMAnomaly is the Schema for the vmanomalies API.
            displayName: VMAnomaly
            kind: VMAnomaly
            name: vmanomalies.operator.victoriametrics.com
            version: v1
          - description: VMAuth is the Schema for the vmauths API
            displayName: VMAuth
            kind: VMAuth
            name: vmauths.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMCluster is fast, cost-effective and scalable time-series database.
              Cluster version with
            displayName: VMCluster
            kind: VMCluster
            name: vmclusters.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMNodeScrape defines discovery for targets placed on kubernetes nodes,
              usually its node-exporters and other host services.
              InternalIP is used as __address__ for scraping.
            displayName: VMNode Scrape
            kind: VMNodeScrape
            name: vmnodescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMPodScrape is scrape configuration for pods,
              it generates vmagent's config for scraping pod targets
              based on selectors.
            displayName: VMPod Scrape
            kind: VMPodScrape
            name: vmpodscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMProbe defines a probe for targets, that will be executed with prober,
              like blackbox exporter.
              It helps to monitor reachability of target with various checks.
            displayName: VMProbe
            kind: VMProbe
            name: vmprobes.operator.victoriametrics.com
            version: v1beta1
          - description: VMRule defines rule records for vmalert application
            displayName: VMRule
            kind: VMRule
            name: vmrules.operator.victoriametrics.com
            version: v1beta1
          - description: VMScrapeConfig specifies a set of targets and parameters
              describing how to scrape them.
            displayName: VMScrape Config
            kind: VMScrapeConfig
            name: vmscrapeconfigs.operator.victoriametrics.com
            version: v1beta1
          - description: |-
              VMServiceScrape is scrape configuration for endpoints associated with
              kubernetes service,
              it generates scrape configuration for vmagent based on selectors.
              result config will scrape service endpoints
            displayName: VMService Scrape
            kind: VMServiceScrape
            name: vmservicescrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMSingle  is fast, cost-effective and scalable time-series
              database.
            displayName: VMSingle
            kind: VMSingle
            name: vmsingles.operator.victoriametrics.com
            version: v1beta1
          - description: VMStaticScrape  defines static targets configuration for
              scraping.
            displayName: VMStatic Scrape
            kind: VMStaticScrape
            name: vmstaticscrapes.operator.victoriametrics.com
            version: v1beta1
          - description: VMUser is the Schema for the vmusers API
            displayName: VMUser
            kind: VMUser
            name: vmusers.operator.victoriametrics.com
            version: v1beta1
          - description: VTCluster is fast, cost-effective and scalable traces database.
            displayName: VTCluster
            kind: VTCluster
            name: vtclusters.operator.victoriametrics.com
            version: v1
          - description: |-
              VTSingle is fast, cost-effective and scalable traces database.
              VTSingle is the Schema for the API
            displayName: VTSingle
            kind: VTSingle
            name: vtsingles.operator.victoriametrics.com
            version: v1
        description: |
          Operator manages VictoriaMetrics applications and provides monitoring features for applications running inside and outside kubernetes cluster. It has support for prometheus-operator objects and
          provides migration mechanism.

          Operator designed to use existing installation of VictoriaMetrics inside or outside kubernetes cluster or it could
          bring own resources and fully manages them.

           You can read more about operator at docs:
            - quick start [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/quick-start.md)
            - high availability [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/high-availability.md)
            - design and description of implementation [design](https://github.com/VictoriaMetrics/operator/blob/master/docs/design.md)
            - operator objects description [doc](https://github.com/VictoriaMetrics/operator/blob/master/docs/api.md)

          # VictoriaMetrics
          * VictoriaMetrics can be used as long-term storage for Prometheus or for [vmagent](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md).
          See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#prometheus-setup) for details.
          * Supports [Prometheus querying API](https://prometheus.io/docs/prometheus/latest/querying/api/), so it can be used as Prometheus drop-in replacement in Grafana.
          VictoriaMetrics implements [MetricsQL](https://github.com/VictoriaMetrics/VictoriaMetrics/wiki/MetricsQL) query language, which is inspired by PromQL.
          * Supports global query view. Multiple Prometheus instances may write data into VictoriaMetrics. Later this data may be used in a single query.
          * High performance and good scalability for both [inserts](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b)
          and [selects](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4).
          [Outperforms InfluxDB and TimescaleDB by up to 20x](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae).
          * [Uses 10x less RAM than InfluxDB](https://medium.com/@valyala/insert-benchmarks-with-inch-influxdb-vs-victoriametrics-e31a41ae2893) when working with millions of unique time series (aka high cardinality).
          * Optimized for time series with high churn rate. Think about [prometheus-operator](https://github.com/coreos/prometheus-operator) metrics from frequent deployments in Kubernetes.
          * High data compression, so [up to 70x more data points](https://medium.com/@valyala/when-size-matters-benchmarking-victoriametrics-vs-timescale-and-influxdb-6035811952d4)
          may be crammed into limited storage comparing to TimescaleDB.
          * Optimized for storage with high-latency IO and low IOPS (HDD and network storage in AWS, Google Cloud, Microsoft Azure, etc). See [graphs from these benchmarks](https://medium.com/@valyala/high-cardinality-tsdb-benchmarks-victoriametrics-vs-timescaledb-vs-influxdb-13e6ee64dd6b).
          * A single-node VictoriaMetrics may substitute moderately sized clusters built with competing solutions such as Thanos, M3DB, Cortex, InfluxDB or TimescaleDB.
          See [vertical scalability benchmarks](https://medium.com/@valyala/measuring-vertical-scalability-for-time-series-databases-in-google-cloud-92550d78d8ae),
          [comparing Thanos to VictoriaMetrics cluster](https://medium.com/@valyala/comparing-thanos-to-victoriametrics-cluster-b193bea1683)
          and [Remote Write Storage Wars](https://promcon.io/2019-munich/talks/remote-write-storage-wars/) talk
          from [PromCon 2019](https://promcon.io/2019-munich/talks/remote-write-storage-wars/).
          * Easy operation:
              * VictoriaMetrics consists of a single [small executable](https://medium.com/@valyala/stripping-dependency-bloat-in-victoriametrics-docker-image-983fb5912b0d) without external dependencies.
              * All the configuration is done via explicit command-line flags with reasonable defaults.
              * All the data is stored in a single directory pointed by `-storageDataPath` flag.
              * Easy and fast backups from [instant snapshots](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282)
              to S3 or GCS with [vmbackup](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmbackup/README.md) / [vmrestore](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmrestore/README.md).
              See [this article](https://medium.com/@valyala/speeding-up-backups-for-big-time-series-databases-533c1a927883) for more details.
          * Storage is protected from corruption on unclean shutdown (i.e. OOM, hardware reset or `kill -9`) thanks to [the storage architecture](https://medium.com/@valyala/how-victoriametrics-makes-instant-snapshots-for-multi-terabyte-time-series-data-e1f3fb0e0282).
          * Supports metrics' scraping, ingestion and [backfilling](https://github.com/VictoriaMetrics/VictoriaMetrics#backfilling) via the following protocols:
            * [Metrics from Prometheus exporters](https://github.com/prometheus/docs/blob/master/content/docs/instrumenting/exposition_formats.md#text-based-format)
              such as [node_exporter](https://github.com/prometheus/node_exporter). See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-scrape-prometheus-exporters-such-as-node-exporter) for details.
              * [Prometheus remote write API](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write)
              * [InfluxDB line protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP.
              * [Graphite plaintext protocol](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon)
              if `-graphiteListenAddr` is set.
              * [OpenTSDB put message](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-data-via-telnet-put-protocol) if `-opentsdbListenAddr` is set.
              * [HTTP OpenTSDB /api/put requests](https://github.com/VictoriaMetrics/VictoriaMetrics#sending-opentsdb-data-via-http-apiput-requests) if `-opentsdbHTTPListenAddr` is set.
              * [How to import time series data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-time-series-data).
              * [Prometheus exposition format](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-data-in-prometheus-exposition-format).
              * [Arbitrary CSV data](https://github.com/VictoriaMetrics/VictoriaMetrics#how-to-import-csv-data).
          * Supports metrics' relabeling. See [these docs](https://github.com/VictoriaMetrics/VictoriaMetrics#relabeling) for details.
          * Ideally works with big amounts of time series data from Kubernetes, IoT sensors, connected cars, industrial telemetry, financial data and various Enterprise workloads.
        displayName: VictoriaMetrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - prometheus
        - victoriametrics
        links:
        - name: VictoriaMetrics
          url: https://victoriametrics.com
        maintainers:
        - email: nik@victoriametrics.com
          name: f41gh7
        - email: info@victoriametrics.com
          name: info
        maturity: beta
        minKubeVersion: 1.23.0
        provider:
          name: VictoriaMetrics
        relatedImages:
        - docker.io/victoriametrics/operator:v0.63.0
        - quay.io/community-operator-pipeline-prod/victoriametrics-operator:0.63.0
        version: 0.63.0
      entries:
      - name: victoriametrics-operator.v0.63.0
        version: 0.63.0
      - name: victoriametrics-operator.v0.62.0
        version: 0.62.0
      - name: victoriametrics-operator.v0.61.2
        version: 0.61.2
      - name: victoriametrics-operator.v0.61.1
        version: 0.61.1
      - name: victoriametrics-operator.v0.61.0
        version: 0.61.0
      - name: victoriametrics-operator.v0.60.2
        version: 0.60.2
      - name: victoriametrics-operator.v0.60.1
        version: 0.60.1
      - name: victoriametrics-operator.v0.60.0
        version: 0.60.0
      - name: victoriametrics-operator.v0.59.2
        version: 0.59.2
      - name: victoriametrics-operator.v0.59.1
        version: 0.59.1
      - name: victoriametrics-operator.v0.59.0
        version: 0.59.0
      - name: victoriametrics-operator.v0.58.1
        version: 0.58.1
      - name: victoriametrics-operator.v0.58.0
        version: 0.58.0
      - name: victoriametrics-operator.v0.57.0
        version: 0.57.0
      - name: victoriametrics-operator.v0.56.0
        version: 0.56.0
      - name: victoriametrics-operator.v0.55.0
        version: 0.55.0
      - name: victoriametrics-operator.v0.54.1
        version: 0.54.1
      - name: victoriametrics-operator.v0.53.0
        version: 0.53.0
      - name: victoriametrics-operator.v0.52.0
        version: 0.52.0
      name: beta
    defaultChannel: beta
    packageName: victoriametrics-operator
    provider:
      name: VictoriaMetrics
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: run-once-duration-override-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: runoncedurationoverrideoperator.v1.2.2
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "RunOnceDurationOverride",
                "metadata": {
                  "name": "cluster"
                },
                "spec": {
                  "runOnceDurationOverride": {
                    "spec": {
                      "activeDeadlineSeconds": 3600
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certifiedLevel: "false"
          console.openshift.io/operator-monitoring-default: "true"
          containerImage: registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
          createdAt: 2025/06/24
          description: An operator to manage the OpenShift RunOnceDurationOverride
            Mutating Admission Webhook Server
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.1.0 <1.2.2'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-run-once-duration-override-operator
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/run-once-duration-override-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: RunOnceDurationOverride is the schema for the runoncedurationoverrides
              API
            displayName: Run Once Duration Override
            kind: RunOnceDurationOverride
            name: runoncedurationoverrides.operator.openshift.io
            version: v1
        displayName: Run Once Duration Override Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mutating
        - webhook
        - workload
        - run-once
        links:
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override
        - name: Source Code
          url: https://github.com/openshift/run-once-duration-override-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.30.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-operator-bundle@sha256:8ff7f1c707b40448a719137f37f6a3e09860812f7f07c3e61b7f73d1216399dc
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9-operator@sha256:bd1c357e4dc6a598f86a8d988212b804dd9a7fa7d864c59ea58394540211ff90
        - registry.redhat.io/run-once-duration-override-operator/run-once-duration-override-rhel9@sha256:42f73634df2c06216fac7b3742b2ac21ef5f12ee98d8694c06f7c9073773acd0
        version: 1.2.2
      entries:
      - name: runoncedurationoverrideoperator.v1.2.2
        version: 1.2.2
      - name: runoncedurationoverrideoperator.v1.2.1
        version: 1.2.1
      - name: runoncedurationoverrideoperator.v1.2.0
        version: 1.2.0
      - name: runoncedurationoverrideoperator.v1.1.2
        version: 1.1.2
      - name: runoncedurationoverrideoperator.v1.1.1
        version: 1.1.1
      - name: runoncedurationoverrideoperator.v1.1.0
        version: 1.1.0
      - name: runoncedurationoverrideoperator.v1.0.2
        version: 1.0.2
      - name: runoncedurationoverrideoperator.v1.0.1
        version: 1.0.1
      - name: runoncedurationoverrideoperator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: run-once-duration-override-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Starburst Data, Inc.
      provider-url: https://www.starburstdata.com
    name: starburst-enterprise-helm-operator-paygo-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: starburst-enterprise-helm-operator-paygo.v402.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstEnterprise",
                "metadata": {
                  "name": "starburstenterprise-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "cache": {
                    "diskUsagePercentage": 80,
                    "enabled": false,
                    "ttl": "7d",
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "catalogs": {
                    "tpch": "connector.name=tpch\ntpch.splits-per-node=4"
                  },
                  "commonLabels": {},
                  "coordinator": {
                    "additionalProperties": "",
                    "affinity": {},
                    "deploymentAnnotations": {},
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=true\nnode-scheduler.include-coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://localhost:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "resources": {
                      "cpu": 16,
                      "memory": "60Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "tolerations": []
                  },
                  "environment": null,
                  "expose": {
                    "clusterIp": {
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      }
                    },
                    "ingress": {
                      "annotations": {},
                      "host": null,
                      "ingressClassName": null,
                      "ingressName": "coordinator-ingress",
                      "path": "/",
                      "pathType": "ImplementationSpecific",
                      "serviceName": "starburst",
                      "servicePort": 8080,
                      "tls": {
                        "enabled": true,
                        "secretName": null
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "port": 8080
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "starburst",
                      "ports": {
                        "http": {
                          "nodePort": 30080,
                          "port": 8080
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "externalSecrets": {
                    "enabled": false,
                    "goDaddy": {
                      "backendType": "secretsManager"
                    },
                    "secretPrefix": "external0",
                    "type": "goDaddy"
                  },
                  "extraArguments": [],
                  "extraSecret": {
                    "file": null,
                    "name": null
                  },
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise",
                    "tag": "402-e.1"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "initImage": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/starburst-enterprise-init",
                    "tag": "1.3.2"
                  },
                  "internal": {
                    "ports": {
                      "http": {
                        "port": 8080
                      },
                      "https": {
                        "port": 8443
                      }
                    }
                  },
                  "internalTls": false,
                  "licenseManager": {
                    "cronjob": {
                      "affinity": {},
                      "nodeSelector": {},
                      "priorityClassName": null,
                      "securityContext": {},
                      "tolerations": []
                    },
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "repository": "registry.connect.redhat.com/starburst/license-manager",
                      "tag": "0.0.2"
                    },
                    "url": "https://license-api.dogfood.eng.starburstdata.net/test/license/checkout"
                  },
                  "livenessProbe": null,
                  "prometheus": {
                    "agent": {
                      "config": "/etc/starburst/telemetry/prometheus.yaml",
                      "port": 8081,
                      "version": "0.16.1"
                    },
                    "enabled": true,
                    "rules": [
                      {
                        "attrNameSnakeCase": true,
                        "name": "$1",
                        "pattern": "trino.execution<name=QueryManager><>(running_queries|queued_queries)",
                        "type": "GAUGE"
                      },
                      {
                        "name": "failed_queries",
                        "pattern": "trino.execution<name=QueryManager><>FailedQueries\\.TotalCount",
                        "type": "COUNTER"
                      }
                    ]
                  },
                  "query": {
                    "maxConcurrentQueries": 3
                  },
                  "readinessProbe": null,
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "securityContext": {},
                  "sharedSecret": null,
                  "siac": {
                    "additionalHeapSizePercentage": 15,
                    "enabled": false
                  },
                  "spilling": {
                    "enabled": false,
                    "volume": {
                      "emptyDir": {}
                    }
                  },
                  "startupProbe": null,
                  "usageMetrics": {
                    "enabled": false,
                    "usageClient": {
                      "initialDelay": "1m",
                      "interval": "1m"
                    }
                  },
                  "userDatabase": {
                    "enabled": false,
                    "users": [
                      {
                        "password": "46991b33f7a75ff79213c0dc0e610610",
                        "username": "admin"
                      }
                    ]
                  },
                  "worker": {
                    "additionalProperties": "",
                    "affinity": {},
                    "autoscaling": {
                      "enabled": false,
                      "maxReplicas": 100,
                      "minReplicas": 1,
                      "targetCPUUtilizationPercentage": 80
                    },
                    "deploymentAnnotations": {},
                    "deploymentTerminationGracePeriodSeconds": 300,
                    "envFrom": [],
                    "etcFiles": {
                      "jvm.config": "-server\n-XX:G1HeapRegionSize=32M\n-XX:+ExplicitGCInvokesConcurrent\n-XX:+ExitOnOutOfMemoryError\n-XX:+HeapDumpOnOutOfMemoryError\n-XX:-OmitStackTraceInFastThrow\n-XX:ReservedCodeCacheSize=512M\n-XX:PerMethodRecompilationCutoff=10000\n-XX:PerBytecodeRecompilationCutoff=10000\n-Djdk.attach.allowAttachSelf=true\n-Djdk.nio.maxCachedBufferSize=2000000\n-XX:+UnlockDiagnosticVMOptions\n-XX:+UseAESCTRIntrinsics\n--add-opens=java.base/sun.nio.ch=ALL-UNNAMED\n--add-opens=java.base/java.nio=ALL-UNNAMED\n--add-opens=java.base/java.lang=ALL-UNNAMED\n--add-opens=java.security.jgss/sun.security.krb5=ALL-UNNAMED\n-XX:-UseBiasedLocking\n-XX:+UseG1GC\n",
                      "other": {},
                      "properties": {
                        "config.properties": "coordinator=false\nhttp-server.http.port=8080\ndiscovery.uri=http://coordinator:8080\n",
                        "log.properties": "# Enable verbose logging from Starburst Enterprise\n#io.trino=DEBUG\n#com.starburstdata.presto=DEBUG\n",
                        "node.properties": "node.environment={{ include \"starburst.environment\" . }}\nnode.data-dir=/data/starburst\nplugin.dir=/usr/lib/starburst/plugin\nnode.server-log-file=/var/log/starburst/server.log\nnode.launcher-log-file=/var/log/starburst/launcher.log\n"
                      }
                    },
                    "heapHeadroomPercentage": 30,
                    "heapSizePercentage": 90,
                    "initContainers": [],
                    "kedaScaler": {
                      "enabled": false,
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.connect.redhat.com/starburst/keda-trino-scaler",
                        "tag": "0.1.7"
                      },
                      "port": 8021,
                      "scaledObjectSpec": {
                        "advanced": {},
                        "cooldownPeriod": 300,
                        "idleReplicaCount": 0,
                        "maxReplicaCount": 100,
                        "minReplicaCount": 1,
                        "pollingInterval": 30,
                        "scaleTargetRef": {
                          "name": "worker"
                        },
                        "triggers": [
                          {
                            "metadata": {
                              "numberOfQueriesPerWorker": "10",
                              "scaleInToIdleReplicaCountIfNoQueuedQueriesLeft": "true",
                              "scaleMethod": "query_queue"
                            },
                            "type": "external"
                          }
                        ]
                      }
                    },
                    "nodeMemoryHeadroom": "2Gi",
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "priorityClassName": null,
                    "replicas": 2,
                    "resources": {
                      "cpu": 16,
                      "memory": "100Gi"
                    },
                    "securityContext": {},
                    "sidecars": [],
                    "starburstWorkerShutdownGracePeriodSeconds": 120,
                    "tolerations": []
                  }
                }
              },
              {
                "apiVersion": "charts.starburstdata.com/v1alpha1",
                "kind": "StarburstHive",
                "metadata": {
                  "name": "starbursthive-sample"
                },
                "spec": {
                  "additionalVolumes": [],
                  "affinity": {},
                  "commonLabels": {},
                  "database": {
                    "external": {
                      "driver": null,
                      "jdbcUrl": null,
                      "password": null,
                      "user": null
                    },
                    "internal": {
                      "databaseName": "hive",
                      "databasePassword": "HivePass1234",
                      "databaseUser": "hive",
                      "driver": "org.postgresql.Driver",
                      "env": [],
                      "envFrom": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "repository": "registry.redhat.io/rhel8/postgresql-10",
                        "tag": "1"
                      },
                      "port": 5432,
                      "resources": {
                        "limits": {
                          "cpu": 2,
                          "memory": "1Gi"
                        },
                        "requests": {
                          "cpu": 2,
                          "memory": "1Gi"
                        }
                      },
                      "volume": {
                        "emptyDir": {}
                      }
                    },
                    "type": "internal"
                  },
                  "deploymentAnnotations": {},
                  "env": [],
                  "envFrom": [],
                  "expose": {
                    "clusterIp": {
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      }
                    },
                    "loadBalancer": {
                      "IP": "",
                      "annotations": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "port": 9083
                        }
                      },
                      "sourceRanges": []
                    },
                    "nodePort": {
                      "extraLabels": {},
                      "name": "hive",
                      "ports": {
                        "http": {
                          "nodePort": 30083,
                          "port": 9083
                        }
                      }
                    },
                    "type": "clusterIp"
                  },
                  "extraArguments": [],
                  "hdfs": {
                    "hadoopUserName": null
                  },
                  "heapSizePercentage": 85,
                  "hiveMetastoreStorageAuthorization": true,
                  "hiveMetastoreWarehouseDir": null,
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "repository": "registry.connect.redhat.com/starburst/hive",
                    "tag": "3.1.2-e.22"
                  },
                  "imagePullSecrets": null,
                  "initFile": "",
                  "nodeSelector": {},
                  "objectStorage": {
                    "adl": {
                      "oauth2": {
                        "clientId": null,
                        "credential": null,
                        "refreshUrl": null
                      }
                    },
                    "awsS3": {
                      "accessKey": null,
                      "endpoint": null,
                      "pathStyleAccess": false,
                      "region": null,
                      "secretKey": null
                    },
                    "azure": {
                      "abfs": {
                        "accessKey": {
                          "accessKey": null,
                          "storageAccount": null
                        },
                        "authType": "accessKey",
                        "oauth": {
                          "clientId": null,
                          "endpoint": null,
                          "secret": null
                        }
                      },
                      "wasb": {
                        "accessKey": null,
                        "storageAccount": null
                      }
                    },
                    "gs": {
                      "cloudKeyFileSecret": null
                    }
                  },
                  "podAnnotations": {},
                  "registryCredentials": {
                    "enabled": false,
                    "password": null,
                    "registry": null,
                    "username": null
                  },
                  "resources": {
                    "limits": {
                      "cpu": 1,
                      "memory": "1Gi"
                    },
                    "requests": {
                      "cpu": 1,
                      "memory": "1Gi"
                    }
                  },
                  "securityContext": {},
                  "tolerations": []
                }
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          certified: "true"
          containerImage: registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
          description: Deploys StarburstEnterprise Helm Operator
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/starburst-enterprise-helm-operator-paygo-rhmp/support?utm_source=openshift_console
          marketplace.redhat.com/meterDefinition: |-
            {
              "apiVersion": "marketplace.redhat.com/v1beta1",
              "kind": "MeterDefinition",
              "metadata": {
                "name": "sep-vcpu-usage"
              },
              "spec": {
                "group": "charts.starburstdata.com",
                "kind": "StarburstEnterprise",
                "resourceFilters": [
                  {
                    "namespace": {
                      "useOperatorGroup": true
                    },
                    "ownerCRD": {
                      "apiVersion": "charts.starburstdata.com/v1alpha1",
                      "kind": "StarburstEnterprise"
                    },
                    "workloadType": "Pod"
                  }
                ],
                "meters": [
                  {
                    "aggregation": "sum",
                    "period": "1h",
                    "metricId": "container_cpu_usage_core_avg",
                    "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=~\"coordinator|worker\"}[5m])*100",
                    "workloadType": "Pod"
                  }
                ]
              }
            }
          operators.operatorframework.io/builder: operator-sdk-v1.13.0+git
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: Starburst Data
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: StarburstEnterprise
            name: starburstenterprises.charts.starburstdata.com
            version: v1alpha1
          - kind: StarburstHive
            name: starbursthives.charts.starburstdata.com
            version: v1alpha1
        description: CRs for deploying starburst enterprise and hive
        displayName: Starburst Enterprise Pay as you go Helm Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - starburst
        - trino
        - sep
        - bigdata
        - data
        links:
        - name: Starburst Enterprise Helm Operator
          url: https://www.starburstdata.com
        maintainers:
        - email: hello@starburstdata.com
          name: Starburst Data, Inc.
        maturity: alpha
        provider:
          name: Starburst Data, Inc.
          url: https://www.starburstdata.com
        relatedImages:
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator-paygo-bundle-certified@sha256:e0a8c4d0de20c91b73957d6ea09fb94005e680a092c54389e6e411edbeb59296
        - registry.connect.redhat.com/starburst/starburst-enterprise-helm-operator@sha256:82733a4367a98fff10fbf3c56efb41af5cd71dc3ddb1610250c0b2053c3ebf23
        version: 402.1.0
      entries:
      - name: starburst-enterprise-helm-operator-paygo.v402.1.0
        version: 402.1.0
      name: alpha
    defaultChannel: alpha
    packageName: starburst-enterprise-helm-operator-paygo-rhmp
    provider:
      name: Starburst Data, Inc.
      url: https://www.starburstdata.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: https://www.ibm.com
    name: ibm-security-verify-access-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-security-verify-access-operator.v25.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ibm.com/v1",
                "kind": "IBMSecurityVerifyAccess",
                "metadata": {
                  "name": "ivia-sample"
                },
                "spec": {
                  "image": "icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2",
                  "instance": "default"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
          createdAt: "2025-05-01T01:27:17Z"
          description: The IBM Verify Identity Access Operator manages the lifecycle
            of IBM Verify Identity Access worker containers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM-Security/verify-access-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMSecurityVerifyAccess is the Schema for the ibmsecurityverifyaccesses
              API.
            displayName: IBMSecurity Verify Access
            kind: IBMSecurityVerifyAccess
            name: ibmsecurityverifyaccesses.ibm.com
            version: v1
        description: |+
          In a world of highly fragmented access management environments, [IBM Verify Identity Access](https://www.ibm.com/au-en/products/verify-access) helps you simplify your users' access while more securely adopting web, mobile and cloud technologies. This solution helps you strike a balance between usability and security through the use of risk-based access, single sign-on, integrated access management control, identity federation and its mobile multi-factor authentication capability, IBM Verify. Take back control of your access management with IBM Verify Identity Access.

          The IBM Verify Identity Access operator provides lifecycle management of the lightweight containers which are used to protect an environment, namely:
          * [Web Reverse Proxy](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-web-reverse-proxy)
          * [Runtime](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-runtime)
          * [Distributed Session Cache](https://www.ibm.com/docs/en/sva/latest?topic=support-docker-image-verify-access-distributed-session-cache)

          The operator will manage the deployment of these lightweight IBM Verify Identity Access worker containers, and also control the rolling restart of these containers when a configuration snapshot is updated.

          See the project [Readme](https://github.com/IBM-Security/verify-access-operator/blob/master/README.md) for further information and details.

        displayName: IBM Verify Identity Access Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - identity and access
        - security
        links:
        - name: Verify Identity Access Product Information
          url: https://www.ibm.com/au-en/products/verify-access
        - name: Verify Identity Access Documentation
          url: https://www.ibm.com/docs/en/sva
        maintainers:
        - email: isamdev@au1.ibm.com
          name: Verify Identity Access Development Team
        maturity: stable
        minKubeVersion: 1.17.0
        provider:
          name: IBM
          url: https://www.ibm.com
        relatedImages:
        - registry.connect.redhat.com/ibm/verify-access-operator@sha256:6e37f565564de371b5797895f3e96e0efd4f592d7d3e899ea8507438bdb60e78
        - icr.io/isva/verify-access-operator@sha256:ca1d7be14b4dd313d4675f18541b2cc670ea0a0adf3ec87e347e5e20a86bc39c
        - icr.io/ivia/ivia-wrp@sha256:c2dd5ec2b362136896e47ccbd0f3ee033af945fc2794ddaf075da603e4d4d7b2
        version: 25.4.0
      entries:
      - name: ibm-security-verify-access-operator.v25.4.0
        version: 25.4.0
      - name: ibm-security-verify-access-operator.v24.12.0
        version: 24.12.0
      name: stable
    defaultChannel: stable
    packageName: ibm-security-verify-access-operator
    provider:
      name: IBM
      url: https://www.ibm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Infinidat
      provider-url: https://infinidat.com
    name: infinibox-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: alpha
    - currentCSV: infinibox-operator-certified.v2.23.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csidriver.infinidat.com/v1alpha1",
                "kind": "InfiniboxCsiDriver",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "infinibox-operator-certified",
                    "app.kubernetes.io/instance": "infiniboxcsidriver-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "infiniboxcsidriver",
                    "app.kubernetes.io/part-of": "infinibox-operator-certified"
                  },
                  "name": "infiniboxcsidriver-sample"
                },
                "spec": {
                  "Ibox_Creds": [
                    {
                      "SecretName": "infinibox-creds",
                      "hostname": "ibox0000",
                      "inbound_secret": "0.0000000000000",
                      "inbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonout",
                      "outbound_secret": "0.0000000000001",
                      "outbound_user": "iqn.2020-06.com.csi-driver-iscsi.infinidat:commonin",
                      "password": "my-pool-admin-passwd",
                      "username": "my-pool-admin"
                    }
                  ],
                  "Infinibox_Protocol_Secrets": [
                    {
                      "SecretName": "infinibox-protocol-secret",
                      "iscsi": {
                        "network_space": "default_iscsi_space",
                        "useCHAP": "none"
                      },
                      "nfs": {
                        "network_space": "default_nas_space",
                        "nfs_export_permissions": ""
                      },
                      "nvme": {
                        "network_space": "default_nvme_space"
                      },
                      "preferred_auto_order": "fc,nvme,iscsi",
                      "storage_protocol": "fc"
                    }
                  ],
                  "autoUpdate": false,
                  "cleanupNFSPerms": false,
                  "controllerserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "controllerservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "createEvents": true,
                  "csiDriverName": "infinibox-csi-driver",
                  "csiDriverVersion": "v2.23.0",
                  "e2etesting": false,
                  "images": {
                    "attachersidecar": "registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901",
                    "attachersidecar_pull_policy": "IfNotPresent",
                    "csidriver": "registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61",
                    "csidriver_pull_policy": "Always",
                    "livenesssidecar": "registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc",
                    "livenesssidecar_pull_policy": "IfNotPresent",
                    "provisionersidecar": "registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2",
                    "provisionersidecar_pull_policy": "IfNotPresent",
                    "registrarsidecar": "registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24",
                    "registrarsidecar_pull_policy": "IfNotPresent",
                    "resizersidecar": "registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d",
                    "resizersidecar_pull_policy": "IfNotPresent",
                    "snapshottersidecar": "registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a",
                    "snapshottersidecar_pull_policy": "IfNotPresent"
                  },
                  "instanceCount": 1,
                  "logLevel": "debug",
                  "nodeSelector": {
                    "kubernetes.io/os": "linux"
                  },
                  "nodeserverresources": {
                    "limits": {
                      "cpu": "0.5",
                      "memory": "450Mi"
                    },
                    "requests": {
                      "cpu": "0.1",
                      "memory": "150Mi"
                    }
                  },
                  "nodeservertolerations": [
                    {
                      "effect": "",
                      "key": "",
                      "operator": "Exists",
                      "tolerationSeconds": 0,
                      "value": ""
                    }
                  ],
                  "protocolSecret": "",
                  "removeDomainName": false,
                  "replicaCount": 1,
                  "skipCredentialsCreation": false,
                  "volumeNamePrefix": "ibox"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "true"
          containerImage: registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
          createdAt: "2025-09-27T16:12:42Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <2.23.0
          operatorframework.io/suggested-namespace: infinidat-csi
          operators.operatorframework.io/builder: operator-sdk-v1.40.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/Infinidat/infinibox-csi-driver
          support: Infinidat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Iboxreplica
            name: iboxreplicas.csidriver.infinidat.com
            version: v1
          - kind: InfiniboxCsiDriver
            name: infiniboxcsidrivers.csidriver.infinidat.com
            version: v1alpha1
        description: "Infinidat InfiniBox Container Storage Interface (CSI) Driver
          is a\n    CNCF-compliant Kubernetes integration for InfiniBox storage systems,
          offering\n    advanced enterprise functionality for petabyte-scale Kubernetes
          deployments including\n    Red Hat OpenShift.\n \n## Features and Benefits\n
          \ \n* **Multi-protocol flexibility** - manage\n    Kubernetes Persistent
          Volumes attached via block and file protocols,\n    including Fibre Channel,
          iSCSI, and NFS, with all Kubernetes PV access modes\n    \n* **Multi-petabyte
          scalability** - support hundreds of thousands of PVs per InfiniBox\n    system
          and control multiple InfiniBox arrays within a single Kubernetes cluster\n
          \   \\ \n* **Advanced enterprise features** - manage native InfiniBox snapshots
          and\n    clones, including restoring from snapshots, and import PVs created
          outside of\n    InfiniBox CSI Driver\n  \n## Required Parameters\n  \n*
          `hostname` - IP address\n    or hostname of the InfiniBox management interface\n*
          `username` / `password` -\n    InfiniBox credentials\n* `SecretName` - secret
          name, to be used in the StorageClass\n    to define a specific InfiniBox
          for persistent volumes\n\n## Optional Parameters\n*\n    `inbound_user`
          / `inbound_secret` / `outbound_user` / `outbound_secret` - credentials \n
          \   for iSCSI CHAP authentication\n\n    \n## Installation Instructions\n
          \n1. Create *infinidat-csi* namespace to install operator into.\n    \n2.
          Delete the following *clusterrolebindings*, if they exist, before installing
          operator\n     * infinidat-csi-operator-infinidat-csi-attacher\n     * infinidat-csi-operator-infinidat-csi-controller
          \                    \n     * infinidat-csi-operator-infinidat-csi-driver
          \                    \n     * infinidat-csi-operator-infinidat-csi-node
          \                              \n     * infinidat-csi-operator-infinidat-csi-provisioner
          \                       \n     * infinidat-csi-operator-infinidat-csi-resizer
          \          \n     * infinidat-csi-operator-infinidat-csi-snapshotter \n
          \   \n3. Install Infinidat Infinibox CSI Driver - Operator (this) into the
          *infinidat-csi* namespace.\n    \n4. Apply the following Security Context
          Constraints to the created service accounts in the *infinidat-csi* \n    namespace\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user privileged -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user anyuid -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-controller-manager\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-driver\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-node\n\n
          \    * oc adm policy add-scc-to-user hostnetwork -z infinidat-csi-operator-infinidat-csi-controller\n\n
          \    \n5. Recreate or edit any existing storage classes, pointing to the
          new namespace\n\n     \n6. If you choose a different namespace other than
          *infinidat-csi* you will need to edit the clusterrolebindings listed above
          to specify your custom namespace\n\n     \n7. Install an instance of Infinidat
          Driver into the *infinidat-csi* namespace"
        displayName: InfiniBox CSI Driver - Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - csi
        - nas
        - san
        - storage
        - infinidat
        - infinibox
        links:
        - name: Infinidat
          url: https://infinidat.com
        maintainers:
        - email: info@infinidat.com
          name: Infinidat
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Infinidat
          url: https://infinidat.com
        relatedImages:
        - registry.k8s.io/sig-storage/csi-resizer@sha256:706f7cdcccd30ca5f0e94d548e2e0c658f69c0fe4b68a5bf37818a04ca618d3d
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:339a83a86e6e1eead14413fe494b43de8bc48243d146c2810b39dd741d26ca6a
        - registry.k8s.io/sig-storage/livenessprobe@sha256:13f6b1f9d0514b859e549e20b731d93c90a144186deb68bfc931e3dbf9041afc
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified@sha256:42ff34a019bc67076515e89735b4bb9921fbd4610c127ac8b61e1fa6a6a08194
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f032a0ca4c699eebe403988a0e217c3dfc82e2cee8b7d9d247a493e5a2425f24
        - registry.connect.redhat.com/infinidat/infinibox-csidriver-certified@sha256:3e6e1c7a7639f50562d4916383c9de28348b03590633973c8c33c4ebba074d61
        - registry.connect.redhat.com/infinidat/infinibox-operator-certified-bundle@sha256:8f72c137e6b0af661eb19f43948a6490f8426b65f424e511d66689d1edeb340a
        - registry.k8s.io/sig-storage/csi-attacher@sha256:47ab8aebebdc59316004ba8d51a903637d808f4e62a6d0f599ed3c2483cea901
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:67ee5137252811fd471b8571efe9e173145ec8af7b520861eeccf7c078a772f2
        version: 2.23.0
      entries:
      - name: infinibox-operator-certified.v2.23.0
        version: 2.23.0
      - name: infinibox-operator-certified.v2.22.0
        version: 2.22.0
      - name: infinibox-operator-certified.v2.21.0
        version: 2.21.0
      name: stable
    defaultChannel: stable
    packageName: infinibox-operator-certified
    provider:
      name: Infinidat
      url: https://infinidat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Grafana Loki SIG Operator
      provider-url: ""
    name: loki-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: loki-operator.v0.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "AlertingRule",
                "metadata": {
                  "name": "alertingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "alerting-rules-group",
                      "rules": [
                        {
                          "alert": "HighPercentageError",
                          "annotations": {
                            "summary": "High request latency"
                          },
                          "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"production\"}[5m])) by (job)\n  \u003e 0.05\n",
                          "for": "10m",
                          "labels": {
                            "severity": "page"
                          }
                        },
                        {
                          "alert": "HttpCredentialsLeaked",
                          "annotations": {
                            "message": "{{ $labels.job }} is leaking http basic auth credentials."
                          },
                          "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e 0)",
                          "for": "10m",
                          "labels": {
                            "severity": "critical"
                          }
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "LokiStack",
                "metadata": {
                  "name": "lokistack-sample"
                },
                "spec": {
                  "size": "1x.small",
                  "storage": {
                    "secret": {
                      "name": "test"
                    }
                  },
                  "storageClassName": "standard"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RecordingRule",
                "metadata": {
                  "name": "recordingrule-sample"
                },
                "spec": {
                  "groups": [
                    {
                      "interval": "10m",
                      "name": "recording-rules-group",
                      "rules": [
                        {
                          "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                          "record": "myservice:requests:rate10m"
                        },
                        {
                          "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                          "record": "otherservice:requests:rate1m"
                        }
                      ]
                    }
                  ],
                  "tenantID": "test-tenant"
                }
              },
              {
                "apiVersion": "loki.grafana.com/v1",
                "kind": "RulerConfig",
                "metadata": {
                  "name": "rulerconfig-sample"
                },
                "spec": {
                  "alertmanager": {
                    "discovery": {
                      "enableSRV": true,
                      "refreshInterval": "1m"
                    },
                    "enableV2": true,
                    "endpoints": [
                      "http://alertmanager-host1.mycompany.org",
                      "http://alertmanager-host2.mycompany.org"
                    ],
                    "externalLabels": {
                      "environment": "production",
                      "region": "us-east-2"
                    },
                    "externalUrl": "http://www.mycompany.org/alerts",
                    "notificationQueue": {
                      "capacity": 1000,
                      "forGracePeriod": "10m",
                      "forOutageTolerance": "1h",
                      "resendDelay": "1m",
                      "timeout": "30s"
                    }
                  },
                  "evaluationInterval": "1m",
                  "pollInterval": "1m",
                  "remoteWrite": {
                    "client": {
                      "authorization": "basic",
                      "authorizationSecretName": "my-secret-resource",
                      "name": "remote-write-log-metrics",
                      "proxyUrl": "http://proxy-host.mycompany.org",
                      "relabelConfigs": [
                        {
                          "action": "replace",
                          "regex": "ALERTS.*",
                          "replacement": "$1",
                          "separator": "",
                          "sourceLabels": [
                            "labelc",
                            "labeld"
                          ],
                          "targetLabel": "labelnew"
                        }
                      ],
                      "timeout": "30s",
                      "url": "http://remote-write-host.mycompany.org"
                    },
                    "enabled": true,
                    "refreshPeriod": "10s"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Logging & Tracing
          certified: "false"
          containerImage: docker.io/grafana/loki-operator:0.8.0
          createdAt: "2025-03-17T11:00:08Z"
          description: The Community Loki Operator provides Kubernetes native deployment
            and management of Loki and related logging components.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "true"
          features.operators.openshift.io/token-auth-gcp: "true"
          operators.operatorframework.io/builder: operator-sdk-unknown
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/grafana/loki/tree/main/operator
          support: Grafana Loki SIG Operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertingRule is the Schema for the alertingrules API
            displayName: AlertingRule
            kind: AlertingRule
            name: alertingrules.loki.grafana.com
            version: v1
          - description: LokiStack is the Schema for the lokistacks API
            displayName: LokiStack
            kind: LokiStack
            name: lokistacks.loki.grafana.com
            version: v1
          - description: RecordingRule is the Schema for the recordingrules API
            displayName: RecordingRule
            kind: RecordingRule
            name: recordingrules.loki.grafana.com
            version: v1
          - description: RulerConfig is the Schema for the rulerconfigs API
            displayName: RulerConfig
            kind: RulerConfig
            name: rulerconfigs.loki.grafana.com
            version: v1
        description: |-
          The Community Loki Operator provides Kubernetes native deployment and management of Loki and related logging components.
          The purpose of this project is to simplify and automate the configuration of a Loki based logging stack for Kubernetes clusters.

          ### Operator features

          The Loki Operator includes, but is not limited to, the following features:

          * Kubernetes Custom Resources: Use Kubernetes custom resources to deploy and manage Loki, Alerting rules, Recording rules, and related components.
          * Simplified Deployment Configuration: Configure the fundamentals of Loki like tenants, limits, replication factor and storage from a native Kubernetes resource.

          ### Feature Gates

          The Loki Operator Bundle provides a set of feature gates that enable/disable specific feature depending on the target Kubernetes distribution. The following feature gates are enabled by default:
          * `serviceMonitors`: Enables creating a Prometheus-Operator managed ServiceMonitor resource per LokiStack component.
          * `serviceMonitorTlsEndpoints`: Enables TLS for the ServiceMonitor endpoints.
          * `lokiStackAlerts`: Enables creating PrometheusRules for common Loki alerts.
          * `httpEncryption`: Enables TLS encryption for all HTTP LokiStack services.
          * `grpcEncryption`: Enables TLS encryption for all GRPC LokiStack services.
          * `builtInCertManagement`: Enables the built-in facility for generating and rotating TLS client and serving certificates for all LokiStack services and internal clients
          * `lokiStackGateway`: Enables reconciling the reverse-proxy lokistack-gateway component for multi-tenant authentication/authorization traffic control to Loki.
          * `runtimeSeccompProfile`: Enables the restricted seccomp profile on all Lokistack components.
          * `defaultNodeAffinity`: Enable the operator will set a default node affinity on all pods. This will limit scheduling of the pods to Nodes with Linux.
          * `lokiStackWebhook`: Enables the LokiStack CR validation and conversion webhooks.
          * `alertingRuleWebhook`: Enables the AlertingRule CR validation webhook.
          * `recordingRuleWebhook`: Enables the RecordingRule CR validation webhook.
          * `rulerConfigWebhook`: Enables the RulerConfig CR validation webhook.

          In addition it enables the following OpenShift-only related feature gates:
          * `servingCertsService`: Enables OpenShift ServiceCA annotations on the lokistack-gateway service only.
          * `ruleExtendedValidation`: Enables extended validation of AlertingRule and RecordingRule to enforce tenancy in an OpenShift context.
          * `clusterTLSPolicy`: Enables usage of TLS policies set in the API Server.
          * `clusterProxy`: Enables usage of the proxy variables set in the proxy resource.

          ### Before you start

          1. Ensure that the appropriate object storage solution, that will be used by Loki, is avaliable and configured.
        displayName: Community Loki Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - logging
        - loki
        links:
        - name: Documentation
          url: https://loki-operator.dev/
        maintainers:
        - email: loki-operator-team@googlegroups.com
          name: Grafana Loki SIG Operator
        maturity: alpha
        minKubeVersion: 1.21.1
        provider:
          name: Grafana Loki SIG Operator
        relatedImages:
        - quay.io/observatorium/opa-openshift:latest
        - quay.io/openshift/origin-kube-rbac-proxy:latest
        - docker.io/grafana/loki-operator:0.8.0
        - docker.io/grafana/loki:3.4.2
        - quay.io/community-operator-pipeline-prod/loki-operator@sha256:b7e52fc9b9a988eed8b8338a2b2eba34b020cb24a189c2fa4db63110118dbbc1
        - quay.io/observatorium/api:latest
        version: 0.8.0
      entries:
      - name: loki-operator.v0.8.0
        version: 0.8.0
      - name: loki-operator.v0.7.1
        version: 0.7.1
      - name: loki-operator.v0.7.0
        version: 0.7.0
      - name: loki-operator.v0.6.2
        version: 0.6.2
      - name: loki-operator.v0.6.1
        version: 0.6.1
      - name: loki-operator.v0.6.0
        version: 0.6.0
      - name: loki-operator.v0.5.0
        version: 0.5.0
      - name: loki-operator.v0.4.0
        version: 0.4.0
      - name: loki-operator.v0.3.0
        version: 0.3.0
      - name: loki-operator.v0.2.0
        version: 0.2.0
      - name: loki-operator.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: loki-operator
    provider:
      name: Grafana Loki SIG Operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sns-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sns-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformApplication",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "PlatformEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sns.services.k8s.aws/v1alpha1",
                "kind": "Topic",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
          createdAt: "2025-09-19T19:13:23Z"
          description: AWS SNS controller is a service controller for managing SNS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PlatformApplication represents the state of an AWS sns PlatformApplication
              resource.
            displayName: PlatformApplication
            kind: PlatformApplication
            name: platformapplications.sns.services.k8s.aws
            version: v1alpha1
          - description: PlatformEndpoint represents the state of an AWS sns PlatformEndpoint
              resource.
            displayName: PlatformEndpoint
            kind: PlatformEndpoint
            name: platformendpoints.sns.services.k8s.aws
            version: v1alpha1
          - description: Subscription represents the state of an AWS sns Subscription
              resource.
            displayName: Subscription
            kind: Subscription
            name: subscriptions.sns.services.k8s.aws
            version: v1alpha1
          - description: Topic represents the state of an AWS sns Topic resource.
            displayName: Topic
            kind: Topic
            name: topics.sns.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Notification Service (SNS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SNS**

          Amazon Simple Notification Service (Amazon SNS) is a managed service that provides message delivery from publishers to subscribers (also known as producers and consumers). Publishers communicate asynchronously with subscribers by sending messages to a topic, which is a logical access point and communication channel. Clients can subscribe to the SNS topic and receive published messages using a supported endpoint type, such as Amazon Kinesis Data Firehose, Amazon SQS, AWS Lambda, HTTP, email, mobile push notifications, and mobile text messages (SMS).

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SNS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sns
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SNS Developer Resources
          url: https://aws.amazon.com/sns/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sns maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sns-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sns-controller@sha256:f535d7861f6ef1488ccf95d0c5629a924033acb133d9452dc86e4c80352e69d2
        version: 1.2.1
      entries:
      - name: ack-sns-controller.v1.2.1
        version: 1.2.1
      - name: ack-sns-controller.v1.2.0
        version: 1.2.0
      - name: ack-sns-controller.v1.1.16
        version: 1.1.16
      - name: ack-sns-controller.v1.1.15
        version: 1.1.15
      - name: ack-sns-controller.v1.1.14
        version: 1.1.14
      - name: ack-sns-controller.v1.1.13
        version: 1.1.13
      - name: ack-sns-controller.v1.1.12
        version: 1.1.12
      - name: ack-sns-controller.v1.1.11
        version: 1.1.11
      - name: ack-sns-controller.v1.1.10
        version: 1.1.10
      - name: ack-sns-controller.v1.1.9
        version: 1.1.9
      - name: ack-sns-controller.v1.1.8
        version: 1.1.8
      - name: ack-sns-controller.v1.1.7
        version: 1.1.7
      - name: ack-sns-controller.v1.1.6
        version: 1.1.6
      - name: ack-sns-controller.v1.1.5
        version: 1.1.5
      - name: ack-sns-controller.v1.1.4
        version: 1.1.4
      - name: ack-sns-controller.v1.1.3
        version: 1.1.3
      - name: ack-sns-controller.v1.1.2
        version: 1.1.2
      - name: ack-sns-controller.v1.1.1
        version: 1.1.1
      - name: ack-sns-controller.v1.1.0
        version: 1.1.0
      - name: ack-sns-controller.v1.0.14
        version: 1.0.14
      - name: ack-sns-controller.v1.0.13
        version: 1.0.13
      - name: ack-sns-controller.v1.0.12
        version: 1.0.12
      - name: ack-sns-controller.v1.0.11
        version: 1.0.11
      - name: ack-sns-controller.v1.0.9
        version: 1.0.9
      - name: ack-sns-controller.v1.0.6
        version: 1.0.6
      - name: ack-sns-controller.v1.0.5
        version: 1.0.5
      - name: ack-sns-controller.v1.0.4
        version: 1.0.4
      - name: ack-sns-controller.v1.0.3
        version: 1.0.3
      - name: ack-sns-controller.v1.0.2
        version: 1.0.2
      - name: ack-sns-controller.v1.0.1
        version: 1.0.1
      - name: ack-sns-controller.v1.0.0
        version: 1.0.0
      - name: ack-sns-controller.v0.0.12
        version: 0.0.12
      - name: ack-sns-controller.v0.0.11
        version: 0.0.11
      - name: ack-sns-controller.v0.0.10
        version: 0.0.10
      - name: ack-sns-controller.v0.0.9
        version: 0.0.9
      - name: ack-sns-controller.v0.0.8
        version: 0.0.8
      - name: ack-sns-controller.v0.0.7
        version: 0.0.7
      - name: ack-sns-controller.v0.0.5
        version: 0.0.5
      - name: ack-sns-controller.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: ack-sns-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel Corporation
      provider-url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
    name: sriov-fec
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sriov-fec.v2.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "acc100SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "acc100": {
                        "downlink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "downlink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        },
                        "maxQueueSize": 1024,
                        "numVfBundles": 16,
                        "pfMode": true,
                        "uplink4G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 0
                        },
                        "uplink5G": {
                          "aqDepthLog2": 4,
                          "numAqsPerGroups": 16,
                          "numQueueGroups": 4
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecClusterConfig",
                "metadata": {
                  "name": "n3000SampleConfig",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "acceleratorSelector": {
                    "deviceID": "someDevice",
                    "driver": "someDriver",
                    "maxVirtualFunctions": 2,
                    "pciAddress": "somePciAddress",
                    "vendorID": "someVendor"
                  },
                  "drainSkip": false,
                  "nodeSelector": {
                    "expectedLabel1": "valueOfExpectedLabel1",
                    "expectedLabelN": "valueOfExpectedLabelN"
                  },
                  "physicalFunction": {
                    "bbDevConfig": {
                      "n3000": {
                        "downlink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        },
                        "flrTimeout": 610,
                        "networkType": "FPGA_5GNR",
                        "pfMode": true,
                        "uplink": {
                          "bandwidth": 3,
                          "loadBalance": 128,
                          "queues": {
                            "vf0": 16,
                            "vf1": 16,
                            "vf2": 0,
                            "vf3": 0,
                            "vf4": 0,
                            "vf5": 0,
                            "vf6": 0,
                            "vf7": 0
                          }
                        }
                      }
                    },
                    "pfDriver": "pci-pf-stub",
                    "vfAmount": 2,
                    "vfDriver": "vfio-pci"
                  },
                  "priority": 100
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "acc100-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "acc100": {
                          "downlink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "downlink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          },
                          "maxQueueSize": 1024,
                          "numVfBundles": 16,
                          "pfMode": true,
                          "uplink4G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 0
                          },
                          "uplink5G": {
                            "aqDepthLog2": 4,
                            "numAqsPerGroups": 16,
                            "numQueueGroups": 4
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              },
              {
                "apiVersion": "sriovfec.intel.com/v2",
                "kind": "SriovFecNodeConfig",
                "metadata": {
                  "name": "n3000-worker",
                  "namespace": "vran-acceleration-operators"
                },
                "spec": {
                  "drainSkip": false,
                  "physicalFunctions": [
                    {
                      "bbDevConfig": {
                        "n3000": {
                          "downlink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          },
                          "flrTimeout": 610,
                          "networkType": "FPGA_5GNR",
                          "pfMode": true,
                          "uplink": {
                            "bandwidth": 3,
                            "loadBalance": 128,
                            "queues": {
                              "vf0": 16,
                              "vf1": 16,
                              "vf2": 0,
                              "vf3": 0,
                              "vf4": 0,
                              "vf5": 0,
                              "vf6": 0,
                              "vf7": 0
                            }
                          }
                        }
                      },
                      "pci_addr": "somePciAddress",
                      "pf_driver": "pci-pf-stub",
                      "vf_amount": 2,
                      "vf_driver": "vfio-pci"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
          createdAt: "2025-01-15"
          description: An operator for Intel Wireless FEC Accelerator to orchestrate
            and manage the resources/devices exposed by a range of Intel's vRAN FEC
            acceleration devices/hardware within the OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/intel/sriov-fec-operator
          support: Intel Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SriovFecClusterConfig is the Schema for the sriovfecclusterconfigs
              API
            displayName: SriovFecClusterConfig
            kind: SriovFecClusterConfig
            name: sriovfecclusterconfigs.sriovfec.intel.com
            version: v2
          - description: SriovFecNodeConfig is the Schema for the sriovfecnodeconfigs
              API
            displayName: SriovFecNodeConfig
            kind: SriovFecNodeConfig
            name: sriovfecnodeconfigs.sriovfec.intel.com
            version: v2
          - description: SriovVrbClusterConfig is the Schema for the sriovvrbclusterconfigs
              API
            displayName: SriovVrbClusterConfig
            kind: SriovVrbClusterConfig
            name: sriovvrbclusterconfigs.sriovvrb.intel.com
            version: v1
          - description: SriovVrbNodeConfig is the Schema for the sriovvrbnodeconfigs
              API
            displayName: SriovVrbNodeConfig
            kind: SriovVrbNodeConfig
            name: sriovvrbnodeconfigs.sriovvrb.intel.com
            version: v1
        description: The Intel vRAN Boost family of accelerators ACC100/VRB1/VRB2,
          to offload and accelerate the computing-intensive process of forward error
          correction (FEC) and more for 4G/LTE and 5G technology, freeing up processing
          power. It allows the optimization of data plane performance to reduce total
          cost of ownership while maintaining a high degree of flexibility.  The Intel
          vRAN Boost accelerators ACC100/VRB1/VRB2 plays a key role in accelerating
          5G and network functions virtualization (NFV) adoption for ecosystem partners
          such as telecommunications equipment manufacturers (TEMs) virtual network
          functions (VNF) vendors, system integrators and telcos, to bring scalable
          and high-performance solutions to market. The Intel ACC100 includes a variant
          that is design to be Network Equipment Building System (NEBS)-friendly,
          and features a Root-of-Trust device that helps protect systems from FPGA
          host security exploits.
        displayName: SRIOV-FEC Operator for Intel vRAN Boost Accelerators
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ACC100
        - vRAN
        - ORAN
        links:
        - name: SRIOV-FEC source code
          url: https://github.com/intel/sriov-fec-operator
        maturity: alpha
        provider:
          name: Intel Corporation
          url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
        relatedImages:
        - registry.connect.redhat.com/intel/n3000-labeler@sha256:b417cfcedc998a35491784005cc7eee3d9c647b84eac5ca41cb459076870b8f6
        - registry.connect.redhat.com/intel/sriov-fec-daemon@sha256:5b0c60b67bcf5d7b2797deab0d5e3fa079577179f0acc6b0062533584ea0bc23
        - registry.connect.redhat.com/intel/sriov-fec-operator-bundle@sha256:c30849cbf451eb016786bc367e3716389b1d45676ad298c4de6a492c7bb00f4f
        - registry.connect.redhat.com/intel/sriov-fec-operator@sha256:adb6a4c2da6399c8ed40c906f790bce13bfcefae3326a047f6a50592b3f5e485
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:1ed3b514bf9af6ac0d386e3739874d1d964e667d54e041c2d67edd769b0009fb
        - registry.redhat.io/openshift4/ose-sriov-network-device-plugin@sha256:d0fcb6e9135d9a19ea26851bcdb4ac72c04bd747706fc5123020460e9dbfc21b
        version: 2.11.1
      entries:
      - name: sriov-fec.v2.11.1
        version: 2.11.1
      - name: sriov-fec.v2.11.0
        version: 2.11.0
      - name: sriov-fec.v2.10.0
        version: 2.10.0
      - name: sriov-fec.v2.9.0
        version: 2.9.0
      - name: sriov-fec.v2.8.0
        version: 2.8.0
      - name: sriov-fec.v2.7.2
        version: 2.7.2
      - name: sriov-fec.v2.7.1
        version: 2.7.1
      name: stable
    defaultChannel: stable
    packageName: sriov-fec
    provider:
      name: Intel Corporation
      url: https://builders.intel.com/docs/networkbuilders/intel-vran-dedicated-accelerator-acc100-product-brief.pdf
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-atlas-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-atlas-kubernetes.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "id": "1",
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "WEEKLY",
                      "id": "2",
                      "retentionUnit": "DAYS",
                      "retentionValue": 6
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": true,
                  "policies": [
                    {
                      "name": "atlas-default-backuppolicy",
                      "namespace": "mongodb-atlas-system"
                    },
                    {
                      "name": "atlas-default-backuppolicy2",
                      "namespace": "mongodb-atlas-system"
                    }
                  ],
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "my-database-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "david"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-cluster"
                },
                "spec": {
                  "deploymentSpec": {
                    "name": "test-cluster",
                    "providerSettings": {
                      "instanceSizeName": "M10",
                      "providerName": "AWS",
                      "regionName": "US_EAST_1"
                    }
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Database
          containerImage: registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasclusters API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB Atlas —
          the only multi-cloud document database service that gives you the versatility you need to build sophisticated and resilient applications that can adapt to changing customer demands and market trends.

          > Current Status: *Beta*. The Operator gives users the ability to provision
          > Atlas projects, clusters and database users using Kubernetes Specifications and bind connection information
          > into applications deployed to Kubernetes.

          ## Quick Start guide
          ### Step 1. Deploy Kubernetes operator by clicking Install button.

          ### Step 2. Create Atlas Cluster

          **1.** Create an Atlas API Key Secret
          In order to work with the Atlas Operator you need to provide [authentication information](https://docs.atlas.mongodb.com/configure-api-access)
           to allow the Atlas Operator to communicate with Atlas API. Once you have generated a Public and Private key in Atlas, you can create a Kuberentes Secret with:
          ```
          kubectl create secret generic mongodb-atlas-operator-api-key \
                   --from-literal="orgId=<the_atlas_organization_id>" \
                   --from-literal="publicApiKey=<the_atlas_api_public_key>" \
                   --from-literal="privateApiKey=<the_atlas_api_private_key>" \
                   -n openshift-operators
          ```
          (Note, that you should use the namespace where the Operator was installed - it's `openshift-operators` by default)

          **2.** Create an `AtlasProject` Custom Resource

          The `AtlasProject` CustomResource represents Atlas Projects in our Kubernetes cluster. You need to specify
          `projectIpAccessList` with the IP addresses or CIDR blocks of any hosts that will connect to the Atlas Cluster.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasProject
          metadata:
            name: my-project
          spec:
            name: Test Atlas Operator Project
            projectIpAccessList:
              - ipAddress: "192.0.2.15"
                comment: "IP address for Application Server A"
              - ipAddress: "203.0.113.0/24"
                comment: "CIDR block for Application Server B - D"
          ```
          **3.** Create an `AtlasDeployment` Custom Resource.
          The example below is a minimal configuration to create an M10 Atlas cluster in the AWS US East region. For a full list of properties, check
          `atlasclusters.atlas.mongodb.com` [CRD specification](config/crd/bases/atlas.mongodb.com_atlasclusters.yaml)):
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDeployment
          metadata:
            name: my-atlas-cluster
          spec:
            name: "Test-cluster"
            projectRef:
              name: my-project
            providerSettings:
              instanceSizeName: M10
              providerName: AWS
              regionName: US_EAST_1
          ```

          **4.** Create a database user password Kubernetes Secret
          The Secret must be created in the same namespace as the `AtlasDeployment` and `AtlasProject` were created.
          ```
          kubectl create secret generic the-user-password --from-literal="password=P@@sword%"
          ```

          **5.** Create an `AtlasDatabaseUser` Custom Resource

          In order to connect to an Atlas Cluster the database user needs to be created. `AtlasDatabaseUser` resource should reference
          the password Kubernetes Secret created in the previous step.
          ```
          apiVersion: atlas.mongodb.com/v1
          kind: AtlasDatabaseUser
          metadata:
            name: my-database-user
          spec:
            roles:
              - roleName: "readWriteAnyDatabase"
                databaseName: "admin"
            projectRef:
              name: my-project
            username: theuser
            passwordSecretRef:
              name: the-user-password
          ```
          **6.** Wait for the `AtlasDatabaseUser` Custom Resource to be ready

          Wait until the AtlasDatabaseUser resource gets to "ready" status (it will wait until the cluster is created that may take around 10 minutes):
          ```
          kubectl get atlasdatabaseusers my-database-user -o=jsonpath='{.status.conditions[?(@.type=="Ready")].status}'
          True
          ```
          ### Step 3. Connect your application to the Atlas Cluster

          The Atlas Operator will create a Kubernetes Secret with the information necessary to connect to the Atlas Cluster created
          in the previous step. An application in the same Kubernetes Cluster can mount and use the Secret:

          ```
          ...
          containers:
                - name: test-app
                  env:
                   - name: "CONNECTION_STRING"
                     valueFrom:
                       secretKeyRef:
                         name: test-atlas-operator-project-test-cluster-theuser
                         key: connectionStringStandardSrv

          ```
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:2338eaab863304250b3b1d267e9f2261c48eef1f97c280b7fa0af0ebc3ba208b
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator@sha256:3cfbd700e4771c2c7e75c2c41cf6de57c9ef71ecd286e96a21d2567e43424bc3
        version: 1.0.0
      entries:
      - name: mongodb-atlas-kubernetes.v1.0.0
        version: 1.0.0
      name: beta
    - currentCSV: mongodb-atlas-kubernetes.v2.11.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupCompliancePolicy",
                "metadata": {
                  "name": "my-backup-compliance-policy"
                },
                "spec": {
                  "authorizedEmail": "john.doe@example.com",
                  "authorizedUserFirstName": "John",
                  "authorizedUserLastName": "Doe",
                  "copyProtectionEnabled": false,
                  "encryptionAtRestEnabled": false,
                  "onDemandPolicy": {
                    "retentionUnit": "weeks",
                    "retentionValue": 3
                  },
                  "overwriteBackupPolicies": false,
                  "pointInTimeEnabled": true,
                  "restoreWindowDays": 42,
                  "scheduledPolicyItems": [
                    {
                      "frequencyInterval": 2,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupPolicy",
                "metadata": {
                  "name": "atlasbackuppolicy-sample"
                },
                "spec": {
                  "items": [
                    {
                      "frequencyInterval": 6,
                      "frequencyType": "hourly",
                      "retentionUnit": "days",
                      "retentionValue": 2
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "daily",
                      "retentionUnit": "days",
                      "retentionValue": 7
                    },
                    {
                      "frequencyInterval": 1,
                      "frequencyType": "weekly",
                      "retentionUnit": "weeks",
                      "retentionValue": 4
                    },
                    {
                      "frequencyInterval": 15,
                      "frequencyType": "monthly",
                      "retentionUnit": "months",
                      "retentionValue": 12
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasBackupSchedule",
                "metadata": {
                  "name": "atlasbackupschedule-sample"
                },
                "spec": {
                  "autoExportEnabled": false,
                  "copySettings": [
                    {
                      "cloudProvider": "AWS",
                      "frequencies": [
                        "WEEKLY"
                      ],
                      "regionName": "US_WEST_1",
                      "shouldCopyOplogs": false
                    }
                  ],
                  "policy": {
                    "name": "atlasbackuppolicy-sample"
                  },
                  "referenceHourOfDay": 10,
                  "referenceMinuteOfHour": 10,
                  "restoreWindowDays": 2,
                  "useOrgAndGroupNamesInExportPrefix": true
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasCustomRole",
                "metadata": {
                  "name": "shard-operator-role"
                },
                "spec": {
                  "projectRef": {
                    "name": "my-project"
                  },
                  "role": {
                    "actions": [
                      {
                        "name": "getShardMap",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "shardingState",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "connPoolStats",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      },
                      {
                        "name": "getLog",
                        "resources": [
                          {
                            "cluster": true
                          }
                        ]
                      }
                    ],
                    "inheritedRoles": [
                      {
                        "database": "admin",
                        "name": "operator-role-1"
                      }
                    ],
                    "name": "my-role"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDataFederation",
                "metadata": {
                  "name": "my-data-federation",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "name": "my-data-federation",
                  "privateEndpoints": [
                    {
                      "endpointId": "vpce-03f9eeaa764e32454",
                      "provider": "AWS",
                      "type": "DATA_LAKE"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project",
                    "namespace": "mongodb-atlas-system"
                  },
                  "storage": {
                    "databases": [
                      {
                        "collections": [
                          {
                            "dataSources": [
                              {
                                "storeName": "http-test",
                                "urls": [
                                  "https://data.cityofnewyork.us/api/views/vfnx-vebw/rows.csv"
                                ]
                              }
                            ],
                            "name": "test-collection-1"
                          }
                        ],
                        "name": "test-db-1"
                      }
                    ],
                    "stores": [
                      {
                        "name": "http-test",
                        "provider": "http"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDatabaseUser",
                "metadata": {
                  "name": "my-database-user"
                },
                "spec": {
                  "databaseName": "admin",
                  "passwordSecretRef": {
                    "name": "the-user-password"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "roles": [
                    {
                      "databaseName": "admin",
                      "roleName": "readWriteAnyDatabase"
                    }
                  ],
                  "username": "theuser"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasDeployment",
                "metadata": {
                  "name": "my-atlas-deployment"
                },
                "spec": {
                  "backupRef": {
                    "name": "atlasbackupschedule-sample"
                  },
                  "deploymentSpec": {
                    "backupEnabled": true,
                    "clusterType": "REPLICASET",
                    "mongoDBMajorVersion": "7.0",
                    "name": "test-deployment",
                    "replicationSpecs": [
                      {
                        "regionConfigs": [
                          {
                            "electableSpecs": {
                              "instanceSize": "M10",
                              "nodeCount": 3
                            },
                            "priority": 7,
                            "providerName": "AWS",
                            "regionName": "US_EAST_1"
                          }
                        ],
                        "zoneName": "Zone 1"
                      }
                    ],
                    "tags": [
                      {
                        "key": "team",
                        "value": "atlas-operator"
                      }
                    ]
                  },
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasFederatedAuth",
                "metadata": {
                  "name": "atlas-default-federated-auth",
                  "namespace": "mongodb-atlas-system"
                },
                "spec": {
                  "connectionSecretRef": {
                    "name": "my-org-secret",
                    "namespace": "mongodb-atlas-system"
                  },
                  "domainAllowList": [
                    "my-org-domain.com"
                  ],
                  "domainRestrictionEnabled": true,
                  "enabled": true,
                  "postAuthRoleGrants": [
                    "ORG_MEMBER"
                  ],
                  "roleMappings": [
                    {
                      "externalGroupName": "org-admin",
                      "roleAssignments": [
                        {
                          "role": "ORG_OWNER"
                        }
                      ]
                    },
                    {
                      "externalGroupName": "dev-team",
                      "roleAssignments": [
                        {
                          "role": "ORG_GROUP_CREATOR"
                        },
                        {
                          "projectName": "dev-project",
                          "role": "GROUP_OWNER"
                        }
                      ]
                    }
                  ],
                  "ssoDebugEnabled": false
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasIPAccessList",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasipaccesslist-sample"
                },
                "spec": {
                  "entries": [
                    {
                      "cidrBlock": "192.168.1.0/24"
                    },
                    {
                      "deleteAfterDate": "2025-01-20T23:59:59+02:00",
                      "ipAddress": "10.0.0.1"
                    }
                  ],
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkContainer",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasnetworkcontainer-sample"
                },
                "spec": {
                  "cidrBlock": "10.11.0.0/16",
                  "connectionSecret": {
                    "name": "atlas-connection-secret"
                  },
                  "externalProjectRef": {
                    "projectId": "66e2f2b621571b7e69a89b66"
                  },
                  "provider": "AWS",
                  "region": "US_EAST_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasNetworkPeering",
                "metadata": {
                  "name": "atlasnetworkpeering-sample"
                },
                "spec": {
                  "atlasCidrBlock": "10.12.0.0/16",
                  "awsConfiguration": {
                    "accepterRegionName": "US_EAST_1",
                    "awsAccountId": "23214235145",
                    "routeTableCidrBlock": "10.11.0.0/16",
                    "vpcId": "vpcid-7123782134"
                  },
                  "containerId": "623412394512350",
                  "containerRegion": "US_EAST_1",
                  "projectRef": {
                    "name": "atlas-project",
                    "namespace": "namespace"
                  },
                  "provider": "AWS"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasPrivateEndpoint",
                "metadata": {
                  "name": "atlasprivateendpoint-sample"
                },
                "spec": {
                  "awsConfiguration": {
                    "id": "vpce-f4k34w51d"
                  },
                  "projectRef": {
                    "name": "my-project"
                  },
                  "provider": "AWS",
                  "region": "EU_CENTRAL_1"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasProject",
                "metadata": {
                  "name": "my-project"
                },
                "spec": {
                  "name": "Test Atlas Operator Project",
                  "projectIpAccessList": [
                    {
                      "comment": "IP address for Application Server A",
                      "ipAddress": "192.0.2.15"
                    },
                    {
                      "cidrBlock": "203.0.113.0/24",
                      "comment": "CIDR block for Application Server B - D"
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasSearchIndexConfig",
                "metadata": {
                  "name": "atlassearchindexconfig-sample"
                },
                "spec": {
                  "analyzer": "lucene.standard",
                  "analyzers": [
                    {
                      "name": "test-analyzer",
                      "tokenFilters": [
                        {
                          "asciiFolding": {
                            "originalTokens": "omit"
                          },
                          "type": "asciiFolding"
                        }
                      ],
                      "tokenizer": {
                        "type": "whiteSpace",
                        "whitespace": {
                          "maxTokenLenght": 200
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamConnection",
                "metadata": {
                  "name": "my-streamconnection-kafka"
                },
                "spec": {
                  "kafkaConfig": {
                    "authentication": {
                      "credentials": {
                        "name": "kafka-userpass"
                      },
                      "mechanism": "SCRAM-512"
                    },
                    "bootstrapServers": "kafka.server1:9001,kafka.server2:9002,kafka.server3:9003",
                    "security": {
                      "certificate": {
                        "name": "kafka-cert"
                      },
                      "protocol": "SSL"
                    }
                  },
                  "name": "kafka-config",
                  "type": "Kafka"
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasStreamInstance",
                "metadata": {
                  "name": "my-streaminstance-sample"
                },
                "spec": {
                  "clusterConfig": {
                    "provider": "AWS",
                    "region": "VIRGINIA_USA",
                    "tier": "SP30"
                  },
                  "connectionRegistry": [
                    {
                      "name": "my-streamconnection-sample"
                    }
                  ],
                  "name": "my-instance",
                  "projectRef": {
                    "name": "my-project"
                  }
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasTeam",
                "metadata": {
                  "name": "atlasteam-sample"
                },
                "spec": {
                  "name": "my-team",
                  "usernames": [
                    "username@mongodb.com"
                  ]
                }
              },
              {
                "apiVersion": "atlas.mongodb.com/v1",
                "kind": "AtlasThirdPartyIntegration",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-atlas-kubernetes"
                  },
                  "name": "atlasintegration-sample"
                },
                "spec": {
                  "datadog": {
                    "apiKeySecretRef": {
                      "name": "secret-ref"
                    },
                    "region": "string",
                    "sendCollectionLatencyMetrics": "string",
                    "sendDatabaseMetrics": "string"
                  },
                  "projectRef": {
                    "name": "my-atlas-project"
                  },
                  "type": "DATADOG"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Database
          containerImage: quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
          createdAt: "2025-09-25T14:34:34Z"
          description: The MongoDB Atlas Kubernetes Operator enables easy management
            of Clusters in MongoDB Atlas
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/mongodb/mongodb-atlas-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: The AtlasBackupCompliancePolicy is a configuration that enforces
              specific backup and retention requirements
            displayName: Atlas Backup Compliance Policy
            kind: AtlasBackupCompliancePolicy
            name: atlasbackupcompliancepolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupPolicy is the Schema for the atlasbackuppolicies
              API
            displayName: Atlas Backup Policy
            kind: AtlasBackupPolicy
            name: atlasbackuppolicies.atlas.mongodb.com
            version: v1
          - description: AtlasBackupSchedule is the Schema for the atlasbackupschedules
              API
            displayName: Atlas Backup Schedule
            kind: AtlasBackupSchedule
            name: atlasbackupschedules.atlas.mongodb.com
            version: v1
          - description: AtlasCustomRole is the Schema for the AtlasCustomRole API
            displayName: Atlas Custom Role
            kind: AtlasCustomRole
            name: atlascustomroles.atlas.mongodb.com
            version: v1
          - description: AtlasDatabaseUser is the Schema for the Atlas Database User
              API
            displayName: Atlas Database User
            kind: AtlasDatabaseUser
            name: atlasdatabaseusers.atlas.mongodb.com
            version: v1
          - description: AtlasDataFederation is the Schema for the Atlas Data Federation
              API
            displayName: Atlas Data Federation
            kind: AtlasDataFederation
            name: atlasdatafederations.atlas.mongodb.com
            version: v1
          - description: AtlasDeployment is the Schema for the atlasdeployments API
            displayName: Atlas Deployment
            kind: AtlasDeployment
            name: atlasdeployments.atlas.mongodb.com
            version: v1
          - description: AtlasFederatedAuth is the Schema for the Atlasfederatedauth
              API
            displayName: Atlas Federated Auth
            kind: AtlasFederatedAuth
            name: atlasfederatedauths.atlas.mongodb.com
            version: v1
          - description: AtlasIPAccessList is the Schema for the atlasipaccesslists
              API.
            displayName: Atlas IPAccess List
            kind: AtlasIPAccessList
            name: atlasipaccesslists.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkContainer is the Schema for the AtlasNetworkContainer
              API
            displayName: Atlas Network Container
            kind: AtlasNetworkContainer
            name: atlasnetworkcontainers.atlas.mongodb.com
            version: v1
          - description: AtlasNetworkPeering is the Schema for the AtlasNetworkPeering
              API
            displayName: Atlas Network Peering
            kind: AtlasNetworkPeering
            name: atlasnetworkpeerings.atlas.mongodb.com
            version: v1
          - displayName: Atlas Org Settings
            kind: AtlasOrgSettings
            name: atlasorgsettings.atlas.mongodb.com
            version: v1
          - description: "The AtlasPrivateEndpoint custom resource definition (CRD)
              defines a desired [Private Endpoint](https://www.mongodb.com/docs/atlas/security-private-endpoint/#std-label-private-endpoint-overview)
              configuration for an Atlas project. It allows a private connection between
              your cloud provider and Atlas that doesn't send information through
              a public network. \n You can use private endpoints to create a unidirectional
              connection to Atlas clusters from your virtual network."
            displayName: Atlas Private Endpoint
            kind: AtlasPrivateEndpoint
            name: atlasprivateendpoints.atlas.mongodb.com
            version: v1
          - description: AtlasProject is the Schema for the atlasprojects API
            displayName: Atlas Project
            kind: AtlasProject
            name: atlasprojects.atlas.mongodb.com
            version: v1
          - description: Atlas Search Index Config is the Schema for the Atlas Search
              API
            displayName: Atlas Search Index Config
            kind: AtlasSearchIndexConfig
            name: atlassearchindexconfigs.atlas.mongodb.com
            version: v1
          - description: AtlasStreamConnection is the Schema for the atlasstreamconnections
              API
            displayName: Atlas Stream Connection
            kind: AtlasStreamConnection
            name: atlasstreamconnections.atlas.mongodb.com
            version: v1
          - description: AtlasStreamInstance is the Schema for the atlasstreaminstances
              API
            displayName: Atlas Stream Instance
            kind: AtlasStreamInstance
            name: atlasstreaminstances.atlas.mongodb.com
            version: v1
          - description: AtlasTeam is the Schema for the Atlas Teams API
            displayName: Atlas Team
            kind: AtlasTeam
            name: atlasteams.atlas.mongodb.com
            version: v1
          - description: AtlasThirdPartyIntegration is the Schema for the atlas 3rd
              party integrations API.
            displayName: Atlas Third Party Integration
            kind: AtlasThirdPartyIntegration
            name: atlasthirdpartyintegrations.atlas.mongodb.com
            version: v1
        description: |
          The MongoDB Atlas Operator provides a native integration between the Kubernetes orchestration platform and MongoDB
          Atlas — the only multi-cloud document database service that gives you the versatility you need to build sophisticated
          and resilient applications that can adapt to changing customer demands and market trends.

          The full documentation for the Operator can be found [here](https://docs.atlas.mongodb.com/atlas-operator/)

          ## Getting Started

          ### Supported features

          * Create and configure a project, or connect to an existing one.
          * Deploy, manage, scale, and tear down clusters.
          * Support for serverless instances.
          * Create and edit database users.
          * Manage IP Access Lists, network peering and private endpoints.
          * Configure and control Atlas’s fully managed cloud backup.
          * Configure federated authentication for your Atlas organization
          * Integrate Atlas monitoring with Prometheus.

          ... and more.

          To view the list of custom resources and each CR's schema, visit our [reference](https://www.mongodb.com/docs/atlas/operator/stable/custom-resources/)
          documentation. See the [Quickstart](https://www.mongodb.com/docs/atlas/operator/stable/ak8so-quick-start/) to get
          started with Atlas Kubernetes Operator.
        displayName: MongoDB Atlas Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - MongoDB
        - Atlas
        - Database
        - Replica Set
        - Cluster
        links:
        - name: MongoDB Atlas Kubernetes
          url: https://github.com/mongodb/mongodb-atlas-kubernetes
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: beta
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:9eb6465e723ede43ac47eed6de22042d26c0efb276301a4424a6a3948a8f5b15
        - registry.connect.redhat.com/mongodb/mongodb-atlas-kubernetes-operator-bundle@sha256:40802ba8bf531637f19adbbaeba3631e0b0557168ecf61f566fb5f2da25d1d4f
        - quay.io/mongodb/mongodb-atlas-kubernetes-operator@sha256:8ac3f6b1c0f31af5f32562ddeed37dd6586ac47e1e7e111947df3ed570c5bed9
        version: 2.11.0
      entries:
      - name: mongodb-atlas-kubernetes.v2.11.0
        version: 2.11.0
      - name: mongodb-atlas-kubernetes.v2.10.0
        version: 2.10.0
      - name: mongodb-atlas-kubernetes.v2.9.1
        version: 2.9.1
      - name: mongodb-atlas-kubernetes.v2.9.0
        version: 2.9.0
      - name: mongodb-atlas-kubernetes.v2.8.2
        version: 2.8.2
      - name: mongodb-atlas-kubernetes.v2.8.1
        version: 2.8.1
      - name: mongodb-atlas-kubernetes.v2.8.0
        version: 2.8.0
      - name: mongodb-atlas-kubernetes.v2.7.1
        version: 2.7.1
      - name: mongodb-atlas-kubernetes.v2.7.0
        version: 2.7.0
      - name: mongodb-atlas-kubernetes.v2.6.1
        version: 2.6.1
      - name: mongodb-atlas-kubernetes.v2.6.0
        version: 2.6.0
      - name: mongodb-atlas-kubernetes.v2.5.0
        version: 2.5.0
      - name: mongodb-atlas-kubernetes.v2.4.1
        version: 2.4.1
      - name: mongodb-atlas-kubernetes.v2.4.0
        version: 2.4.0
      - name: mongodb-atlas-kubernetes.v2.3.1
        version: 2.3.1
      - name: mongodb-atlas-kubernetes.v2.3.0
        version: 2.3.0
      - name: mongodb-atlas-kubernetes.v2.2.2
        version: 2.2.2
      - name: mongodb-atlas-kubernetes.v2.2.1
        version: 2.2.1
      - name: mongodb-atlas-kubernetes.v2.2.0
        version: 2.2.0
      - name: mongodb-atlas-kubernetes.v2.1.0
        version: 2.1.0
      - name: mongodb-atlas-kubernetes.v2.0.1
        version: 2.0.1
      - name: mongodb-atlas-kubernetes.v2.0.0
        version: 2.0.0
      - name: mongodb-atlas-kubernetes.v1.9.1
        version: 1.9.1
      - name: mongodb-atlas-kubernetes.v1.9.0
        version: 1.9.0
      - name: mongodb-atlas-kubernetes.v1.8.2
        version: 1.8.2
      - name: mongodb-atlas-kubernetes.v1.8.1
        version: 1.8.1
      - name: mongodb-atlas-kubernetes.v1.8.0
        version: 1.8.0
      - name: mongodb-atlas-kubernetes.v1.7.3
        version: 1.7.3
      - name: mongodb-atlas-kubernetes.v1.7.2
        version: 1.7.2
      - name: mongodb-atlas-kubernetes.v1.7.1
        version: 1.7.1
      - name: mongodb-atlas-kubernetes.v1.7.0
        version: 1.7.0
      - name: mongodb-atlas-kubernetes.v1.6.1
        version: 1.6.1
      - name: mongodb-atlas-kubernetes.v1.6.0
        version: 1.6.0
      - name: mongodb-atlas-kubernetes.v1.5.1
        version: 1.5.1
      - name: mongodb-atlas-kubernetes.v1.5.0
        version: 1.5.0
      - name: mongodb-atlas-kubernetes.v1.4.1
        version: 1.4.1
      - name: mongodb-atlas-kubernetes.v1.3.0
        version: 1.3.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-atlas-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-db
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-db.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Db",
                "metadata": {
                  "name": "db-sample"
                },
                "spec": {
                  "clusterDomain": "cluster.local",
                  "databases": [
                    {
                      "name": "wapdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "grafanadb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "brokerdb",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    },
                    {
                      "name": "nimbusmosaic",
                      "secret": {
                        "key": "password",
                        "name": "wap-db-postgresql-token"
                      },
                      "user": "user"
                    }
                  ],
                  "image": {
                    "migrationJobImage": "wap-ose-cli",
                    "migrationJobRepository": "prod.harbor.keysight.digital/loadcore",
                    "migrationJobTag": "v4.13.0-3747-35-ocp",
                    "pgImage": "postgresql-ubi-image",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "pgTag": "14.2.0-3747-32"
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "postgresql": {
                    "image": {
                      "pullSecrets": [
                        "load-core-secret"
                      ],
                      "registry": "prod.harbor.keysight.digital/loadcore",
                      "repository": "postgresql-ubi-image",
                      "tag": "14.2.0-3747-32"
                    },
                    "persistence": {
                      "storageClass": "managed-loadcore-storage"
                    },
                    "pgHbaConfiguration": "local all all trust\nhost all all 0.0.0.0/0 trust",
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "storageClass": "managed-loadcore-storage"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "volumePermissions": {
                        "containerSecurityContext": {
                          "runAsUser": "auto"
                        }
                      }
                    },
                    "serviceAccount": {
                      "automountServiceAccountToken": false,
                      "name": "load-core-service-account"
                    }
                  },
                  "storage": "8Gi"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:30:46Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Db
            name: dbs.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP DB Operator
        displayName: Keysight WAP DB Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Db
          url: https://keysight-wap-db.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        - prod.harbor.keysight.digital/loadcore/keysight-wap-db@sha256:5689d6a4da2fdd97223bc64aa61a8574bba810f93d09ab804cb45b4475f80ff5
        - registry.connect.redhat.com/loadcore/keysight-wap-db-bundle@sha256:3940b8cbf7579d9ce888f176f96c4d191a7cdfaec4be9251e8da4942e7897e97
        version: 0.1.21
      entries:
      - name: keysight-wap-db.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-db
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: kubeturbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubeturbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:7111c7c4002a447dc6977dbb73e96507bae56766fc0552fb2d036b20583722ba
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:2647a21dfb60364cfb7c4d1b81508adc13d859e97553bb2f710901e6ad59f4a3
        version: 8.17.40-beta.1
      entries:
      - name: kubeturbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: kubeturbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Kubeturbo","metadata":{"name":"kubeturbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"},"targetConfig":{"targetName":"Cluster_Name"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBM Turbonomic software automates and simultaneously optimizes
              performance, compliance, and cost in real time. Turbonomic discovers
              and automates resource optimization for the complete application stack.
              Applications are continually resourced to perform while satisfying technical
              and business constraints.
            displayName: Kubeturbo Operator
            kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1
          - kind: Kubeturbo
            name: kubeturbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Optimization for Kubernetes
          IBM Turbonomic software automates and simultaneously optimizes performance, compliance, and cost in real time.
          Turbonomic discovers and automates resource optimization for the complete application stack. Applications are continually resourced to perform while satisfying technical and business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (proactive rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          Turbonomic assures application performance by giving workloads the resources they need when they need them.

          ### How Does It Work?
          Turbonomic uses a client agent called — KubeTurbo — that runs in your cluster to discover, monitor and optimize resource usage within your environment.
          KubeTurbo runs together with the default Kubernetes scheduler and sends data back to the Turbonomic for analysis.
          Turbonomic determines the right actions that drive continuous health for your workloads and your cluster.

          Turbonomic discovers Red Hat OpenShift Virtualization resources through a new or existing target configuration with Red Hat OpenShift. For more information refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=targets-red-hat-openshift-virtualization)

          ### Get Started
          Please follow the [Kubeturbo Documentation Page](https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets) for all deployment options and an overview of typical use cases.

          ### Known Issues
          For known issues refer to this [link](https://www.ibm.com/docs/en/tarm/8.16.x?topic=notes-known-issues)
        displayName: Kubeturbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Kubeturbo Documentation Page
          url: https://www.ibm.com/docs/en/tarm/latest?topic=configuration-container-platform-targets
        - name: Kubeturbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/kubeturbo
        - name: IBM Turbonomic Application Resource Management
          url: https://www.ibm.com/products/turbonomic
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/kubeturbo-operator@sha256:dfaeb7ac2ccfed03042febea709f0ab66a4c57d62fa041b7ea887a0783d60a36
        - registry.connect.redhat.com/turbonomic/kubeturbo-operator-bundle@sha256:8d20352c0091985cf93e39cd18b2bb092bc15fc5e9900fe46ed5c6fc3c0af885
        version: 8.17.30
      entries:
      - name: kubeturbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: kubeturbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ec2-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ec2-controller.v1.7.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "DHCPOptions",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "ElasticIPAddress",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Instance",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "InternetGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "NATGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "RouteTable",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "SecurityGroup",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "Subnet",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "TransitGateway",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPC",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "ec2.services.k8s.aws/v1alpha1",
                "kind": "VPCEndpoint",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
          createdAt: "2025-09-27T00:04:55Z"
          description: AWS EC2 controller is a service controller for managing EC2
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CapacityReservation represents the state of an AWS ec2 CapacityReservation
              resource.
            displayName: CapacityReservation
            kind: CapacityReservation
            name: capacityreservations.ec2.services.k8s.aws
            version: v1alpha1
          - description: DHCPOptions represents the state of an AWS ec2 DHCPOptions
              resource.
            displayName: DHCPOptions
            kind: DHCPOptions
            name: dhcpoptions.ec2.services.k8s.aws
            version: v1alpha1
          - description: ElasticIPAddress represents the state of an AWS ec2 ElasticIPAddress
              resource.
            displayName: ElasticIPAddress
            kind: ElasticIPAddress
            name: elasticipaddresses.ec2.services.k8s.aws
            version: v1alpha1
          - description: FlowLog represents the state of an AWS ec2 FlowLog resource.
            displayName: FlowLog
            kind: FlowLog
            name: flowlogs.ec2.services.k8s.aws
            version: v1alpha1
          - description: Instance represents the state of an AWS ec2 Instance resource.
            displayName: Instance
            kind: Instance
            name: instances.ec2.services.k8s.aws
            version: v1alpha1
          - description: InternetGateway represents the state of an AWS ec2 InternetGateway
              resource.
            displayName: InternetGateway
            kind: InternetGateway
            name: internetgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: LaunchTemplate represents the state of an AWS ec2 LaunchTemplate
              resource.
            displayName: LaunchTemplate
            kind: LaunchTemplate
            name: launchtemplates.ec2.services.k8s.aws
            version: v1alpha1
          - description: NATGateway represents the state of an AWS ec2 NATGateway
              resource.
            displayName: NATGateway
            kind: NATGateway
            name: natgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: NetworkACL represents the state of an AWS ec2 NetworkACL
              resource.
            displayName: NetworkACL
            kind: NetworkACL
            name: networkacls.ec2.services.k8s.aws
            version: v1alpha1
          - description: RouteTable represents the state of an AWS ec2 RouteTable
              resource.
            displayName: RouteTable
            kind: RouteTable
            name: routetables.ec2.services.k8s.aws
            version: v1alpha1
          - description: SecurityGroup represents the state of an AWS ec2 SecurityGroup
              resource.
            displayName: SecurityGroup
            kind: SecurityGroup
            name: securitygroups.ec2.services.k8s.aws
            version: v1alpha1
          - description: Subnet represents the state of an AWS ec2 Subnet resource.
            displayName: Subnet
            kind: Subnet
            name: subnets.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGateway represents the state of an AWS ec2 TransitGateway
              resource.
            displayName: TransitGateway
            kind: TransitGateway
            name: transitgateways.ec2.services.k8s.aws
            version: v1alpha1
          - description: TransitGatewayVPCAttachment represents the state of an AWS
              ec2 TransitGatewayVPCAttachment resource.
            displayName: TransitGatewayVPCAttachment
            kind: TransitGatewayVPCAttachment
            name: transitgatewayvpcattachments.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpoint represents the state of an AWS ec2 VPCEndpoint
              resource.
            displayName: VPCEndpoint
            kind: VPCEndpoint
            name: vpcendpoints.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCEndpointServiceConfiguration represents the state of an
              AWS ec2 VPCEndpointServiceConfiguration resource.
            displayName: VPCEndpointServiceConfiguration
            kind: VPCEndpointServiceConfiguration
            name: vpcendpointserviceconfigurations.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPCPeeringConnection represents the state of an AWS ec2 VPCPeeringConnection
              resource.
            displayName: VPCPeeringConnection
            kind: VPCPeeringConnection
            name: vpcpeeringconnections.ec2.services.k8s.aws
            version: v1alpha1
          - description: VPC represents the state of an AWS ec2 VPC resource.
            displayName: VPC
            kind: VPC
            name: vpcs.ec2.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Elastic Compute Cloud (EC2) resources in AWS from within your Kubernetes cluster.

          **About Amazon EC2**

          Amazon Elastic Compute Cloud (Amazon EC2) provides scalable computing capacity in the Amazon Web Services (AWS) Cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster. You can use Amazon EC2 to launch as many or as few virtual servers as you need, configure security and networking, and manage storage. Amazon EC2 enables you to scale up or down to handle changes in requirements or spikes in popularity, reducing your need to forecast traffic.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon EC2
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ec2
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon EC2 Developer Resources
          url: https://aws.amazon.com/ec2/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ec2 maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ec2-controller:1.7.0
        - quay.io/community-operator-pipeline-prod/ack-ec2-controller@sha256:58d3249744b82a7f75ea2b1461adb03259350930da791da62841f05d91e94d01
        version: 1.7.0
      entries:
      - name: ack-ec2-controller.v1.7.0
        version: 1.7.0
      - name: ack-ec2-controller.v1.6.2
        version: 1.6.2
      - name: ack-ec2-controller.v1.6.1
        version: 1.6.1
      - name: ack-ec2-controller.v1.6.0
        version: 1.6.0
      - name: ack-ec2-controller.v1.4.11
        version: 1.4.11
      - name: ack-ec2-controller.v1.4.10
        version: 1.4.10
      - name: ack-ec2-controller.v1.4.9
        version: 1.4.9
      - name: ack-ec2-controller.v1.4.8
        version: 1.4.8
      - name: ack-ec2-controller.v1.4.7
        version: 1.4.7
      - name: ack-ec2-controller.v1.4.6
        version: 1.4.6
      - name: ack-ec2-controller.v1.4.5
        version: 1.4.5
      - name: ack-ec2-controller.v1.4.4
        version: 1.4.4
      - name: ack-ec2-controller.v1.4.2
        version: 1.4.2
      - name: ack-ec2-controller.v1.4.1
        version: 1.4.1
      - name: ack-ec2-controller.v1.4.0
        version: 1.4.0
      - name: ack-ec2-controller.v1.3.7
        version: 1.3.7
      - name: ack-ec2-controller.v1.3.6
        version: 1.3.6
      - name: ack-ec2-controller.v1.3.5
        version: 1.3.5
      - name: ack-ec2-controller.v1.3.4
        version: 1.3.4
      - name: ack-ec2-controller.v1.3.3
        version: 1.3.3
      - name: ack-ec2-controller.v1.3.1
        version: 1.3.1
      - name: ack-ec2-controller.v1.3.0
        version: 1.3.0
      - name: ack-ec2-controller.v1.2.28
        version: 1.2.28
      - name: ack-ec2-controller.v1.2.27
        version: 1.2.27
      - name: ack-ec2-controller.v1.2.26
        version: 1.2.26
      - name: ack-ec2-controller.v1.2.25
        version: 1.2.25
      - name: ack-ec2-controller.v1.2.24
        version: 1.2.24
      - name: ack-ec2-controller.v1.2.23
        version: 1.2.23
      - name: ack-ec2-controller.v1.2.22
        version: 1.2.22
      - name: ack-ec2-controller.v1.2.21
        version: 1.2.21
      - name: ack-ec2-controller.v1.2.20
        version: 1.2.20
      - name: ack-ec2-controller.v1.2.17
        version: 1.2.17
      - name: ack-ec2-controller.v1.2.16
        version: 1.2.16
      - name: ack-ec2-controller.v1.2.15
        version: 1.2.15
      - name: ack-ec2-controller.v1.2.14
        version: 1.2.14
      - name: ack-ec2-controller.v1.2.13
        version: 1.2.13
      - name: ack-ec2-controller.v1.2.12
        version: 1.2.12
      - name: ack-ec2-controller.v1.2.11
        version: 1.2.11
      - name: ack-ec2-controller.v1.2.10
        version: 1.2.10
      - name: ack-ec2-controller.v1.2.9
        version: 1.2.9
      - name: ack-ec2-controller.v1.2.8
        version: 1.2.8
      - name: ack-ec2-controller.v1.2.7
        version: 1.2.7
      - name: ack-ec2-controller.v1.2.6
        version: 1.2.6
      - name: ack-ec2-controller.v1.2.4
        version: 1.2.4
      - name: ack-ec2-controller.v1.2.1
        version: 1.2.1
      - name: ack-ec2-controller.v1.2.0
        version: 1.2.0
      - name: ack-ec2-controller.v1.1.2
        version: 1.1.2
      - name: ack-ec2-controller.v1.1.1
        version: 1.1.1
      - name: ack-ec2-controller.v1.1.0
        version: 1.1.0
      - name: ack-ec2-controller.v1.0.7
        version: 1.0.7
      - name: ack-ec2-controller.v1.0.6
        version: 1.0.6
      - name: ack-ec2-controller.v1.0.4
        version: 1.0.4
      - name: ack-ec2-controller.v1.0.3
        version: 1.0.3
      - name: ack-ec2-controller.v1.0.2
        version: 1.0.2
      - name: ack-ec2-controller.v1.0.1
        version: 1.0.1
      - name: ack-ec2-controller.v1.0.0
        version: 1.0.0
      - name: ack-ec2-controller.v0.1.1
        version: 0.1.1
      - name: ack-ec2-controller.v0.1.0
        version: 0.1.0
      - name: ack-ec2-controller.v0.0.21
        version: 0.0.21
      - name: ack-ec2-controller.v0.0.20
        version: 0.0.20
      - name: ack-ec2-controller.v0.0.19
        version: 0.0.19
      - name: ack-ec2-controller.v0.0.18
        version: 0.0.18
      - name: ack-ec2-controller.v0.0.17
        version: 0.0.17
      - name: ack-ec2-controller.v0.0.16
        version: 0.0.16
      - name: ack-ec2-controller.v0.0.15
        version: 0.0.15
      - name: ack-ec2-controller.v0.0.14
        version: 0.0.14
      - name: ack-ec2-controller.v0.0.13
        version: 0.0.13
      - name: ack-ec2-controller.v0.0.12
        version: 0.0.12
      - name: ack-ec2-controller.v0.0.11
        version: 0.0.11
      - name: ack-ec2-controller.v0.0.10
        version: 0.0.10
      - name: ack-ec2-controller.v0.0.9
        version: 0.0.9
      - name: ack-ec2-controller.v0.0.8
        version: 0.0.8
      - name: ack-ec2-controller.v0.0.7
        version: 0.0.7
      - name: ack-ec2-controller.v0.0.6
        version: 0.0.6
      - name: ack-ec2-controller.v0.0.5
        version: 0.0.5
      - name: ack-ec2-controller.v0.0.4
        version: 0.0.4
      - name: ack-ec2-controller.v0.0.3
        version: 0.0.3
      - name: ack-ec2-controller.v0.0.2
        version: 0.0.2
      name: alpha
    defaultChannel: alpha
    packageName: ack-ec2-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: egressip-ipam-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: egressip-ipam-operator.v1.2.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-aws"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "10.0.128.0/19",
                      "labelValue": "us-west-2a"
                    },
                    {
                      "CIDR": "10.0.160.0/19",
                      "labelValue": "us-west-2b",
                      "reservedIPs": [
                        "10.0.160.126"
                      ]
                    },
                    {
                      "CIDR": "10.0.192.0/19",
                      "labelValue": "us-west-2c"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "failure-domain.beta.kubernetes.io/zone"
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "EgressIPAM",
                "metadata": {
                  "name": "egressipam-baremetal"
                },
                "spec": {
                  "cidrAssignments": [
                    {
                      "CIDR": "192.169.0.0/24",
                      "labelValue": "true"
                    }
                  ],
                  "nodeSelector": {
                    "matchLabels": {
                      "node-role.kubernetes.io/worker": ""
                    }
                  },
                  "topologyLabel": "egressGateway"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Networking, Cloud Provider, Security
          certified: "false"
          containerImage: quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
          createdAt: "2022-07-25T15:07:22Z"
          description: This operator provides automation around managing egressIPs.
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: egressip-ipam-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.8.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/egressip-ipam-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: EgressIPAM is the Schema for the egressipams API
            displayName: Egress IPAM
            kind: EgressIPAM
            name: egressipams.redhatcop.redhat.io
            version: v1alpha1
        description: |
          This operator automates the assignment of egressIPs to namespaces.
          Namespaces can opt in to receiving one or more egressIPs with the following annotation `egressip-ipam-operator.redhat-cop.io/egressipam:<egressIPAM>`, where `egressIPAM` is the CRD that controls how egressIPs are assigned.
          IPs assigned to the namespace can be looked up in the following annotation: `egressip-ipam-operator.redhat-cop.io/egressips`.

          EgressIP assignments is managed by the EgressIPAM CRD. Here is an example of it:

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: example-egressipam
          spec:
            cidrAssignments:
              - labelValue: "true"
                CIDR: 192.169.0.0/24
                reservedIPs:
                - "192.159.0.5"
            topologyLabel: egressGateway
            nodeSelector:
              node-role.kubernetes.io/worker: ""
          ```

          This EgressIPAM specifies that all nodes that comply with the specified node selector and that also have labels `egressGateway: "true"` will be assigned egressIP from the specified CIDR.

          Note that the `cidrAssigments` field is an array and therefore, multiple groups of nodes can be identified with the labelValue and different CIDRs can be assigned to them. This is usually not necessary on a bare metal deployment.

          In the bare metal scenario, when this egressCRD is created, all the `hostsubnet` relative to the nodes selected by this EgressIPAM will be update to have the EgressCIDRs field equal to the specified CIDR (see below for cloud provider scenarios).

          When a namespace is created with the opt-in annotation: `egressip-ipam-operator.redhat-cop.io/egressipam=<egressIPAM>`, an available egressIP is selected from the CIDR and assigned to the namespace.The `netnamespace` associated with this namespace is updated to use that egressIP.

          It is possible to specify a set of reserved IPs. These IPs must belong to the CIDR and will never be assigned.

          ## Passing EgressIPs as input

          The normal mode of operation of this operator is to pick a random IP from the configured CIDR. However, it also supports a scenario where egressIPs are picked by an external process and passed as input.

          In this case IPs must me passed as an annotation to the namespace, like this: `egressip-ipam-operator.redhat-cop.io/egressips=IP1,IP2...`. The value of the annotation is a comma separated array of ip with no spaces.

          There must be exactly one IP per CIDR defined in the referenced egressIPAM. Moreover, each IP must belong to the corresponding CIDR. Because this situation can lead to inconsistencies, failing to have correspondence between IPs in the namespace annotation and CIDRs in the egressIPAM CR will cause the operator to error out and stop processing all the namespaces associated with the given EgressIPAM CR.

          It also also a responsibility of the progress picking the IPs to ensure that those IPs are actually available.

          ## Assumptions

          The following assumptions apply when using this operator:

          1. If multiple EgressIPAMs are defined, their selected nodes MUST NOT overlap.
          2. When using a cloud provider the topology label MUST be `failure-domain.beta.kubernetes.io/zone`.

          ## Support for AWS

          In AWS as well as other cloud providers, one cannot freely assign IPs to machines. Additional steps need to be performed in this case. Considering this EgressIPAM

          ```yaml
          apiVersion: redhatcop.redhat.io/v1alpha1
          kind: EgressIPAM
          metadata:
            name: egressipam-aws
          spec:
            cidrAssignments:
              - labelValue: "eu-central-1a"
                CIDR: 10.0.128.0/20
              - labelValue: "eu-central-1b"
                CIDR: 10.0.144.0/20
              - labelValue: "eu-central-1c"
                CIDR: 10.0.160.0/20
            topologyLabel: failure-domain.beta.kubernetes.io/zone
            nodeSelector:
              matchLabels:
                node-role.kubernetes.io/worker: ""
          ```

          When a namespace with the opt-in annotation is created, the following happens:

          1. for each of the CIDRs, an available IP is selected and assigned to the namespace
          2. the relative `netnamespace` is update to reflect the assignment (multiple IPs will be assigned in this case).
          3. one node per zone is selected to carry the egressIP
          4. the relative aws machines are assigned the additional IP on the main interface (support for secondary interfaces in not available)
          5. the relative `hostsubnets` are updated to reflect the assigned IP, the `egressIP` field is updated.
        displayName: Egressip Ipam Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - egressIPs
        - high availability
        - firewall support
        - traffic management
        - ipam
        links:
        - name: Egressip Ipam Operator
          url: https://github.com/redhat-cop/egressip-ipam-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/egressip-ipam-operator:v0.1.2
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/egressip-ipam-operator@sha256:1e2443488e0aad361cccece51689e2b8da882862cb62c1208a14ab1b5ec2922d
        - quay.io/redhat-cop/egressip-ipam-operator@sha256:d6129bd287b87696a23e0da64dcd598958f779dc88ae33b46b3c909dbab80adf
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        version: 1.2.4
      entries:
      - name: egressip-ipam-operator.v1.2.4
        version: 1.2.4
      - name: egressip-ipam-operator.v1.2.3
        version: 1.2.3
      - name: egressip-ipam-operator.v1.2.2
        version: 1.2.2
      - name: egressip-ipam-operator.v1.2.1
        version: 1.2.1
      - name: egressip-ipam-operator.v1.2.0
        version: 1.2.0
      - name: egressip-ipam-operator.v1.1.0
        version: 1.1.0
      - name: egressip-ipam-operator.v1.0.8
        version: 1.0.8
      - name: egressip-ipam-operator.v1.0.7
        version: 1.0.7
      - name: egressip-ipam-operator.v1.0.6
        version: 1.0.6
      - name: egressip-ipam-operator.v1.0.5
        version: 1.0.5
      - name: egressip-ipam-operator.v1.0.4
        version: 1.0.4
      - name: egressip-ipam-operator.v1.0.3
        version: 1.0.3
      - name: egressip-ipam-operator.v1.0.2
        version: 1.0.2
      - name: egressip-ipam-operator.v1.0.1
        version: 1.0.1
      - name: egressip-ipam-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: egressip-ipam-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Silicom Ltd.
      provider-url: http://www.silicom-usa.com
    name: silicom-sts-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: silicom-sts-operator.v0.0.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "openshift-operators",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "images": null,
                  "sro": null
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.6
          createdAt: 2021-11-17T10:21:24 CET
          description: |
            This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of
            Silicom STS device plugins for Kubernetes.
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io.bundle.channels.v1: alpha
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/tsyncd@sha256:78be78ec8939f025f9a8b899693ab77b3990116d1c7ffb2115c620b7c3382d9d
        - quay.io/silicom/gpsd@sha256:649e8a6c524920e1c564d8f797b8d64235b2527aa923f94bc9c7a0b7aba8ae7f
        - quay.io/silicom/ice-driver-src@sha256:ca728f648769d31501cdedda92554bcfa890754405b0b06d41629cea47707da8
        - quay.io/silicom/phc2sys@sha256:0f3b63a3059b7d6158a916814281097f0d418b99a5fac9e8e5fd9bae58fb0b74
        - quay.io/silicom/sts-operator:0.0.6
        - quay.io/silicom/sts-plugin@sha256:44339fd400d46d9dddb02c2da47c6f9f96ad2605a5c6de8a7dce2f8c3da3bb57
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/silicom/grpc-tsyncd@sha256:aac7233423439e352d8eb39cef63e479c30c2f80b063c84ffc8797cc5a045117
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:5d4e4e62b03888bcbf6336934f2ddf5c5cb3d91bc8482a3aa1697e09625c6865
        - quay.io/silicom/tsync_extts@sha256:be6a5950a5fdc6fdf8015767901856bf173d01bdc803be7141e4f9795ab1478a
        version: 0.0.6
      entries:
      - name: silicom-sts-operator.v0.0.6
        version: 0.0.6
      - name: silicom-sts-operator.v0.0.5
        version: 0.0.5
      - name: silicom-sts-operator.v0.0.4
        version: 0.0.4
      - name: silicom-sts-operator.v0.0.3
        version: 0.0.3
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: silicom-sts-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "gm-1",
                  "namespace": "sts-silicom"
                },
                "spec": {
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "mode": "Master",
                      "synce": 1
                    }
                  ],
                  "mode": "T-GM.8275.1",
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "gm-1"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsConfig",
                "metadata": {
                  "name": "bc-1"
                },
                "spec": {
                  "aprLevel": 0,
                  "domainNumber": 24,
                  "esmcMode": 1,
                  "forwardable": 0,
                  "interfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Slave",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "holdoff": 500,
                      "ipv4": 1,
                      "ipv6": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "holdoff": 500,
                      "ipv4": 1,
                      "mode": "Master",
                      "ql": 4,
                      "qlEnable": 1,
                      "synce": 1
                    }
                  ],
                  "mode": "T-TBC.8275.1",
                  "mode10MHz": 2,
                  "modePPS": 2,
                  "namespace": "sts-silicom",
                  "nodeSelector": {
                    "sts.silicom.com/config": "bc-1"
                  },
                  "phyLedsCtl": 0,
                  "priority2": 128,
                  "src10MHz": 1,
                  "srcPPS": 1,
                  "syncOption": 1,
                  "synceCpu": 10,
                  "synceRecClkPort": 0,
                  "tracePtpMsg": -1,
                  "twoStep": 0
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsNode",
                "metadata": {
                  "name": "worker2"
                },
                "spec": null,
                "status": {
                  "ethInterfaces": [
                    {
                      "ethName": "enp2s0f0",
                      "ethPort": 0,
                      "pciAddr": "02:00.0",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f1",
                      "ethPort": 1,
                      "pciAddr": "02:00.1",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f2",
                      "ethPort": 2,
                      "pciAddr": "02:00.2",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f3",
                      "ethPort": 3,
                      "pciAddr": "02:00.3",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f4",
                      "ethPort": 4,
                      "pciAddr": "02:00.4",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f5",
                      "ethPort": 5,
                      "pciAddr": "02:00.5",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f6",
                      "ethPort": 6,
                      "pciAddr": "02:00.6",
                      "status": "down"
                    },
                    {
                      "ethName": "enp2s0f7",
                      "ethPort": 7,
                      "pciAddr": "02:00.7",
                      "status": "down"
                    }
                  ],
                  "gpsStatus": {
                    "active": 0,
                    "device": "",
                    "lat": "",
                    "lon": "",
                    "mode": 0,
                    "time": "2021-12-01T09:57:10.968Z"
                  },
                  "tsyncStatus": {
                    "mode": "PTP Master Mode",
                    "status": "Normal Status",
                    "time": "Wed, 01 Dec 2021 09:57:10 UTC"
                  }
                }
              },
              {
                "apiVersion": "sts.silicom.com/v1alpha1",
                "kind": "StsOperatorConfig",
                "metadata": {
                  "name": "sts-operator-config"
                },
                "spec": {
                  "gpsSvcPort": 2947,
                  "grpcSvcPort": 50051,
                  "images": {
                    "gpsd": "quay.io/silicom/gpsd:3.23.1",
                    "grpcTsyncd": "quay.io/silicom/grpc-tsyncd:2.1.0.0",
                    "phcs2Sys": "quay.io/silicom/phcs2sys:3.1.1",
                    "stsPlugin": "quay.io/silicom/sts-plugin:0.0.2",
                    "ts2Phcs": "quay.io/silicom/ts2phcs:1.0.0",
                    "tsyncd": "quay.io/silicom/tsyncd:2.1.0.0"
                  },
                  "namespace": "sts-silicom",
                  "sro": {
                    "chart": {
                      "name": "ice-special-resource",
                      "repository": {
                        "name": "ice-special-resource",
                        "url": "http://ice-driver-src"
                      },
                      "version": "0.0.1"
                    },
                    "iceVersion": "1.7.16",
                    "namespace": "openshift-operators",
                    "srcImage": "quay.io/silicom/ice-driver-src:1.7.16",
                    "srcSvcPort": 3000
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Drivers and plugins
          certified: "false"
          containerImage: quay.io/silicom/sts-operator:0.0.2
          createdAt: 2021-11-17T10:21:24 CET
          description: This Operator is a Kubernetes custom controller whose goal
            is to serve the installation and lifecycle management of Silicom STS device
            plugins for Kubernetes.
          operators.operatorframework.io.bundle.channels.v1: stable,development
          operators.operatorframework.io/builder: operator-sdk-v1.15.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: github.com/silicomdk/sts-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: StsConfig is the Schema for the stsconfigs API
            displayName: Sts Config
            kind: StsConfig
            name: stsconfigs.sts.silicom.com
            version: v1alpha1
          - description: StsNode is the Schema for the stsnodes API
            displayName: Sts Node
            kind: StsNode
            name: stsnodes.sts.silicom.com
            version: v1alpha1
          - description: StsOperatorConfig is the Schema for the stsoperatorconfigs
              API
            displayName: Sts Operator Config
            kind: StsOperatorConfig
            name: stsoperatorconfigs.sts.silicom.com
            version: v1alpha1
        description: |
          Silicom STS line card for 4G and 5G NIC enable real-time data transmission
          with high timing accuracy at the lowest cost to power 5G DRAN and CRAN edge
          deployments.

          * Support 1588/PTP over IPv4 / IPV6, IEEE1588v2
          * Support SyncE /ITU-T G.8262
          * T-BC/T-TSC Boundary Clock and TSC Slave Clock /G.8273.2
          * T-GM Grand Master /G.8273.1 per G.8275.1 PTP Profile
          * PRTC Primary Reference Time Clock Class B/G.8272
          * T-TC Transparent Clock /G.8273.3
          * 1588 Software Stack and Servo Software in x86
          * [Silicom STS2 Card](https://www.silicom-usa.com/pr/server-adapters/networking-adapters/10-gigabit-ethernet-networking-adapters/p410g8ts81-timesync-server-adapter)

          ### Prequisites:
          * [NFD Operator](https://operatorhub.io/operator/nfd-operator "NFD operator needed for the discovery of STS cards").
          * [SRO Operator](https://github.com/openshift/special-resource-operator)
        displayName: Silicom STS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sts
        - silicom
        - ptp
        links:
        - name: GitHub Repository
          url: https://github.com/silicomDK/sts-operator
        - name: Github SRO Helm Charts
          url: https://github.com/silicomDK/sts2-charts
        - name: Silicom Ltd.
          url: http://www.silicom-usa.com
        maintainers:
        - email: rmr@silicom.dk
          name: Ryan M. Raasch
        maturity: beta
        provider:
          name: Silicom Ltd.
          url: http://www.silicom-usa.com
        relatedImages:
        - quay.io/silicom/sts-operator:0.0.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/openshift-community-operators/silicom-sts-operator@sha256:df0939534c6aaa988714e64728ecdb387ff7dc4b32a5d97d56703ab143787171
        version: 0.0.2
      entries:
      - name: silicom-sts-operator.v0.0.2
        version: 0.0.2
      - name: silicom-sts-operator.v0.0.1
        version: 0.0.1
      name: beta
    defaultChannel: alpha
    packageName: silicom-sts-operator
    provider:
      name: Silicom Ltd.
      url: http://www.silicom-usa.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.7.1-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:ef65de2143df2c93aa73f099fa66d39e3239c30a9d547ed7bf258da0a19e8e35
        - turbonomic/prometurbo-operator:8.7.1-SNAPSHOT
        version: 8.7.1-beta.1
      entries:
      - name: prometurbo-operator.v8.7.1-beta.1
        version: 8.7.1-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        - turbonomic/prometurbo-operator:8.6
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: stable
    - currentCSV: prometurbo-operator.v8.5.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.5
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.5
        - quay.io/openshift-community-operators/prometurbo@sha256:83444ab464bce4fb5021a2fca6d4d6ff72b40e90a6cfc9c5ac6c74357c775ad5
        version: 8.5.0
      entries:
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.5.0
    - currentCSV: prometurbo-operator.v8.6.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: turbonomic/prometurbo-operator:8.6
          createdAt: "2019-05-01 00:00:00"
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
        maintainers:
        - email: endre.sara@turbonomic.com
          name: Endre Sara
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - turbonomic/prometurbo-operator:8.6
        - quay.io/openshift-community-operators/prometurbo@sha256:cb8b13fece10b8217b79e63584f8df4975721f2cec82aada2cf6c266d72dbf75
        version: 8.6.0
      entries:
      - name: prometurbo-operator.v8.6.0
        version: 8.6.0
      - name: prometurbo-operator.v8.5.0
        version: 8.5.0
      name: v8.6.0
    defaultChannel: stable
    packageName: prometurbo
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-automation-platform-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aap-operator.v2.4.0-0.1758079821
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635279521 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:032f053f21e273714e9f8a8f6854c83bc71cb84f5b32174704b490eca09eebc8
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        version: 2.4.0+0.1758079821
      entries:
      - name: aap-operator.v2.4.0-0.1749069319
        version: 2.4.0+0.1749069319
      - name: aap-operator.v2.4.0-0.1750877268
        version: 2.4.0+0.1750877268
      - name: aap-operator.v2.4.0-0.1753232084
        version: 2.4.0+0.1753232084
      - name: aap-operator.v2.4.0-0.1755833431
        version: 2.4.0+0.1755833431
      - name: aap-operator.v2.4.0-0.1758079821
        version: 2.4.0+0.1758079821
      name: stable-2.4
    - currentCSV: aap-operator.v2.4.0-0.1758080293
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.0.1-0.1635283332 <2.4.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the hub, including all hosted Ansible
              content, secrets and the database. By default, a persistent volume claim
              will be created using the default StorageClass on your cluster to store
              the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous hub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Automation Hub
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-24

          ## How to Upgrade

          Starting with AAP 2.4 in the May 2025 async release, the managed database was upgraded to PostgreSQL 15. It is recommended to take a backup of your deployment before upgrading.

          To take a backup, create AutomationControllerBackup and AutomationHubBackup objects for your deployments.
          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel.

          Note: Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-24/hub-web-rhel8@sha256:6ff1f602f0d9cad201b3addb65acc570a9d527de005ca85fc861d3cb98138710
        - registry.redhat.io/ansible-automation-platform-24/ee-supported-rhel8@sha256:f1384eb558f7da1e4a35d51408b3cf5a5a9be08f4d50503f473b26909bc246fc
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8-operator@sha256:e20106a8a04abd568a4c56861cea39ae444c0bd204bc807ac01f1705e12937d3
        - registry.redhat.io/ansible-automation-platform-24/lightspeed-rhel8@sha256:cccb4db08735551c697a7b2cbf75812f34500e691f9cdbaea02b30edea292e47
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8@sha256:04fb5cb6b0d7c79cf5d4b754fe917224ddfcb2b9a4994462db404c833ed69483
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-ui-rhel8@sha256:d229d4e09b2b788d1ba3413ea7ec15374bd1613d507c285ac1ddc10570b9473f
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8@sha256:7ae8db2b5d12b619f40db1e581db1e982a695bc1a29a2628222fff09e7c25249
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8-operator@sha256:5b957031609c8aabf7575e675bc636c1152babc9564d503802301aff4642a404
        - registry.redhat.io/ansible-automation-platform-24/eda-controller-rhel8@sha256:bfd13d6b39c5ec3a67adbafb05a0482328f0b9b97a517b45c093f6c35a3a92d3
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-24/hub-rhel8-operator@sha256:3c28a1326ecb7495952a5bb8be13b0972a4378af82611b0b9d990c5fa9fa82c3
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-rhel8-operator@sha256:b120e6a5ef376662b4324f0458eae31627087dc91fc613788143bc3954a15318
        - registry.redhat.io/ansible-automation-platform-24/ee-minimal-rhel8@sha256:7670e41037f023e3fc6ee95d2342787a80077273ab8d30ba6804029ff05b4edc
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:40b961a28ea9761c052a5cd17d2452fb72ea4786a92e13a0a1b589c7de31bf26
        - registry.redhat.io/ansible-automation-platform-24/controller-rhel8-operator@sha256:5253c9fd587ff2ca94d251f2b5fc7f284080adc5eb64643df5299beb2015592f
        - registry.redhat.io/ansible-automation-platform-24/platform-resource-runner-rhel8@sha256:e14286dc9605be3a8c6cb144d0b2e3f3bb3211129005009e7566ea6c16c91f52
        version: 2.4.0+0.1758080293
      entries:
      - name: aap-operator.v2.4.0-0.1749069693
        version: 2.4.0+0.1749069693
      - name: aap-operator.v2.4.0-0.1750878254
        version: 2.4.0+0.1750878254
      - name: aap-operator.v2.4.0-0.1753232791
        version: 2.4.0+0.1753232791
      - name: aap-operator.v2.4.0-0.1755833968
        version: 2.4.0+0.1755833968
      - name: aap-operator.v2.4.0-0.1758080293
        version: 2.4.0+0.1758080293
      name: stable-2.4-cluster-scoped
    - currentCSV: aap-operator.v2.5.0-0.1758147230
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727257740 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:e4d902722d83f82372021d49c80f9880dfdb8bb312d4635e1d97980b60ca1fb6
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        version: 2.5.0+0.1758147230
      entries:
      - name: aap-operator.v2.5.0-0.1727240614
        version: 2.5.0+0.1727240614
      - name: aap-operator.v2.5.0-0.1727875185
        version: 2.5.0+0.1727875185
      - name: aap-operator.v2.5.0-0.1728520175
        version: 2.5.0+0.1728520175
      - name: aap-operator.v2.5.0-0.1729741547
        version: 2.5.0+0.1729741547
      - name: aap-operator.v2.5.0-0.1731456196
        version: 2.5.0+0.1731456196
      - name: aap-operator.v2.5.0-0.1733193164
        version: 2.5.0+0.1733193164
      - name: aap-operator.v2.5.0-0.1733969526
        version: 2.5.0+0.1733969526
      - name: aap-operator.v2.5.0-0.1736893446
        version: 2.5.0+0.1736893446
      - name: aap-operator.v2.5.0-0.1737101025
        version: 2.5.0+0.1737101025
      - name: aap-operator.v2.5.0-0.1737675968
        version: 2.5.0+0.1737675968
      - name: aap-operator.v2.5.0-0.1739248042
        version: 2.5.0+0.1739248042
      - name: aap-operator.v2.5.0-0.1740093573
        version: 2.5.0+0.1740093573
      - name: aap-operator.v2.5.0-0.1740773472
        version: 2.5.0+0.1740773472
      - name: aap-operator.v2.5.0-0.1741368605
        version: 2.5.0+0.1741368605
      - name: aap-operator.v2.5.0-0.1742434024
        version: 2.5.0+0.1742434024
      - name: aap-operator.v2.5.0-0.1743660124
        version: 2.5.0+0.1743660124
      - name: aap-operator.v2.5.0-0.1746137767
        version: 2.5.0+0.1746137767
      - name: aap-operator.v2.5.0-0.1747343762
        version: 2.5.0+0.1747343762
      - name: aap-operator.v2.5.0-0.1749074128
        version: 2.5.0+0.1749074128
      - name: aap-operator.v2.5.0-0.1749604727
        version: 2.5.0+0.1749604727
      - name: aap-operator.v2.5.0-0.1750901111
        version: 2.5.0+0.1750901111
      - name: aap-operator.v2.5.0-0.1753900156
        version: 2.5.0+0.1753900156
      - name: aap-operator.v2.5.0-0.1755835086
        version: 2.5.0+0.1755835086
      - name: aap-operator.v2.5.0-0.1758147230
        version: 2.5.0+0.1758147230
      name: stable-2.5
    - currentCSV: aap-operator.v2.5.0-0.1758147817
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Automation Platform Resource Operator manages everything
            Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.4.0-0.1727259094 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Create a complete backup of an Ansible Automation Platform
              instance.
            displayName: Ansible Automation Platform Backup
            kind: AnsibleAutomationPlatformBackup
            name: ansibleautomationplatformbackups.aap.ansible.com
            version: v1alpha1
          - description: Restore a backup of an Ansible Automation Platform instance.
            displayName: Ansible Automation Platform Restore
            kind: AnsibleAutomationPlatformRestore
            name: ansibleautomationplatformrestores.aap.ansible.com
            version: v1alpha1
          - description: |-
              Deploy the complete platform:
              - Automation-Controller
              - Automation-Hub
              - EDA
              - Ansible-Lightspeed
            displayName: Ansible Automation Platform
            kind: AnsibleAutomationPlatform
            name: ansibleautomationplatforms.aap.ansible.com
            version: v1alpha1
          - description: Back up a deployment of the controller, including jobs, inventories,
              and credentials
            displayName: Automation Controller Backup
            kind: AutomationControllerBackup
            name: automationcontrollerbackups.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a instance of AutomationController Mesh ingress to
              allow inbound connection to the AutomationController Receptor Mesh.
            displayName: Automation Controller Mesh Ingress
            kind: AutomationControllerMeshIngress
            name: automationcontrollermeshingresses.automationcontroller.ansible.com
            version: v1alpha1
          - description: Restore a previous controller deployment from an AutomationControllerBackup.
              The deployment name you provide will be the name of the new AutomationController
              CR that will be created.
            displayName: Automation Controller Restore
            kind: AutomationControllerRestore
            name: automationcontrollerrestores.automationcontroller.ansible.com
            version: v1beta1
          - description: Deploy a new instance of AutomationController. A standardized
              way to define, operate and scale automation with Ansible.
            displayName: Automation Controller
            kind: AutomationController
            name: automationcontrollers.automationcontroller.ansible.com
            version: v1beta1
          - description: Back up a deployment of the eda, including jobs, inventories,
              and credentials
            displayName: EDA Backup
            kind: EDABackup
            name: edabackups.eda.ansible.com
            version: v1alpha1
          - description: Restore a previous eda deployment from an EDABackup. The
              deployment name you provide will be the name of the new EDA CR that
              will be created.
            displayName: EDA Restore
            kind: EDARestore
            name: edarestores.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new instance of Event-Driven Ansible Server. A standardized
              way to configure and run rulebooks to trigger automation.
            displayName: EDA
            kind: EDA
            name: edas.eda.ansible.com
            version: v1alpha1
          - description: Deploy a new AutomationHub instance
            displayName: Automation Hub
            kind: AutomationHub
            name: automationhubs.automationhub.ansible.com
            version: v1beta1
          - description: Back up a deployment of the AutomationHub, including all
              hosted Ansible content, secrets and the database. By default, a persistent
              volume claim will be created using the default StorageClass on your
              cluster to store the backup on.
            displayName: Automation Hub Backup
            kind: AutomationHubBackup
            name: automationhubbackups.automationhub.ansible.com
            version: v1beta1
          - description: Restore a previous AutomationHub deployment into the namespace
            displayName: Automation Hub Restore
            kind: AutomationHubRestore
            name: automationhubrestores.automationhub.ansible.com
            version: v1beta1
          - description: Deploy a new instance of Ansible Lightspeed.
            displayName: Ansible Lightspeed
            kind: AnsibleLightspeed
            name: ansiblelightspeeds.lightspeed.ansible.com
            version: v1alpha1
          - description: Define a Credential
            displayName: Automation Controller Credential
            kind: AnsibleCredential
            name: ansiblecredentials.tower.ansible.com
            version: v1alpha1
          - description: Define an Instance Group
            displayName: Automation Controller Instance Group
            kind: AnsibleInstanceGroup
            name: ansibleinstancegroups.tower.ansible.com
            version: v1alpha1
          - description: Define a new Inventory in controller
            displayName: Automation Controller Inventory
            kind: AnsibleInventory
            name: ansibleinventories.tower.ansible.com
            version: v1alpha1
          - description: Launch a new job via controller
            displayName: Automation Controller Job
            kind: AnsibleJob
            name: ansiblejobs.tower.ansible.com
            version: v1alpha1
          - description: Define a new project in controller
            displayName: Automation Controller Project
            kind: AnsibleProject
            name: ansibleprojects.tower.ansible.com
            version: v1alpha1
          - description: Define a new Schedule in controller
            displayName: Automation Controller Schedule
            kind: AnsibleSchedule
            name: ansibleschedules.tower.ansible.com
            version: v1alpha1
          - description: Launch a new workflow via controller
            displayName: Automation Controller Workflow
            kind: AnsibleWorkflow
            name: ansibleworkflows.tower.ansible.com
            version: v1alpha1
          - description: Define a new job template in controller
            displayName: Automation Controller Job Template
            kind: JobTemplate
            name: jobtemplates.tower.ansible.com
            version: v1alpha1
          - description: Define a new workflow template
            displayName: Automation Controller Workflow Template
            kind: WorkflowTemplate
            name: workflowtemplates.tower.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Automation Platform is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Automation Platform Operator provides for push-button deployments and seamless upgrades of the Platform into your OpenShift environment. Ansible Automation Platform deployments include:
            * Ansible lightspeed: Deploy and manage the lifecycle of an ansible lightspeed application.
            * Automation gateway - Unify the management of AAP from a central location.
            * Automation controller - Define, operate, scale, and delegate automation across the enterprise.
            * Event-Driven Ansible controller - Configure Ansible rulebooks to trigger automation when specific conditions are met, and manage rulebook lifecycles
            * Private automation hub - Synchronize, manage and publish automation content to streamline Ansible code within your organization
            * Platform resource operator - Define, launch, and monitor automation and AAP resources in a cloud-native manner


          ## New in AAP 2.5

          In AAP 2.5, we are introducing a unified experience for all of the services provided in the Ansible Automation Platform. You can now manage content from Hub, rulebooks from EDA and automation jobs from Controller in a single pane.

          Instead of creating individual AutomationController, AutomationHub, and EDA custom resources, you will now instead create a single AnsibleAutomationPlatform custom resource.

          Ansible Lightspeed and it's integration with Watson Code Assistant is now available.  To read more about configuring it, please read https://red.ht/ansible-lightspeed-2x-docs.

          The managed database has been upgraded to PostgreSQL 15.

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Automation Platform operator by following instructions presented when you click the Install button.
          You can find additional installation guidance in the install documentation: https://red.ht/AAP-25

          ## How to Upgrade

          ### Backup before upgrading

          It is recommended to take a backup of your deployment before upgrading.
          * If you are still on AAP 2.4, to take a backup, create AutomationControllerBackup, AutomationHubBackup and EDABackup objects for your deployments.
          * If you are on AAP 2.5, create an AnsibleAutomationPlatformBackup object for your deployment to take a backup.

          ### Upgrading to 2.5

          #### Change the Operator Channel

          When you are ready to upgrade to the next major or minor version, you can do so by changing the channel in the Subscription tab in the Openshift Console. For example, change the channel from `stable-2.4` to `stable-2.5`.

          *Note:* Namespace-scoped and Cluster-scoped deployments are not compatible.
          If you originally installed the AAP Operator from the stable-2.4 channel for example, you will want to upgrade to the next minor version by selecting the stable-2.5 channel.

          After changing to the new 2.5 channel, the AAP Operator will update your existing Controller, Hub and EDA deployments in place.

          #### Create the AnsibleAutomationPlatform custom resource

          Once that is complete, create an AnsibleAutomationPlatform custom resource to enable the unified UI. On the AnsibleAutomationPlatform custom resource, you can specify the name of an existing component (controller, hub, eda, lightspeed) in the same namespace and it will register that deployment with the new UI. If a component by that name does not exist, one will be created. Here is an example spec which uses an existing controller, deploys a fresh hub and eda, and does not deploy lightspeed:

          ```
          ---
          apiVersion: aap.ansible.com/v1alpha1
          kind: AnsibleAutomationPlatform
          metadata:
            name: myaap
            namespace: aap
          spec:
            # Platform
            image_pull_policy: IfNotPresent

            # Components
            controller:
              disabled: false
              name: existing-controller-name

            hub:
              disabled: false
              name: existing-hub-name

              ## uncomment if using file storage for Content pod
              storage_type: file
              file_storage_storage_class: your-rwx-storage-class
              file_storage_size: 10Gi

              ## uncomment if using S3 storage for Content pod
              # storage_type: S3
              # object_storage_s3_secret: example-galaxy-object-storage

              ## uncomment if using Azure storage for Content pod
              # storage_type: azure
              # object_storage_azure_secret: azure-secret-name

            eda:
              disabled: false

            lightspeed:
              disabled: true
          ```

          ### Accessing the AAP UI

          Once the upgrade is complete, you can access the AAP Platform UI by navigating to the route created for the AnsibleAutomationPlatform custom resource. The route will be named the same as your deployment name (`myaap` in the example above).

          This Platform URL is the single pane of glass for all of your Ansible Automation Platform services.  The Rest API for each service in the platform is also available at this URL at `/api/controller`, `/api/galaxy`, `/api/eda`.  For backwards compatibility, the API endpoints on the existing routes for each service are still available in upgraded environments for a time.
        displayName: Ansible Automation Platform
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-controller
        - automation-hub
        - automation-platform
        - ansible-automation-platform
        - aap
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-15@sha256:b70af4767a5b34c4a67761aa28bee72b4f9cd1ce31245596640371f670d0dbba
        - registry.redhat.io/ansible-automation-platform-25/ee-supported-rhel8@sha256:52d2c73fa2015af5f1b9a88ee48085c7c7bab44330fc9f95914f48b58e4b48bc
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8-operator@sha256:503fc7a20e2dd7de2cb55b9be5f97efa0d2b31f054ce326f23f3a11b4d683e64
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8@sha256:aed3973eb953dda3571260dcf1b33bc7dcbc18c41e3350b8c361b7e635fb9527
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8-operator@sha256:43e731b9a435b3a412108170b4b39e6b6215be51d362a9f03c33fa412e97da53
        - registry.redhat.io/ansible-automation-platform-25/gateway-rhel8@sha256:697343aaea153ab22d7efa8f1bb5aed37a55cbb353fc206f4a355e8b878ae418
        - registry.redhat.io/ansible-automation-platform/platform-operator-bundle@sha256:961aa224525116fd575cff3a37a6465e9c652ee446927dd2bd03249b29469472
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-chatbot-rhel8@sha256:9a37c9739c488ccf211ec5efc6debb045223d508c9cb3fa5a8256eca98ff2f0c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        - registry.redhat.io/ansible-automation-platform-25/ee-minimal-rhel8@sha256:41f944eb1450db56c3b14fc7baece8f148d5f10886a57bad730bec60fb28ac04
        - registry.redhat.io/ansible-automation-platform-25/hub-web-rhel8@sha256:e96832efcc03c6754031cb70f56c808135f62d1b80a882dd6d6fc9e1a4cfc69b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8@sha256:ac0de2b7d354eb0c15a6c12861a4de3577a982c84af661fbab25517d8105c085
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-rhel8-operator@sha256:74be9182723b41f87b3d13b3d86b6a9049bba88e2902cc9453264aa8a0f584e4
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8-operator@sha256:8d8182e3a87c55eacc222e71df4a9bf7a2e51102d913a5824cc91e36056d4249
        - registry.redhat.io/rhel8/redis-6@sha256:72ee6edb5cc470104fd29e0ee0876ee5e811b6aac03f0ee2e946a2aad24304e6
        - registry.redhat.io/ubi8/ubi-minimal@sha256:43dde01be4e94afd22d8d95ee8abcc9f610b4e50aff5bcc141b558c74d4c68b5
        - registry.redhat.io/ansible-automation-platform-25/gateway-proxy-rhel8@sha256:789d55f355fa0d90f9509a912d199123bfc5a4041930a72a8fd4ed1d819f0d84
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-ui-rhel8@sha256:9ab35e80a6172018ed3f5cd179e0bcb6befc1378583d9e73c3a64a7cf9e8050f
        - registry.redhat.io/ansible-automation-platform-25/platform-resource-runner-rhel8@sha256:ace1d2ab49ba5ff9f11ab18e8edce27c9e73ae7298bf7335569fa1740a44899b
        - registry.redhat.io/ansible-automation-platform-25/lightspeed-rhel8-operator@sha256:acc0fcd509eeac46db307ab744dd0a0c26770a93e495cf2ec0ccb61a91c6fbc5
        - registry.redhat.io/ansible-automation-platform-25/eda-controller-rhel8-operator@sha256:07aa7e257d04ae0eda6fcaa4af9121f1aece439231fce4f924e6467da91f0022
        - registry.redhat.io/ansible-automation-platform-25/hub-rhel8@sha256:9f2867603b9747c2fe37862c2c5efac4a4530902f323a8bf7cc0717de4cdb5aa
        - registry.redhat.io/ansible-automation-platform-25/controller-rhel8@sha256:3a989f148648a5a62e687058ec3ca772fa22aa10faec89453626d2b17b3fd9c2
        version: 2.5.0+0.1758147817
      entries:
      - name: aap-operator.v2.5.0-0.1727241223
        version: 2.5.0+0.1727241223
      - name: aap-operator.v2.5.0-0.1727875896
        version: 2.5.0+0.1727875896
      - name: aap-operator.v2.5.0-0.1728520687
        version: 2.5.0+0.1728520687
      - name: aap-operator.v2.5.0-0.1729742145
        version: 2.5.0+0.1729742145
      - name: aap-operator.v2.5.0-0.1731456854
        version: 2.5.0+0.1731456854
      - name: aap-operator.v2.5.0-0.1733193761
        version: 2.5.0+0.1733193761
      - name: aap-operator.v2.5.0-0.1733970210
        version: 2.5.0+0.1733970210
      - name: aap-operator.v2.5.0-0.1736894220
        version: 2.5.0+0.1736894220
      - name: aap-operator.v2.5.0-0.1737101781
        version: 2.5.0+0.1737101781
      - name: aap-operator.v2.5.0-0.1737676792
        version: 2.5.0+0.1737676792
      - name: aap-operator.v2.5.0-0.1739248649
        version: 2.5.0+0.1739248649
      - name: aap-operator.v2.5.0-0.1740094176
        version: 2.5.0+0.1740094176
      - name: aap-operator.v2.5.0-0.1740774104
        version: 2.5.0+0.1740774104
      - name: aap-operator.v2.5.0-0.1741369251
        version: 2.5.0+0.1741369251
      - name: aap-operator.v2.5.0-0.1742434756
        version: 2.5.0+0.1742434756
      - name: aap-operator.v2.5.0-0.1743660958
        version: 2.5.0+0.1743660958
      - name: aap-operator.v2.5.0-0.1746138413
        version: 2.5.0+0.1746138413
      - name: aap-operator.v2.5.0-0.1747345055
        version: 2.5.0+0.1747345055
      - name: aap-operator.v2.5.0-0.1749074612
        version: 2.5.0+0.1749074612
      - name: aap-operator.v2.5.0-0.1749607543
        version: 2.5.0+0.1749607543
      - name: aap-operator.v2.5.0-0.1750901870
        version: 2.5.0+0.1750901870
      - name: aap-operator.v2.5.0-0.1753900674
        version: 2.5.0+0.1753900674
      - name: aap-operator.v2.5.0-0.1755835623
        version: 2.5.0+0.1755835623
      - name: aap-operator.v2.5.0-0.1758147817
        version: 2.5.0+0.1758147817
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5
    packageName: ansible-automation-platform-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx8-openshift-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx8-openshift-operator.v0.0.14
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX8","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
          createdAt: "2023-06-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX 8 on JBoss EAP 7
            displayName: Joget DX 8 on JBoss EAP 7
            kind: JogetDX8
            name: jogetdx8.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX8/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX 8 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx8-operator-bundle@sha256:ad94f962d7aa064dc2b5b12ac2ea66ae5973909c82e6a96829165eaeff7ea952
        - registry.connect.redhat.com/joget/joget-dx8-operator@sha256:0fe3194771d9ab11e6a3b7f17beed3aabcf8a1bc1864083ca991a9458ab1b884
        - quay.io/joget/joget-dx8-eap7@sha256:154f85cc6721a70909dee755ddd88897ad51038afe2538b6dee4d26e7d8b9afb
        version: 0.0.14
      entries:
      - name: joget-dx8-openshift-operator.v0.0.14
        version: 0.0.14
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx8-openshift-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Intel
      provider-url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
    name: aikit-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aikit-operator.v2021.2.102120
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aikit.intel/v1alpha1",
                "kind": "AIKitContainer",
                "metadata": {
                  "name": "intel-aikit-container",
                  "namespace": "redhat-ods-applications"
                },
                "spec": {
                  "fullnameOverride": "",
                  "imagestream": {
                    "name": "oneapi-aikit",
                    "namespace": "redhat-ods-applications",
                    "registry": {
                      "root": "registry.connect.redhat.com",
                      "repo": "intel",
                      "name": "oneapi-aikit",
                      "version": "latest"
                    }
                  },
                  "nameOverride": ""
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          support: Intel
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AIKitContainer
            name: aikitcontainers.aikit.intel
            version: v1alpha1
        description: |
          The Intel® oneAPI AI Analytics Toolkit gives data scientists, AI developers, and researchers familiar Python* tools and frameworks to accelerate end-to-end data science and machine learning pipelines on Intel® architectures. The components are built using oneAPI libraries for low-level compute optimizations. This toolkit maximizes performance from preprocessing through machine and deep learning phases and provides interoperability for efficient model development and deployment across single and multinodes.
        displayName: Intel® oneAPI AI Analytics Toolkit Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - aikit-operator
        links:
        - name: Aikit Operator
          url: https://aikit-operator.domain
        maintainers:
        - email: kam.d.kasravi@intel.com
          name: Kam Kasravi
        - email: abolfazl.shahbazi@intel.com
          name: Abolfazl Shahbazi
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Intel
          url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
        relatedImages:
        - registry.connect.redhat.com/intel/aikit-operator-bundle@sha256:504672b2db7fd25d1c85c9614f8242f98b4d2d04184a947dbaaada41ec205b45
        - registry.connect.redhat.com/intel/aikit-operator@sha256:7abb5e6164ffb3e4b87c39061f45d09848218fcfb74b48bf5e1035d56b41ee0c
        version: 2021.2.102120
      entries:
      - name: aikit-operator.v2021.2.102120
        version: 2021.2.102120
      - name: aikit-operator.v2021.2.100410
        version: 2021.2.100410
      - name: aikit-operator.v2021.2.92911
        version: 2021.2.92911
      - name: aikit-operator.v2021.2.92814
        version: 2021.2.92814
      - name: aikit-operator.v2021.2.92808-dev-a60e5ba
        version: 2021.2.92808-dev-a60e5ba
      name: alpha
    defaultChannel: alpha
    packageName: aikit-operator
    provider:
      name: Intel
      url: https://software.intel.com/content/www/us/en/develop/tools/oneapi/application-catalog.html#gs.407z40
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-codeartifact-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-codeartifact-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
          createdAt: "2025-09-19T22:00:24Z"
          description: AWS codeartifact controller is a service controller for managing
            codeartifact resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Domain represents the state of an AWS codeartifact Domain
              resource.
            displayName: Domain
            kind: Domain
            name: domains.codeartifact.services.k8s.aws
            version: v1alpha1
          - description: PackageGroup represents the state of an AWS codeartifact
              PackageGroup resource.
            displayName: PackageGroup
            kind: PackageGroup
            name: packagegroups.codeartifact.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon codeartifact resources in AWS from within your Kubernetes cluster.

          **About Amazon codeartifact**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon codeartifact
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - codeartifact
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon codeartifact Developer Resources
          url: https://aws.amazon.com/codeartifact/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: codeartifact maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/codeartifact-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-codeartifact-controller@sha256:9bc95a0db7be356d2f657ae61c1c3098ff6e7e8bf1a675810dfb1b18805c6c8b
        version: 1.1.1
      entries:
      - name: ack-codeartifact-controller.v1.1.1
        version: 1.1.1
      - name: ack-codeartifact-controller.v1.1.0
        version: 1.1.0
      - name: ack-codeartifact-controller.v1.0.13
        version: 1.0.13
      - name: ack-codeartifact-controller.v1.0.11
        version: 1.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-codeartifact-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: opdev
      provider-url: opdev.github.io
    name: airflow-helm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-helm-operator.v0.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "workflow.apache.org/v1alpha1",
                "kind": "AirFlow",
                "metadata": {
                  "name": "airflow-helm"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Big Data
          containerImage: quay.io/opdev/airflow-helm-operator:0.0.1
          createdAt: "2023-06-28T16:57:21Z"
          description: An experimental operator that installs Apache Airflow.
          operatorframework.io/suggested-namespace: airflow-helm
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: hybrid.helm.sdk.operatorframework.io/v1-alpha
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Apache Airflow based on the official helm chart in
              openshift
            kind: AirFlow
            name: airflows.workflow.apache.org
            version: v1alpha1
        description: Deploys Apache Airflow based on the official helm chart with
          openshift specific modifications applied by default
        displayName: Airflow Helm Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        - data
        - batch
        - scheduler
        - data workflow
        links:
        - name: Airflow Helm Operator
          url: https://airflow-helm-operator.domain
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: opdev
          url: opdev.github.io
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/opdev/airflow-helm-operator:0.0.1
        - quay.io/openshift-community-operators/airflow-helm-operator@sha256:b3d05d2e2b4ff8ae64e4fc0338fcfa1ffdc345c6fdda8eaefed4e9108750fe58
        version: 0.0.1
      entries:
      - name: airflow-helm-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: airflow-helm-operator
    provider:
      name: opdev
      url: opdev.github.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: https://github.com/che-incubator/kubernetes-image-puller-operator
    name: kubernetes-imagepuller-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubernetes-imagepuller-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "che.eclipse.org/v1alpha1",
                "kind": "KubernetesImagePuller",
                "metadata": {
                  "name": "image-puller"
                },
                "spec": {
                  "daemonsetName": "k8s-image-puller",
                  "images": "java11-maven=quay.io/eclipse/che-java11-maven:next;che-theia=quay.io/eclipse/che-theia:next"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "false"
          containerImage: quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
          createdAt: "2024-01-15T16:23:04Z"
          description: Create and manage kubernetes-image-puller instances.
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/che-incubator/kubernetes-image-puller-operator
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubernetesImagePuller is the Schema for the kubernetesimagepullers
              API
            displayName: Kubernetes Image Puller
            kind: KubernetesImagePuller
            name: kubernetesimagepullers.che.eclipse.org
            version: v1alpha1
        description: "## About the managed application\n\nCreate a `KubernetesImagePuller`
          custom resource to automatically configure and run an instance of the [kubernetes-image-puller.](https://github.com/che-incubator/kubernetes-image-puller)\n##
          About this Operator\n\nThe `KubernetesImagePuller` custom resource understands
          the following fields in the `spec`:\n\n1. `configMapName` - The name of
          the `ConfigMap` to create\t\n2. `daemonsetName` - The name of the `DaemonSet`
          to be create.\n3. `deploymentName` - The name of the `kubernetes-image-puller`
          `Deployment` to create.\n4. `images` - A list of key-value pairs separated
          by semicolons of images to pull.  For example: `java=quay.io/eclipse/che-java8-maven:latest;theia=quay.io/eclipse/che-theia:next`\n5.
          `cachingIntervalHours` - The amount of time, in hours, between `DaemonSet`
          health checks.\n6. `cachingMemoryRequest` - The memory request for each
          cached image when the puller is running.\n7. `cachingMemoryLimit` - The
          memory limit for each cached image when the puller is running.\n8. `nodeSelector`
          - Node selector applied to pods created by the `DaemonSet`.\n## Prerequisites
          for enabling this Operator\n\nThe operator requires an existing namespace
          to be installed in.\nCheck the [README](https://github.com/che-incubator/kubernetes-image-puller-operator#installing-the-operator)
          for more details.\n"
        displayName: Kubernetes Image Puller Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - che
        - kubernetes-image-puller
        links:
        - name: Repository
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        maintainers:
        - email: ibuziuk@redhat.com
          name: Ilya Buziuk
        - email: abazko@redhat.com
          name: Anatolii Bazko
        maturity: stable
        provider:
          name: Red Hat, Inc.
          url: https://github.com/che-incubator/kubernetes-image-puller-operator
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/kubernetes-imagepuller-operator@sha256:a72e3728dcf395667820c10c4c5ccb3f3d040ae5d04f50680e193398d3b72b88
        - quay.io/eclipse/kubernetes-image-puller-operator:1.1.0
        version: 1.1.0
      entries:
      - name: kubernetes-imagepuller-operator.v1.1.0
        version: 1.1.0
      - name: kubernetes-imagepuller-operator.v1.0.6
        version: 1.0.6
      - name: kubernetes-imagepuller-operator.v1.0.5
        version: 1.0.5
      - name: kubernetes-imagepuller-operator.v1.0.4
        version: 1.0.4
      - name: kubernetes-imagepuller-operator.v1.0.3
        version: 1.0.3
      - name: kubernetes-imagepuller-operator.v1.0.2
        version: 1.0.2
      - name: kubernetes-imagepuller-operator.v1.0.1
        version: 1.0.1
      - name: kubernetes-imagepuller-operator.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: kubernetes-imagepuller-operator
    provider:
      name: Red Hat, Inc.
      url: https://github.com/che-incubator/kubernetes-image-puller-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/os.linux: supported
      provider: Trilio
      provider-url: ""
    name: k8s-triliovault
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: k8s-triliovault-stable.3.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          containerImage: registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          olm.skipRange: '>=3.0.0 <3.1.3'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          TrilioVault for Kubernetes is an enterprise-grade, cloud-native platform purpose built for data protection and management of Kubernetes applications for IT managers, administrators and developers. TrilioVault supports upstream Kubernetes and OpenShift environments and offers the following features:
            * Application-Centric - protects both data and metadata for Helm, Operator or custom Label based applications.
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any Storage (CSI, NFS, S3), or any Cloud (Private or Public).
            * Application Deployment & Tooling: Helm, Operators, Labels, Prometheus, Fluentd.
          Tutorials
          ------------
          Please click the link below to access the TrilioVault for Kubernetes "How-To" series for deployment, best practice and use-case videos
          <a href="https://www.trilio.io/triliovault-for-kubernetes">TrilioVault for Kubernetes - Tutorials and How-To </a>”
          Licensing
          --------
          Customers can download a 30-day Free Trial or a 10-node Basic Edition at no cost.  You can also connect with the Trilio team for an Enterprise Edition license with no limitations and Premium Support.
          For more information on license plans please vist:
          <a href="www.trilio.io/plans"> Trilio Vault for Kubernetes licensing and plans </a>
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our TrilioVault technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. www.trilio.io and @triliodata on Twitter.
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://www.trilio.io/
        - name: Product Licensing
          url: https://www.trilio.io/plans
        - name: Tutorials
          url: https://www.trilio.io/triliovault-for-kubernetes/
        - name: Support for Free/Basic Users
          url: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.18.0
        provider:
          name: Trilio
        relatedImages:
        - registry.connect.redhat.com/trilio/control-plane@sha256:01e36ed5b07104865f922d2996d2211b2b6000adc90266197d9bac74d908db97
        - registry.connect.redhat.com/trilio/web@sha256:76865d8afa8059d40810cb94df29420950a7593a40e87d142d771c57b3319e1d
        - registry.connect.redhat.com/trilio/event-stack@sha256:37c33ae1adddfe739c98fd2b2183c49ec7cfdafc88dec2970ad7c670016ebaac
        - registry.connect.redhat.com/trilio/dex@sha256:b6331cf3cb3a6a66d9b9ea639cda3c2559fbf1f92f7174032352c1c7bb5176d9
        - registry.connect.redhat.com/trilio/k8s-triliovault-operator@sha256:0e3094945427231ecbb884b903dc4d347dfaa73e5a04eb4c5cbaed6fce1cbbb8
        - registry.connect.redhat.com/trilio/minio@sha256:cee0f976edc79ff9e7a3f3574a7115dd960042241f3ed5b4636240ad609f9bca
        - registry.connect.redhat.com/trilio/nats@sha256:47396bfcff3e65efb3410fa8369f11f20d794ab543f9ada87b451448b1813178
        - registry.connect.redhat.com/trilio/trilio-webhook-init@sha256:c45e9b6b75862451d873d3eba29d6bd11cafc64823855f82ccb8c0b3d779ebb0
        - registry.connect.redhat.com/trilio/datamover@sha256:c7bc15f8217899ed73a812d278eb118633edbee1bf284132054b6f9ca2afd801
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:e80f87f289ea944a78b5c38df600c0a22fa470a2ce90f0b1e9453cd63586d75d
        version: 3.1.3
      entries:
      - name: k8s-triliovault-stable.3.1.3
        version: 3.1.3
      - name: k8s-triliovault-stable.3.1.2
        version: 3.1.2
      - name: k8s-triliovault-stable.3.1.1
        version: 3.1.1
      - name: k8s-triliovault-stable.3.1.0
        version: 3.1.0
      name: 3.1.x
    - currentCSV: k8s-triliovault-stable.4.0.5
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <4.0.5'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/k8s-triliovault-operator@sha256:69871613ceecefff7f691f21bb0b9147236af5a5dc2c7b36ad5e0cd8551f0e6a
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/control-plane@sha256:53368a72a60a38d3ce4293236a841346da6da2380ffd2525e5ffdb8dbcf11c7f
        - quay.io/triliodata/event-stack@sha256:c793418df5a1fc1a49f629d335f4684d9cde1a5f2b462f7c9483ffba904456d2
        - quay.io/triliodata/web@sha256:20de1a2f9477ad238bf24f48c8c16a38521c3a86164f2eb6b10d20f09839016e
        - quay.io/triliodata/datamover@sha256:e6848d94ba9f81843bde882a030b0f2556aae4bf16d79e2c5b199417cf626c85
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/operator-webhook-init@sha256:aa85b787f7f37a71e909f049d2abdda746cdd0627e2f74aaf7476d5e5a7f4401
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:bddfd4ebd973089f7446cb16ed9668d418f91c3907a5ede22052be4ccbea6d0c
        version: 4.0.5
      entries:
      - name: k8s-triliovault-stable.4.0.5
        version: 4.0.5
      - name: k8s-triliovault-stable.4.0.4
        version: 4.0.4
      - name: k8s-triliovault-stable.4.0.3
        version: 4.0.3
      - name: k8s-triliovault-stable.4.0.2
        version: 4.0.2
      - name: k8s-triliovault-stable.4.0.1
        version: 4.0.1
      - name: k8s-triliovault-stable.4.0.0
        version: 4.0.0
      name: 4.0.x
    - currentCSV: k8s-triliovault-stable.5.0.4
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "800Mi",
                      "cpu": "100m"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.0.4'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/dex@sha256:b4b2372fd03be586b9ff8cd269b9001797d813a34155ab37de31a44fe3cab1ea
        - quay.io/triliodata/k8s-triliovault-operator@sha256:8d582842dbef1c64bbe3b0ce5eeb64790910bdc0416c5e5561557642dede7022
        - quay.io/triliodata/control-plane@sha256:e60ef0a2c7a5b7879c9718e9a14c223f5290b4710b598fae19466a7558484824
        - quay.io/triliodata/datamover@sha256:104bdca9513c78f81c4340e9a3219df0a3bc8a5b026e4a252363a7e2c10977ba
        - quay.io/triliodata/nats@sha256:e6723595bcc5674b6201091d45d711d29a5cd1421c8b97c76cf00e468379fddf
        - quay.io/triliodata/operator-webhook-init@sha256:ffcc55e85f884b884947b73fa7d2fb257704ff70ccd7f2dc37d07e02292abed6
        - quay.io/triliodata/web@sha256:04c4baadae9c05f4add098ff62774d00102d56581e261626840d17f74effe438
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:b31ebfab28ec75c0d7110d6f01e3b26fd973c36013308e12f69584ff38ab3779
        - quay.io/triliodata/event-stack@sha256:859535e19e0ebebc8261c649c2446f8ca5fe4b4e5b6236e69b780dcc0d0bb5b9
        version: 5.0.4
      entries:
      - name: k8s-triliovault-stable.5.0.4
        version: 5.0.4
      - name: k8s-triliovault-stable.5.0.3
        version: 5.0.3
      - name: k8s-triliovault-stable.5.0.2
        version: 5.0.2
      - name: k8s-triliovault-stable.5.0.1
        version: 5.0.1
      - name: k8s-triliovault-stable.5.0.0
        version: 5.0.0
      name: 5.0.x
    - currentCSV: k8s-triliovault-stable.5.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "triliovault.trilio.io/v1",
                "kind": "TrilioVaultManager",
                "metadata": {
                  "name": "triliovault-manager"
                },
                "spec": {
                  "applicationScope": "Cluster",
                  "tvkInstanceName": "tvk-instance",
                  "logLevel": "Info",
                  "componentConfiguration": {
                    "ingress-controller": {
                      "enabled": false
                    }
                  },
                  "dataJobResources": {
                    "requests": {
                      "memory": "2Gi",
                      "cpu": "1"
                    }
                  },
                  "metadataJobResources": {
                    "requests": {
                      "memory": "10Mi",
                      "cpu": "10m"
                    }
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,Database,Monitoring,Security,Developer Tools
          certified: "false"
          console.openshift.io/plugins: '["tvk-console-plugin"]'
          containerImage: quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
          createdAt: "2020-04-27 17:12:00"
          description: Cloud-Native Data Protection for Kubernetes
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.0.0 <5.1.1'
          operatorframework.io/suggested-namespace: trilio-system
          repository: https://github.com/triliovault-k8s-issues/triliovault-k8s-issues
          support: Trilio
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents TrilioVault Manager which is an instance of TrilioVault
              for Kubernetes.
            displayName: TrilioVault Manager
            kind: TrilioVaultManager
            name: triliovaultmanagers.triliovault.trilio.io
            version: v1
        description: |-
          Trilio for Red Hat OpenShift provides fast and easy backup of application containers and virtual machines, disaster and ransomware recovery and application mobility across any cluster, anywhere.
          Trilio for Red Hat OpenShift:
            * Red Hat Certified - first backup and recovery solution with OpenShift Operator Certification.
            * Application-Centric - protects both data and metadata for Helm or Label based applications.
            * Namespace protection: Protect single or multi-namespace protection for complete cluster recovery
            * Fully multi-tenant tool, adhering to OpenShift RBAC policies, allowing you to operate hybrid, on-prem, air-gapped, public cloud (ROSA or ARO) or OpenShift on public cloud instances.
            * OpenShift Virtualization - protects data and metadata of OpenShift Virtualization Virtual Machines
            * Operators - protect and recover Red Hat OpenShift Lifecycle Management (OLM) Operators
            * Native to Kubernetes & OpenShift: Packaged and deployed as an Operator, provides new CRDs, integrated via Kubernetes API and with all features that it provides.
            * Infrastructure Agnostic: Compatible with any CSI Persistent Volume storage; backup storage based on NFS, S3, Blob/Container; and run on any Cloud (Private or Public).
            * Avoid vendor lock-in with industry proven Open Source backup format
            * Application Deployment & Tooling protection: Helm, Operators, Labels, Namespaces
            * Observability: Prometheus, Grafana
            * Fully Automated: Use our Ansible Certified Collection and our seamless integration with Red Hat Advanced Cluster Manager to fully automate your backup strategy and enable automated Disaster Recovery testing.
            * Unattended Recovery: Integrate Trilio with your Event Drive Ansible architecture for unattended recovery.
            * Fast Disaster Recovery: Utilize Trilio’s Continuous Restore to continually copy Persistent Volume data to any number of Kubernetes and OpenShift clusters, shortening recovery times and remote site failover
          More Information
          ----------------------
          Please click the link below to access more information about Trilio for Red Hat OpenShift:
          <a target="_blank" href="https://trilio.io/products/openshift-backup-and-recovery/">More Information</a>"
          About Trilio
          ----------------
          Trilio is trusted by global cloud infrastructure operators to deliver data protection, application resiliency, infrastructure migration and infrastructure version management. Our Trilio technology supports Kubernetes, OpenStack and Virtualization environments to recover from disasters, migrate tenant workloads, move workloads to new infrastructures and migrate to new infrastructure software distributions. Visit our website https://trilio.io
        displayName: Trilio for Kubernetes
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - backup
        - restore
        - data protection
        - disaster recovery
        - protect
        - recover
        - migrate
        - migration
        - DR
        links:
        - name: Trilio Homepage
          url: https://trilio.io/
        - name: Tutorials
          url: https://trilio.io/products/trilio-for-red-hat-openshift/
        - name: Documentation page
          url: https://docs.trilio.io/kubernetes/getting-started-3/red-hat-openshift
        maintainers:
        - email: info@trilio.io
          name: Trilio
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: Trilio
        relatedImages:
        - quay.io/triliodata/nats@sha256:72283871b6b9343aeabc3ff090305979422f192d30a6cbc4b0f20b2de7533996
        - quay.io/triliodata/control-plane@sha256:2e7e44cf6bc70dd9b978dce2cb2d9a36b8886c0f55286d5a6f5eb0a5182c4350
        - quay.io/triliodata/datamover@sha256:8a1350c8975610f23dd0b40d604d218f59dca79db6beb81ffe78a2682e40864f
        - quay.io/triliodata/filerecovery-vm-ssh-server@sha256:09186be228bd0b59e83119b28f3239c6173e410a3cf36957fe2d9b56d25200f8
        - quay.io/triliodata/web@sha256:3908a39eed44d702c6b6804d392d08a570b3f5562590d050f17ddc7a5c87b37f
        - registry.connect.redhat.com/trilio/r-3445381-bundle@sha256:7edf59729c2ae8b4506d5361eda01a98f33243987ac082abf758719c76a4804a
        - quay.io/triliodata/filerecovery@sha256:42559ac5d949d5ecf819cd4768a71fa688c72230502ff0b56539a8bb5e2f5838
        - quay.io/triliodata/dex@sha256:6b9d1667916968fb1f187ad5945fcc08db702dcab4c947e309dd9d2881a7b9cf
        - quay.io/triliodata/k8s-triliovault-operator@sha256:f0cf3f586561414671668210c59ac44071fde6abc954917e6ca04df5777a9a2f
        - quay.io/triliodata/event-stack@sha256:548ae1a748919f518e811774ec87f128aa8857be234837e09308c295de165726
        version: 5.1.1
      entries:
      - name: k8s-triliovault-stable.5.1.1
        version: 5.1.1
      - name: k8s-triliovault-stable.5.1.0
        version: 5.1.0
      name: 5.1.x
    defaultChannel: 5.1.x
    packageName: k8s-triliovault
    provider:
      name: Trilio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Thales Group
      provider-url: https://cpl.thalesgroup.com/data-protection
    name: cte-k8s-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: candidate
    - currentCSV: cte-k8s-operator.v1.6.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cte-k8s-operator.csi.cte.cpl.thalesgroup.com/v1",
                "kind": "CteK8sOperator",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "cte-k8s-operator"
                  },
                  "name": "ctek8soperator-sample"
                },
                "spec": null
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-06-05T08:24:49Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.39.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CteK8sOperator is the Schema for the ctek8soperators API
            displayName: Cte K8s Operator
            kind: CteK8sOperator
            name: ctek8soperators.cte-k8s-operator.csi.cte.cpl.thalesgroup.com
            version: v1
        description: |-
          # Kubernetes Operator to deploy CTE for Kubernetes on an OCP/K8s Cluster.
          This Operator deploys CTE for Kubernetes on the OCP/K8s Cluster.

          ## About this operator

          It is a GO based operator that manages the deployment, upgrade and
          deletion of the CTE for Kubernetes on all the nodes of the Cluster on
          which it is installed.

          ## Prerequisites

          Applicable only on OCP cluster
            For an OCP cluster the operator requires a SecurityContextConstraints (SCC)
            to be created before installation. To create the SecurityContextConstraints
            you can use a definition provided by the deployment scripts at
            https://github.com/thalescpl-io/ciphertrust-transparent-encryption-kubernetes

            Look for cte-csi-scc.yaml file in deploy/kubernetes/X.Y.Z/operator-deploy/
            Where X.Y.Z is the last released version of the deploy scripts.

          For more information refer to the guide at:
            https://thalesdocs.com/ctp/cte-con/cte-k8s/latest/cte-4-k8s/index.html
        displayName: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Encryption
        - Data Protection
        - Security
        - Kubernetes
        - Openshift
        links:
        - name: Thales CipherTrust Transparent Encryption for Kubernetes (CTE-K8s)
          url: https://cpl.thalesgroup.com/encryption/ciphertrust-transparent-data-encryption-kubernetes
        maintainers:
        - email: technical.support.DIS@thalesgroup.com
          name: Support
        maturity: alpha
        minKubeVersion: 1.26.0
        provider:
          name: Thales Group
          url: https://cpl.thalesgroup.com/data-protection
        relatedImages:
        - docker.io/thalesciphertrust/ciphertrust-transparent-encryption-kubernetes-operator@sha256:d58cbb65d758fd0b9d4cfba4cb48a35bdaebe2e239fb25edb9b491cd8c2c3f73
        - registry.connect.redhat.com/thalescpl/ciphertrust-transparent-encryption-kubernetes-operator-bundle@sha256:2e46e5c542fb29c6e787be3e6214a35494aa9c6d7c12ed07a88118ca6801df02
        version: 1.6.5
      entries:
      - name: cte-k8s-operator.v1.6.5
        version: 1.6.5
      name: stable
    defaultChannel: stable
    packageName: cte-k8s-operator
    provider:
      name: Thales Group
      url: https://cpl.thalesgroup.com/data-protection
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: fence-agents-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: fence-agents-remediation.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
          createdAt: "2023-10-30 14:06:57"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          olm.skipRange: '>=0.2.0 <0.2.1'
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.30.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:94bade4574c3679a425628f68067f9aead57bdd78e60f635041f421ed30341de
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:00cc420250d8e25d9da7c43d9c4cff35c7edeb7d390af754c48c0ce00e9e7770
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        version: 0.2.1
      entries:
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.14-eus
    - currentCSV: fence-agents-remediation.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
          createdAt: "2024-08-10 22:49:23"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.2.0 <0.4.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/24.3/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:f5dda99364d07df3a3680a2d7031c27df071c8e80746b1ececb5487240e9bbb0
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:a367e8fdd80938e97ac195631a996e0c537bf3666aef0199147534762dcf739f
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:77df668a9591bbaae675d0553f8dca5423c0f257317bc08fe821d965f44ed019
        version: 0.4.1
      entries:
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.16-eus
    - currentCSV: fence-agents-remediation.v0.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-07-16 10:25:57"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.4.0 <0.5.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.5/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:da5975557078fbfa6882df87cff5bda5671b9838b5fb0a738fea724469d03576
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:a4704a3977fccceb145e8972d2d7de074455bc228f2e925e5b427ffe3dcc623d
        version: 0.5.1
      entries:
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: 4.18-eus
    - currentCSV: fence-agents-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--password": "password",
                    "--username": "admin"
                  }
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {}
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
          createdAt: "2023-04-20 15:36:05"
          description: Fence Agents Remediation Operator for remediating nodes using
            upstream fence-agents.
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.26.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: "Fence Agents Remediation Operator (FAR)\n\nThe operator will
          remediate/fence a node when a FenceAgentsRemediation (far) custom resource
          exists with the node's name.\nFAR uses an existing set of upstream fencing
          agents for environments with a \ntraditional API end-point (e.g., IPMI)
          for power cycling cluster nodes.\n\nThe operator is recommended to be used
          with Node Healthcheck Operator for easier and smoother experience,\nbut
          it can also work as a standalonde remediator for the more advanced user.\n"
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/23.2/html/remediation_fencing_and_maintenance/index
        - name: Source Code
          url: https://github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:3af07bfe0fb73b1a7f00487eba302d6da442e742ab4007fc50e52efd4d45dae5
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:4f8f9dae947e9d34e86dc4237aec9167a9a25f83cecfae4982dc952494502486
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel8-operator@sha256:5c3778c7aae45f54e949fbae3d133777f7aa78fe1251ac3d5510aa28b71e2ba4
        version: 0.1.0
      entries:
      - name: fence-agents-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: fence-agents-remediation.v0.6.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediation",
                "metadata": {
                  "name": "worker-1"
                },
                "spec": {
                  "agent": "fence_ipmilan",
                  "nodeSecretNames": {
                    "master-0": "fence-agents-credentials-master0",
                    "master-1": "fence-agents-credentials-master1",
                    "master-2": "fence-agents-credentials-master2",
                    "worker-0": "fence-agents-credentials-worker0",
                    "worker-1": "fence-agents-credentials-worker1",
                    "worker-2": "fence-agents-credentials-worker2"
                  },
                  "nodeparameters": {
                    "--ipport": {
                      "master-0": "6230",
                      "master-1": "6231",
                      "master-2": "6232",
                      "worker-0": "6233",
                      "worker-1": "6234",
                      "worker-2": "6235"
                    }
                  },
                  "remediationStrategy": "ResourceDeletion",
                  "retrycount": 5,
                  "retryinterval": "5s",
                  "sharedSecretName": "fence-agents-credentials-shared",
                  "sharedparameters": {
                    "--action": "reboot",
                    "--ip": "192.168.111.1",
                    "--lanplus": "",
                    "--username": "admin"
                  },
                  "timeout": "60s"
                }
              },
              {
                "apiVersion": "fence-agents-remediation.medik8s.io/v1alpha1",
                "kind": "FenceAgentsRemediationTemplate",
                "metadata": {
                  "name": "fenceagentsremediationtemplate-default"
                },
                "spec": {
                  "template": {
                    "spec": {
                      "agent": "fence_ipmilan",
                      "nodeSecretNames": {
                        "master-0": "fence-agents-credentials-master0",
                        "master-1": "fence-agents-credentials-master1",
                        "master-2": "fence-agents-credentials-master2",
                        "worker-0": "fence-agents-credentials-worker0",
                        "worker-1": "fence-agents-credentials-worker1",
                        "worker-2": "fence-agents-credentials-worker2"
                      },
                      "nodeparameters": {
                        "--ipport": {
                          "master-0": "6230",
                          "master-1": "6231",
                          "master-2": "6232",
                          "worker-0": "6233",
                          "worker-1": "6234",
                          "worker-2": "6235"
                        }
                      },
                      "remediationStrategy": "ResourceDeletion",
                      "retrycount": 5,
                      "retryinterval": "5s",
                      "sharedSecretName": "fence-agents-credentials-shared",
                      "sharedparameters": {
                        "--action": "reboot",
                        "--ip": "192.168.111.1",
                        "--lanplus": "",
                        "--username": "admin"
                      },
                      "timeout": "60s"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: ""
          createdAt: "2025-09-08 09:19:00"
          description: Fence Agents Remediation Operator uses well-known agents to
            fence and remediate unhealthy nodes. The remediation includes rebooting
            the unhealthy node using a fence agent, and then evicting workloads from
            the unhealthy node.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.5.0 <0.6.0'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/fence-agents-remediation
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: FenceAgentsRemediation is the Schema for the fenceagentsremediations
              API
            displayName: Fence Agents Remediation
            kind: FenceAgentsRemediation
            name: fenceagentsremediations.fence-agents-remediation.medik8s.io
            version: v1alpha1
          - description: FenceAgentsRemediationTemplate is the Schema for the fenceagentsremediationtemplates
              API
            displayName: Fence Agents Remediation Template
            kind: FenceAgentsRemediationTemplate
            name: fenceagentsremediationtemplates.fence-agents-remediation.medik8s.io
            version: v1alpha1
        description: |
          ### Introduction
          Fence Agents Remediation (FAR) is a Kubernetes operator that uses well-known agents to fence and remediate unhealthy nodes.
          The remediation includes rebooting the unhealthy node using a fence agent and then evicting workloads from the unhealthy node.

          ### Compatibility
          FAR is one of the remediator operators by [Medik8s](https://www.medik8s.io/remediation/remediation/),
          such as [Self Node Remediation](https://github.com/medik8s/self-node-remediation) and [Machine Deletion Remediation](https://github.com/medik8s/machine-deletion-remediation),
          that were designed to run with the Node HealthCheck Operator [(NHC)](https://github.com/medik8s/node-healthcheck-operator) which detects an unhealthy node and creates remediation CR.
          It is recommended to use FAR with NHC for an easier and smoother experience by fully automating the remediation process, but it can be used as a standalone remediator for the more experienced user.

          ### Advantages
          - Robustness - FAR has direct feedback from the traditional Application Programming Interface (API) call (e.g., IPMI) about the result of the fence action without using the Kubernetes API.
          - Speed - FAR is rapid since it can reboot a node and receive an acknowledgment from the API call while other remediators might need to wait a safe time till they can expect the node to be rebooted.
          - Diversity - FAR includes several fence agents from a large known set of upstream fencing agents for bare metal servers, virtual machines, cloud platforms, etc.
          - Adjustability - FAR allows to set up different parameters for running the API call that remediates the node.
        displayName: Fence Agents Remediation Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - fencing
        - auto-healing
        - recovery
        - high-availability
        - fence-agents
        - fence-agents-remediation
        - remediation
        - far
        - baremetal
        links:
        - name: Fence Agents Remediation
          url: https://access.redhat.com/documentation/en-us/workload_availability_for_red_hat_openshift/25.8/html/remediation_fencing_and_maintenance/about-remediation-fencing-maintenance
        - name: Source Code
          url: https://www.github.com/medik8s/fence-agents-remediation
        maintainers:
        - email: team-dragonfly@redhat.com
          name: Dragonfly Team
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/workload-availability/fence-agents-remediation-operator-bundle@sha256:6c0899610a78d2737c2096552e5146d9d5602b5568fab24b51b409dbffcf1961
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d37a6d10b0fa07370066a31fdaffe2ea553faf4e4e98be7fcef5ec40d62ffe29
        - registry.redhat.io/workload-availability/fence-agents-remediation-rhel9-operator@sha256:eef1a298718650a22a3da4a07c140ce21b6f287d87af79801610317fa8f63d52
        version: 0.6.0
      entries:
      - name: fence-agents-remediation.v0.6.0
        version: 0.6.0
      - name: fence-agents-remediation.v0.5.1
        version: 0.5.1
      - name: fence-agents-remediation.v0.5.0
        version: 0.5.0
      - name: fence-agents-remediation.v0.4.1
        version: 0.4.1
      - name: fence-agents-remediation.v0.4.0
        version: 0.4.0
      - name: fence-agents-remediation.v0.3.0
        version: 0.3.0
      - name: fence-agents-remediation.v0.2.1
        version: 0.2.1
      name: stable
    defaultChannel: stable
    packageName: fence-agents-remediation
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: SmartX
      provider-url: https://www.smartx.com
    name: iomesh-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: iomesh-operator.0.10.1-rc4
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
          createdAt: 2021-08-17 00:00:00+00:00
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/iomesh-operator-rhmp/support?utm_source=openshift_console
          repository: https://github.com/iomesh/docs
          support: SmartX, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: iomesh
            displayName: iomesh
            kind: IOMeshCluster
            name: iomeshclusters.iomesh.com
            version: v1alpha1
        description: "IOMesh (https://www.iomesh.com/) is a cloud-native storage product
          specifically designed and developed for Kubernetes. It takes ZBS as its
          core, a “production-ready” distributed storage system developed independently
          by SmartX, to provide production-level high performance and reliable persistent
          storage for business-critical applications such as MySQL, Cassandra, MongoDB,
          etc., contributing to the containerized transformation of stateful applications.
          \     \n## Pre-Requisites\n* According to https://docs.iomesh.com/deploy/prerequisites,
          make sure that your openshift cluster meets the \"Installation Requirements\"
          and set up the worker nodes correctly according to the \"Setup Worker Node\"
          chapter.\n* Run IOMesh Operator pre-installation script in an environment
          where oc or kubectl can be used to access the openshift cluster, the script
          will install the dependencies of IOMesh Operator and specific IOMesh settings
          for the openshift cluster:\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-pre-install-openshift.sh
          | sh -\n```\n## Install IOMesh Operator and IOMesh Cluster\n* Click install
          button on the current web page to install IOMesh Operator\n* On the **Installed
          Operators** -> **IOMesh Operator** -> **Create instance** -> **YAML view**,
          fill in an IOMesh Custom Resources according https://iomesh.run/iomesh.yaml,
          change the `spec.*.dataCIDR` to your own data network CIDR:\n## Post Install\n*
          Run IOMesh Operator post-installation script in an environment where oc
          or kubectl can be used to access the openshift cluster, the script will
          install IOMesh CSI Driver\n```shell\ncurl -sSL https://iomesh.run/iomesh-operator-post-install-openshift.sh
          | sh -\n```\n## Getting Started\n* Try to quickstart setup and use IOMesh:
          https://docs.iomesh.com/deploy/setup-iomesh\n"
        displayName: IOMesh Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iomesh-operator
        links:
        - name: IOMesh Operator
          url: https://iomesh.com
        maintainers:
        - email: na.wang@smartx.com
          name: Na Wang
        maturity: alpha
        provider:
          name: SmartX
          url: https://www.smartx.com
        relatedImages:
        - registry.connect.redhat.com/iomesh/iomesh@sha256:84c6c826473b844293bd2048a498c5e0e8a558224a09f8e2e0cc85a2f10fcacf
        - iomesh/operator@sha256:52b8718ce46fa764f802f7709f2ca39ebc78cf25e4af78a4c8011e4c0ff1d8f3
        version: 0.10.1-rc4
      entries:
      - name: iomesh-operator.0.10.1-rc4
        version: 0.10.1-rc4
      name: stable
    defaultChannel: stable
    packageName: iomesh-operator-rhmp
    provider:
      name: SmartX
      url: https://www.smartx.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Marketplace
      provider-url: ""
    name: ibm-metrics-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: beta
    - currentCSV: ibm-metrics-operator.v2.20.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MarketplaceConfig",
                "metadata": {
                  "name": "marketplaceconfig"
                },
                "spec": {
                  "features": {
                    "deployment": true,
                    "meterDefinitionCatalogServer": false,
                    "registration": true
                  },
                  "installIBMCatalogSource": true,
                  "isDisconnected": false,
                  "license": {
                    "accept": false
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterBase",
                "metadata": {
                  "name": "meterbase"
                },
                "spec": {
                  "enabled": true,
                  "prometheus": {
                    "resources": {
                      "limits": {
                        "cpu": "2",
                        "memory": "2Gi"
                      },
                      "requests": {
                        "cpu": "1",
                        "memory": "1G"
                      }
                    },
                    "storage": {
                      "size": "20Gi"
                    }
                  }
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1alpha1",
                "kind": "MeterReport",
                "metadata": {
                  "name": "example-meterreport"
                },
                "spec": {
                  "endTime": "2020-08-05T00:00:00Z",
                  "prometheusService": {
                    "name": "rhm-prometheus-meterbase",
                    "namespace": "openshift-redhat-marketplace",
                    "targetPort": "rbac"
                  },
                  "startTime": "2020-08-04T00:00:00Z"
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "example-meterdefinition-4"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              },
              {
                "apiVersion": "marketplace.redhat.com/v1beta1",
                "kind": "MeterDefinition",
                "metadata": {
                  "name": "meterdefinition-sample"
                },
                "spec": {
                  "group": "partner.metering.com",
                  "kind": "App",
                  "meters": [
                    {
                      "aggregation": "sum",
                      "metricId": "container_cpu_usage_core_avg",
                      "period": "1h",
                      "query": "rate(container_cpu_usage_seconds_total{cpu=\"total\",container=\"db\"}[5m])*100",
                      "workloadType": "Pod"
                    }
                  ],
                  "resourceFilters": [
                    {
                      "namespace": {
                        "useOperatorGroup": true
                      },
                      "ownerCRD": {
                        "apiVersion": "marketplace.redhat.com/v1alpha1",
                        "kind": "RazeeDeployment"
                      },
                      "workloadType": "Pod"
                    }
                  ]
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          certified: "true"
          containerImage: quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
          createdAt: "2025-08-28T14:18:55Z"
          description: The IBM Metrics Operator provides workload metering and reporting
            for IBM and Red Hat Marketplace customers.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <2.20.0'
          operatorframework.io/suggested-namespace: ibm-software-central
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-marketplace/redhat-marketplace-operator
          support: IBM Corporation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MarketplaceConfig is configuration manager for our Red Hat
              Marketplace controllers By installing this product you accept the license
              terms https://ibm.biz/BdfaAY.
            displayName: Marketplace Config
            kind: MarketplaceConfig
            name: marketplaceconfigs.marketplace.redhat.com
            version: v1alpha1
          - description: MeterBase is the resource that sets up Metering for Red Hat
              Marketplace. This is an internal resource not meant to be modified directly.
            displayName: Meter Base
            kind: MeterBase
            name: meterbases.marketplace.redhat.com
            version: v1alpha1
          - description: MeterDefinition defines the meter workloads used to enable
              pay for use billing.
            displayName: Meter Definition
            kind: MeterDefinition
            name: meterdefinitions.marketplace.redhat.com
            version: v1beta1
          - kind: MeterReport
            name: meterreports.marketplace.redhat.com
            version: v1alpha1
        description: "The IBM Metrics Operator provides workload metering and reporting
          for IBM Software Central and Red Hat Marketplace customers.\n### **Important
          Note**\nA set of instructions for onboarding is provided here. For more
          detailed onboarding instructions or information about what is installed
          please visit [swc.saas.ibm.com](https://swc.saas.ibm.com).\n\nUsage metrics
          may be monitored through [https://swc.saas.ibm.com](https://swc.saas.ibm.com)
          with only IBM Metrics Operator and a Red Hat Marketplace account.\n\n###
          Upgrade Policy\n\nThe operator releases adhere to semantic versioning and
          provides a seamless upgrade path for minor and patch releases within the
          current stable channel.\n\n### Prerequisites\n1. Installations are required
          to [enable monitoring for user-defined projects](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)
          as the Prometheus provider.\n2. Edit the cluster-monitoring-config ConfigMap
          object:\n\n   ```sh\n   $ oc -n openshift-monitoring edit configmap cluster-monitoring-config\n
          \   ```\n\n3. Add enableUserWorkload: true under data/config.yaml:\n  \n
          \   ```sh\n    apiVersion: v1\n    kind: ConfigMap\n    metadata:\n      name:
          cluster-monitoring-config\n        namespace: openshift-monitoring\n    data:\n
          \     config.yaml: |\n        enableUserWorkload: true\n    ```\n\n4. Configure
          the user-workload-monitoring-config ConfigMap object:\n\n    ```sh\n    $
          oc -n openshift-user-workload-monitoring edit configmap user-workload-monitoring-config\n
          \   ```\n\n5. Configure a minimum retention time of 168h and minimum storage
          capacity of 40Gi\n  \n    ```sh\n    apiVersion: v1\n    kind: ConfigMap\n
          \   metadata:\n      name: user-workload-monitoring-config\n      namespace:
          openshift-user-workload-monitoring\n\n    data:\n      config.yaml: |\n
          \       prometheus:\n          retention: 168h\n          volumeClaimTemplate:\n
          \           spec:\n              resources:\n                requests:\n
          \                 storage: 40Gi\n    ```\n\n### Resources Required\n\nMinimum
          system resources required:\n\n| Operator                | Memory (MB) |
          CPU (cores) | Disk (GB) | Nodes |\n| ----------------------- | -----------
          | ----------- | --------- | ----- |\n| **Metrics**   |        750  |     0.25
          \   | 3x1       |    3  |\n| **Deployment** |        250  |     0.25    |
          -         |    1  |\n\n| Prometheus Provider  | Memory (GB) | CPU (cores)
          | Disk (GB) | Nodes |\n| --------- | ----------- | ----------- | ---------
          | ----- |\n| **[Openshift User Workload Monitoring](https://docs.openshift.com/container-platform/latest/monitoring/enabling-monitoring-for-user-defined-projects.html)**
          |          1  |     0.1       | 2x40        |   2    |\n\nMultiple nodes
          are required to provide pod scheduling for high availability for IBM Metrics
          Operator Data Service and Prometheus.\n\nThe IBM Metrics Operator automatically
          creates 3 x 1Gi PersistentVolumeClaims to store reports as part of the data
          service, with _ReadWriteOnce_ access mode. Te PersistentVolumeClaims are
          automatically created by the ibm-metrics-operator after creating a `redhat-marketplace-pull-secret`
          and accepting the license in `marketplaceconfig`.\n\n| NAME                                |
          CAPACITY | ACCESS MODES |\n| ----------------------------------- | --------
          | ------------ |\n| rhm-data-service-rhm-data-service-0 | 1Gi | RWO |\n|
          rhm-data-service-rhm-data-service-1 | 1Gi | RWO |\n| rhm-data-service-rhm-data-service-2
          | 1Gi | RWO |\n\n### Supported Storage Providers\n\n- OpenShift Container
          Storage / OpenShift Data Foundation version 4.x, from version 4.2 or higher\n-
          IBM Cloud Block storage and IBM Cloud File storage\n- IBM Storage Suite
          for IBM Cloud Paks:\n  - File storage from IBM Spectrum Fusion/Scale \n
          \ - Block storage from IBM Spectrum Virtualize, FlashSystem or DS8K\n- Portworx
          Storage, version 2.5.5 or above\n- Amazon Elastic File Storage\n\n### Access
          Modes required\n\n - ReadWriteOnce (RWO)\n\n### Provisioning Options supported\n\nChoose
          one of the following options to provision storage for the ibm-metrics-operator
          data-service\n\n#### Dynamic provisioning using a default StorageClass\n
          \  - A StorageClass is defined with a `metadata.annotations: storageclass.kubernetes.io/is-default-class:
          \"true\"`\n   - PersistentVolumes will be provisioned automatically for
          the generated PersistentVolumeClaims\n--- \n#### Manually create each PersistentVolumeClaim
          with a specific StorageClass\n   - Must be performed before creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`. Otherwise, the automatically
          generated PersistentVolumeClaims are immutable.\n```\napiVersion: v1\nkind:
          PersistentVolumeClaim\nmetadata:\n  labels:\n    app: rhm-data-service\n
          \ name: rhm-data-service-rhm-data-service-0\n  namespace: ibm-software-central\nspec:\n
          \ storageClassName: rook-cephfs\n  accessModes:\n  - ReadWriteOnce\n  resources:\n
          \   requests:\n      storage: 1Gi\n```\n---\n#### Manually provision each
          PersistentVolume for the generated PersistentVolumeClaims with a specific
          StorageClass\n  - May be performed before or after creating a `redhat-marketplace-pull-secret`
          or accepting the license in `marketplaceconfig`.  \n```\napiVersion: v1\nkind:
          PersistentVolume\nmetadata:\n  name: rhm-data-service-rhm-data-service-0\nspec:\n
          \ csi:\n    driver: rook-ceph.cephfs.csi.ceph.com\n    volumeHandle: rhm-data-service-rhm-data-service-0\n
          \ capacity:\n    storage: 1Gi\n  accessModes:\n    - ReadWriteOnce\n  persistentVolumeReclaimPolicy:
          Delete\n  storageClassName: rook-cephfs\n  volumeMode: Filesystem\n  claimRef:\n
          \   kind: PersistentVolumeClaim\n    namespace: ibm-software-central\n    name:
          rhm-data-service-rhm-data-service-0\n```\n\n### Installation\n1. Create
          or get your [pull secret](https://swc.saas.ibm.com/en-us/documentation/clusters#get-pull-secret).\n2.
          Install the IBM Metrics Operator\n3. Create a Kubernetes secret in the installed
          namespace with the name `redhat-marketplace-pull-secret` and key `PULL_SECRET`
          with the value of the pull secret.\n    ```sh\n    # Replace ${PULL_SECRET}
          with your pull secret\n    oc create secret generic redhat-marketplace-pull-secret
          -n ibm-software-central --from-literal=PULL_SECRET=${PULL_SECRET}\n    ```\n4.
          Use of the platform is governed by the:\n\n    [IBM Cloud Services Agreement](https://www.ibm.com/support/customer/csol/terms/?id=Z126-6304_WS&_ga=2.116312197.2046730452.1684328846-812467790.1684328846)
          (or other base agreement between you and IBM such as a [Passport Advantage
          Agreement](https://www.ibm.com/software/passportadvantage/pa_agreements.html?_ga=2.116312197.2046730452.1684328846-812467790.1684328846))
          and the [Service Description for the Red Hat Marketplace](https://www.ibm.com/support/customer/csol/terms/?id=i126-8719&_ga=2.83289621.2046730452.1684328846-812467790.1684328846).\n
          \   \n5. Update MarketplaceConfig to accept the license.\n    ```\n    oc
          patch marketplaceconfig marketplaceconfig -n ibm-software-central --type='merge'
          -p '{\"spec\": {\"license\": {\"accept\": true}}}'\n    ```\n6. Install
          the pull secret as a global pull secret on the cluster.\n\n    These steps
          require `oc`, `jq`, and `base64` to be available on your machine.\n\n    ```sh\n
          \   # Create the docker pull secret file using your pull secret.\n    #
          Store it in a file called entitledregistryconfigjson.\n    oc create secret
          docker-registry entitled-registry --docker-server=registry.marketplace.redhat.com
          --docker-username \"cp\" --docker-password \"${PULL_SECRET}\" --dry-run=client
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > entitledregistryconfigjson\n
          \   # Get the current global secrets on the cluster and store it as a file
          named dockerconfigjson\n    oc get secret pull-secret -n openshift-config
          --output=\"jsonpath={.data.\\.dockerconfigjson}\" | base64 --decode > dockerconfigjson\n
          \   # Merge the two dockerconfigs together into a file called dockerconfigjson-merged.\n
          \   jq -s '.[0] * .[1]' dockerconfigjson entitledregistryconfigjson > dockerconfigjson-merged\n
          \   # Set the cluster's dockerconfig file to the new merged version.\n    oc
          set data secret/pull-secret -n openshift-config --from-file=.dockerconfigjson=dockerconfigjson-merged\n
          \   ```\n  7. (Optional) Your IBM product may have installed IBM Metrics
          Operator using an `ibm-entitlement-key` for authentication to IBM Software
          Central. To authorize the automatic creation of an IBM Software Central
          account for the IBM account owning the `ibm-entitlement-key`\n      ```\n
          \     oc patch marketplaceconfig marketplaceconfig -n ibm-software-central
          --type='merge' -p '{\"spec\": {\"authorizeAccountCreation\": true}}'\n      ```\n\n###
          Why is a global pull secret required?\nIn order to successfully install
          the products hosted by the container image registry, you will need to make
          the pull secret available across the cluster. This can be achieved by applying
          the pull as a [global pull secret](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html#images-update-global-pull-secret_using-image-pull-secrets).
          For alternative approaches, please see the official OpenShift [documentation](https://docs.openshift.com/container-platform/latest/openshift_images/managing_images/using-image-pull-secrets.html).\n\n\n###
          SecurityContextConstraints requirements\n\nThe Operators and their components
          support running under the OpenShift Container Platform default restricted
          and restricted-v2 security context constraints.\n\n### Installation Namespace
          and ClusterRoleBinding requirements\n\nThe IBM Metrics Operator components
          require specific ClusterRoleBindings.\n- The metric-state component requires
          a ClusterRoleBinding for the the `view` ClusterRole.\n  - Ability to read
          non-sensitive CustomResources. \n  - The `view` ClusterRole dynamically
          adds new CustomResourceDefinitions.\n- The operator & reporter component
          requires a ClusterRoleBinding for the the `cluster-monitoring-view` ClusterRole.\n
          \ - Ability to view Prometheus metrics. \n  - The underlying ClusterRole
          RBAC often updated between OpenShift versions.\n\nDue to limitations of
          Operator Lifecycle Manager (OLM), this ClusterRoleBinding can not be provided
          dynamically for arbitrary installation target namespaces.\n\nA static ClusterRoleBinding
          is included for installation to the default namespace of `ibm-software-central`,
          and namespaces `openshift-redhat-marketplace`,  `redhat-marketplace`, `ibm-common-services`.\n\nTo
          create the ClusterRoleBindings for installation to an alternate namespace\n```\noc
          project INSTALL-NAMESPACE\noc adm policy add-cluster-role-to-user view -z
          ibm-metrics-operator-metric-state\noc adm policy add-cluster-role-to-user
          cluster-monitoring-view -z ibm-metrics-operator-controller-manager,ibm-metrics-operator-reporter\n```\n\n###
          Metric State scoping requirements\nThe metric-state Deployment obtains `get/list/watch`
          access to metered resources via the `view` ClusterRole. For operators deployed
          using Operator Lifecycle Manager (OLM), permissions are added to `clusterrole/view`
          dynamically via a generated and annotated `-view` ClusterRole. If you wish
          to meter an operator, and its Custom Resource Definitions (CRDs) are not
          deployed through OLM, one of two options are required\n1. Add the following
          label to a clusterrole that has get/list/watch access to your CRD: `rbac.authorization.k8s.io/aggregate-to-view:
          \"true\"`, thereby dynamically adding it to `clusterrole/view`\n2. Create
          a ClusterRole that has get/list/watch access to your CRD, and create a ClusterRoleBinding
          for the metric-state ServiceAccount\n\nAttempting to meter a resource with
          a MeterDefinition without the required permissions will log an `AccessDeniedError`
          in metric-state.\n\n### Disaster Recovery\n\nTo plan for disaster recovery,
          note the PhysicalVolumeClaims `rhm-data-service-rhm-data-service-N`. \n-
          In connected environments, MeterReport data upload attempts occur hourly,
          and are then removed from data-service. There is a low risk of losing much
          unreported data.\n- In an airgap environment, MeterReport data must be pulled
          from data-service and uploaded manually using [datactl](https://github.com/redhat-marketplace/datactl/).
          To prevent data loss in a disaster scenario, the data-service volumes should
          be considered in a recovery plan.\n\n### Subscription Config\n\nIt is possible
          to [configure](https://github.com/operator-framework/operator-lifecycle-manager/blob/master/doc/design/subscription-config.md)
          how OLM deploys an Operator via the `config` field in the [Subscription](https://github.com/operator-framework/olm-book/blob/master/docs/subscriptions.md)
          object.\n\nThe IBM Metrics Operator will also read the `config` and append
          it to the operands. The primary use case is to control scheduling using
          [Tolerations](https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/)
          and [NodeSelectors](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/).\n\nA
          limitation is that the `config` elements are only appended to the operands.
          The elements in the operands are not removed if the `config` is removed
          from the `Subscripion`. The operand must be modified manually, or deleted
          and recreated by the controller.\n\n### Cluster permission requirements\n\n|API
          group             |Resources              |Verbs                                     |\n|----------------------|-----------------------|------------------------------------------|\n|apps
          \                 |statefulsets           |get;list;watch                            |\n|authentication.k8s.io
          |tokenreviews           |create                                    |\n|authorization.k8s.io
          \ |subjectaccessreviews   |create                                    |\n|config.openshift.io
          \  |clusterversions        |get;list;watch                            |\n|marketplace.redhat.com|meterdefinitions
          \      |get;list;watch;create;update;patch;delete |\n|operators.coreos.com
          \ |clusterserviceversions |get;list;watch                            |\n|operators.coreos.com
          \ |subscriptions          |get;list;watch                            |\n|operators.coreos.com
          \ |operatorgroups         |get;list                                  |\n|storage.k8s.io
          \       |storageclasses         |get;list;watch                            |\n|
          \                     |namespaces             |get;list;watch                            |\n|
          \                     |configmaps             |get;list;watch                            |\n|
          \                     |services               |get;list;watch                            |\n|
          \                     |**clusterrole/view**   |                                          |\n\n\n###
          Documentation\nYou can find our documentation [here.](https://swc.saas.ibm.com/en-us/documentation/)\n\n###
          Getting help\nIf you encounter any issues while using the operators, you
          can create an issue on our [Github\nrepo](https://github.com/redhat-marketplace/redhat-marketplace-operator)
          for bugs, enhancements, or other requests. You can also visit our main page
          and\nreview our [support](https://swc.saas.ibm.com/en-us/support) and [documentation](https://swc.saas.ibm.com/en-us/documentation/).\n\n###
          Readme\nYou can find our readme [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/README.md)\n\n###
          License information\nYou can find our license information [here.](https://github.com/redhat-marketplace/redhat-marketplace-operator/blob/develop/LICENSE)\n"
        displayName: IBM Metrics Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - red hat marketplace
        - operators
        - buy
        links:
        - name: Visit the marketplace!
          url: https://marketplace.redhat.com
        - name: About
          url: https://marketplace.redhat.com/en-us/about
        - name: Support
          url: https://marketplace.redhat.com/en-us/support
        maintainers:
        - email: rhmoper@us.ibm.com
          name: RHM Operator Team
        maturity: stable
        provider:
          name: Red Hat Marketplace
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - quay.io/rh-marketplace/ibm-metrics-operator@sha256:a5a9303fe105d1c7cc6e4e839ba94f79e276321bd181db203fc9626c24c6e436
        - quay.io/rh-marketplace/redhat-marketplace-authcheck@sha256:13a81b8aca79ae8c89ee24dcc30f3322a4cfdedf929e7bb984295c1e7e631fa9
        - quay.io/rh-marketplace/redhat-marketplace-data-service@sha256:30fd8d2996b921cfce51f697ee4490a6d9a836021952429b877336f5edf2749d
        - quay.io/rh-marketplace/redhat-marketplace-metric-state@sha256:e4014c9a546c10e84b13a09db6a18431ee9489ad1296808635e48d9ef4963cf1
        - quay.io/rh-marketplace/redhat-marketplace-reporter@sha256:2bd098f057ca07f2b14be9b60ea6ad8ca6b0cae1d226b1ca3b27e23d04ccdf20
        - registry.connect.redhat.com/rh-marketplace/ibm-metrics-operator-bundle@sha256:b3b3d1c6774b25759438b0db96fcedf1907e2e4d2759cc8025933391bb74a2f1
        version: 2.20.0
      entries:
      - name: ibm-metrics-operator.v2.20.0
        version: 2.20.0
      name: stable
    defaultChannel: stable
    packageName: ibm-metrics-operator
    provider:
      name: Red Hat Marketplace
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Zettaset
      provider-url: ""
    name: xcrypt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: zts-xcrypt-operator.v0.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xcrypt.zettaset.com/v1alpha1",
                "kind": "XCrypt",
                "metadata": {
                  "name": "zts-masterset"
                },
                "spec": {
                  "status": {
                    "podNames": "zts-ca",
                    "replicas": "1,"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
          createdAt: "2021-06-16T12:29:13.97Z"
          description: XCrypt Operator deploys Zettaset XCrypt Container Encryption  -
            the only software-defined solution that offers transparent high performance
            data-at-rest encryption of critical data in container environments.
          support: Zettaset
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: XCrypt is the Schema for the xcrypts API
            kind: XCrypt
            name: xcrypts.xcrypt.zettaset.com
            version: v1alpha1
        description: "Zettaset’s RedHat Certified XCrypt Operator allows users to
          deploy multiple XCrypt components on OpenShift with just a few simple commands.
          The XCrypt Operator deployment provides software-defined encryption that
          transparently protects container data throughout your OpenShift environment.\n\n##
          Core Features\n* **Automated XCrypt provisioning** - Deploying a Zettaset
          XCrypt has never been easier. Adjust cluster specific configuration and
          let the Zettaset Operator take care of provisioning XCrypt services and
          setting up the XCrypt encryption to your exact specification. \n* **CSI
          Driver deployment**: Zettaset XCrypt Operator automatically deploys Zettaset
          XCrypt CSI Driver in order to provide automated and transparent data-at-rest
          \ encryption functionality compatible with Openshift and any other container
          orchestrator. Once the Operator is provisioned, the driver runs fine tuned
          and ready to work with other XCrypt components.\n* **Multi-node topology**:
          Encrypt as many nodes and volumes as you need to. XCrypt components can
          be installed on multiple nodes across the cluster. This is defined by two
          labels: zts-master and zts-worker. Normally zts-master node is a single
          node that runs Zettaset XCrypt major services required for the product operation.
          Multiple nodes can be marked as zts-worker and those are the nodes that
          can have encrypted volumes.\n* **Seamless Encrypted Volume mount**: Volumes
          are not only being encrypted, but also are mounted/unmounted to containers
          with Persistent Volume Claims. This happens as seamlessly as regular volumes
          provision and mount/unmount.\n* **Robust failover**: A node failure doesn’t
          affect XCrypt operations on other nodes of the cluster. Encryption on a
          restarted/recovered node gets back to normal functioning if the node stays
          healthy and data hasn’t been corrupted during the restart\n\n## About Zettaset
          XCrypt Kubernetes Encryption for OpenShift\nXCrypt Kubernetes Encryption
          for OpenShift is a software-only solution that enables encryption of data
          at rest stored in Red Hat OpenShift container environments without slowing
          down the DevSecOps process. In addition to providing a transparent and high-performance
          layer of security that flexibly protects container data across on-premise,
          in hybrid, and multi-cloud environments, it also helps your organization
          efficiently transition from DevOps to DevSecOps.  \n\nMore information about
          Zettaset XCrypt Kubernetes Encryption for OpenShift can be found in the
          [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)\n\n##
          RedHat OpenShift + Zettaset: The next generation of secure enterprise infrastructure
          for DevSecOps  \nBy combining RedHat’s Container Platform with Zettaset’s
          certified software-defined encryption, you can create the next generation
          of secure enterprise infrastructure that DevSecOps requires. Together they
          create the stable, consistent, and supported base that your organization
          needs to develop applications faster, simplify updates, drive innovation,
          and reduce the risk of potential security breaches and data theft.\n###
          Core Features\n* Software-only for simple and scalable deployments \n* Negligible
          performance impact\n* Transparent to developers and administrators \n* Works
          with AWS, Google, and Azure \n* Unique encryption key per container volume
          \n* Encrypted volumes are only available when in use\n* Direct integration
          with OpenShift \n* Automated encryption policy management Secure erase of
          volumes\n* Ability to securely decommission a node\n\n## How does XCrypt
          Kubernetes Encryption fit into OpenShift? \nXCrypt XCrypt Kubernetes Encryption
          for OpenShift Encryption makes it incredibly easy to protect container data
          in your OpenShift environment. \n### Use cases include:\n* Transitioning
          from DevOps to DevSecOps \n* Ensuring data protection in OpenShift container
          environments with negligible performance impact \n* Achieving compliance
          in regulated industries like healthcare and finance\n\n## Other Information\nRole-based
          access control needs to be configured prior to the Operator installation.
          Please find details in the [Deployment and Administration Guide](https://1f2aca073706bb52f501-133f3466c38fdb0417539cdf095d8336.ssl.cf1.rackcdn.com/ZettasetXCryptContainerEncryption.8.5.0.Final.pdf)
          \   "
        displayName: Zettaset XCrypt Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - security
        - encryption
        - storage
        links:
        - name: Zettaset
          url: https://www.zettaset.com/
        maintainers:
        - email: info@zettaset.com
          name: Zettaset
        maturity: alpha
        provider:
          name: Zettaset
        relatedImages:
        - registry.connect.redhat.com/zts/xcrypt-host-manager-1@sha256:9faa4b597eb74df412761dffa057b64fb69dbf9fa016414962eb2099a343e734
        - registry.connect.redhat.com/zts/xcrypt-kmip-server-1@sha256:9af5d21dd748dfbf4e107f27a5f10fa3c0ba928ade9b989a744b620ce34ec757
        - registry.connect.redhat.com/zts/xcrypt-license-server-1@sha256:bc5486892a41e70f5dc4c4e1d82fca8db2149598fb8b5950ba5da1ee2d9d245d
        - registry.connect.redhat.com/zts/xcrypt-operator-1-bundle@sha256:c3913778f986ae4f3a7d94da2a9d0157233dbb676e3c1eecacbdbc02a864685a
        - registry.connect.redhat.com/zts/xcrypt-operator-1@sha256:7e6316a6c41dd71c4d9025bd958a475929369284eee1a02d9e870e883acfd9ac
        - registry.connect.redhat.com/zts/zts-csi-driver-1@sha256:c15fa9a7cc40891103f426206663dcebe1210c8ee3ab6037cbf64e1f86b9c7e5
        - registry.connect.redhat.com/zts/xcrypt-ca-1@sha256:9f1015dfb459b69f5dbf449efa0517452095a6cf4f20826ebae278f9f800ffe7
        version: 0.0.17
      entries:
      - name: zts-xcrypt-operator.v0.0.17
        version: 0.0.17
      name: alpha
    defaultChannel: alpha
    packageName: xcrypt-operator
    provider:
      name: Zettaset
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-commons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: commons-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
          description: Stackable Commons Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/commons-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AuthenticationClassSpec via
              `CustomResource`
            displayName: authenticationclasses.authentication.stackable.tech
            kind: AuthenticationClass
            name: authenticationclasses.authentication.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for ConnectionSpec via `CustomResource`
            displayName: s3connections.s3.stackable.tech
            kind: S3Connection
            name: s3connections.s3.stackable.tech
            version: v1alpha1
          - description: Auto-generated derived type for BucketSpec via `CustomResource`
            displayName: s3buckets.s3.stackable.tech
            kind: S3Bucket
            name: s3buckets.s3.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Commons Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - commons
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/commons-operator@sha256:26b2098ef3530098b8acd9e4d8091686197b99af10aa883716786a8e29e99ea5
        - registry.connect.redhat.com/stackable/stackable-commons-operator@sha256:ca2d0ea011ee4d02a8043df932890291bcf7f6ab15edac5e8b1e74bc2b29ff85
        version: 25.7.0
      entries:
      - name: commons-operator.v25.7.0
        version: 25.7.0
      - name: commons-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-commons-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: NVIDIA
      provider-url: https://github.com/Mellanox/maintenance-operator/
    name: nvidia-maintenance-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: stable
    - currentCSV: nvidia-maintenance-operator.v0.2.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "MaintenanceOperatorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "maintenanceoperatorconfig-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "maintenanceoperatorconfig",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "maintenanceoperatorconfig-sample"
                },
                "spec": {
                  "logLevel": "info",
                  "maxNodeMaintenanceTimeSeconds": 1600,
                  "maxParallelOperations": 1,
                  "maxUnavailable": "30%"
                }
              },
              {
                "apiVersion": "maintenance.nvidia.com/v1alpha1",
                "kind": "NodeMaintenance",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "maintenance-operator",
                    "app.kubernetes.io/instance": "nodemaintenance-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "nodemaintenance",
                    "app.kubernetes.io/part-of": "maintenance-operator"
                  },
                  "name": "nodemaintenance-sample"
                },
                "spec": {
                  "cordon": true,
                  "drainSpec": {
                    "deleteEmptyDir": true,
                    "force": true,
                    "podEvictionFilters": [
                      {
                        "byResourceNameRegex": "nvidia.com/gpu-*"
                      },
                      {
                        "byResourceNameRegex": "nvidia.com/rdma*"
                      }
                    ],
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  },
                  "nodeName": "worker-node-01",
                  "requestorID": "some-cluster-component.vendor.com",
                  "waitForPodCompletion": {
                    "podSelector": "app=rdma-workload",
                    "timeoutSeconds": 0
                  }
                }
              }
            ]
          capabilities: Basic Install
          containerImage: ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
          createdAt: "2025-07-13T12:33:06Z"
          description: Node maintenance in K8s cluster in a coordinated manner
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: nvidia-maintenance-operator
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          provider: NVIDIA
          repository: https://github.com/Mellanox/maintenance-operator/
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MaintenanceOperatorConfig is the Schema for the maintenanceoperatorconfigs
              API
            displayName: Maintenance Operator Config
            kind: MaintenanceOperatorConfig
            name: maintenanceoperatorconfigs.maintenance.nvidia.com
            version: v1alpha1
          - description: NodeMaintenance is the Schema for the nodemaintenances API
            displayName: Node Maintenance
            kind: NodeMaintenance
            name: nodemaintenances.maintenance.nvidia.com
            version: v1alpha1
        description: Coordinates And Performs Common Node Preparations For Maintenance
        displayName: NVIDIA Maintenance Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - node
        - maintenance
        - drain
        links:
        - name: NVIDIA Maintenance Operator
          url: https://github.com/Mellanox/maintenance-operator/
        maintainers:
        - email: nvidia-network-operator-support@nvidia.com
          name: NVIDIA
        maturity: alpha
        minKubeVersion: 1.30.0
        provider:
          name: NVIDIA
          url: https://github.com/Mellanox/maintenance-operator/
        relatedImages:
        - ghcr.io/mellanox/maintenance-operator@sha256:cd24d18137d5a31dd8c791e8eb52324b8e880333fb03f1fcbc08fc5e94637746
        - registry.connect.redhat.com/nvidia-network-operator/nvidia-maintenance-operator@sha256:c9867753cd85ead32702783c880d75118549025ec4793805381aa3bc1ba1d221
        version: 0.2.3
      entries:
      - name: nvidia-maintenance-operator.v0.2.3
        version: 0.2.3
      name: v0.2
    defaultChannel: v0.2
    packageName: nvidia-maintenance-operator
    provider:
      name: NVIDIA
      url: https://github.com/Mellanox/maintenance-operator/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-s3control-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-s3control-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
          createdAt: "2025-09-19T20:52:02Z"
          description: AWS S3 Control controller is a service controller for managing
            S3 Control resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AccessPoint represents the state of an AWS s3control AccessPoint
              resource.
            displayName: AccessPoint
            kind: AccessPoint
            name: accesspoints.s3control.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon S3 Control resources in AWS from within your Kubernetes cluster.

          **About Amazon S3 Control**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon S3 Control
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - s3control
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon S3 Control Developer Resources
          url: https://aws.amazon.com/S3 Control/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: s3control maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-s3control-controller@sha256:114ea0669459ae148ab32f5366f239da9cfde23cd6bd6a67c14e5844856837a4
        - public.ecr.aws/aws-controllers-k8s/s3control-controller:1.1.1
        version: 1.1.1
      entries:
      - name: ack-s3control-controller.v1.1.1
        version: 1.1.1
      - name: ack-s3control-controller.v1.1.0
        version: 1.1.0
      - name: ack-s3control-controller.v1.0.11
        version: 1.0.11
      - name: ack-s3control-controller.v1.0.10
        version: 1.0.10
      name: alpha
    defaultChannel: alpha
    packageName: ack-s3control-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Backube
      provider-url: https://github.com/backube
    name: snapscheduler
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        - quay.io/backube/snapscheduler:3.5.0
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: candidate
    - currentCSV: snapscheduler.v3.5.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "snapscheduler.backube/v1",
                "kind": "SnapshotSchedule",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "snapscheduler",
                    "app.kubernetes.io/instance": "daily",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "snapshotschedule",
                    "app.kubernetes.io/part-of": "snapscheduler"
                  },
                  "name": "daily"
                },
                "spec": {
                  "retention": {
                    "maxCount": 7
                  },
                  "schedule": "0 0 * * *"
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage
          certified: "false"
          containerImage: quay.io/backube/snapscheduler:3.5.0
          createdAt: "2025-05-14T14:46:43Z"
          description: The SnapScheduler operator takes snapshots of Kubernetes CSI-based
            persistent volumes according to user-supplied schedules.
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/backube/snapscheduler
          support: https://github.com/backube/snapscheduler
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SnapshotSchedule defines a schedule for taking automated
              snapshots of PVC(s)
            displayName: Snapshot Schedule
            kind: SnapshotSchedule
            name: snapshotschedules.snapscheduler.backube
            version: v1
        description: |-
          ## About this operator

          The SnapScheduler operator takes snapshots of CSI-based PersistentVolumes
          according to a configurable
          [Cron-like](https://en.wikipedia.org/wiki/Cron#Overview) schedule. The
          schedules include configurable retention policies for snapshots as well as
          selectors to limit the volumes that are snapshotted. An example schedule
          could be:

          > *Snapshot **all volumes** in a namespace **daily at midnight**, retaining
          > the most recent **7** snapshots for each volume.*

          Multiple schedules can be combined to provide more elaborate protection
          schemes. For example, a given volume (or collection of volumes) could be
          protected with:

          - 6 hourly snapshots
          - 7 daily snapshots
          - 4 weekly snapshots
          - 12 monthly snapshots

          ### How it works

          The operator watches for `SnapshotSchedule` CRs in each namespace. When the
          current time matches the schedule's cronspec, the operator creates a
          `VolumeSnapshot` object for each `PersistentVolumeClaim` in the namespace
          (or subset thereof if a label selector is provided). The `VolumeSnapshot`
          objects are named according to the template:
          `<pvcname>-<schedulename>-<timestamp>`. After creating the new snapshots,
          the oldest snapshots are removed if necessary, according to the retention
          policy of the schedule.

          Please see the [full
          documentation](https://backube.github.io/snapscheduler/) for more
          information.

          ## Requirements

          - OpenShift >= 4.7 (Kubernetes >= 1.20)
          - CSI-based storage driver that supports snapshots (i.e. has the
            `CREATE_DELETE_SNAPSHOT` capability)

          ## Examples

          The schedule for snapshotting is controlled by the
          `snapshotschedules.snapscheduler.backube` Custom Resource. This is a
          namespaced resource that applies only to the PersistentVolumeClaims in its
          namespace.

          Below is a simple example that keeps 7 daily (taken at midnight) snapshots
          of all PVCs in a given namespace:

          ```yaml
          ---
          apiVersion: snapscheduler.backube/v1
          kind: SnapshotSchedule
          metadata:
            name: daily
          spec:
            retention:
              maxCount: 7
            schedule: "0 0 * * *"
          ```

          See the [usage
          documentation](https://backube.github.io/snapscheduler/usage.html) for full
          details, including how to:

          - add label selectors to restrict which PVCs this schedule applies to
          - set the VolumeSnapshotClass used by the schedule
          - apply custom labels to the automatically created VolumeSnapshot objects
        displayName: SnapScheduler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - csi
        - snapshot
        - storage
        - scheduler
        links:
        - name: Documentation
          url: https://backube.github.io/snapscheduler/
        - name: Source
          url: https://github.com/backube/snapscheduler
        - name: Blog
          url: https://backube.github.io/
        maintainers:
        - email: jstrunk@redhat.com
          name: John Strunk
        maturity: stable
        minKubeVersion: 1.20.0-0
        provider:
          name: Backube
          url: https://github.com/backube
        relatedImages:
        - quay.io/backube/snapscheduler:3.5.0
        - quay.io/brancz/kube-rbac-proxy:v0.19.1
        - quay.io/community-operator-pipeline-prod/snapscheduler@sha256:53902ba6ef33ee750285bdeacc6701027d90dada0184fd162c7a587cfdb57585
        version: 3.5.0
      entries:
      - name: snapscheduler.v3.5.0
        version: 3.5.0
      - name: snapscheduler.v3.4.0
        version: 3.4.0
      - name: snapscheduler.v3.3.0
        version: 3.3.0
      - name: snapscheduler.v3.2.0
        version: 3.2.0
      - name: snapscheduler.v3.1.0
        version: 3.1.0
      - name: snapscheduler.v3.0.0
        version: 3.0.0
      - name: snapscheduler.v2.1.0
        version: 2.1.0
      - name: snapscheduler.v2.0.0
        version: 2.0.0
      name: stable
    defaultChannel: stable
    packageName: snapscheduler
    provider:
      name: Backube
      url: https://github.com/backube
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: networkservicemesh.io
      provider-url: networkservicemesh.io
    name: nsm-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nsm-operator.v1.8.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nsm.networkservicemesh.io/v1alpha1",
                "kind": "NSM",
                "metadata": {
                  "name": "nsm-sample",
                  "namespace": "nsm"
                },
                "spec": {
                  "forwarders": [
                    {
                      "image": "ghcr.io/networkservicemesh/cmd-forwarder-vpp:v1.8.0",
                      "type": "vpp"
                    }
                  ],
                  "nsmLogLevel": "TRACE",
                  "nsmPullPolicy": "IfNotPresent",
                  "registry": {
                    "image": "ghcr.io/networkservicemesh/cmd-registry-memory:v1.8.0",
                    "type": "memory"
                  },
                  "version": "v1.8.0",
                  "webhook": {
                    "image": "ghcr.io/networkservicemesh/admission-webhook-k8s:v1.8.0"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: quay.io/acmenezes/nsm-operator:v1.8.0
          createdAt: "2023-03-02T19:36:46Z"
          operatorframework.io/suggested-namespace: nsm
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/networkservicemesh/nsm-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: NetworkServiceEndpoint
            name: networkserviceendpoints.networkservicemesh.io
            version: v1
          - kind: NetworkService
            name: networkservices.networkservicemesh.io
            version: v1
          - kind: NSM
            name: nsms.nsm.networkservicemesh.io
            version: v1alpha1
        description: |-
          The Network Service Mesh Operator manages and install network service meshes.

          Check networkservicemesh.io

          Install requirements:

          Before installing the operator it's necessary to have installed [spire](https://github.com/spiffe/spire)

          We have a ready to install yaml file for spire [here](https://github.com/networkservicemesh/nsm-operator/blob/master/config/spire/spire.yaml)
        displayName: nsm-operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - network
        - service mesh
        links:
        - name: Nsm Operator
          url: https://nsm-operator.domain
        maintainers:
        - email: adcmenezes@gmail.com
          name: Alexandre Menezes
        maturity: alpha
        provider:
          name: networkservicemesh.io
          url: networkservicemesh.io
        relatedImages:
        - quay.io/openshift-community-operators/nsm-operator@sha256:330a930f885dacf65537beb600fe7f428e02687e83778e47dcda3dc2df15ae24
        - quay.io/acmenezes/nsm-operator:v1.8.0
        version: 1.8.0
      entries:
      - name: nsm-operator.v1.8.0
        version: 1.8.0
      name: alpha
    defaultChannel: alpha
    packageName: nsm-operator
    provider:
      name: networkservicemesh.io
      url: networkservicemesh.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      operatorframework.io/os.zos: supported
      provider: IBM
      provider-url: ""
    name: ibm-block-csi-operator-community
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12
    - currentCSV: ibm-block-csi-operator.v1.12.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v3.6.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.4.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v6.3.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.12.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.0",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:30375ef2cab46b967f6d2cd46862377c83e58bc6489eb0e75b495a173ebfd84f
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e555a7ee221b1654e5c1911a895a2a26bf4774cc1de6ebadfcf083aa1d9563e7
        version: 1.12.1
      entries:
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.1
    - currentCSV: ibm-block-csi-operator.v1.12.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.2",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:466b9c30c02a3dfcd9cc94fe8eb432dca7076b8c79eb97ebe5802d716fd963f8
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:3fc8df547138a2009225d63c2a3111a672340bdf9f3cb24a889cbb69f2c33c2d
        version: 1.12.2
      entries:
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.2
    - currentCSV: ibm-block-csi-operator.v1.12.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.3",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:2bd461a1e40a1366c871585aa7a4e8039e686a66bf7192efdeb1c215209a071d
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:d5121af03b32a4b2286ff4f39499931a1ec04e05af45dccc6599dcc1b49e27ad
        version: 1.12.3
      entries:
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.3
    - currentCSV: ibm-block-csi-operator.v1.12.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.4",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:a0739ee7420b64e6195ec53057b032c356faf0ea4b8c1943725052f925eaeb19
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:e29db476de3778222bf049926223dee20f6bc34e55a04df5f0f3df38741ff23e
        version: 1.12.4
      entries:
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.4
    - currentCSV: ibm-block-csi-operator.v1.12.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "IBMBlockCSI",
                "metadata": {
                  "name": "ibm-block-csi"
                },
                "spec": {
                  "controller": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-controller",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "node": {
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-driver-node",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  },
                  "sidecars": [
                    {
                      "name": "csi-node-driver-registrar",
                      "repository": "registry.k8s.io/sig-storage/csi-node-driver-registrar",
                      "tag": "v2.13.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-provisioner",
                      "repository": "registry.k8s.io/sig-storage/csi-provisioner",
                      "tag": "v4.0.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-attacher",
                      "repository": "registry.k8s.io/sig-storage/csi-attacher",
                      "tag": "v4.8.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-snapshotter",
                      "repository": "registry.k8s.io/sig-storage/csi-snapshotter",
                      "tag": "v8.2.0",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-resizer",
                      "repository": "registry.k8s.io/sig-storage/csi-resizer",
                      "tag": "v1.13.1",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-addons-replicator",
                      "repository": "quay.io/ibmcsiblock/csi-block-volumereplication-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "csi-volume-group",
                      "repository": "quay.io/ibmcsiblock/csi-volume-group-operator",
                      "tag": "v0.9.2",
                      "imagePullPolicy": "IfNotPresent"
                    },
                    {
                      "name": "livenessprobe",
                      "repository": "registry.k8s.io/sig-storage/livenessprobe",
                      "tag": "v2.15.0",
                      "imagePullPolicy": "IfNotPresent"
                    }
                  ]
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefinition",
                "metadata": {
                  "name": "host-definition"
                },
                "spec": {
                  "hostDefinition": {
                    "nodeName": "node-name"
                  }
                }
              },
              {
                "apiVersion": "csi.ibm.com/v1",
                "kind": "HostDefiner",
                "metadata": {
                  "name": "host-definer"
                },
                "spec": {
                  "hostDefiner": {
                    "allowDelete": true,
                    "dynamicNodeLabeling": false,
                    "repository": "quay.io/ibmcsiblock/ibm-block-csi-host-definer",
                    "tag": "1.12.5",
                    "imagePullPolicy": "IfNotPresent",
                    "affinity": {
                      "nodeAffinity": {
                        "requiredDuringSchedulingIgnoredDuringExecution": {
                          "nodeSelectorTerms": [
                            {
                              "matchExpressions": [
                                {
                                  "key": "kubernetes.io/arch",
                                  "operator": "In",
                                  "values": [
                                    "amd64",
                                    "s390x",
                                    "ppc64le"
                                  ]
                                }
                              ]
                            }
                          ]
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage,Cloud Provider
          certified: "true"
          containerImage: quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
          createdAt: "2024-08-27T11:00:00Z"
          description: Run IBM block storage CSI driver.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: ""
          repository: https://github.com/IBM/ibm-block-csi-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a block storage CSI driver
            displayName: IBM block storage CSI driver
            kind: IBMBlockCSI
            name: ibmblockcsis.csi.ibm.com
            version: v1
          - description: Represents a Host Definition for IBM block storage
            displayName: IBM block storage Host Definition
            kind: HostDefinition
            name: hostdefinitions.csi.ibm.com
            version: v1
          - description: Represents Host Definer for block storage CSI driver
            displayName: IBM block storage Host Definer
            kind: HostDefiner
            name: hostdefiners.csi.ibm.com
            version: v1
        description: |
          The Container Storage Interface (CSI) Driver for IBM block storage systems enables container orchestrators such as Kubernetes to manage the life cycle of persistent storage.

          This is the official operator to deploy and manage IBM block storage CSI driver.

          For compatibility, prerequisites, release notes, and other user information, see [IBM block storage CSI driver documentation](https://www.ibm.com/docs/en/stg-block-csi-driver).
        displayName: IBM block storage CSI driver operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - IBM
        - BlockStorage
        - CSI
        links:
        - name: Source Code
          url: https://github.com/IBM/ibm-block-csi-operator
        maintainers:
        - email: csi.block1@il.ibm.com
          name: IBM Block CSI Team
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: IBM
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ibm-block-csi-operator-community@sha256:e7a8c0311df988481de08d8c74785a7e14d3581ece1448b142dfa099349b244d
        - quay.io/ibmcsiblock/ibm-block-csi-operator@sha256:0ffb93e305f042d3e04e35662338d220ee2b578e89917dad4ab7a311f8517047
        version: 1.12.5
      entries:
      - name: ibm-block-csi-operator.v1.12.5
        version: 1.12.5
      - name: ibm-block-csi-operator.v1.12.4
        version: 1.12.4
      - name: ibm-block-csi-operator.v1.12.3
        version: 1.12.3
      - name: ibm-block-csi-operator.v1.12.2
        version: 1.12.2
      - name: ibm-block-csi-operator.v1.12.1
        version: 1.12.1
      name: stable-v1.12.5
    defaultChannel: stable
    packageName: ibm-block-csi-operator-community
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-reporter-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-reporter-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1alpha1",
                "kind": "IBMLicenseServiceReporter",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-license-service-reporter-operator",
                    "app.kubernetes.io/instance": "ibmlicenseservicereporter-instance",
                    "app.kubernetes.io/managed-by": "operator",
                    "app.kubernetes.io/name": "ibmlicenseservicereporter",
                    "app.kubernetes.io/part-of": "ibm-license-service-reporter-operator"
                  },
                  "name": "instance"
                },
                "spec": {
                  "authentication": {
                    "useradmin": {
                      "enabled": true
                    }
                  },
                  "license": {
                    "accept": false
                  },
                  "version": "4.2.14"
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
          createdAt: "2025-02-26T12:10:52Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.14'
          operatorframework.io/suggested-namespace: ibm-ls-reporter
          operators.operatorframework.io/builder: operator-sdk-v1.27.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "IBM License Service Reporter is a singleton which may be
              used in a single or multi-cluster environment. It aggregates data pushed
              from IBM License Services, deployed on clusters and from ILMT. \n Documentation:
              https://ibm.biz/lsvc-rprtr. \n License: Please refer to the IBM Terms
              website (ibm.biz/lsvc-lic) to find the license terms for the particular
              IBM product for which you are deploying this component. \n IBM License
              Service Reporter is a free, optionally installed add-on – one of services
              of Cloud Pak Foundational Services. Thanks to the IBM License Service
              Reporter customer can: \n - see IBM software deployments on a single
              dashboard, verify and maintain license compliance and avoid audit infractions,
              \n - see historical data on IBM software deployments to support making
              informed decisions for future purchases, \n - see details of software
              deployments, including source (ILMT for VMs, License Service for containers),
              cluster and quantities of licenses deployed, \n - identify which software
              deployments are VMs vs. containers and can use this information to evaluate
              workloads for modernization."
            displayName: IBMLicense Service Reporter
            kind: IBMLicenseServiceReporter
            name: ibmlicenseservicereporters.operator.ibm.com
            version: v1alpha1
        description: |-
          IBM License Service Reporter is a free, optionally installed add-on – one of services of Cloud Pak Foundational Services. Thanks to the IBM License Service Reporter customer can:
           * see IBM software deployments on a single dashboard, verify and maintain license compliance and avoid audit infractions,
           * see historical data on IBM software deployments to support making informed decisions for future purchases,
           * see details of software deployments, including source (ILMT for VMs, License Service for containers), cluster and quantities of licenses deployed,
           * identify which software deployments are VMs vs. containers and can use this information to evaluate workloads for modernization.

          IBM License Service Reporter is a singleton which may be used in a single or multi-cluster environment. It aggregates data pushed from IBM License Services, deployed on clusters and from ILMT.

          **Supported platforms**

          Red Hat OpenShift Container Platform 4.10 or newer installed on Linux x86_64.

          **Prerequisites**

          Prerequisites depend on the integration of the License Service Reporter with an IBM Cloud Pak. For more information, see the applicable IBM Cloud Pak documentation.
        displayName: IBM License Service Reporter
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceReporter
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Reporter
        - License Service Reporter
        - License Reporter
        - Service Reporter
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: alpha
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:277e2ab849907553e4992c2dccc11f03ae1e0302f20aae48c620968aa97d6afc
        - registry.connect.redhat.com/cpopen/ibm-license-service-reporter-operator@sha256:02d101a5929b60e612bf64a944e282ad9fd353c47d4ff28d2c9d5c0b84704ce5
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-oauth2-proxy@sha256:c403c24857e29ddc306ba5d9bc28c3d556b181c5265fee16879fd2b4d946f899
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter-ui@sha256:213889e3c4c1d7bb10c88926ac8276d7e78c558397993d098e36f6c5da617de1
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-license-service-reporter@sha256:87e62695d733a60bfc2084cd751810df6801a5fbe1ab87b0a421f6814f1b887a
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-postgresql@sha256:b6e11080774863f56dd16d5748d03c3f89f6f4872f662ee218d6bf600491a35f
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-reporter-operator-bundle@sha256:648f92bacd52fb51655925e34da1d0bdd0da720da8489b79064e73a47dcdd54a
        version: 4.2.14
      entries:
      - name: ibm-license-service-reporter-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-reporter-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Coroot Inc.
      provider-url: https://coroot.com
    name: coroot-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: coroot-operator.1.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "coroot.com/v1",
                "kind": "Coroot",
                "metadata": {
                  "name": "coroot",
                  "namespace": "coroot"
                },
                "spec": {
                  "enterpriseEdition": {
                    "licenseKey": "<COROOT_LICENSE_KEY> from https://coroot.com/account"
                  },
                  "ingress": {
                    "host": "coroot.example.com"
                  },
                  "storage": {
                    "size": "50Gi"
                  },
                  "clickhouse": {
                    "replicas": 2,
                    "shards": 2,
                    "storage": {
                      "size": "100Gi"
                    }
                  },
                  "prometheus": {
                    "storage": {
                      "size": "100Gi"
                    }
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: Monitoring, Logging & Tracing
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/coroot/coroot-operator
          support: support@coroot.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Coroot is the Schema for the coroots API
            displayName: Coroot
            kind: Coroot
            name: coroots.coroot.com
            version: v1
        description: An open-source observability platform using eBPF to gather telemetry
          data, providing actionable insights for quick issue detection and resolution
          in Kubernetes and VM environments.
        displayName: Coroot
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - monitoring
        - coroot
        - observability
        links:
        - name: Documentation
          url: https://docs.coroot.com
        maintainers:
        - email: support@coroot.com
          name: Coroot Inc.
        maturity: stable
        provider:
          name: Coroot Inc.
          url: https://coroot.com
        relatedImages:
        - ghcr.io/coroot/coroot-node-agent@sha256:3d552e98ca2ff8c6e125496857306172ec6f8e134ff13886c7f799abef29d1b8
        - ghcr.io/coroot/coroot-operator@sha256:13f1a4a07877c8fe8230a0bb7282ff47451007c9cf689a995a37d16697b9268a
        - ghcr.io/coroot/kube-state-metrics@sha256:080a85e77b1e530df56f84c8e0ee2523c1f740adea152015367b87edbfac078d
        - ghcr.io/coroot/prometheus@sha256:e7117bd86e1088fc02ac12758bc94dba1029de54a5bc1ee04954569d8b79625d
        - registry.connect.redhat.com/coroot/coroot-operator@sha256:44aa4ea98d468f1d1ced671bd2e2bd5bf79ce467f8c572c2248abdce2b7c5acf
        - ghcr.io/coroot/clickhouse@sha256:afed1054626598371975c948e5b9e4dcaac54e8059b7ad9f0b905ef1c99f2260
        - ghcr.io/coroot/coroot-cluster-agent@sha256:674bcb147354873eaa0ee9c1328ddf758d43adbbe532c9fbbe41ddae7addc114
        - ghcr.io/coroot/coroot-ee@sha256:09b9177ac07c2d26f4bd16050d07bb1e2132431055c32e047563083ee09194a8
        version: 1.1.5
      entries:
      - name: coroot-operator.1.1.5
        version: 1.1.5
      name: stable
    defaultChannel: stable
    packageName: coroot-operator
    provider:
      name: Coroot Inc.
      url: https://coroot.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-acmpca-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-acmpca-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "acmpca.services.k8s.aws/v1alpha1",
                "kind": "CertificateAuthority",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
          createdAt: "2025-09-19T20:51:07Z"
          description: AWS ACM PCA controller is a service controller for managing
            ACM PCA resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CertificateAuthority represents the state of an AWS acmpca
              CertificateAuthority resource.
            displayName: CertificateAuthority
            kind: CertificateAuthority
            name: certificateauthorities.acmpca.services.k8s.aws
            version: v1alpha1
          - description: CertificateAuthorityActivation represents the state of an
              AWS acmpca CertificateAuthorityActivation resource.
            displayName: CertificateAuthorityActivation
            kind: CertificateAuthorityActivation
            name: certificateauthorityactivations.acmpca.services.k8s.aws
            version: v1alpha1
          - description: Certificate represents the state of an AWS acmpca Certificate
              resource.
            displayName: Certificate
            kind: Certificate
            name: certificates.acmpca.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ACM PCA resources in AWS from within your Kubernetes cluster.

          **About Amazon ACM PCA**

          AWS Private CA enables creation of private certificate authority (CA) hierarchies, including root and subordinate CAs, without the investment and maintenance costs of operating an on-premises CA. Your private CAs can issue end-entity X.509 certificates useful in scenarios including:
          - Creating encrypted TLS communication channels
          - Authenticating users, computers, API endpoints, and IoT devices
          - Cryptographically signing code
          - Implementing Online Certificate Status Protocol (OCSP) for obtaining certificate revocation status

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ACM PCA
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acmpca
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ACM PCA Developer Resources
          url: https://aws.amazon.com/private-ca/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: acmpca maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/acmpca-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-acmpca-controller@sha256:c3929eb9a46af8feb2c830fd0b9225b6e3ad2559a9b5e04fda9239ef4cc07d9a
        version: 1.1.1
      entries:
      - name: ack-acmpca-controller.v1.1.1
        version: 1.1.1
      - name: ack-acmpca-controller.v1.1.0
        version: 1.1.0
      - name: ack-acmpca-controller.v1.0.10
        version: 1.0.10
      - name: ack-acmpca-controller.v1.0.9
        version: 1.0.9
      - name: ack-acmpca-controller.v1.0.8
        version: 1.0.8
      - name: ack-acmpca-controller.v1.0.7
        version: 1.0.7
      - name: ack-acmpca-controller.v1.0.6
        version: 1.0.6
      - name: ack-acmpca-controller.v1.0.5
        version: 1.0.5
      - name: ack-acmpca-controller.v1.0.4
        version: 1.0.4
      - name: ack-acmpca-controller.v1.0.3
        version: 1.0.3
      - name: ack-acmpca-controller.v1.0.2
        version: 1.0.2
      - name: ack-acmpca-controller.v1.0.1
        version: 1.0.1
      - name: ack-acmpca-controller.v1.0.0
        version: 1.0.0
      - name: ack-acmpca-controller.v0.0.20
        version: 0.0.20
      - name: ack-acmpca-controller.v0.0.16
        version: 0.0.16
      - name: ack-acmpca-controller.v0.0.15
        version: 0.0.15
      - name: ack-acmpca-controller.v0.0.14
        version: 0.0.14
      - name: ack-acmpca-controller.v0.0.13
        version: 0.0.13
      - name: ack-acmpca-controller.v0.0.12
        version: 0.0.12
      - name: ack-acmpca-controller.v0.0.11
        version: 0.0.11
      - name: ack-acmpca-controller.v0.0.10
        version: 0.0.10
      - name: ack-acmpca-controller.v0.0.9
        version: 0.0.9
      - name: ack-acmpca-controller.v0.0.8
        version: 0.0.8
      - name: ack-acmpca-controller.v0.0.5
        version: 0.0.5
      - name: ack-acmpca-controller.v0.0.4
        version: 0.0.4
      - name: ack-acmpca-controller.v0.0.3
        version: 0.0.3
      - name: ack-acmpca-controller.v0.0.2
        version: 0.0.2
      - name: ack-acmpca-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-acmpca-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Medik8s
      provider-url: https://github.com/medik8s
    name: machine-deletion-remediation
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: machine-deletion-remediation.v0.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
          createdAt: "2023-05-11 10:46:12"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          operators.operatorframework.io/builder: operator-sdk-v1.25.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.1.0
        - quay.io/openshift-community-operators/machine-deletion-remediation@sha256:34d1e6cbfaf248a7b28c2a6e6a39f3fe46c183b9c38ac5fbb98eb3f421ba903c
        version: 0.1.0
      entries:
      - name: machine-deletion-remediation.v0.1.0
        version: 0.1.0
      name: candidate
    - currentCSV: machine-deletion-remediation.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediation",
                "metadata": {
                  "name": "machinedeletionremediation-sample"
                },
                "spec": {},
                "status": {
                  "conditions": [
                    {
                      "reason": "RemediationFinished",
                      "status": false,
                      "type": "Processing"
                    },
                    {
                      "reason": "RemediationFinished",
                      "status": true,
                      "type": "Succeeded"
                    },
                    {
                      "reason": "MachineDeletionOnBareMetalProviderReason",
                      "status": false,
                      "type": "PermanentNodeDeletionExpected"
                    }
                  ]
                }
              },
              {
                "apiVersion": "machine-deletion-remediation.medik8s.io/v1alpha1",
                "kind": "MachineDeletionRemediationTemplate",
                "metadata": {
                  "name": "machinedeletionremediationtemplate-sample"
                },
                "spec": {
                  "template": {
                    "spec": {}
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          containerImage: quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
          createdAt: "2024-03-05 18:45:16"
          description: Machine Deletion Remediation operator for reprovisioning unhealthy
            nodes using the Machine API.
          olm.skipRange: '>=0.0.1'
          operatorframework.io/suggested-namespace: openshift-workload-availability
          operatorframework.io/suggested-namespace-template: '{"kind":"Namespace","apiVersion":"v1","metadata":{"name":"openshift-workload-availability","annotations":{"openshift.io/node-selector":""}}}'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/medik8s/machine-deletion-remediation
          support: Medik8s
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: MachineDeletionRemediation is the Schema for the machinedeletionremediations
              API
            displayName: Machine Deletion Remediation
            kind: MachineDeletionRemediation
            name: machinedeletionremediations.machine-deletion-remediation.medik8s.io
            version: v1alpha1
          - description: MachineDeletionRemediationTemplate is the Schema for the
              machinedeletionremediationtemplates API
            displayName: Machine Deletion Remediation Template
            kind: MachineDeletionRemediationTemplate
            name: machinedeletionremediationtemplates.machine-deletion-remediation.medik8s.io
            version: v1alpha1
        description: |
          Machine Deletion Remediation (MDR) is a remediator designed to reprovision unhealthy
          nodes using the Machine API. MDR can be used together with NodeHealthCheck (NHC),
          which detects the unhealthy Node and creates the Custom Resource for MDR.

          # How it works
          MDR follows the annotation on the Node to the associated Machine object, confirms
          that it has an owning controller (e.g. MachineSetController), that will recreate the
          Machine once deleted, and then proceeds with deleting the Machine.

          # How to use MDR with NHC
          Since NHC comes with Self Node Remediation as the default remediation mechanism,
          it needs to be configured to use MDR with the following steps

          - Create the MDR Template CR
          - Modify NHC's `.spec.remediationTemplate` to use MDR as it's remediator

          see https://github.com/medik8s/machine-deletion-remediation#readme for configuration examples.
        displayName: Machine Deletion Remediation Operator - Community Edition
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - medik8s
        - machine
        - deletion
        - remediation
        - mdr
        links:
        - name: Machine Deletion Remediation
          url: https://medik8s.io
        - name: Source Code
          url: https://github.com/medik8s/machine-deletion-remediation
        maintainers:
        - email: medik8s@googlegroups.com
          name: Medik8s Team
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Medik8s
          url: https://github.com/medik8s
        relatedImages:
        - quay.io/community-operator-pipeline-prod/machine-deletion-remediation@sha256:99b0b439d37ecdac8ec74433acf1a0ef787d0d6d1d7ef7cfe5f0a29f9dfe6071
        - quay.io/medik8s/machine-deletion-remediation-operator:v0.3.0
        - quay.io/brancz/kube-rbac-proxy:v0.15.0
        version: 0.3.0
      entries:
      - name: machine-deletion-remediation.v0.3.0
        version: 0.3.0
      - name: machine-deletion-remediation.v0.2.1
        version: 0.2.1
      - name: machine-deletion-remediation.v0.2.0
        version: 0.2.0
      name: stable
    defaultChannel: stable
    packageName: machine-deletion-remediation
    provider:
      name: Medik8s
      url: https://github.com/medik8s
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: container-security-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: container-security-operator.v3.4.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                "kind": "ImageManifestVuln",
                "metadata": {
                  "selfLink": "/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "resourceVersion": "14565395",
                  "name": "sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18",
                  "uid": "3f16a188-e69a-406c-aa06-df340195409c",
                  "creationTimestamp": "2021-03-17T12:07:16Z",
                  "generation": 2,
                  "managedFields": [
                    {
                      "apiVersion": "secscan.quay.redhat.com/v1alpha1",
                      "fieldsType": "FieldsV1",
                      "fieldsV1": {
                        "f:metadata": {
                          "f:labels": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          }
                        },
                        "f:spec": {
                          ".": {},
                          "f:features": {},
                          "f:image": {},
                          "f:manifest": {}
                        },
                        "f:status": {
                          ".": {},
                          "f:affectedPods": {
                            ".": {},
                            "f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": {}
                          },
                          "f:fixableCount": {},
                          "f:highCount": {},
                          "f:highestSeverity": {},
                          "f:lastUpdate": {},
                          "f:lowCount": {},
                          "f:mediumCount": {}
                        }
                      },
                      "manager": "security-labeller",
                      "operation": "Update",
                      "time": "2021-03-17T13:07:20Z"
                    }
                  ],
                  "namespace": "openshift-cluster-version",
                  "labels": {
                    "openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6": "true"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
          createdAt: 2021-11-02 11:11 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:246712ce23cdd590ff83d84e9800b4ed761886ed5bab9f9cda4125a5d930e7e5
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:447e05859134a0455ca55f8aae16c8a27ce06af4c18c191b41f6339df553f1bf
        version: 3.4.7
      entries:
      - name: container-security-operator.v3.4.7
        version: 3.4.7
      name: quay-v3.4
    - currentCSV: container-security-operator.v3.5.7
      currentCSVDesc:
        annotations:
          alm-examples: "[\n  {\n    \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \   \"kind\": \"ImageManifestVuln\",\n    \"metadata\": {\n      \"selfLink\":
            \"/apis/secscan.quay.redhat.com/v1alpha1/namespaces/openshift-cluster-version/imagemanifestvulns/sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"resourceVersion\": \"14565395\",\n      \"name\": \"sha256.a7b23f38d1e5be975a6b516739689673011bdfa59a7158dc6ca36cefae169c18\",\n
            \     \"uid\": \"3f16a188-e69a-406c-aa06-df340195409c\",\n      \"creationTimestamp\":
            \"2021-03-17T12:07:16Z\",\n      \"generation\": 2,\n      \"managedFields\":
            [\n        {\n          \"apiVersion\": \"secscan.quay.redhat.com/v1alpha1\",\n
            \         \"fieldsType\": \"FieldsV1\",\n          \"fieldsV1\": {\n            \"f:metadata\":
            {\n              \"f:labels\": {\n                \".\": {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              }\n            },\n            \"f:spec\": {\n              \".\":
            {},\n              \"f:features\": {},\n              \"f:image\": {},\n
            \             \"f:manifest\": {}\n            },\n            \"f:status\":
            {\n              \".\": {},\n              \"f:affectedPods\": {\n                \".\":
            {},\n                \"f:openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            {}\n              },\n              \"f:fixableCount\": {},\n              \"f:highCount\":
            {},\n              \"f:highestSeverity\": {},\n              \"f:lastUpdate\":
            {},\n              \"f:lowCount\": {},\n              \"f:mediumCount\":
            {}\n            }\n          },\n          \"manager\": \"security-labeller\",\n
            \         \"operation\": \"Update\",\n          \"time\": \"2021-03-17T13:07:20Z\"\n
            \       }\n      ],\n      \"namespace\": \"openshift-cluster-version\",\n
            \     \"labels\": {\n        \"openshift-cluster-version/cluster-version-operator-b98ccdb7d-t5sq6\":
            \"true\"\n      }\n    },\n    \"spec\": {\n      \"features\": [\n        {\n
            \         \"name\": \"musl\",\n          \"namespaceName\": \"alpine:v3.14\",\n
            \         \"version\": \"1.2.2-r3\",\n          \"versionformat\": \"dpkg\",\n
            \         \"vulnerabilities\": [\n            {\n              \"fixedby\":
            \"1.2.2_pre2-r0\",\n              \"link\": \"https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28928\",\n
            \             \"metadata\": \"{\\\"NVD\\\": {\\\"CVSSv3\\\": {\\\"Vectors\\\":
            \\\"CVSS:3.0/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H\\\", \\\"ImpactScore\\\":
            3.6, \\\"Score\\\": 5.5, \\\"ExploitabilityScore\\\": 1.8}, \\\"CVSSv2\\\":
            {\\\"Score\\\": 2.1, \\\"PublishedDateTime\\\": \\\"2020-11-24T18:15Z\\\",
            \\\"Vectors\\\": \\\"AV:L/AC:L/Au:N/C:N/I:N/A:P\\\"}}}\",\n              \"name\":
            \"CVE-2020-28928\",\n              \"namespaceName\": \"alpine:v3.14\",\n
            \             \"severity\": \"Low\"\n            }\n          ]\n        }\n
            \     ],\n      \"image\": \"quay.io/harishg/pruned-redhat-operators-index\",\n
            \     \"manifest\": \"sha256:19631c86642609254777ff59fc9e128aa5d0a94a7dbb504ef1fcfb2bbc6e486e\"\n
            \   }\n  }\n]    "
          capabilities: Full Lifecycle
          categories: OpenShift Optional, Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
          createdAt: 2021-09-10 23:09 UTC
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.5.7'
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Quay Container Security Operator (CSO) brings Quay and Clair
          metadata to Kubernetes / OpenShift. Starting with vulnerability information
          the scope will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Quay Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Quay Container Security
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: quay-devel@redhat.com
          name: Quay Engineering Team
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:68e766235bdcd8f3335e2ab6e09942d98b183b3e498abb65acab6481664d46ed
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:6b3581e62550d6a615d36eee6eaf4c89b9c35ba53cab8433bd93c68393a9d115
        version: 3.5.7
      entries:
      - name: container-security-operator.v3.5.7
        version: 3.5.7
      name: quay-v3.5
    - currentCSV: container-security-operator.v3.10.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.7.x <3.10.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.10.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:91afe3cccd5eabbe4933459e63fd64a6bd87e399ea1aa81ae30dc3e8341c83e4
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:de7c66a0004f3c0643c09d4c3bbb4a0f1f8f2cbf264250802cdc58028fc72b50
        version: 3.10.15
      entries:
      - name: container-security-operator.v3.10.15
        version: 3.10.15
      - name: container-security-operator.v3.10.14
        version: 3.10.14
      - name: container-security-operator.v3.10.13
        version: 3.10.13
      - name: container-security-operator.v3.10.12
        version: 3.10.12
      - name: container-security-operator.v3.10.11
        version: 3.10.11
      - name: container-security-operator.v3.10.10
        version: 3.10.10
      - name: container-security-operator.v3.10.9
        version: 3.10.9
      - name: container-security-operator.v3.10.8
        version: 3.10.8
      - name: container-security-operator.v3.10.7
        version: 3.10.7
      - name: container-security-operator.v3.10.6
        version: 3.10.6
      - name: container-security-operator.v3.10.5
        version: 3.10.5
      - name: container-security-operator.v3.10.4
        version: 3.10.4
      - name: container-security-operator.v3.10.3
        version: 3.10.3
      - name: container-security-operator.v3.10.2
        version: 3.10.2
      - name: container-security-operator.v3.10.1
        version: 3.10.1
      - name: container-security-operator.v3.10.0
        version: 3.10.0
      name: stable-3.10
    - currentCSV: container-security-operator.v3.11.13
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.8.x <3.11.13'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.11.13
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:56db2fd54dbe982dde6a4558b664b19a676a8ac99bfbc8082d097cd6d71a03ea
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:3914389fd53c371627921298c5dc3564999d92bc720f27a240f74c034ebbf5ec
        version: 3.11.13
      entries:
      - name: container-security-operator.v3.11.13
        version: 3.11.13
      - name: container-security-operator.v3.11.12
        version: 3.11.12
      - name: container-security-operator.v3.11.11
        version: 3.11.11
      - name: container-security-operator.v3.11.10
        version: 3.11.10
      - name: container-security-operator.v3.11.9
        version: 3.11.9
      - name: container-security-operator.v3.11.8
        version: 3.11.8
      - name: container-security-operator.v3.11.7
        version: 3.11.7
      - name: container-security-operator.v3.11.6
        version: 3.11.6
      - name: container-security-operator.v3.11.5
        version: 3.11.5
      - name: container-security-operator.v3.11.4
        version: 3.11.4
      - name: container-security-operator.v3.11.3
        version: 3.11.3
      - name: container-security-operator.v3.11.2
        version: 3.11.2
      - name: container-security-operator.v3.11.1
        version: 3.11.1
      - name: container-security-operator.v3.11.0
        version: 3.11.0
      name: stable-3.11
    - currentCSV: container-security-operator.v3.12.12
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.9.x <3.12.12'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.12.12
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:d17af97462de060dc21f997dac973d9f895ce91f80821f1a810d93dbcbda31e3
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:1a15328f0b2c64d3a857f2f155ad4def095027e809610de98545e303eec63225
        version: 3.12.12
      entries:
      - name: container-security-operator.v3.12.12
        version: 3.12.12
      - name: container-security-operator.v3.12.11
        version: 3.12.11
      - name: container-security-operator.v3.12.10
        version: 3.12.10
      - name: container-security-operator.v3.12.9
        version: 3.12.9
      - name: container-security-operator.v3.12.8
        version: 3.12.8
      - name: container-security-operator.v3.12.7
        version: 3.12.7
      - name: container-security-operator.v3.12.6
        version: 3.12.6
      - name: container-security-operator.v3.12.5
        version: 3.12.5
      - name: container-security-operator.v3.12.4
        version: 3.12.4
      - name: container-security-operator.v3.12.3
        version: 3.12.3
      - name: container-security-operator.v3.12.2
        version: 3.12.2
      - name: container-security-operator.v3.12.1
        version: 3.12.1
      - name: container-security-operator.v3.12.0
        version: 3.12.0
      name: stable-3.12
    - currentCSV: container-security-operator.v3.13.8
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.10.x <3.13.8'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.13.8
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:783380361431859301477dfe63b6915240e934ffd07a2af2651094579be47a15
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:ba50621739581da9a661c2c2e355849c328cfa33b7311b06abb148e1183a6321
        version: 3.13.8
      entries:
      - name: container-security-operator.v3.13.8
        version: 3.13.8
      - name: container-security-operator.v3.13.7
        version: 3.13.7
      - name: container-security-operator.v3.13.6
        version: 3.13.6
      - name: container-security-operator.v3.13.5
        version: 3.13.5
      - name: container-security-operator.v3.13.4
        version: 3.13.4
      - name: container-security-operator.v3.13.3
        version: 3.13.3
      - name: container-security-operator.v3.13.2
        version: 3.13.2
      - name: container-security-operator.v3.13.1
        version: 3.13.1
      - name: container-security-operator.v3.13.0
        version: 3.13.0
      name: stable-3.13
    - currentCSV: container-security-operator.v3.14.5
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.11.x <3.14.5'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.14.5
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:46f8c0ea6e383b742ab9cdaed1c8f624c2cf405bb1ce39c0ea6977940d52de07
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:c42487ccf098df3ec6dbeb5d409346bdd30e92781cf3ffea5fdb6dd0b196d0cf
        version: 3.14.5
      entries:
      - name: container-security-operator.v3.14.5
        version: 3.14.5
      - name: container-security-operator.v3.14.4
        version: 3.14.4
      - name: container-security-operator.v3.14.3
        version: 3.14.3
      - name: container-security-operator.v3.14.2
        version: 3.14.2
      - name: container-security-operator.v3.14.1
        version: 3.14.1
      - name: container-security-operator.v3.14.0
        version: 3.14.0
      name: stable-3.14
    - currentCSV: container-security-operator.v3.15.2
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.12.x <3.15.2'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.15.2
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:e4ea95a86e77aa64da0b1800f00c2e0e598776cde7cfc7bb9941a60bac207ed9
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:81e929ec095f791a83310ee7fcda239d5c348ba69a3cf5adcf1b3a77c349ce77
        version: 3.15.2
      entries:
      - name: container-security-operator.v3.15.2
        version: 3.15.2
      - name: container-security-operator.v3.15.1
        version: 3.15.1
      - name: container-security-operator.v3.15.0
        version: 3.15.0
      name: stable-3.15
    - currentCSV: container-security-operator.v3.6.10
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.3.x <3.6.10'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.6.10
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:aca3d91b7b9f8c1e8a1e0d045c42ba5f27444e0c1c0bc4a87c697b428c2fff78
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:641b80df8d7102d49a5dfd5b99245669f60393880a6f1bc8ba8450b88a640d96
        version: 3.6.10
      entries:
      - name: container-security-operator.v3.6.10
        version: 3.6.10
      - name: container-security-operator.v3.6.9
        version: 3.6.9
      - name: container-security-operator.v3.6.8
        version: 3.6.8
      - name: container-security-operator.v3.6.7
        version: 3.6.7
      - name: container-security-operator.v3.6.6
        version: 3.6.6
      - name: container-security-operator.v3.6.5
        version: 3.6.5
      - name: container-security-operator.v3.6.4
        version: 3.6.4
      - name: container-security-operator.v3.6.2
        version: 3.6.2
      - name: container-security-operator.v3.6.1
        version: 3.6.1
      - name: container-security-operator.v3.6.0
        version: 3.6.0
      name: stable-3.6
    - currentCSV: container-security-operator.v3.7.14
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.4.x <3.7.14'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.7.14
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:5af7a32da67d39e7a011672de333792e63604cbfee997c19a495fc759d2fd739
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:10736a06beb503b4eb73260a87a2cde9bd83353374ee0ac9e8dd0a6260cf2665
        version: 3.7.14
      entries:
      - name: container-security-operator.v3.7.14
        version: 3.7.14
      - name: container-security-operator.v3.7.13
        version: 3.7.13
      - name: container-security-operator.v3.7.12
        version: 3.7.12
      - name: container-security-operator.v3.7.11
        version: 3.7.11
      - name: container-security-operator.v3.7.10
        version: 3.7.10
      - name: container-security-operator.v3.7.9
        version: 3.7.9
      - name: container-security-operator.v3.7.8
        version: 3.7.8
      - name: container-security-operator.v3.7.7
        version: 3.7.7
      - name: container-security-operator.v3.7.6
        version: 3.7.6
      - name: container-security-operator.v3.7.5
        version: 3.7.5
      - name: container-security-operator.v3.7.4
        version: 3.7.4
      - name: container-security-operator.v3.7.3
        version: 3.7.3
      - name: container-security-operator.v3.7.2
        version: 3.7.2
      - name: container-security-operator.v3.7.1
        version: 3.7.1
      - name: container-security-operator.v3.7.0
        version: 3.7.0
      name: stable-3.7
    - currentCSV: container-security-operator.v3.8.15
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          olm.skipRange: '>=3.5.x <3.8.15'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.8.15
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:108bd033543beb7cce077130c96f992c202f7a258229c239795f8f19e5fae5e5
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:74f0cf91e2feb60fe794f71b9f56fa8afc60e53379a0aef104701dd6692272e5
        version: 3.8.15
      entries:
      - name: container-security-operator.v3.8.15
        version: 3.8.15
      - name: container-security-operator.v3.8.14
        version: 3.8.14
      - name: container-security-operator.v3.8.13
        version: 3.8.13
      - name: container-security-operator.v3.8.12
        version: 3.8.12
      - name: container-security-operator.v3.8.11
        version: 3.8.11
      - name: container-security-operator.v3.8.10
        version: 3.8.10
      - name: container-security-operator.v3.8.9
        version: 3.8.9
      - name: container-security-operator.v3.8.8
        version: 3.8.8
      - name: container-security-operator.v3.8.7
        version: 3.8.7
      - name: container-security-operator.v3.8.6
        version: 3.8.6
      - name: container-security-operator.v3.8.5
        version: 3.8.5
      - name: container-security-operator.v3.8.4
        version: 3.8.4
      - name: container-security-operator.v3.8.3
        version: 3.8.3
      - name: container-security-operator.v3.8.2
        version: 3.8.2
      - name: container-security-operator.v3.8.1
        version: 3.8.1
      - name: container-security-operator.v3.8.0
        version: 3.8.0
      name: stable-3.8
    - currentCSV: container-security-operator.v3.9.16
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Security
          containerImage: registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
          createdAt: "2020-04-13 00:00:00"
          description: Identify image vulnerabilities in Kubernetes pods
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=3.6.x <3.9.16'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus",
            "Red Hat Quay"]'
          quay-version: 3.9.16
          repository: https://github.com/quay/container-security-operator
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a set of vulnerabilities in an image manifest.
            displayName: Image Manifest Vulnerability
            kind: ImageManifestVuln
            name: imagemanifestvulns.secscan.quay.redhat.com
            version: v1alpha1
        description: The Container Security Operator (CSO) brings Quay and Clair metadata
          to Kubernetes / OpenShift. Starting with vulnerability information the scope
          will get expanded over time. If it runs on OpenShift, the corresponding
          vulnerability information is shown inside the OCP Console. The Container
          Security Operator enables cluster administrators to monitor known container
          image vulnerabilities in pods running on their Kubernetes cluster. The controller
          sets up a watch on pods in the specified namespace(s) and queries the container
          registry for vulnerability information. If the container registry supports
          image scanning, such as [Quay](https://github.com/quay/quay) with [Clair](https://github.com/quay/clair),
          then the Operator will expose any vulnerabilities found via the Kubernetes
          API in an `ImageManifestVuln` object.  This Operator requires no additional
          configuration after deployment, and will begin watching pods and populating
          `ImageManifestVulns` immediately once installed.
        displayName: Red Hat Quay Container Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - open source
        - containers
        - security
        links:
        - name: Source Code
          url: https://github.com/quay/container-security-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/quay/quay-container-security-operator-bundle@sha256:72724358b360ab74cc0857e2f18749d3d5b4c76bb6e930f31c5b04c7716e4c2d
        - registry.redhat.io/quay/quay-container-security-operator-rhel8@sha256:df8e1d246273d21049c6bb8169a2e5f3fc8d69c85c564359abea7364969c6acc
        version: 3.9.16
      entries:
      - name: container-security-operator.v3.9.16
        version: 3.9.16
      - name: container-security-operator.v3.9.15
        version: 3.9.15
      - name: container-security-operator.v3.9.14
        version: 3.9.14
      - name: container-security-operator.v3.9.13
        version: 3.9.13
      - name: container-security-operator.v3.9.12
        version: 3.9.12
      - name: container-security-operator.v3.9.11
        version: 3.9.11
      - name: container-security-operator.v3.9.10
        version: 3.9.10
      - name: container-security-operator.v3.9.9
        version: 3.9.9
      - name: container-security-operator.v3.9.8
        version: 3.9.8
      - name: container-security-operator.v3.9.7
        version: 3.9.7
      - name: container-security-operator.v3.9.6
        version: 3.9.6
      - name: container-security-operator.v3.9.5
        version: 3.9.5
      - name: container-security-operator.v3.9.4
        version: 3.9.4
      - name: container-security-operator.v3.9.3
        version: 3.9.3
      - name: container-security-operator.v3.9.2
        version: 3.9.2
      - name: container-security-operator.v3.9.1
        version: 3.9.1
      - name: container-security-operator.v3.9.0
        version: 3.9.0
      name: stable-3.9
    defaultChannel: stable-3.15
    packageName: container-security-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-license-service-scanner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-license-service-scanner-operator.v4.2.16
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "IBMLicenseServiceScanner",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ibm-licensing-scanner-operator",
                    "app.kubernetes.io/instance": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "ibm-licensing-scanner-instance",
                    "app.kubernetes.io/part-of": "ibm-licensing-scanner-operator"
                  },
                  "name": "ibm-licensing-scanner-instance"
                },
                "spec": {
                  "license": {
                    "accept": false
                  },
                  "license-service-upload-secret": "ibm-licensing-scanner-upload-secret",
                  "log-level": "INFO",
                  "registry-pull-secret": "ibm-licensing-scanner-registry-pull-secret",
                  "scan": {
                    "namespaces": []
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
          createdAt: "2025-05-14T12:59:00Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.0.0 <4.2.16'
          operatorframework.io/suggested-namespace: ibm-licensing-scanner
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: IBMLicenseServiceScanner is the Schema for the ibmlicenseservicescanners
              API
            displayName: IBMLicense Service Scanner
            kind: IBMLicenseServiceScanner
            name: ibmlicenseservicescanners.operator.ibm.com
            version: v1
        description: IBM License Service Scanner
        displayName: IBM License Service Scanner
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - licensing
        - usage
        - IBMLicenseServiceScanner
        - IBM
        - Cloud
        - License
        - License Service
        - Service
        - Scanner
        - License Service Scanner
        - License Scanner
        - Service Scanner
        links:
        - name: Ibm License Service Scanner Operator
          url: https://ibm-license-service-scanner-operator.domain
        maintainers:
        - email: talk2sam@us.ibm.com
          name: IBM
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:77c9db26a13dba25438a6fc5a41290606418481e08ea2ecc0a9a397062af74fb
        - registry.connect.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:61f935c5985994994eb2f454fb12d74312f7d7d1627fc2b55f3d6ec9bdeb4bc7
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:f3503f72a90ff8abbadaef3e78a16db6e1fbff4c90b475c33ead17bf1e69990c
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.16
      entries:
      - name: ibm-license-service-scanner-operator.v4.2.16
        version: 4.2.16
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-license-service-scanner-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: MongoDB, Inc
      provider-url: ""
    name: mongodb-kubernetes
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-kubernetes.v1.4.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "ClusterMongoDBRole",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "mongodb-enterprise"
                  },
                  "name": "clustermongodbrole-sample"
                },
                "spec": {
                  "db": "admin",
                  "role": "rootMonitor",
                  "roles": [
                    {
                      "db": "admin",
                      "role": "root"
                    },
                    {
                      "db": "admin",
                      "role": "clusterMonitor"
                    }
                  ]
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "my-replica-set"
                },
                "spec": {
                  "credentials": "my-credentials",
                  "members": 3,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDB",
                "metadata": {
                  "name": "sample-sharded-cluster"
                },
                "spec": {
                  "configServerCount": 3,
                  "credentials": "my-credentials",
                  "mongodsPerShardCount": 3,
                  "mongosCount": 2,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "shardCount": 1,
                  "type": "ShardedCluster",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBMultiCluster",
                "metadata": {
                  "name": "multi-replica-set"
                },
                "spec": {
                  "clusterSpecList": [
                    {
                      "clusterName": "e2e.cluster1.mongokubernetes.com",
                      "members": 2
                    },
                    {
                      "clusterName": "e2e.cluster2.mongokubernetes.com",
                      "members": 1
                    },
                    {
                      "clusterName": "e2e.cluster3.mongokubernetes.com",
                      "members": 2
                    }
                  ],
                  "credentials": "my-credentials",
                  "duplicateServiceObjects": false,
                  "opsManager": {
                    "configMapRef": {
                      "name": "my-project"
                    }
                  },
                  "persistent": true,
                  "type": "ReplicaSet",
                  "version": "4.4.0-ent"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBOpsManager",
                "metadata": {
                  "name": "ops-manager"
                },
                "spec": {
                  "adminCredentials": "ops-manager-admin",
                  "applicationDatabase": {
                    "members": 3,
                    "podSpec": {
                      "cpu": 1
                    }
                  },
                  "configuration": {
                    "mms.fromEmailAddr": "admin@thecompany.com"
                  },
                  "externalConnectivity": {
                    "type": "LoadBalancer"
                  },
                  "version": "6.0.3"
                }
              },
              {
                "apiVersion": "mongodb.com/v1",
                "kind": "MongoDBUser",
                "metadata": {
                  "name": "my-replica-set-x509-user"
                },
                "spec": {
                  "db": "$external",
                  "mongodbResourceRef": {
                    "name": "my-replica-set"
                  },
                  "roles": [
                    {
                      "db": "admin",
                      "name": "dbOwner"
                    }
                  ],
                  "username": "CN=my-replica-set-x509-user,OU=cloud,O=MongoDB,L=New York,ST=New York,C=US"
                },
                "status": {
                  "phase": "Updated"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
          createdAt: "2025-09-16T14:56:17Z"
          description: The MongoDB Controllers for Kubernetes enable easy deploys
            of MongoDB into Kubernetes clusters, using our management, monitoring
            and backup platforms, Ops Manager and Cloud Manager.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.26.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/mongodb/mongodb-kubernetes
          support: support@mongodb.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: ClusterMongoDBRole
            name: clustermongodbroles.mongodb.com
            version: v1
          - description: MongoDB Deployment
            displayName: MongoDB Deployment
            kind: MongoDB
            name: mongodb.mongodb.com
            version: v1
          - kind: MongoDBCommunity
            name: mongodbcommunity.mongodbcommunity.mongodb.com
            version: v1
          - description: MongoDB Multi Deployment
            displayName: MongoDB Multi Deployment
            kind: MongoDBMultiCluster
            name: mongodbmulticluster.mongodb.com
            version: v1
          - kind: MongoDBSearch
            name: mongodbsearch.mongodb.com
            version: v1
          - description: MongoDB x509 User
            displayName: MongoDB User
            kind: MongoDBUser
            name: mongodbusers.mongodb.com
            version: v1
          - description: MongoDB Ops Manager
            displayName: MongoDB Ops Manager
            kind: MongoDBOpsManager
            name: opsmanagers.mongodb.com
            version: v1
        description: |
          MongoDB Controllers for Kubernetes (MCK) is an open-source project. It represents a merge of the previous
          MongoDB Community Operator (MCO) and MongoDB Enterprise Kubernetes Operator (MEKO).

          This release brings MongoDB Community and Enterprise editions together under a single, unified operator,
          making it easier to manage, scale, and upgrade your deployments. While the first version simply brings
          the capabilities of both into a single Operator, future changes will build on this to more closely align
          how Community and Enterprise are managed in Kubernetes, to offer an even more seamless and streamlined experience.
          As an open-source project, it now allows for community contributions, helping drive quicker bug fixes and ongoing innovation.
          The MongoDB Controllers for Kubernetes enable easy deploys of MongoDB
          into Kubernetes clusters, using our management, monitoring and backup
          platforms, Ops Manager and Cloud Manager.

          ## Before You Start

          To start using the operator you''ll need an account in MongoDB Cloud Manager or
          a MongoDB Ops Manager deployment.

          * [Create a Secret with your OpsManager API key](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-operator-credentials/#procedure)

          * [Create a ConfigMap with your OpsManager project ID and URL](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/create-project-using-configmap/)

          By installing this integration, you will be able to deploy MongoDB instances
          with a single simple command.

          ## Required Parameters

          * `opsManager` or `cloudManager` - Enter the name of the ConfigMap containing project information
          * `credentials` - Enter the name of the Secret containing your OpsManager credentials
          * `type` - Enter MongoDB Deployment Types ("Standalone", "ReplicaSet", "ShardedCluster"

          ## Supported MongoDB Deployment Types ##

          * Standalone: An instance of mongod that is running as a single server and
          not as part of a replica set, this is, it does not do any kind of
          replication.

          * Replica Set: A replica set in MongoDB is a group of mongod processes that
          maintain the same data set. Replica sets provide redundancy and high
          availability, and are the basis for all production deployments. This section
          introduces replication in MongoDB as well as the components and architecture
          of replica sets. The section also provides tutorials for common tasks
          related to replica sets.

          * Sharded Cluster: The set of nodes comprising a sharded MongoDB deployment.
          A sharded cluster consists of config servers, shards, and one or more mongos
          routing processes. Sharding is a A database architecture that partitions
          data by key ranges and distributes the data among two or more database
          instances. Sharding enables horizontal scaling.

          ## Requirements for deploying MongoDB OpsManager

          * In order to deploy resources of type MongoDB OpsManager, you will need to
          create a secret containing the [credentials](https://docs.mongodb.com/kubernetes-operator/stable/tutorial/plan-om-resource/#om-rsrc-prereqs)
          for the Global Owner user

          ## Security ##

          The operator can enable TLS for all traffic between servers and also between
          clients and servers. Before enabling `security.tls.enabled` to `true` you
          should create your certificates.  or you can leave the operator to create all
          the certificates for you. The operator ability to create certs is been
          deprecated due to Kubernetes API changes.

          For more information, please read the official MongoDB
          Kubernetes Operator  [docs](https://docs.mongodb.com/kubernetes-operator/stable/).
        displayName: MongoDB Controllers for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - mongodb
        - databases
        - nosql
        links:
        - name: Documentation
          url: https://docs.opsmanager.mongodb.com/current/tutorial/install-k8s-operator/index.html
        maintainers:
        - email: support@mongodb.com
          name: MongoDB, Inc
        maturity: stable
        provider:
          name: MongoDB, Inc
        relatedImages:
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:146c4f5dc443f94f54d72808b4e4df93bb904e23589b222d17700166780776b4
        - registry.connect.redhat.com/mongodb/mongodb-kubernetes-bundle@sha256:354b16ea43cfcb9bff704b6265fb10d94037beb231138e1fb44963f3b348bb40
        - quay.io/mongodb/mongodb-enterprise-server@sha256:376ad7e5fe42651d946cba5705c0eee44aa636c1b4ff258210e4ce2fe4205b4d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa8683500e8d3632c4588bbcdddb4e3186a01f48aefe1b6d95f8860dae8b6c76
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c752a7f0726d21d55f204a359b85fa29cd727a5fe571b37289ae3ea52d6206b0
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c8aec83b76dfecc05b8f5d8f40fd61b24980c0287a07294f75f321c2d42c32b4
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8445d59d491a6d3890ea9c8b421214bcd0fdcfbc420ed40497a790b79f0bc89e
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9bef77e905dadd1e61d0439dd53b21ed1f3606e49affd9ffc4ba8156578e8f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:11ab324698f919f84a5dab5a964d93fe63f98a48146bb3b86974cba7606d30ec
        - quay.io/mongodb/mongodb-enterprise-server@sha256:c48631742ad1fec8a9ee745fc3cc9ca1fe7d86ef7fa68ad8e8337c7a4566dc87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:454b1585c342375dea643b10ead688d0f62347841f5f30059fcd4f92e3297ee9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a86c4d8a0b9ca3c20cc524809676eb03a944ada43378ec98c107cd72cc64700f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3add8b68bfbaa42596fe5dbbee4623ba5ff1a826b18bb71301525adf574dcb94
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:adeab0dc766a27138db7df5f37cbb1da860c14179ef631b2e001e5bb875325e4
        - quay.io/mongodb/mongodb-agent@sha256:6cc4e953df3f09d793f0dbc6e57e55a3c5277ef63177a6d8f691a41227facf64
        - quay.io/mongodb/mongodb-enterprise-server@sha256:dfd4277f0c16ed476ac7de51894739e8c8d13a65ea5024b9a97b3dce0f6c6975
        - quay.io/mongodb/mongodb-agent@sha256:6c23febaa9dd52bc29ac1e52d087fc4b2b47c578f4d6ac3a1674bb750a38cb07
        - quay.io/mongodb/mongodb-search@sha256:7149a7b6f1ee84646d3972837920f80fc5db49503c598406ce92e2f794bd5f0d
        - quay.io/mongodb/mongodb-agent@sha256:3fa5bcfda3228e0ea880434a60ced829e3872e74763b347636c18475d866d0e1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7885ab6e88214f9289d8babf5189136dde0dd4c84a15a5481ccd3ef512b2e66f
        - quay.io/mongodb/mongodb-agent@sha256:5f5031e4c2a3729f618d026b0a7f34fd4509ea73d5a4e3068b551041dc291583
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:439d6fefb469613432e2d05149dfc96c32b16469d743835d5f402a2ada8be8cc
        - quay.io/mongodb/mongodb-enterprise-server@sha256:15e31285c8949a4666f7702df53416b0905d76945a28575a6028a876cc8a6f0f
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7266b405ebb2dd24a9f7c12d3921024f547c3d8c01831e6bcfc1d551d63be1
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:f71b027fccf649f18f63f8df7a59b1bac421a2cdab6712e0c985a93c08ed970a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:720056b985c1a16d48b2aecb7a88a15cba3d8186e55b3c6c24729699a85c8232
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b8aa1edbc5fb02af312a863bc7db8535a282dfd692a07b1e321e5a1acbfeb6a1
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1a77851485ad3b2566e765b4bb051ff0ddfa48703492cef6c6db021dde84d87f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:78f11c9cd5e63f2b0e10c67dc1491f4ce1279fa90178c108814725414defab9a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d49417ea85ee1ddb84429552a5aac87a82c2843904d8c0300a43b657b7a49e0e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:09788976bdc1327acbf18cedf8efd5fca2159e43b506faa40e03ee42ad94928a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:85c1d669c073caf9cbaac815e4c8ea2ca18b71deefa03bf84cd6d8b8ebd53e12
        - quay.io/mongodb/mongodb-enterprise-server@sha256:fb8397cbcdbf7c33f6ad142492792771127f0d9a8c46388f350b6927405ee590
        - quay.io/mongodb/mongodb-agent@sha256:da77dcb60b927bf6059c2c9bf008f8c3c0eec45cf7d1b40c34a0c263e8cc18b8
        - quay.io/mongodb/mongodb-enterprise-server@sha256:aa300abf739e26350a258f643fa8e57aedacd041055c7b923bd22b21ef554cdb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:523566f44cbb6bbf0cd12b1adae939a883f2254c78dc4b8c9d0098581337b1db
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d2d8b59c1f39c41951e7878963f8496eb494688720366f8edc7060887cdbe6a0
        - quay.io/mongodb/mongodb-agent@sha256:2d5652d0e263aa94554e7a068c93e4fd94b84ec223f077361d030d172bc93f67
        - quay.io/mongodb/mongodb-agent@sha256:b422829aac63f068aed4a6948fd14cbd3c3bfad6d47ab76523f6839bf20b0203
        - quay.io/mongodb/mongodb-enterprise-server@sha256:bfbcfcfc5f0de22e20520b2d96f826638e424adb028a909ccc192013d11f5833
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:cd869ff7d661059fad7593e30141d74e8cc751295b1579577f7a853396eb5d0b
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:702f529cb9d8d9042e54263f8bc37b40c062ac2695efc3e7150819779886e323
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f1e72679b2468aed2f3c3106568327911e6594c9ee15f7348e953eaa0d8fd44a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:a48a4d5c275fd669a27469649fb39ea38c728dbcf9f0a9da4e7cf27a2ce912ed
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:d81eac2129b6ab8346be1749c56ea67f6da64d4ee0fb9f3ff4171ba0f61918eb
        - quay.io/mongodb/mongodb-enterprise-server@sha256:67238129df09d33ee4746ff1acb5515d699a29d9f42350eae3f3f69ccb7fbf54
        - quay.io/mongodb/mongodb-kubernetes@sha256:745cc2c6bb893c1cb1e75d79b6f3bfc7b125a6a62f7431bdc43cf6cacd373998
        - quay.io/mongodb/mongodb-enterprise-server@sha256:9a3a6496be0d2d446e6d654eb56386c7726e0cd2d7232f6dcec579923edf0014
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:a4590af027182305913f0de1a2e37ad8612bd73b8b31f4df11068b80089021ff
        - quay.io/mongodb/mongodb-kubernetes-init-database@sha256:3ea9106c0e35c015bc8f15657c76d974bb1cadb9220dc58be2076c9699f26fb7
        - quay.io/mongodb/mongodb-kubernetes-init-appdb@sha256:afb0d463304fc0bfb3d8dade91459414daac6bf6fcf742347a2012cbc9fea354
        - quay.io/mongodb/mongodb-enterprise-server@sha256:e7aeedb752dda77cf0e49df3430843b5a2bba64c8404628a983f46640d316808
        - quay.io/mongodb/mongodb-enterprise-server@sha256:0fde236fe26d5b20210d6f246aadacd7845874712def516cbe2eec11d79d5181
        - quay.io/mongodb/mongodb-enterprise-server@sha256:1d418c65d82ccad205b7889be4ba161d3cbfa8b1fb75b21c8fcf02de38fad9ca
        - quay.io/mongodb/mongodb-enterprise-server@sha256:655896bac9e82de8e9729d59a36c3a31f66235f22b93449260a2ccb7709de084
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3316015ac0699d66ac73f714e5dc0784738654e0349e80b1e1f4e953b3c88964
        - quay.io/mongodb/mongodb-enterprise-server@sha256:3fe62f3ad52707ccf0e9993c69891e44c569a4b4e1a3244cf60e3b5fdde57bd2
        - quay.io/mongodb/mongodb-enterprise-server@sha256:778f9e29e36d6687cd17dd95ec1f8127d2a732bde1a2dd5821212eb55c1c998a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d323c1bcd14118cc537ac282996ea5a58756d20b6b73f1eec0efa8710d246077
        - quay.io/mongodb/mongodb-agent@sha256:d8a2ef79d767f03162ec57fe028f9072294b292d376af8eb97d2967eb6d7047b
        - quay.io/mongodb/mongodb-enterprise-server@sha256:df6081325ea2fe01d1b27d0c70005dec8232aaf51a43d91661b722b8a1761263
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5c8553e5ba115657fb878c0fe08704af9fd5decec572a87d89d2c8c156ffc9f9
        - quay.io/mongodb/mongodb-agent@sha256:fa30370a75e8e4ac648650ceb39a7a259a754aab8a48d7e2f2801cd1b777e91b
        - quay.io/mongodb/mongodb-agent@sha256:6e826bbe7dd812c101464e75295370c4b10e4e9c31477db89cfceb96759853e9
        - quay.io/mongodb/mongodb-enterprise-server@sha256:33a30d11969eb21952681108092fe7f916e25c8e1de43e6a4e40e126025a2bec
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b7764c316eeacffd3f9ddae62d9fd2a74eb58484d2cfd5021e460095d604764f
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:33bbd29b74f95baad522b380d70bcc9f933a3fee57642c37b09376c86179c40f
        - quay.io/mongodb/mongodb-agent@sha256:6394391f78da5e3218bb2c694e7ec9f6682fd17a6dd7e7250b0c0b5e1729796d
        - quay.io/mongodb/mongodb-enterprise-server@sha256:8d7e5361c2a1604dec3fc671d629806fe6ddc6df524b69a441b86933a4b30093
        - quay.io/mongodb/mongodb-kubernetes-database@sha256:f8c981a6223f70d7c6bbd30c503fc8b2497f351830b2b74c30e6f81863d78d2f
        - quay.io/mongodb/mongodb-agent@sha256:89819704d70a3b6b4d2358745a22060ef740b33ad5c6debf8e6ac4d4a6fd63ba
        - quay.io/mongodb/mongodb-kubernetes-init-ops-manager@sha256:93c4633cbc84e9bfa3563771ac2efbf5e6515f380dec2c085494cb8326546377
        - quay.io/mongodb/mongodb-agent@sha256:5b064c470335df840b050b714b28b75b6061444843250f1eb9215cb2b8c1a149
        - quay.io/mongodb/mongodb-agent@sha256:ee9f3906d893004bb2aaaa6412646f58288b0684502b9e90f33cafadcfce4349
        - quay.io/mongodb/mongodb-enterprise-server@sha256:19925e1297ab652db59c35f53e7b7c8b2b277f8ecdc4215682961aaf2b3e924e
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:fa4aa97767bd19cf3711a74bf21ecb9cbbc25fed6b6b09a2eb241ebd3fbc8f3a
        - quay.io/mongodb/mongodb-enterprise-server@sha256:172e338a43df01ee5408d8d1e3cb038d2c55e77761c1a2ac6ed37959d1525140
        - quay.io/mongodb/mongodb-enterprise-server@sha256:37af0913742b6de339c7e664b30f35dcef04506529e4a09a9bbbdd57643f7149
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:85b4e954c9c4bbe1019774bad1e05e9c8dc175648b01d3badf0b64143f182019
        - quay.io/mongodb/mongodb-enterprise-server@sha256:5145c97d2281aabcd8538ff515c3833aa48acb66b0b739fb1881120c1985d88a
        - quay.io/mongodb/mongodb-agent@sha256:a8358ce465adf463d8e73916d7944765d7aaf2b0ed10bd3b8119a87fc9080667
        - quay.io/mongodb/mongodb-enterprise-server@sha256:d21c65780cadc9d8668d2c28935d492625d6993f13ba562722200a819281cbcc
        - quay.io/mongodb/mongodb-agent@sha256:b387b7458f5640f907c9074fcb08ae4d4785be31d4bb65f2d7c459cd6762b2d6
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:3bbd3d7fd45d107621e7770f2b4c2a758e148cd472e0ee21719442f71bf6ade0
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:895dbdfb87ddc2e9693671c7108a31ab28e16fec2e5be6d046bdd2e30319eade
        - quay.io/mongodb/mongodb-agent@sha256:af3de4f631d0b00cae7f7e80c3900378ba7e449e01830bf9dc643b873d2e6d33
        - quay.io/mongodb/mongodb-enterprise-server@sha256:7a93a0276531ff9be4c90bb8fe8d104e0a9e930c29792aafe03cc6a76a9fa89c
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:b9ab7147b59b62e40a69c6400f87d0c7faecfa5462bed6e68ca653b00e50b887
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5e3982493e85ac1e40536758c3145b7c2fff9a7b4be9d053ee41201363a89920
        - quay.io/mongodb/mongodb-enterprise-ops-manager-ubi@sha256:5d299820bb903f0cdfdf2626a3c44099d2eaee16188cd8234d2ff3a57259b3a7
        - quay.io/mongodb/mongodb-enterprise-server@sha256:f2d9a3264a8814f33143b73c56624572fb80afeab184133d3a901713a31c1a87
        - quay.io/mongodb/mongodb-enterprise-server@sha256:b00ac773623e1bf422b47e6d783263d2d08737865abe9abbe106d86d1ea25165
        - quay.io/mongodb/mongodb-enterprise-server@sha256:ef732e3eea91cc86caf2eabedc0c9115ed235a8d204f33e74bbf090a0b0259fb
        version: 1.4.0
      entries:
      - name: mongodb-kubernetes.v1.4.0
        version: 1.4.0
      - name: mongodb-kubernetes.v1.3.0
        version: 1.3.0
      - name: mongodb-kubernetes.v1.2.0
        version: 1.2.0
      - name: mongodb-kubernetes.v1.1.0
        version: 1.1.0
      - name: mongodb-kubernetes.v1.0.1
        version: 1.0.1
      - name: mongodb-kubernetes.v1.0.0
        version: 1.0.0
      name: stable
    defaultChannel: stable
    packageName: mongodb-kubernetes
    provider:
      name: MongoDB, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sqs-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sqs-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sqs.services.k8s.aws/v1alpha1",
                "kind": "Queue",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
          createdAt: "2025-09-19T19:14:25Z"
          description: AWS SQS controller is a service controller for managing SQS
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Queue represents the state of an AWS sqs Queue resource.
            displayName: Queue
            kind: Queue
            name: queues.sqs.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Simple Queue Service (SQS) resources in AWS from within your Kubernetes cluster.

          **About Amazon SQS**

          Amazon Simple Queue Service (Amazon SQS) offers a secure, durable, and available hosted queue that lets you integrate and decouple distributed software systems and components. Amazon SQS offers common constructs such as [dead-letter queues](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-dead-letter-queues.html) and [cost allocation tags](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/sqs-queue-tags.html). It provides a generic web services API that you can access using any programming language that the AWS SDK supports.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SQS
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sqs
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SQS Developer Resources
          url: https://aws.amazon.com/sqs/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: sqs maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sqs-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-sqs-controller@sha256:a22009a5d2037aacf52d223e45b74e6aa08020d4cfd1836ff137254b73e19bb9
        version: 1.2.1
      entries:
      - name: ack-sqs-controller.v1.2.1
        version: 1.2.1
      - name: ack-sqs-controller.v1.2.0
        version: 1.2.0
      - name: ack-sqs-controller.v1.1.16
        version: 1.1.16
      - name: ack-sqs-controller.v1.1.14
        version: 1.1.14
      - name: ack-sqs-controller.v1.1.13
        version: 1.1.13
      - name: ack-sqs-controller.v1.1.12
        version: 1.1.12
      - name: ack-sqs-controller.v1.1.11
        version: 1.1.11
      - name: ack-sqs-controller.v1.1.10
        version: 1.1.10
      - name: ack-sqs-controller.v1.1.9
        version: 1.1.9
      - name: ack-sqs-controller.v1.1.8
        version: 1.1.8
      - name: ack-sqs-controller.v1.1.7
        version: 1.1.7
      - name: ack-sqs-controller.v1.1.6
        version: 1.1.6
      - name: ack-sqs-controller.v1.1.5
        version: 1.1.5
      - name: ack-sqs-controller.v1.1.4
        version: 1.1.4
      - name: ack-sqs-controller.v1.1.3
        version: 1.1.3
      - name: ack-sqs-controller.v1.1.2
        version: 1.1.2
      - name: ack-sqs-controller.v1.1.1
        version: 1.1.1
      - name: ack-sqs-controller.v1.1.0
        version: 1.1.0
      - name: ack-sqs-controller.v1.0.17
        version: 1.0.17
      - name: ack-sqs-controller.v1.0.16
        version: 1.0.16
      - name: ack-sqs-controller.v1.0.15
        version: 1.0.15
      - name: ack-sqs-controller.v1.0.14
        version: 1.0.14
      - name: ack-sqs-controller.v1.0.13
        version: 1.0.13
      - name: ack-sqs-controller.v1.0.12
        version: 1.0.12
      - name: ack-sqs-controller.v1.0.10
        version: 1.0.10
      - name: ack-sqs-controller.v1.0.7
        version: 1.0.7
      - name: ack-sqs-controller.v1.0.5
        version: 1.0.5
      - name: ack-sqs-controller.v1.0.3
        version: 1.0.3
      - name: ack-sqs-controller.v1.0.2
        version: 1.0.2
      - name: ack-sqs-controller.v1.0.1
        version: 1.0.1
      - name: ack-sqs-controller.v1.0.0
        version: 1.0.0
      - name: ack-sqs-controller.v0.0.7
        version: 0.0.7
      - name: ack-sqs-controller.v0.0.6
        version: 0.0.6
      - name: ack-sqs-controller.v0.0.5
        version: 0.0.5
      - name: ack-sqs-controller.v0.0.4
        version: 0.0.4
      - name: ack-sqs-controller.v0.0.2
        version: 0.0.2
      - name: ack-sqs-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-sqs-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Managed Openshift Black Belt
      provider-url: ""
    name: ecr-secret-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ecr-secret-operator.v0.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.3.2
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.19.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.3.2
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:90f512ce90433a33b50c58613d9b7136e00f116c48e179c4a1be89d6a35f9abc
        version: 0.3.2
      entries:
      - name: ecr-secret-operator.v0.3.2
        version: 0.3.2
      name: v0.3
    - currentCSV: ecr-secret-operator.v0.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "Secret",
                "metadata": {
                  "name": "ecr-secret-sample",
                  "namespace": "test-ecr-secret-operator"
                },
                "spec": {
                  "ecr_registry": "ACCOUNT_ID.dkr.ecr.us-east-2.amazonaws.com",
                  "frequency": "10h",
                  "generated_secret_name": "ecr-docker-secret",
                  "region": "us-east-2"
                }
              },
              {
                "apiVersion": "ecr.mobb.redhat.com/v1alpha1",
                "kind": "ArgoHelmRepoSecret",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "ecr-secret-operator",
                    "app.kubernetes.io/instance": "argohelmreposecret-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "argohelmreposecret",
                    "app.kubernetes.io/part-of": "ecr-secret-operator"
                  },
                  "name": "argohelmreposecret-sample"
                },
                "spec": null
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/mobb/ecr-secret-operator:v0.4.1
          createdAt: "2023-04-14T14:52:13Z"
          operatorframework.io/suggested-namespace: ecr-secret-operator
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/rh-mobb/ecr-secret-operator
          support: Managed Openshift Black Belt
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ArgoHelmRepoSecret is the Schema for the argohelmreposecrets
              API
            displayName: Argo Helm Repo Secret
            kind: ArgoHelmRepoSecret
            name: argohelmreposecrets.ecr.mobb.redhat.com
            version: v1alpha1
          - description: Secret is the Schema for the secrets API
            displayName: Secret
            kind: Secret
            name: secrets.ecr.mobb.redhat.com
            version: v1alpha1
        description: "# ECR Secret Operator\n\nAmazon Elastic Container Registry [Private
          Registry Authentication](https://docs.aws.amazon.com/AmazonECR/latest/userguide/registry_auth.html)
          provides a temporary token that is valid only for 12 hours. It is a challenge
          for automatic container image build process to refresh the token or secret
          in a timely manner.\n\nThis operators frequently talks with AWS ECR GetAuthroization
          Token and create/update the secret, so that the service account can perform
          docker image build.\n\n\n## How to use this operator\n\n### Prerequisites\n\n*
          [Create an ECR private repository](https://docs.aws.amazon.com/AmazonECR/latest/userguide/repository-create.html)\n*
          Create An Openshift Cluster\n* Provide AWS Authentication to the operator.
          Two Options:\n  **Note: You have to provide a secret aws-ecr-cloud-credentials
          in the installed namespace**\n  * [IAM User](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_user.md)\n
          \ * [STS Assume Role](https://github.com/rh-mobb/ecr-secret-operator/tree/main/docs/iam_assume_role.md)\n\n###
          Install the operator\n\n### Create the ECR Secret CRD\n\n```\napiVersion:
          ecr.mobb.redhat.com/v1alpha1\nkind: Secret\nmetadata:\n  name: ecr-secret\n
          \ namespace: test-ecr-secret-operator\nspec:\n  generated_secret_name: ecr-docker-secret\n
          \ ecr_registry: [ACCOUNT_ID].dkr.ecr.us-east-2.amazonaws.com\n  frequency:
          10h\n  region: us-east-2\n```\n\n```\noc create -f CRD_FILE\n```\n\nA docker
          registry secret is created by the operator momentally and the token is patched
          every 10 hours\n\n```\noc get secret ecr-docker-secret   \nNAME                TYPE
          \                            DATA   AGE\necr-docker-secret   kubernetes.io/dockerconfigjson
          \  1      16h\n```\n\n### A sample build process with generated secret\n\n\nLink
          the secret to builder\n\n```\noc secrets link builder ecr-docker-secret
          \n```\n\nConfigure [build config](https://github.com/rh-mobb/ecr-secret-operator/tree/main/samples/build-config.yaml)
          to point to your ECR Container repository\n\n```\noc create imagestream
          ruby\noc tag openshift/ruby:2.5-ubi8 ruby:2.5\noc create -f samples/build-config.yaml\noc
          start-build ruby-sample-build --wait\n```\n\nBuild should succeed and push
          the image to the the private ECR Container repository\n\n"
        displayName: ECR Secret Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - redhat
        links:
        - name: Ecr Secret Operator
          url: https://github.com/rh-mobb/ecr-secret-operator
        maintainers:
        - email: shading@redhat.com
          name: Shaozhen Ding
        maturity: alpha
        provider:
          name: Managed Openshift Black Belt
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - quay.io/mobb/ecr-secret-operator:v0.4.1
        - quay.io/openshift-community-operators/ecr-secret-operator@sha256:53e73c53c31e0e06c990d3d966c2b4053d559a58a17a781155a3ceff3efac831
        version: 0.4.1
      entries:
      - name: ecr-secret-operator.v0.4.1
        version: 0.4.1
      - name: ecr-secret-operator.v0.4.0
        version: 0.4.0
      name: v0.4
    defaultChannel: v0.4
    packageName: ecr-secret-operator
    provider:
      name: Managed Openshift Black Belt
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat Community of Practice
      provider-url: ""
    name: namespace-configuration-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: namespace-configuration-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "GroupConfig",
                "metadata": {
                  "name": "test-groupconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "team"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-dev\n  labels:\n    group: {{ .Name }}  \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-qa\n  labels:\n    group: {{ .Name }}          \n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-prod\n  labels:\n    group: {{ .Name }}                           \n"
                    },
                    {
                      "objectTemplate": "apiVersion: quota.openshift.io/v1\nkind: ClusterResourceQuota\nmetadata:\n  name: {{ .Name }}-quota\nspec:\n  quota:\n    hard:\n      pods: \"4\" \n      requests.cpu: \"1\" \n      requests.memory: 1Gi \n      requests.ephemeral-storage: 2Gi \n      limits.cpu: \"2\" \n      limits.memory: 2Gi \n      limits.ephemeral-storage: 4Gi\n  selector:\n    labels:\n      matchLabels:\n        group: {{ .Name }}\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "NamespaceConfig",
                "metadata": {
                  "name": "test-namespaceconfig"
                },
                "spec": {
                  "labelSelector": {
                    "matchLabels": {
                      "type": "secure"
                    }
                  },
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    },
                    {
                      "objectTemplate": "kind: EgressNetworkPolicy\napiVersion: network.openshift.io/v1\nmetadata:\n  name: air-gapped\n  namespace: {{ .Name }}  \nspec:\n  egress:\n  - type: Deny\n    to:\n      cidrSelector: 0.0.0.0/0\n"
                    }
                  ]
                }
              },
              {
                "apiVersion": "redhatcop.redhat.io/v1alpha1",
                "kind": "UserConfig",
                "metadata": {
                  "name": "test-user-config"
                },
                "spec": {
                  "providerName": "my-provider",
                  "templates": [
                    {
                      "objectTemplate": "apiVersion: v1\nkind: Namespace\nmetadata:\n  name: {{ .Name }}-sandbox\n"
                    },
                    {
                      "objectTemplate": "apiVersion: v1\nkind: ResourceQuota\nmetadata:\n  name: standard-sandbox\n  namespace: {{ .Name }}-sandbox\nspec:\n  hard:\n    pods: \"4\" \n    requests.cpu: \"1\" \n    requests.memory: 1Gi \n    requests.ephemeral-storage: 2Gi \n    limits.cpu: \"2\" \n    limits.memory: 2Gi \n    limits.ephemeral-storage: 4Gi\n"
                    }
                  ]
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Security
          certified: "false"
          containerImage: quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
          createdAt: "2023-11-22T15:58:50Z"
          description: This operator provides a facility to define and enforce namespace
            configurations
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: namespace-configuration-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/redhat-cop/namespace-configuration-operator
          support: Best Effort
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: GroupConfig is the Schema for the groupconfigs API
            displayName: Group Config
            kind: GroupConfig
            name: groupconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: NamespaceConfig is the Schema for the namespaceconfigs API
            displayName: Namespace Config
            kind: NamespaceConfig
            name: namespaceconfigs.redhatcop.redhat.io
            version: v1alpha1
          - description: UserConfig is the Schema for the userconfigs API
            displayName: User Config
            kind: UserConfig
            name: userconfigs.redhatcop.redhat.io
            version: v1alpha1
        description: "The `namespace-configuration-operator` helps keeping configurations
          related to Users, Groups and Namespaces aligned with one of more policies
          specified as a CRs. The purpose is to provide the foundational building
          block to create an end-to-end onboarding process. \nBy onboarding process
          we mean all the provisioning steps needed to a developer team working on
          one or more applications to OpenShift.\n This usually involves configuring
          resources such as: Groups, RoleBindings, Namespaces, ResourceQuotas, NetworkPolicies,
          EgressNetworkPolicies, etc.... . Depending on the specific environment the
          list could continue.\nNaturally such a process should be as automatic and
          scalable as possible.\n\nWith the namespace-configuration-operator one can
          create rules that will react to the creation of Users, Groups and Namespace
          and will create and enforce a set of resources.\n\nHere are some examples
          of the type of onboarding processes that one could support:\n\n1. [developer
          sandbox](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/user-sandbox/readme.md)\n
          2. [team onboarding](https://github.com/redhat-cop/namespace-configuration-operator//blob/master/examples/team-onboarding/readme.md)
          with support of the entire SDLC in a multitentant environment.\n\nPolicies
          can be expressed with the following CRDs:\n\n| Watched Resource | CRD |\n|--|--|\n
          | Groups | [GroupConfig](#GroupConfig) |\n| Users | [UserConfig](#UserConfig)
          |\n| Namespace | [NamespaceConfig](#NamespaceConfig) |\n\nThese CRDs all
          share some commonalities:\n\n1. Templated Resources\n2. List of ignored
          jason path\n \n### Templated Resources\n\nEach has a parameter called `templatedResources`,
          which is an array. Each element of the array has two fields `objectTemplate`
          and `excludedPaths` (see below).\n\nThe `objectTemplate` field must contains
          a [go template](https://golang.org/pkg/text/template/) that resolves to
          a single API Resource expressed in `yaml`. The template is merged with the
          object selected by the CR. For example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n
          kind: UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  provider:
          corp-ldap\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n```\n\nThis
          create a rule in which every time a user from the `corp-ldap` provider is
          created, a namespace called `<username>-sandbox` is also created.\n\n###
          Excluded Paths\n\nThe logic of the `namespace-configuration-operator` is
          to enforce that the resources resolved by processing the templates \"stays
          in place\". In other words if those resources are changed and/or deleted
          they will be reset by the operator.\nBut there are situations in which at
          least part of a resource is allowed to change. Common use cases are: annotations
          and in general the metadata section of a resource can be updated by the
          various operators watching that resources. The status field is often updated
          by the main operator managing that resources. Finally, when applicable the
          `spec.replicas` field should also be allowed to change.\n\nTo handle special
          use case, one can also specify additional *jsonpaths* that should be ignored
          when comparing the desired resource and the current resource and making
          a decision on whether that resource should be reset.\n\nThe following paths
          are always included:\n \n1. `.metadata`\n2. `.status`\n3. `.spec.replicas`\n\n##
          NamespaceConfig\n\n The `NamespaceConfig` CR allows specifying one or more
          objects that will be created in the selected namespaces.\n\nNamespaces can
          be selected by labels or annotations via a label selector for example:\n\n```yaml\napiVersion:
          redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n  name:
          small-namespace\n  selector:\n    matchLabels:\n      size: small  \n  resources:\n
          \ - apiVersion: v1\n    kind: ResourceQuota\n    metadata:\n      name:
          small-size  \n    spec:\n      hard:\n        requests.cpu: \"4\"\n        requests.memory:
          \"2Gi\"\n```\n\nHere is a `NamespaceConfig` object using a `matchExpressions`
          selector:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\n kind: NamespaceConfig\nmetadata:\n
          \ name: tier-config\nspec:\n  annotationSelector:\n    matchExpressions:\n
          \     - {key: tier, operator: In, values: [gold,silver]}\n ```\n\nAlthough
          not enforced by the operator the general expectation is that the NamespaceConfig
          CR will be used to create objects inside the selected namespace.\n \nExamples
          of NamespaceConfig usages can be found [here](https://github.com/redhat-cop/namespace-configuration-operator/blob/master/examples/namespace-config/readme.md)\n
          \n## GroupConfig\n\nThe `GroupConfig` CR allows specifying one or more objects
          that will be created in the selected Group.\nGroups can be selected by labels
          or annotations via a label selector, similarly to the `NamespaceConfig`.\n\nOften
          groups are created in OpenShift by a job that synchronizes an Identity Provider
          with OCP. So the idea is that when new groups are added or deleted the configuration
          in OpenShift will adapt automatically.\n\nAlthough not enforced by the operator,
          GroupConfig are expected to create cluster-scoped resources like Namespaces,
          ClusterResourceQuotas and potentially some namespaced resources like RoleBindings.\n\n##
          UserConfig\n \nIn OpenShift an external user is defined by two entities:
          Users and Identities. There is a relationship of on to many between Users
          and Identities. Given one user, there can be one Identity per authentication
          mechanism.\n\nThe `UserConfig` CR allows specifying one or more objects
          that will be created in the selected User.\nUsers can be selected by label
          or annotation like `NamespaceConfig` and `UserConfig`.\nUSers can also be
          selected by provider name (the name of the authentication mechanism) and
          identity extra field.\n\nHere is an example:\n\n```yaml\napiVersion: redhatcop.redhat.io/v1alpha1\nkind:
          UserConfig\nmetadata:\n  name: test-user-config\nspec:\n  providerName:
          okta-provider\n  identityExtraFieldSelector:\n    matchLabels:\n      sandbox_enabled:
          \"true\"\n  templates:\n  - objectTemplate: |\n      apiVersion: v1\n      kind:
          Namespace\n      metadata:\n        name: {{ .Name }}-sandbox\n ```\n\nUser
          will be selected by this `UserConfig` only if they login via the *okta-provider*
          and if the extra field was populate with the label `sandbox_enabled: \"true\"
          `. Note that not all authentication provider allow populating the extra
          fields in the Identity object.\n"
        displayName: Namespace Configuration Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - namespace
        - user
        - group
        - configuration
        - policy
        - management
        links:
        - name: Namespace Configuration Operator
          url: https://github.com/redhat-cop/namespace-configuration-operator
        - name: Container Image
          url: https://quay.io/redhat-cop/namespace-configuration-operator
        maintainers:
        - email: rspazzol@redhat.com
          name: Raffaele Spazzoli
        maturity: alpha
        provider:
          name: Red Hat Community of Practice
        relatedImages:
        - quay.io/openshift-community-operators/namespace-configuration-operator@sha256:971991d2bf717001a98b175c50580cd7db9d230492e72670606c2ed50c53feb0
        - quay.io/redhat-cop/kube-rbac-proxy@sha256:c68135620167c41e3d9f6c1d2ca1eb8fa24312b86186d09b8010656b9d25fb47
        - quay.io/redhat-cop/namespace-configuration-operator@sha256:49ed7d6155342adaa2b12fd80c6761c3081d8e6149d187cb7ff91a247cdf2e7a
        version: 1.2.6
      entries:
      - name: namespace-configuration-operator.v1.2.6
        version: 1.2.6
      - name: namespace-configuration-operator.v1.2.5
        version: 1.2.5
      - name: namespace-configuration-operator.v1.2.4
        version: 1.2.4
      - name: namespace-configuration-operator.v1.2.3
        version: 1.2.3
      - name: namespace-configuration-operator.v1.2.2
        version: 1.2.2
      - name: namespace-configuration-operator.v1.2.1
        version: 1.2.1
      - name: namespace-configuration-operator.v1.2.0
        version: 1.2.0
      - name: namespace-configuration-operator.v1.1.0
        version: 1.1.0
      - name: namespace-configuration-operator.v1.0.7
        version: 1.0.7
      - name: namespace-configuration-operator.v1.0.6
        version: 1.0.6
      - name: namespace-configuration-operator.v1.0.5
        version: 1.0.5
      - name: namespace-configuration-operator.v1.0.4
        version: 1.0.4
      - name: namespace-configuration-operator.v1.0.3
        version: 1.0.3
      - name: namespace-configuration-operator.v1.0.2
        version: 1.0.2
      - name: namespace-configuration-operator.v1.0.1
        version: 1.0.1
      - name: namespace-configuration-operator.v1.0.0
        version: 1.0.0
      name: alpha
    defaultChannel: alpha
    packageName: namespace-configuration-operator
    provider:
      name: Red Hat Community of Practice
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecr-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecr-controller.v1.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ecr.services.k8s.aws/v1alpha1",
                "kind": "Repository",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
          createdAt: "2025-09-20T01:21:24Z"
          description: AWS ECR controller is a service controller for managing ECR
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PullThroughCacheRule represents the state of an AWS ecr PullThroughCacheRule
              resource.
            displayName: PullThroughCacheRule
            kind: PullThroughCacheRule
            name: pullthroughcacherules.ecr.services.k8s.aws
            version: v1alpha1
          - description: Repository represents the state of an AWS ecr Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecr.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Elastic Container Registry (ECR) resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR**

          Amazon Elastic Container Registry (Amazon ECR) is an AWS managed container image registry service that is secure, scalable, and reliable. Amazon ECR supports private repositories with resource-based permissions using AWS IAM. This is so that specified users or Amazon EC2 instances can access your container repositories and images. You can use your preferred CLI to push, pull, and manage Docker images, Open Container Initiative (OCI) images, and OCI compatible artifacts.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecr
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR Developer Resources
          url: https://aws.amazon.com/ecr/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecr maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-ecr-controller@sha256:03ca7f898cfbf14f18a06cf3b510cddade3863275df934be891734b8c177c1e6
        - public.ecr.aws/aws-controllers-k8s/ecr-controller:1.3.1
        version: 1.3.1
      entries:
      - name: ack-ecr-controller.v1.3.1
        version: 1.3.1
      - name: ack-ecr-controller.v1.3.0
        version: 1.3.0
      - name: ack-ecr-controller.v1.2.0
        version: 1.2.0
      - name: ack-ecr-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecr-controller.v1.0.33
        version: 1.0.33
      - name: ack-ecr-controller.v1.0.32
        version: 1.0.32
      - name: ack-ecr-controller.v1.0.31
        version: 1.0.31
      - name: ack-ecr-controller.v1.0.30
        version: 1.0.30
      - name: ack-ecr-controller.v1.0.29
        version: 1.0.29
      - name: ack-ecr-controller.v1.0.28
        version: 1.0.28
      - name: ack-ecr-controller.v1.0.27
        version: 1.0.27
      - name: ack-ecr-controller.v1.0.25
        version: 1.0.25
      - name: ack-ecr-controller.v1.0.24
        version: 1.0.24
      - name: ack-ecr-controller.v1.0.23
        version: 1.0.23
      - name: ack-ecr-controller.v1.0.22
        version: 1.0.22
      - name: ack-ecr-controller.v1.0.21
        version: 1.0.21
      - name: ack-ecr-controller.v1.0.20
        version: 1.0.20
      - name: ack-ecr-controller.v1.0.19
        version: 1.0.19
      - name: ack-ecr-controller.v1.0.17
        version: 1.0.17
      - name: ack-ecr-controller.v1.0.16
        version: 1.0.16
      - name: ack-ecr-controller.v1.0.15
        version: 1.0.15
      - name: ack-ecr-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecr-controller.v1.0.12
        version: 1.0.12
      - name: ack-ecr-controller.v1.0.10
        version: 1.0.10
      - name: ack-ecr-controller.v1.0.9
        version: 1.0.9
      - name: ack-ecr-controller.v1.0.8
        version: 1.0.8
      - name: ack-ecr-controller.v1.0.7
        version: 1.0.7
      - name: ack-ecr-controller.v1.0.6
        version: 1.0.6
      - name: ack-ecr-controller.v1.0.4
        version: 1.0.4
      - name: ack-ecr-controller.v1.0.3
        version: 1.0.3
      - name: ack-ecr-controller.v1.0.2
        version: 1.0.2
      - name: ack-ecr-controller.v1.0.0
        version: 1.0.0
      - name: ack-ecr-controller.v0.1.9
        version: 0.1.9
      - name: ack-ecr-controller.v0.1.7
        version: 0.1.7
      - name: ack-ecr-controller.v0.1.6
        version: 0.1.6
      - name: ack-ecr-controller.v0.1.5
        version: 0.1.5
      - name: ack-ecr-controller.v0.1.4
        version: 0.1.4
      - name: ack-ecr-controller.v0.1.3
        version: 0.1.3
      - name: ack-ecr-controller.v0.1.2
        version: 0.1.2
      - name: ack-ecr-controller.v0.1.1
        version: 0.1.1
      - name: ack-ecr-controller.v0.1.0
        version: 0.1.0
      - name: ack-ecr-controller.v0.0.27
        version: 0.0.27
      - name: ack-ecr-controller.v0.0.26
        version: 0.0.26
      - name: ack-ecr-controller.v0.0.25
        version: 0.0.25
      - name: ack-ecr-controller.v0.0.24
        version: 0.0.24
      - name: ack-ecr-controller.v0.0.23
        version: 0.0.23
      - name: ack-ecr-controller.v0.0.22
        version: 0.0.22
      - name: ack-ecr-controller.v0.0.21
        version: 0.0.21
      - name: ack-ecr-controller.v0.0.20
        version: 0.0.20
      - name: ack-ecr-controller.v0.0.19
        version: 0.0.19
      - name: ack-ecr-controller.v0.0.18
        version: 0.0.18
      - name: ack-ecr-controller.v0.0.17
        version: 0.0.17
      - name: ack-ecr-controller.v0.0.16
        version: 0.0.16
      - name: ack-ecr-controller.v0.0.15
        version: 0.0.15
      - name: ack-ecr-controller.v0.0.14
        version: 0.0.14
      - name: ack-ecr-controller.v0.0.13
        version: 0.0.13
      - name: ack-ecr-controller.v0.0.12
        version: 0.0.12
      - name: ack-ecr-controller.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecr-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: eventing-kogito
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eventing-kogito.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kogito.knative.dev/v1alpha1",
                "kind": "KogitoSource",
                "metadata": {
                  "name": "kogito-order-processing"
                },
                "spec": {
                  "sink": {
                    "ref": {
                      "apiVersion": "eventing.knative.dev/v1",
                      "kind": "Broker",
                      "name": "default"
                    }
                  },
                  "subject": {
                    "apiVersion": "serving.knative.dev/v1",
                    "kind": "Service",
                    "name": "serverless-workflow-order-processing"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
          createdAt: "2021-08-25T12:00:00Z"
          olm.targetNamespaces: knative-kogito
          operators.operatorframework.io/builder: operator-sdk-v1.16.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/knative-sandbox/eventing-kogito
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoSource is the representation of a Knative Eventing
              Source for a Kogito custom Service.
            displayName: Kogito Source
            kind: KogitoSource
            name: kogitosources.kogito.knative.dev
            version: v1alpha1
        description: |-
          Knative Eventing Source for Kogito Services. This controller will enable easy configuration of any Kogito custom Service
          with the Knative Eventing platform. Please find more information about using this source in the [Official Kogito Documentation](https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on)

          **Note:** The admission webhooks are not installed via OLM at this time. Because of that, please ensure that you create the `KogitoSource` objects with the subbject and sink information with the appropriate namespace.
        displayName: Kogito Knative Eventing Source
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless workflow
        - kogito
        - knative
        - dmn
        - bpmn
        - eventing
        - cloudevents
        - messaging
        links:
        - name: Source Repository
          url: https://github.com/knative-sandbox/eventing-kogito
        - name: Kogito Knative Add-On Documentation
          url: https://docs.jboss.org/kogito/release/latest/html_single/#_kogito_knative_eventing_add_on
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: Red Hat
        relatedImages:
        - gcr.io/knative-releases/knative.dev/eventing-kogito/cmd/controller:v1.2.0
        - quay.io/openshift-community-operators/eventing-kogito@sha256:abe8d5803d2d96ebbb9fe4cda0feaa0217e92def6a669404362bcf741d934690
        version: 1.2.0
      entries:
      - name: eventing-kogito.v1.2.0
        version: 1.2.0
      - name: eventing-kogito.v1.1.0
        version: 1.1.0
      name: alpha
    defaultChannel: alpha
    packageName: eventing-kogito
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kepler Operator Contributors
      provider-url: https://sustainable-computing.io/
    name: kepler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kepler-operator.v0.21.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kepler.system.sustainable.computing.io/v1alpha1",
                "kind": "PowerMonitor",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/instance": "powermonitor",
                    "app.kubernetes.io/name": "powermonitor",
                    "app.kubernetes.io/part-of": "kepler-operator"
                  },
                  "name": "power-monitor"
                },
                "spec": {
                  "kepler": {
                    "config": {
                      "logLevel": "info"
                    },
                    "deployment": {
                      "security": {
                        "mode": "none"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Monitoring
          containerImage: quay.io/sustainable_computing_io/kepler-operator:0.21.0
          createdAt: "2025-09-15T11:39:08Z"
          description: Deploys and Manages Kepler on Kubernetes
          operators.operatorframework.io/builder: operator-sdk-v1.39.1
          operators.operatorframework.io/internal-objects: |-
            [
              "powermonitorinternals.kepler.system.sustainable.computing.io"
            ]
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/sustainable-computing-io/kepler-operator
          support: https://github.com/sustainable-computing-io/kepler-operator/issues
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PowerMonitorInternal is the Schema for internal/unsupported
              API
            displayName: PowerMonitorInternal
            kind: PowerMonitorInternal
            name: powermonitorinternals.kepler.system.sustainable.computing.io
            version: v1alpha1
          - description: PowerMonitor is the Schema for the powermonitors API
            displayName: PowerMonitor
            kind: PowerMonitor
            name: powermonitors.kepler.system.sustainable.computing.io
            version: v1alpha1
        description: |-
          Kepler (Kubernetes-based Efficient Power Level Exporter) is a Prometheus exporter that
          measures energy consumption metrics at the container, pod, and node level in Kubernetes clusters.

          ### Prerequisites
          To view the metrics exposed by Kepler, **[User Workload Monitoring](https://docs.redhat.com/documentation/openshift_container_platform/latest/html/monitoring/configuring-user-workload-monitoring)**
          MUST be enabled in your cluster.

          ### Documentation
          * [User Documentation](https://github.com/sustainable-computing-io/kepler-operator/blob/v1alpha1/README.md)
          * [Developer Guide](https://github.com/sustainable-computing-io/kepler-operator/tree/v1alpha1/docs/developer)

          ### Contributing
          * [Report Bugs](https://github.com/sustainable-computing-io/kepler-operator/issues/new?template=bug_report.md) - Found a bug? Let us know!
          * [Browse Issues](https://github.com/sustainable-computing-io/kepler-operator/issues) - See what we're working on

          ### License
          Licensed under [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)
        displayName: Kepler
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Kepler
        - power
        - energy
        - monitoring
        links:
        - name: Installation Guide
          url: https://sustainable-computing.io/installation/community-operator/
        - name: Kepler Project
          url: https://sustainable-computing.io/
        maintainers:
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: vprashar@redhat.com
          name: Vibhu Prashar
        - email: vimalkum@redhat.com
          name: Vimal Kumar
        - email: kaliu@redhat.com
          name: Kaiyi Liu
        maturity: alpha
        minKubeVersion: 1.24.0
        provider:
          name: Kepler Operator Contributors
          url: https://sustainable-computing.io/
        relatedImages:
        - quay.io/brancz/kube-rbac-proxy:v0.19.0
        - quay.io/community-operator-pipeline-prod/kepler-operator@sha256:37fe88e7234e5eb07c493a6412f27bb8df33896f7eab0a94b0e396fcdf870715
        - quay.io/sustainable_computing_io/kepler-operator:0.21.0
        - quay.io/sustainable_computing_io/kepler:v0.11.0
        version: 0.21.0
      entries:
      - name: kepler-operator.v0.21.0
        version: 0.21.0
      - name: kepler-operator.v0.20.0
        version: 0.20.0
      name: alpha
    defaultChannel: alpha
    packageName: kepler-operator
    provider:
      name: Kepler Operator Contributors
      url: https://sustainable-computing.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ian Boden
      provider-url: ""
    name: eclipse-amlen-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: eclipse-amlen-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "_amlen_image_tag": "latest",
                  "_amlen_monitor_image_tag": "latest",
                  "_amlen_name": "amlen-sample",
                  "_amlen_namespace": "osdk-test",
                  "_amlen_persistence": "false",
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.0.1a
          createdAt: "2022-11-24T15:31:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse/amlen
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: Operator for the eclipse Amlen MQTT Broker\n\n### Supported Feature\n\nCurrently
          supports:\n* Install - simple install process\n* Upgrade - non-disruptive
          upgrade\n*High Availability - can be enabled/disabled via the CRD\n\n###
          Documentation\nThe main Amlen website has links to product documentation,
          forum, blog and the slack channel\nhttp://www.eclipse.org/amlen\n\n### Getting
          help\nThe quickest way to get help is to use the slack channel but the forum
          is monitored as well\n\nIssues can also be added to the github repository
          https://github.com/eclipse/amlen\n\n### License\nEclipse Amlen Operator
          is licensed under the [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Ian Boden
        relatedImages:
        - quay.io/openshift-community-operators/eclipse-amlen-operator@sha256:bb89e2a7cb1c9172b080c72ad3227017d135fc30076d2a68b4e215fa8fa55846
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/amlen/operator:v1.0.1a
        version: 1.0.0
      entries:
      - name: eclipse-amlen-operator.v1.0.0
        version: 1.0.0
      - name: eclipse-amlen-operator.v0.0.1
        version: 0.0.1
      name: alpha
    - currentCSV: eclipse-amlen-operator.v1.1.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "eclipse.org/v1",
                "kind": "Amlen",
                "metadata": {
                  "name": "amlen-sample"
                },
                "spec": {
                  "config": "amlen-config",
                  "device_cert_issuer": {
                    "mode": "automatic",
                    "name": "amlen-selfsigned-issuer"
                  },
                  "ha": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "ldap": {
                    "cert_issuer": {
                      "name": "amlen-selfsigned-issuer"
                    },
                    "enabled": true
                  },
                  "size": 2
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Streaming & Messaging
          containerImage: quay.io/amlen/operator:v1.1.3
          createdAt: "2025-05-22T10:00:00Z"
          description: An operator to run the Eclipse Amlen MQTT Broker
          operators.operatorframework.io/builder: operator-sdk-v1.25.2
          operators.operatorframework.io/project_layout: ansible.sdk.operatorframework.io/v1
          repository: https://github.com/eclipse-amlen/amlen
          support: Eclipse
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Amlen
            name: amlens.eclipse.org
            version: v1
        description: |-
          Operator for the eclipse Amlen MQTT Broker

          ### Supported Feature

          Currently supports:
          * Install - simple install process
          * Upgrade - non-disruptive upgrade
          * High Availability - can be enabled/disabled via the CRD

          ### Documentation
          The main Amlen website has links to product documentation, forum, blog and the slack channel
          http://www.eclipse.org/amlen

          ### Getting help
          The quickest way to get help is to use the slack channel but the forum is monitored as well
          Issues can also be added to the github repository https://github.com/eclipse-amlen/amlen

          ### License
          Eclipse Amlen Operator is licensed under [Eclipse Public License - v 2.0](https://www.eclipse.org/legal/epl-2.0/)
        displayName: eclipse-amlen-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - amlen
        - mqtt
        - broker
        links:
        - name: Eclipse Amlen Operator
          url: https://eclipse.org/amlen
        maintainers:
        - email: ianboden@uk.ibm.com
          name: Ian Boden
        maturity: stable
        provider:
          name: Eclipse Amlen
          url: https://eclipse.org/amlen
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.11.0
        - quay.io/community-operator-pipeline-prod/eclipse-amlen-operator@sha256:5fbfb0bd4f764643cdc1021e99e489514a1f035d85131b13e56b11a3045883c2
        - controller:latest
        version: 1.1.3
      entries:
      - name: eclipse-amlen-operator.v1.1.3
        version: 1.1.3
      - name: eclipse-amlen-operator.v1.1.0
        version: 1.1.0
      - name: eclipse-amlen-operator.v1.0.2
        version: 1.0.2
      - name: eclipse-amlen-operator.v1.0.1
        version: 1.0.1
      name: stable
    defaultChannel: alpha
    packageName: eclipse-amlen-operator
    provider:
      name: Ian Boden
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Microcks.io
      provider-url: ""
    name: microcks
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: microcks-operator.v1.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall"
                },
                "spec": {
                  "name": "my-microcksinstall",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 1
                  },
                  "postman": {
                    "replicas": 1
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              },
              {
                "apiVersion": "microcks.github.io/v1alpha1",
                "kind": "MicrocksInstall",
                "metadata": {
                  "name": "my-microcksinstall-minikube"
                },
                "spec": {
                  "name": "my-microcksinstall-minikube",
                  "version": "1.10.0",
                  "microcks": {
                    "replicas": 2,
                    "url": "microcks.192.168.99.100.nip.io"
                  },
                  "postman": {
                    "replicas": 2
                  },
                  "keycloak": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "1Gi",
                    "url": "keycloak.192.168.99.100.nip.io"
                  },
                  "mongodb": {
                    "install": true,
                    "persistent": true,
                    "volumeSize": "2Gi"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "false"
          containerImage: quay.io/microcks/microcks-ansible-operator:1.10.0
          createdAt: "2024-08-01T09:14:55Z"
          description: Open Source mocking and testing platform for API and microservices
          repository: https://github.com/microcks/microcks-ansible-operator
          support: microcks.github.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents a Microcks installation
            displayName: MicrocksInstall
            kind: MicrocksInstall
            name: microcksinstalls.microcks.github.io
            version: v1alpha1
        description: |
          Microcks is an Open source tool for API and microservices mocking and testing. It provides a platform for referencing, deploying mocks and allow contract-testing of your APIs. It can also be considered as a Service Virtualization solution because it will allow you to provide fake API or Service implementation before development being actually done. It supports Async APIs, REST APIs, gRPC, GraphQL and SOAP WebServices. It perfectly integrates into an iterative, contract-first delivery process.
          Microcks tries not reinventing the wheel and let you capitalize on already known tools of standards. It supports [AsyncAPI Specifications](https://asyncapi.com), [OpenAPI Specification](https://www.openapis.org/) `3.x`, [gRPC Protobuf](https://grpc.io/), [GraphQL Schema](https://graphql.org) out-of-the-box and can also relies on popular tools like [SoapUI](http://www.soapui.com) or [Postman](http://www.getpostman.io) to edit your API request / response samples as well as your test scripts.
          ### Operator features
          * **Create/Destroy**: Easily launch Microcks installations (`MicrocksInstall` custom resource) for your Kubernetes namespace
          ### MicrocksInstall options
          * **keycloak** component can be deployed as part of the install or you can reuse an existing instance. You can reference an existing `Secret` for providing TLS certificates for securing the `Ingress`,
          * **mongodb** component can be deployed as part of the install or you can reuse an existing instance, You can reference an existing `Secret` for providing connection credentials,
          * **features** component allow to activate optional features like the mocking of AsyncAPIs. In case you turn `features.async.enabled` to `true` and `features.async.kafka.install` to `true`, you must have the [Strimzi Operator](https://operatorhub.io/operator/strimzi-kafka-operator) already installed on your cluster/namespace.
          ### Documentation
          Documentation to the current _master_ branch as well as all releases can be found on our [GitHub](https://github.com/microcks/microcks-ansible-operator).
          ### License
          Microcks is licensed under the [Apache License, Version 2.0](https://github.com/microcks/microcks/blob/master/LICENSE).
        displayName: Microcks Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mock
        - test
        - API
        - microservice
        links:
        - name: Product Page
          url: https://microcks.io
        - name: Microcks Operator Source Code
          url: https://github.com/microcks/microcks-ansible-operator
        maintainers:
        - email: laurent.broudoux@gmail.com
          name: lbroudoux
        maturity: alpha
        provider:
          name: Microcks.io
        relatedImages:
        - quay.io/community-operator-pipeline-prod/microcks@sha256:1313bb08037e57cde02d9d48c4d8c9663eed08651a3c29f872b5ef85ac764112
        - quay.io/microcks/microcks-ansible-operator:1.10.0
        version: 1.10.0
      entries:
      - name: microcks-operator.v1.10.0
        version: 1.10.0
      - name: microcks-operator.v1.9.0
        version: 1.9.0
      - name: microcks-operator.v1.8.1
        version: 1.8.1
      - name: microcks-operator.v1.8.0
        version: 1.8.0
      - name: microcks-operator.v1.7.1
        version: 1.7.1
      - name: microcks-operator.v1.7.0
        version: 1.7.0
      - name: microcks-operator.v1.6.1
        version: 1.6.1
      - name: microcks-operator.v1.6.0
        version: 1.6.0
      - name: microcks-operator.v1.5.2
        version: 1.5.2
      - name: microcks-operator.v1.5.1
        version: 1.5.1
      - name: microcks-operator.v1.5.0
        version: 1.5.0
      - name: microcks-operator.v1.4.1
        version: 1.4.1
      name: stable
    defaultChannel: stable
    packageName: microcks
    provider:
      name: Microcks.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/openshift/cluster-nfd-operator
    name: openshift-nfd-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfd.v4.10.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nfd.openshift.io/v1",
                "kind": "NodeFeatureDiscovery",
                "metadata": {
                  "name": "nfd-instance",
                  "namespace": "openshift-nfd"
                },
                "spec": {
                  "customConfig": {
                    "configData": "#    - name: \"more.kernel.features\"\n#      matchOn:\n#      - loadedKMod: [\"example_kmod3\"]\n#    - name: \"more.features.by.nodename\"\n#      value: customValue\n#      matchOn:\n#      - nodename: [\"special-.*-node-.*\"]\n"
                  },
                  "operand": {
                    "image": "quay.io/openshift/origin-node-feature-discovery:4.10",
                    "servicePort": 12000
                  },
                  "workerConfig": {
                    "configData": "core:\n#  labelWhiteList:\n#  noPublish: false\n  sleepInterval: 60s\n#  sources: [all]\n#  klog:\n#    addDirHeader: false\n#    alsologtostderr: false\n#    logBacktraceAt:\n#    logtostderr: true\n#    skipHeaders: false\n#    stderrthreshold: 2\n#    v: 0\n#    vmodule:\n##   NOTE: the following options are not dynamically run-time \n##          configurable and require a nfd-worker restart to take effect\n##          after being changed\n#    logDir:\n#    logFile:\n#    logFileMaxSize: 1800\n#    skipLogHeaders: false\nsources:\n#  cpu:\n#    cpuid:\n##     NOTE: whitelist has priority over blacklist\n#      attributeBlacklist:\n#        - \"BMI1\"\n#        - \"BMI2\"\n#        - \"CLMUL\"\n#        - \"CMOV\"\n#        - \"CX16\"\n#        - \"ERMS\"\n#        - \"F16C\"\n#        - \"HTT\"\n#        - \"LZCNT\"\n#        - \"MMX\"\n#        - \"MMXEXT\"\n#        - \"NX\"\n#        - \"POPCNT\"\n#        - \"RDRAND\"\n#        - \"RDSEED\"\n#        - \"RDTSCP\"\n#        - \"SGX\"\n#        - \"SSE\"\n#        - \"SSE2\"\n#        - \"SSE3\"\n#        - \"SSE4.1\"\n#        - \"SSE4.2\"\n#        - \"SSSE3\"\n#      attributeWhitelist:\n#  kernel:\n#    kconfigFile: \"/path/to/kconfig\"\n#    configOpts:\n#      - \"NO_HZ\"\n#      - \"X86\"\n#      - \"DMI\"\n  pci:\n    deviceClassWhitelist:\n      - \"0200\"\n      - \"03\"\n      - \"12\"\n    deviceLabelFields:\n#      - \"class\"\n      - \"vendor\"\n#      - \"device\"\n#      - \"subsystem_vendor\"\n#      - \"subsystem_device\"\n#  usb:\n#    deviceClassWhitelist:\n#      - \"0e\"\n#      - \"ef\"\n#      - \"fe\"\n#      - \"ff\"\n#    deviceLabelFields:\n#      - \"class\"\n#      - \"vendor\"\n#      - \"device\"\n#  custom:\n#    - name: \"my.kernel.feature\"\n#      matchOn:\n#        - loadedKMod: [\"example_kmod1\", \"example_kmod2\"]\n#    - name: \"my.pci.feature\"\n#      matchOn:\n#        - pciId:\n#            class: [\"0200\"]\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#        - pciId :\n#            vendor: [\"8086\"]\n#            device: [\"1000\", \"1100\"]\n#    - name: \"my.usb.feature\"\n#      matchOn:\n#        - usbId:\n#          class: [\"ff\"]\n#          vendor: [\"03e7\"]\n#          device: [\"2485\"]\n#        - usbId:\n#          class: [\"fe\"]\n#          vendor: [\"1a6e\"]\n#          device: [\"089a\"]\n#    - name: \"my.combined.feature\"\n#      matchOn:\n#        - pciId:\n#            vendor: [\"15b3\"]\n#            device: [\"1014\", \"1017\"]\n#          loadedKMod : [\"vendor_kmod1\", \"vendor_kmod2\"]\n"
                  }
                }
              },
              {
                "apiVersion": "nfd.openshift.io/v1alpha1",
                "kind": "NodeFeatureRule",
                "metadata": {
                  "name": "my-sample-rule-object"
                },
                "spec": {
                  "rules": [
                    {
                      "labels": {
                        "my-sample-feature": "true"
                      },
                      "matchFeatures": [
                        {
                          "feature": "kernel.loadedmodule",
                          "matchExpressions": {
                            "dummy": {
                              "op": "Exists"
                            }
                          }
                        },
                        {
                          "feature": "kernel.config",
                          "matchExpressions": {
                            "X86": {
                              "op": "In",
                              "value": [
                                "y"
                              ]
                            }
                          }
                        }
                      ],
                      "name": "my sample rule"
                    }
                  ]
                }
              }
            ]
          capabilities: Deep Insights
          categories: Integration & Delivery,OpenShift Optional
          containerImage: quay.io/openshift/origin-cluster-nfd-operator:4.10
          description: |-
            The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

            NFD consists  of the following software components:

            The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

            ##NFD-Master
            NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

            ##NFD-Worker
            NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

            ##NFD-Topology-Updater
            NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
          olm.skipRange: '>=4.6.0 <4.11.0'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-nfd
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.16.0+git
          operators.operatorframework.io/internal-objects: '["noderesourcetopologies.topology.node.k8s.io"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift/cluster-nfd-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: |
              The NodeFeatureDiscovery instance is the CustomResource being watched by the NFD-Operator, and holds all the needed information to setup the behaviour of the master and worker pods
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
          - description: |
              NodeFeatureRule resource specifies a configuration for feature-based customization of node objects, such as node labeling.
            kind: NodeFeatureRule
            name: nodefeaturerules.nfd.openshift.io
            version: v1alpha1
          - description: |
              NodeResourceTopology resource describes node resources and their topology.
            kind: NodeResourceTopology
            name: noderesourcetopologies.topology.node.k8s.io
            version: v1alpha1
        description: |-
          The Node Feature Discovery Operator manages the detection of hardware features and configuration in a Kubernetes cluster by labeling the nodes with hardware-specific information. The Node Feature Discovery (NFD) will label the host with node-specific attributes, like PCI cards, kernel, or OS version, and many more.

          NFD consists  of the following software components:

          The NFD Operator is based on the Operator Framework an open source toolkit to manage Kubernetes native applications, called Operators, in an effective, automated, and scalable way.

          ##NFD-Master
          NFD-Master is the daemon responsible for communication towards the Kubernetes API. That is, it receives labeling requests from the worker and modifies node objects accordingly.

          ##NFD-Worker
          NFD-Worker is a daemon responsible for feature detection. It then communicates the information to nfd-master which does the actual node labeling. One instance of nfd-worker is supposed to be running on each node of the cluster.

          ##NFD-Topology-Updater
          NFD-Topology-Updater is a daemon responsible for examining allocated resources on a worker node to account for resources available to be allocated to new pod on a per-zone basis (where a zone can be a NUMA node). It then communicates the information to nfd-master which does the NodeResourceTopology CR creation corresponding to all the nodes in the cluster. One instance of nfd-topology-updater is supposed to be running on each node of the cluster.
        displayName: Node Feature Discovery Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - feature-discovery
        - feature-detection
        - node-labels
        links:
        - name: Node Feature Discovery Operator
          url: https://docs.openshift.com/container-platform/4.9/hardware_enablement/psap-node-feature-discovery-operator.html
        - name: Node Feature Discovery Documentation
          url: https://kubernetes-sigs.github.io/node-feature-discovery/stable/get-started/index.html
        maintainers:
        - email: openshift-psap@redhat.com
          name: openshift-psap
        maturity: stable
        minKubeVersion: 1.22.0
        provider:
          name: Red Hat
          url: https://github.com/openshift/cluster-nfd-operator
        relatedImages:
        - quay.io/openshift/origin-cluster-nfd-operator:4.10
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        - quay.io/openshift-community-operators/openshift-nfd-operator@sha256:880517267f12e0ca4dd9621aa196c901eb1f754e5ec990a1459d0869a8c17451
        version: 4.10.0
      entries:
      - name: nfd.v4.10.0
        version: 4.10.0
      - name: nfd.v4.9.0
        version: 4.9.0
      - name: nfd.v4.8.0
        version: 4.8.0
      name: stable
    defaultChannel: stable
    packageName: openshift-nfd-operator
    provider:
      name: Red Hat
      url: https://github.com/openshift/cluster-nfd-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeVirt project
      provider-url: ""
    name: community-kubevirt-hyperconverged
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubevirt-hyperconverged-operator.v1.10.7
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
          createdAt: "2024-02-09 14:34:30"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:16ecabb1d733aa17084cd16d88963c5255d04f389e0f4d41941d67e8be973594
        - quay.io/kubevirt/virt-operator@sha256:cbca7f93afbb18a840f7e763e4c03aeeb1fd97967f381c12c443cc179b51da29
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f6717ce72a2615c7fbc746b4068f788e78579c54c43b8716e5ce650d97af2df1
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/virt-api@sha256:707003b221496b4432da2f507d1e36e528b45888b5d321e06d460f0678da44ae
        - quay.io/kubevirt/hostpath-provisioner@sha256:eae30f37b3e837478a4500144b7bc8eb4f5ddf15f49fa7354bc9d5e240ea8aea
        - registry.k8s.io/sig-storage/livenessprobe@sha256:4dc0b87ccd69f9865b89234d8555d3a614ab0a16ed94a3016ffd27f8106132ce
        - quay.io/kubevirt/mtq-controller@sha256:96dd73ea0d7847d9f6322192559747bd85655e25ee25482d112e6e3c6fef21b2
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/virt-launcher@sha256:4c5fce3de2e2589197de72fb0c9436490ea318aca952c05a622c43e067023f35
        - quay.io/kubevirt/kubevirt-template-validator@sha256:a148b6e812b70f326b48d0439b1e44a967a6e1f8cf2f98b1a08e9294a93674eb
        - quay.io/kubevirt/cdi-uploadproxy@sha256:551221d79902a5053d1c734b81163d69f087217e2ac13c49bdf6900336ef0786
        - quay.io/kubevirt/mtq-operator@sha256:ef5c4bb1fa72f9e6e75d5585ca352f7563720389e6c9876aa8b07d12d1be6194
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:f7567bea7f24184da06c3ed87b396c804342fcfd89f54989622780ce0b7b0724
        - quay.io/kubevirt/cdi-importer@sha256:3143bbc67cdc6267eb48b7eaac664b8551ac4c11401dfbf4921efd3f233e6ce9
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:86c10c122767661e31ee26a9e9de7ea606c2c1168b32793b4ea567ede2ae3b21
        - quay.io/kubevirt/virt-handler@sha256:138dfda5fea8622f3da0d6413fe214fef80c2fd6a6f9533592a0dbfa7e1865b5
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:f64ab5981d9ddef1662dd0a05888783f5dbadda8ede596585ba80bfc97a0f51e
        - quay.io/kubevirt/cdi-apiserver@sha256:e9e39408413b1478d2e98eba68913f9e20c93000558b190b47de73bdfd1d9ac4
        - quay.io/kubevirt/virt-artifacts-server:1.10.7-unstable
        - quay.io/kubevirt/hostpath-csi-driver@sha256:0ca8b50b0adb20afd4e20da9974200f3290b6d51aa7c1129de6e1b953bf26594
        - quay.io/kubevirt/macvtap-cni@sha256:434420511e09b2b5ede785a2c9062b6658ffbc26fbdd4629ce06110f9039c600
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:0ee30557d5d5a42dec19121e70ed77fbdacd8932f55b11e2c7c1a3d5ad5df4aa
        - quay.io/kubevirt/virt-controller@sha256:0789fafed2913b35a771e3db882748502b3250be04ece86d97f30201779b4e54
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5ff5392cb762a41ee9bc4767e3a693249ae08b46a4322f623f24ad6b63dd06b5
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:e5e0c3b3a25c13ea916d70fa6f6ddc20f2c730bd9909aa2654672280c27afda2
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:e79027973b09aac75860f267c7d7f830599978c38a081001f404b9a1c3d2990f
        - quay.io/kubevirt/ovs-cni-plugin@sha256:08f72edf2bef876bba0b0f5513d30225304ad5e7ad6912a61c083664acdb99ff
        - quay.io/kubevirt/libguestfs-tools@sha256:b7e248be783691daad1326cf41641d5042be80c5c057ebc5315ca7a1619b0581
        - quay.io/kubevirt/cni-default-plugins@sha256:825e3f9fec1996c54a52cec806154945b38f76476b160d554c36e38dfffe5e61
        - quay.io/kubevirt/cdi-uploadserver@sha256:30f1827d3696cf996b081c22c3267ca78e7219c872fdb54950198fa54359f6ee
        - quay.io/kubevirt/kubemacpool@sha256:3214020c21204c189cdf2d767beb8efb61ee3ef65dd8de7f0f1edc61e13e8adf
        - quay.io/kubevirt/virt-exportproxy@sha256:f14444b0200a85efb4b3c176caefe70aabba7d295e33d2af14027a59ce297e24
        - quay.io/kubevirt/virt-exportserver@sha256:73311f79a9c71007f8572b3cc40cd6f6da404c7ef0a9c6509fb717d979546582
        - quay.io/kubevirt/cdi-operator@sha256:6c63521d835578b0fbb77d3145b648f95ab0932d5d1b36ff2068ed8fcd91bc5a
        - quay.io/kubevirt/ssp-operator@sha256:49cac00844f091d6e80301573e00235d685415185ec29c89b4bd7361f938711e
        - quay.io/kubevirt/cdi-controller@sha256:27c47883a08226f83757971d3adafb0cd9bcb26e58fbcf7208236070e0adf37e
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cdi-cloner@sha256:9d31b14f23259398c5bac636f5ead13ad0afd6fe8eeab4499e8e047b4d85074f
        - quay.io/kubevirt/mtq-lock-server@sha256:9614df0d5af7c998bcceaf73b548f2e3b29ff8ee4ca64c32613bfbdd6e452f6c
        version: 1.10.7
      entries:
      - name: kubevirt-hyperconverged-operator.v1.10.7
        version: 1.10.7
      name: 1.10.7
    - currentCSV: kubevirt-hyperconverged-operator.v1.11.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployTektonTaskResources":true,"deployVmConsoleProxy":true},"tektonPipelines":{"namespace":"kubevirt"},"tektonTasks":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"mtq.kubevirt.io/v1alpha1","kind":"MTQ","metadata":{"name":"mtq","namespace":"mtq"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
          createdAt: "2024-03-06 15:11:00"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","mtqs.mtq.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a MTQ deployment
            displayName: MTQ deployment
            kind: MTQ
            name: mtqs.mtq.kubevirt.io
            version: v1alpha1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virt-controller@sha256:4f576e30b83db7b9f96dde77896898bde87db3517963c4227634d0c3e96bb86d
        - quay.io/kubevirt/ssp-operator@sha256:d57ef05fea09c4690dbe728ebe6a39d9df1a3574dfa7d18346bf3fe66afed702
        - quay.io/kubevirt/bridge-marker@sha256:bba066e3b5ff3fb8c5e20861fe8abe51e3c9b50ad6ce3b2616af9cb5479a06d0
        - quay.io/kubevirt/cni-default-plugins@sha256:c884d6d08f8c0db98964f1eb3877b44ade41fa106083802a9914775df17d5291
        - quay.io/kubevirt/cdi-importer@sha256:7320dce01f4628363992997b592200b4804daf15a9f237cd50a4a6acab7798e8
        - quay.io/kubevirt/kubemacpool@sha256:cf8daa57ae6603b776d3af512331b143fa03bc2f4b72f28420fddcf5e4156d0a
        - quay.io/kubevirt/aaq-server@sha256:0c6b01c48030e03e99b3ff2b319658c2bf2b2001f4e3e030ba70d1143b25717a
        - quay.io/kubevirt/kubevirt-template-validator@sha256:f6db539fe4d4b752644c6bc8a308f5a80ae09a71da5c5d535114dc4c3cf18e0a
        - quay.io/kubevirt/virt-handler@sha256:238550bb263608ba334b44a15647cbf875482ccfb46049713c14af5d5f0cc788
        - quay.io/kubevirt/mtq-lock-server@sha256:8e686107dd9990dab1a3d90ee1068541938f673bf54ea9e1308b2ed7218546b0
        - quay.io/kubevirt/virt-exportserver@sha256:f8473fc824f8bac5f21c64b71cdb4227de28f9bb25c049932f1259216b874a86
        - ghcr.io/kubevirt/kubesecondarydns@sha256:e87e829380a1e576384145f78ccaa885ba1d5690d5de7d0b73d40cfb804ea24d
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:bf17da674ed6dac95aa4fd7637173fd4e026ebb4523b66f529c0eff394a25bcc
        - quay.io/kubevirt/virt-launcher@sha256:4fbfd76561667e98fafc75c0b804f77df251f96471d1121d69f149cf8fb9a4d8
        - registry.k8s.io/sig-storage/livenessprobe@sha256:5baeb4a6d7d517434292758928bb33efc6397368cbb48c8a4cf29496abf4e987
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:dd2948ca8e62d17eafba40bdcaad8367a97f092fd3b8d9849b2d62aec3152a72
        - quay.io/kubevirt/cdi-uploadserver@sha256:16d257e575b2bd504f148f63b57dc181fa6058d33a11ada64888183c3e4d4863
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:3fbcc32bd4e4d15bd93c96def784a229cd84cca27942bf4858b581f31c97ee02
        - quay.io/kubevirt/libguestfs-tools@sha256:4971af1ee42959d366a3db6129ab684d11064e00bbd3c2b36324ad14e90cb903
        - quay.io/kubevirt/mtq-controller@sha256:ae7dd07e4018c63414b021eba34aa4776a6c7d6ff4fcdf9fd2d7ca30aae98a29
        - quay.io/kubevirt/virt-artifacts-server@sha256:22fa3218deca38dc98fa0e905df16173a47ccec85707eb91404a367da6480817
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:8c65e65bbaf0ebb10dba03655c5854c69c82ddc0e6b2a29293eee029222e5efa
        - quay.io/kubevirt/cdi-apiserver@sha256:95f6e71b36883abcb9ec463cfea2332f5490ec9eb3e4868bdf0a4da769279094
        - quay.io/kubevirt/cdi-controller@sha256:606de89e68be2ac911466e6713345932bda496c1e4f20e34ef5efa270456c9fd
        - quay.io/kubevirt/hostpath-provisioner@sha256:adb1d33e95e8991dc47107926172469b5ecbeace509391ce021ca42d305a41d3
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:57573a24923e5588bca6bc337a8b2b08406c5b77583974365d2cf063c0dd5d06
        - quay.io/kubevirt/ovs-cni-plugin@sha256:e16ac74343da21abb8fb668ce71e728053d00503a992dae2164b9e94a280113e
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:4ed1f37116cbaf883edd8783201d091fbbe121fcb84e7184ee531d3b1abd77ab
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:70efc16ab2bfec04f73ccc292064d080870659cfd47c2e0c9017ff7b6fd78863
        - quay.io/kubevirt/aaq-operator@sha256:590d24742401eeff3fb6a9c8ed977c63442f8f0a0505509fae69aa1bf8aa84ae
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:dd6538e720f18b723b0807037ca7cc1014cba09e57b20706e10e6eb641ac6acc
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:f1bd6ee18c4021c1c94f29edfab89b49b6a4d1b800936c19dbef2d75f8202f2d
        - quay.io/kubevirt/hostpath-csi-driver@sha256:8d84b9a33e0ec3749aece7c5f9879d1e0d248e8028ab3f776528a044d387033c
        - quay.io/kubevirt/virt-operator@sha256:0e6d56ea0ebdf9bb832f29ab0dc4603f0f619474bb8bc02a01e0a06e9d8d6f4a
        - quay.io/kubevirt/cdi-uploadproxy@sha256:b89104ef5513d63cab43b34ac52f752c3d935cb986d5527d920c2c31d3b6f239
        - quay.io/kubevirt/virtio-container-disk@sha256:f8e71a3a930ce95d646d80862d5ff6f9d3ed7ef2541dd2e2acec0bce26274734
        - quay.io/kubevirt/mtq-operator@sha256:90c7608b0ba085582a65ccdc11d089049c5e6b0438d018dd9dc76eb3490383c7
        - quay.io/kubevirt/virt-exportproxy@sha256:942c548d4848a87945eba51683002797eb4e1f5c463a1873115b2ec0e1e6e169
        - quay.io/kubevirt/cdi-cloner@sha256:b22985af0ee7b1ce4e1dedded08c853d90d9e7209f98c5d470897bbefce2a063
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/aaq-controller@sha256:cb2fa52ae3f6a1a99b2de77b58c5db54c43ba8ee86c6540b4844c901180ce611
        - quay.io/kubevirt/virt-api@sha256:735662c824da3d6069624d40674163d758c2256f8043cf6f02f505c3a724afa3
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:c53535af8a7f7e3164609838c4b191b42b2d81238d75c1b2a2b582ada62a9780
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:38293a1050bce9c7cb786cbe8719be1edb91a922a6c583eedb61e56729ed4c72
        - quay.io/kubevirt/cdi-operator@sha256:8a27e1450c3bf8c1d8924e74470a62419d02ed2202924b082a74e2414b1fc21a
        version: 1.11.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.11.0
        version: 1.11.0
      name: 1.11.0
    - currentCSV: kubevirt-hyperconverged-operator.v1.14.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}},{"apiVersion":"networkaddonsoperator.network.kubevirt.io/v1","kind":"NetworkAddonsConfig","metadata":{"name":"cluster"},"spec":{"imagePullPolicy":"IfNotPresent","kubeMacPool":{"rangeEnd":"FD:FF:FF:FF:FF:FF","rangeStart":"02:00:00:00:00:00"},"linuxBridge":{},"macvtap":{},"multus":{},"ovs":{}}},{"apiVersion":"kubevirt.io/v1","kind":"KubeVirt","metadata":{"name":"kubevirt","namespace":"kubevirt"},"spec":{"imagePullPolicy":"Always"}},{"apiVersion":"ssp.kubevirt.io/v1beta2","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"featureGates":{"deployVmConsoleProxy":true},"templateValidator":{"replicas":2}}},{"apiVersion":"ssp.kubevirt.io/v1beta3","kind":"SSP","metadata":{"name":"ssp-sample","namespace":"kubevirt"},"spec":{"commonTemplates":{"namespace":"kubevirt"},"templateValidator":{"replicas":2}}},{"apiVersion":"cdi.kubevirt.io/v1beta1","kind":"CDI","metadata":{"name":"cdi","namespace":"cdi"},"spec":{"imagePullPolicy":"IfNotPresent"}},{"apiVersion":"hostpathprovisioner.kubevirt.io/v1beta1","kind":"HostPathProvisioner","metadata":{"name":"hostpath-provisioner"},"spec":{"imagePullPolicy":"IfNotPresent","storagePools":[{"name":"local","path":"/var/hpvolumes","pvcTemplate":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"50Gi"}}}}],"workload":{"nodeSelector":{"kubernetes.io/os":"linux"}}}},{"apiVersion":"aaq.kubevirt.io/v1alpha1","kind":"AAQ","metadata":{"name":"aaq","namespace":"aaq"},"spec":{"imagePullPolicy":"IfNotPresent"}}]'
          capabilities: Deep Insights
          categories: OpenShift Optional
          certified: "false"
          console.openshift.io/disable-operand-delete: "true"
          containerImage: quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
          createdAt: "2025-03-03 15:18:52"
          description: A unified operator deploying and controlling KubeVirt and its
            supporting operators with opinionated defaults
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "true"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          openshift.io/required-scc: restricted-v2
          operatorframework.io/initialization-resource: '{"apiVersion":"hco.kubevirt.io/v1beta1","kind":"HyperConverged","metadata":{"annotations":{"deployOVS":"false"},"name":"kubevirt-hyperconverged","namespace":"kubevirt-hyperconverged"},"spec":{}}'
          operatorframework.io/suggested-namespace: kubevirt-hyperconverged
          operators.openshift.io/infrastructure-features: '["disconnected","proxy-aware"]'
          operators.openshift.io/must-gather-image: quay.io/kubevirt/must-gather
          operators.operatorframework.io/internal-objects: '["networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io","kubevirts.kubevirt.io","ssps.ssp.kubevirt.io","ssps.ssp.kubevirt.io","cdis.cdi.kubevirt.io","aaqs.aaq.kubevirt.io"]'
          repository: https://github.com/kubevirt/hyperconverged-cluster-operator
          support: "false"
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents the deployment of HyperConverged Cluster Operator
            displayName: HyperConverged Cluster Operator Deployment
            kind: HyperConverged
            name: hyperconvergeds.hco.kubevirt.io
            version: v1beta1
          - description: Cluster Network Addons
            displayName: Cluster Network Addons
            kind: NetworkAddonsConfig
            name: networkaddonsconfigs.networkaddonsoperator.network.kubevirt.io
            version: v1
          - description: Represents a KubeVirt deployment
            displayName: KubeVirt deployment
            kind: KubeVirt
            name: kubevirts.kubevirt.io
            version: v1
          - description: SSP is the Schema for the ssps API
            displayName: SSP
            kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta2
          - kind: SSP
            name: ssps.ssp.kubevirt.io
            version: v1beta3
          - description: Represents a CDI deployment
            displayName: CDI deployment
            kind: CDI
            name: cdis.cdi.kubevirt.io
            version: v1beta1
          - description: Represents the deployment of HostPathProvisioner
            displayName: HostPathProvisioner Deployment
            kind: HostPathProvisioner
            name: hostpathprovisioners.hostpathprovisioner.kubevirt.io
            version: v1beta1
          - description: Represents a AAQ deployment
            displayName: AAQ deployment
            kind: AAQ
            name: aaqs.aaq.kubevirt.io
            version: v1alpha1
        description: |-
          **HyperConverged Cluster Operator** is an Operator pattern for managing multi-operator products.
          Specifically, the HyperConverged Cluster Operator manages the deployment of KubeVirt,
          Containerized Data Importer (CDI), Virtual Machine import operator and Cluster Network Addons (CNA) operators.

          **KubeVirt** is a virtual machine management add-on for Kubernetes.
          The aim is to provide a common ground for virtualization solutions on top of
          Kubernetes.

          # Virtualization extension for Kubernetes

          At its core, KubeVirt extends [Kubernetes](https://kubernetes.io) by adding
          additional virtualization resource types (especially the `VirtualMachine` type) through
          [Kubernetes's Custom Resource Definitions API](https://kubernetes.io/docs/tasks/access-kubernetes-api/extend-api-custom-resource-definitions/).
          By using this mechanism, the Kubernetes API can be used to manage these `VirtualMachine`
          resources alongside all other resources Kubernetes provides.

          The resources themselves are not enough to launch virtual machines.
          For this to happen the _functionality and business logic_ needs to be added to
          the cluster. The functionality is not added to Kubernetes itself, but rather
          added to a Kubernetes cluster by _running_ additional controllers and agents
          on an existing cluster.

          The necessary controllers and agents are provided by KubeVirt.

          As of today KubeVirt can be used to declaratively

            * Create a predefined VM
            * Schedule a VM on a Kubernetes cluster
            * Launch a VM
            * Migrate a VM
            * Stop a VM
            * Delete a VM

          # Start using KubeVirt

            * Try our quickstart at [kubevirt.io](http://kubevirt.io/get_kubevirt/).
            * See our user documentation at [kubevirt.io/docs](http://kubevirt.io/user-guide).

          # Start developing KubeVirt

          To set up a development environment please read our
          [Getting Started Guide](https://github.com/kubevirt/kubevirt/blob/master/docs/getting-started.md).
          To learn how to contribute, please read our [contribution guide](https://github.com/kubevirt/kubevirt/blob/master/CONTRIBUTING.md).

          You can learn more about how KubeVirt is designed (and why it is that way),
          and learn more about the major components by taking a look at our developer documentation:

            * [Architecture](https://github.com/kubevirt/kubevirt/blob/master/docs/architecture.md) - High-level view on the architecture
            * [Components](https://github.com/kubevirt/kubevirt/blob/master/docs/components.md) - Detailed look at all components
            * [API Reference](https://github.com/kubevirt/kubevirt/blob/master/https://www.kubevirt.io/api-reference/)

          # Community

          If you got enough of code and want to speak to people, then you got a couple of options:

            * Follow us on [Twitter](https://twitter.com/kubevirt)
            * Chat with us in the #virtualization channel of the [Kubernetes Slack](https://slack.k8s.io/)
            * Discuss with us on the [kubevirt-dev Google Group](https://groups.google.com/forum/#!forum/kubevirt-dev)
            * Stay informed about designs and upcoming events by watching our [community content](https://github.com/kubevirt/community/)

          # License

          KubeVirt is distributed under the
          [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0.txt).
        displayName: KubeVirt HyperConverged Cluster Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - KubeVirt
        - Virtualization
        - VM
        links:
        - name: KubeVirt project
          url: https://kubevirt.io
        - name: Source Code
          url: https://github.com/kubevirt/hyperconverged-cluster-operator
        maintainers:
        - email: kubevirt-dev@googlegroups.com
          name: KubeVirt project
        maturity: alpha
        provider:
          name: KubeVirt project
        relatedImages:
        - quay.io/kubevirt/virtio-container-disk@sha256:3cff6989c7d7f72d2f0ac27e512d99873768766ba420d65a6b7794676d0c1e88
        - quay.io/kubevirt/virt-operator@sha256:d5fce85dd4232745408daebfbf70c2bc879a06f8340d1f7448b141d5ba3f1821
        - registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e
        - quay.io/kubevirt/cdi-uploadserver@sha256:2b59d3c0a7c154fb6915cf3439b42eb4aaf48f94025ddb75fa9b73b16148fc13
        - quay.io/kubevirt/ssp-operator@sha256:61ac0c182528a9b54d84755fc0931ad8d408ae74e7e4b644316a5e9e630ab360
        - quay.io/kubevirt/libguestfs-tools@sha256:d655b9abbe8cce45bce3934e11d01715e170ddda4da78d7c51f356e7ee1e3263
        - quay.io/kubevirt/bridge-marker@sha256:e492ca4a6d1234781928aedefb096941d95babee4116baaba4d2a3834813826a
        - quay.io/kubevirt/virt-exportserver@sha256:ec1693cc6e8a14ab83f2feea87427643a968d65832c44b25a1bb911482403ff2
        - ghcr.io/k8snetworkplumbingwg/ovs-cni-plugin@sha256:516791acf430bc747f01366e2b748ac76c0b5e39ad8592be2b3dcc809429fa1e
        - ghcr.io/kubevirt/kubesecondarydns@sha256:8273cdbc438e06864eaa8e47947bea18fa5118a97cdaddc41b5dfa6e13474c79
        - quay.io/kubevirt/hyperconverged-cluster-operator@sha256:704677db9ca9f2027e03d54685c2608c0503654c94889b9249114f9f29798c8a
        - quay.io/kubevirt/hostpath-provisioner-operator@sha256:66c55f5ffa4b44f2e58990baa7c672ec9776612a5f1c57fee5a5ab1c080f17bb
        - quay.io/openshift/origin-kube-rbac-proxy@sha256:e2def4213ec0657e72eb790ae8a115511d5b8f164a62d3568d2f1bff189917e8
        - quay.io/kubevirt-ui/kubevirt-plugin@sha256:902af624e05eca764c17137c27813b1618f388e4b1a53b27b54665e8735251a2
        - quay.io/kubevirt/aaq-operator@sha256:5e53b755adaf6c66f59812b590a2eae2d84554fa658e93621ac977316843d918
        - quay.io/kubevirt/hostpath-csi-driver@sha256:f8d44aabe0024058b4380e1af42a57a90b7f8c66c8627091bf425fc384c0880f
        - quay.io/kubevirt/virt-exportproxy@sha256:55e38274e59bfe5a2cb5671534b8b6e75d90e43107abecc95fd463f7be8f0026
        - quay.io/kubevirt/cdi-apiserver@sha256:1849bc14687449d36561eb1c768fdbf0cd14ababf9c254b33a4ddb14abfba01f
        - registry.k8s.io/sig-storage/livenessprobe@sha256:33692aed26aaf105b4d6e66280cceca9e0463f500c81b5d8c955428a75438f32
        - quay.io/kubevirt/network-passt-binding@sha256:c794cc1b9173267f1e3d34e666d9e2f782b88d93b47fb4449675346b467f802c
        - quay.io/kubevirt/virt-controller@sha256:68a4967ef5925d4c24fb5eab0aa5acf0a82946beb7f1139bd45c85efe3653cec
        - quay.io/kubevirt/cdi-controller@sha256:94d16a6bd0d9776eddedf64165f39e63e4dfe4f8a26e858bd099db5d09136047
        - quay.io/kubevirt/cluster-network-addons-operator@sha256:92f2ac10f28a7e80699bd984dbc4de897011b1f68c2be95f3fcb90df410b119d
        - quay.io/kubevirt/virt-artifacts-server@sha256:789f68f4e27cf947cdd4f1c0732042176e67b7faa5e37e1b503b8dace3a46125
        - quay.io/kubevirt/hostpath-provisioner@sha256:156ae46a03e2ff5550967773100122f9abc82e25bd7acaa55dbdec22238dd2ca
        - quay.io/kubevirt/sidecar-shim@sha256:cb4025f7275f8de2891c2195c5faf128ac1b68595cdb3eff776a2e5360b8f034
        - quay.io/kubevirt/cdi-importer@sha256:f9fff83c8307bca61a22e8296cd103b32e63d7c8b2a170c6781c6cd962211d26
        - quay.io/kubevirt/pr-helper@sha256:fd434b4ea1ca5682fc5b05f23de01130c0c10f77a0a2d9d089d8f3cb5b545aa6
        - quay.io/kubevirt/kubevirt-template-validator@sha256:b227679f8ccc7c53469574ee5f58c9cc39eb6c54ec01377e4b3df3205a2f4ab4
        - registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0d23a6fd60c421054deec5e6d0405dc3498095a5a597e175236c0692f4adee0f
        - quay.io/kubevirt/hyperconverged-cluster-webhook@sha256:89ddc40700d52cda23104978498fb4d74a643e807450b630beaa71ffdfa1fc57
        - ghcr.io/kubevirt/ipam-controller@sha256:fd7c38e129ea4ed8a86d1801c9e1ca387802376fc4abc0b0d228d4278f448ecd
        - quay.io/kubevirt/aaq-controller@sha256:b9fd0818e4019cccfb38f0923db058166eafdaea8bb35a720aa22b5ce55daf26
        - quay.io/kubevirt/cni-default-plugins@sha256:0c354fa9d695b8cab97b459e8afea2f7662407a987e83f6f6f1a8af4b45726be
        - ghcr.io/k8snetworkplumbingwg/multus-cni@sha256:16030e8320088cf74dc5fbc7dccfea40169f09722dfad15765fa28bceb8de439
        - quay.io/community-operator-pipeline-prod/community-kubevirt-hyperconverged@sha256:ac955c4ad03c40d74341623b22f99e8f796a6ed2ef5cc539fba3c1504d159955
        - quay.io/kubevirt/macvtap-cni@sha256:850b89343ace7c7ea6b18dd8e11964613974e9d1f7377af03854d407fb15230a
        - registry.k8s.io/sig-storage/csi-snapshotter@sha256:5f4bb469fec51147ce157329dab598c758da1b018bad6dad26f0ff469326d769
        - ghcr.io/k8snetworkplumbingwg/multus-dynamic-networks-controller@sha256:8061bd1276ff022fe52a0b07bc6fa8d27e5f6f20cf3bf764e76d347d2e3c929b
        - quay.io/kubevirt/cdi-uploadproxy@sha256:6a98ef32c3778e46812c698970e14736f178fd2a11b1f74ccc36c34ce2e34466
        - registry.k8s.io/sig-storage/csi-provisioner@sha256:d078dc174323407e8cc6f0f9abd4efaac5db27838f1564d0253d5e3233e3f17f
        - ghcr.io/kubevirt/passt-binding-cni@sha256:f3fea2ec66a26850489197edf7c85d074eb78bbac52916b10016ac0de8ef7c14
        - quay.io/kubevirt/cdi-cloner@sha256:feefb3933780bf352ba03b62b900907d02c64e73f7e74b6d336f89daf2752f62
        - quay.io/kubevirt/cdi-operator@sha256:8297771144f2f358d893771a9dde208a6bedf6bfd67ab2927b16c27424ca3f26
        - quay.io/kubevirt/kubemacpool@sha256:6cc8ca860960c003b36754c6dde249f26f78d32eb22159a25b5647e197dbc9f9
        - quay.io/kubevirt-ui/kubevirt-apiserver-proxy@sha256:935475c2850466aa5ac57e4de627fb177515cb2c402a95842ead095d82b6df5f
        - quay.io/kubevirt/aaq-server@sha256:18c23eb4e32ed083047c8808b35fa81a22294a7e8bbabc74e2612fd2eba97dd7
        - quay.io/kubevirt/virt-launcher@sha256:18508d8767f8530df4d6aca4ffe467c40f9891fbe052548b3b2316f8bfb1bc38
        - quay.io/kubevirt/virt-api@sha256:d92c78b3048756fd11d3c6f1a48f4ce8555dd39e866039313dc23bfa70668443
        - quay.io/kubevirt/virt-handler@sha256:4ddd39539282e79cf3602700b791eb91a6d79991bf2e4dc4cba24e3da9db3d1c
        version: 1.14.0
      entries:
      - name: kubevirt-hyperconverged-operator.v1.14.0
        version: 1.14.0
      - name: kubevirt-hyperconverged-operator.v1.13.0
        version: 1.13.0
      - name: kubevirt-hyperconverged-operator.v1.12.0
        version: 1.12.0
      - name: kubevirt-hyperconverged-operator.v1.11.1
        version: 1.11.1
      - name: kubevirt-hyperconverged-operator.v1.10.0
        version: 1.10.0
      - name: kubevirt-hyperconverged-operator.v1.9.0
        version: 1.9.0
      - name: kubevirt-hyperconverged-operator.v1.8.2
        version: 1.8.2
      - name: kubevirt-hyperconverged-operator.v1.8.1
        version: 1.8.1
      - name: kubevirt-hyperconverged-operator.v1.8.0
        version: 1.8.0
      - name: kubevirt-hyperconverged-operator.v1.7.0
        version: 1.7.0
      - name: kubevirt-hyperconverged-operator.v1.6.0
        version: 1.6.0
      - name: kubevirt-hyperconverged-operator.v1.5.2
        version: 1.5.2
      - name: kubevirt-hyperconverged-operator.v1.5.1
        version: 1.5.1
      - name: kubevirt-hyperconverged-operator.v1.5.0
        version: 1.5.0
      - name: kubevirt-hyperconverged-operator.v1.4.4
        version: 1.4.4
      - name: kubevirt-hyperconverged-operator.v1.4.3
        version: 1.4.3
      - name: kubevirt-hyperconverged-operator.v1.4.2
        version: 1.4.2
      - name: kubevirt-hyperconverged-operator.v1.4.1
        version: 1.4.1
      - name: kubevirt-hyperconverged-operator.v1.4.0
        version: 1.4.0
      name: stable
    defaultChannel: stable
    packageName: community-kubevirt-hyperconverged
    provider:
      name: KubeVirt project
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://www.redhat.com
    name: ansible-cloud-addons-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aca-operator.v2.5.0-0.1758143647
      currentCSVDesc:
        annotations:
          alm-examples: ""
          capabilities: Full Lifecycle
          categories: Integration & Delivery
          description: The Ansible Cloud Addons Operator manages everything Automation
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "true"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=2.1.1-0.1645500829 <2.5.1'
          operatorframework.io/suggested-namespace: aap
          operators.operatorframework.io/builder: operator-sdk-v1.11.0+git
          operators.operatorframework.io/operator-type: non-standalone
          operators.operatorframework.io/project_layout: ansible
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploy an AAP AWS Billing instance
            displayName: AnsibleAutomationPlatformAwsBilling
            kind: AnsibleAutomationPlatformAwsBilling
            name: ansibleautomationplatformawsbillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Azure Billing instance
            displayName: AnsibleAutomationPlatformAzureBilling
            kind: AnsibleAutomationPlatformAzureBilling
            name: ansibleautomationplatformazurebillings.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP Metrics instance
            displayName: AnsibleAutomationPlatformMetrics
            kind: AnsibleAutomationPlatformMetrics
            name: ansibleautomationplatformmetrics.platform.ansible.com
            version: v1alpha1
          - description: Deploy an AAP UI instance
            displayName: AutomationPlatform
            kind: AutomationPlatform
            name: automationplatforms.platform.ansible.com
            version: v1alpha1
        description: |
          Red Hat Ansible Cloud Addons is designed to help your organization accelerate, orchestrate and innovate with automation. Scale automation with control and insight. Collaborate across teams. Manage policy and governance. Empower multiple IT domains without slowing down development. With Ansible Automation Platform you can drive business outcomes and unleash the full potential of your teams and technology.

          ## About this Operator

          The Ansible Cloud Addons Operator provides for push-button deployments of Cloud Add-ons into your OpenShift environment. Ansible Cloud Addons deployments include:
            * Cloud Billing - Cloud billing connector for Ansible Automation Platform
            * Cloud UI - Unified welcome page UI for Ansible Automation Platform

          ## How to Install

          Use of this Red Hat product requires a licensing and subscription agreement.
          Install the Red Hat Ansible Cloud Addons operator by following instructions presented when you click the Install button.
        displayName: Ansible Cloud Addons
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ansible
        - automation
        - automation-platform
        - ansible-automation-platform
        - aap
        - cloud
        - management
        links:
        - name: About Ansible Automation Platform
          url: https://www.redhat.com/en/technologies/management/ansible
        maintainers:
        - email: aap-maintainers@redhat.com
          name: Ansible Automation Platform Productization Team
        maturity: stable
        provider:
          name: Red Hat
          url: https://www.redhat.com
        relatedImages:
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8-operator@sha256:f4b29ec59498bb6c9fdb2790ffc9d8c878efa3fd9da3416f7fe9f2a253171508
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-billing-rhel8@sha256:0e50f539bfeaec91f0b18409cbc1888297a26af5114252a8ca07e1b97436001d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-metrics-collector-rhel8@sha256:c6ab25b3839f8c3134e7d5e1f0b7716ea5fd66fe8d131261ff08aabd97c3cd2d
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8-operator@sha256:47e2c40cdcfe218d5bf7abbf6fc35340bacdc2d02dac7f81deee7876b5570c53
        - registry.redhat.io/ansible-automation-platform-25/aap-cloud-ui-rhel8@sha256:869efe830cdf23f9344263f0e9b5a209cb13b84d9d7588a1744fd3ea4cd3f83b
        - registry.redhat.io/ansible-automation-platform/cloud-addons-operator-bundle@sha256:2cb3b1887e8eeae6708f5f46c9bfb304d7c3df1537ea433b0b23772acacf9f90
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6a9e10a4aad59f7e01dac0407641e9cb5cd5b8ed21f62ea4b2271f8aede4cbec
        version: 2.5.0+0.1758143647
      entries:
      - name: aca-operator.v2.5.0-0.1727240231
        version: 2.5.0+0.1727240231
      - name: aca-operator.v2.5.0-0.1727874815
        version: 2.5.0+0.1727874815
      - name: aca-operator.v2.5.0-0.1728516369
        version: 2.5.0+0.1728516369
      - name: aca-operator.v2.5.0-0.1729737962
        version: 2.5.0+0.1729737962
      - name: aca-operator.v2.5.0-0.1731452082
        version: 2.5.0+0.1731452082
      - name: aca-operator.v2.5.0-0.1733188778
        version: 2.5.0+0.1733188778
      - name: aca-operator.v2.5.0-0.1733964232
        version: 2.5.0+0.1733964232
      - name: aca-operator.v2.5.0-0.1736892777
        version: 2.5.0+0.1736892777
      - name: aca-operator.v2.5.0-0.1737095650
        version: 2.5.0+0.1737095650
      - name: aca-operator.v2.5.0-0.1737671171
        version: 2.5.0+0.1737671171
      - name: aca-operator.v2.5.0-0.1739244024
        version: 2.5.0+0.1739244024
      - name: aca-operator.v2.5.0-0.1740088272
        version: 2.5.0+0.1740088272
      - name: aca-operator.v2.5.0-0.1740769078
        version: 2.5.0+0.1740769078
      - name: aca-operator.v2.5.0-0.1741367998
        version: 2.5.0+0.1741367998
      - name: aca-operator.v2.5.0-0.1742427652
        version: 2.5.0+0.1742427652
      - name: aca-operator.v2.5.0-0.1743655743
        version: 2.5.0+0.1743655743
      - name: aca-operator.v2.5.0-0.1746134365
        version: 2.5.0+0.1746134365
      - name: aca-operator.v2.5.0-0.1747339736
        version: 2.5.0+0.1747339736
      - name: aca-operator.v2.5.0-0.1749072064
        version: 2.5.0+0.1749072064
      - name: aca-operator.v2.5.0-0.1750895274
        version: 2.5.0+0.1750895274
      - name: aca-operator.v2.5.0-0.1753899647
        version: 2.5.0+0.1753899647
      - name: aca-operator.v2.5.0-0.1755834478
        version: 2.5.0+0.1755834478
      - name: aca-operator.v2.5.0-0.1758143647
        version: 2.5.0+0.1758143647
      name: stable-2.5-cluster-scoped
    defaultChannel: stable-2.5-cluster-scoped
    packageName: ansible-cloud-addons-operator
    provider:
      name: Red Hat
      url: https://www.redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: openshift-secondary-scheduler-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: secondaryscheduleroperator.v1.4.1
      currentCSVDesc:
        annotations:
          alm-examples: |
            [
              {
                "apiVersion": "operator.openshift.io/v1",
                "kind": "SecondaryScheduler",
                "metadata": {
                  "name": "cluster",
                  "namespace": "openshift-secondary-scheduler-operator"
                },
                "spec": {
                  "managementState": "Managed",
                  "schedulerConfig": "secondary-scheduler-config"
                }
              }
            ]
          capabilities: Basic Install
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
          createdAt: 2025/05/14
          description: Runs a secondary scheduler in an OpenShift cluster.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.3.0 <1.4.1'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift/secondary-scheduler-operator
          support: Red Hat, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SecondaryScheduler is the schema for the secondaryschedulers
              API
            displayName: Secondary Scheduler
            kind: SecondaryScheduler
            name: secondaryschedulers.operator.openshift.io
            version: v1
        description: |
          The Secondary Scheduler Operator provides the ability to use a customized scheduler image that was developed using the scheduler plugin framework as a secondary scheduler in OpenShift.

          ## Secondary Scheduler Configuration

          Once the Secondary Scheduler Operator is installed, you can configure the plugins to run for your secondary scheduler.

          The configuration of the secondary scheduler is defined through a config map that wraps the KubeSchedulerConfiguration YAML file in "config.yaml".

          ## Additional Parameters

          Additionally, the following parameters can be configured:

          * `schedulerConfig` - Set the config map configuration for the secondary scheduler.
          * `schedulerImage` - Set the default operand image for the secondary scheduler.
        displayName: Secondary Scheduler Operator for Red Hat OpenShift
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - secondary
        - scheduling
        - workload
        links:
        - name: Source Code
          url: https://github.com/openshift/secondary-scheduler-operator
        maintainers:
        - email: support@redhat.com
          name: Red Hat
        maturity: beta
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-rhel9-operator@sha256:3a3921cb05cf8bfdca6f7a3ccbe59f9dfda8c697b53d9be0464cf3b7a961162f
        - registry.redhat.io/openshift-secondary-scheduler-operator/secondary-scheduler-operator-bundle@sha256:3804fcb77893d9198527b4801e2a778bf612596c9a0b125a0f0f872b23138434
        version: 1.4.1
      entries:
      - name: secondaryscheduleroperator.v1.4.1
        version: 1.4.1
      - name: secondaryscheduleroperator.v1.4.0
        version: 1.4.0
      - name: secondaryscheduleroperator.v1.3.1
        version: 1.3.1
      - name: secondaryscheduleroperator.v1.3.0
        version: 1.3.0
      - name: secondaryscheduleroperator.v1.2.2
        version: 1.2.2
      - name: secondaryscheduleroperator.v1.2.1
        version: 1.2.1
      - name: secondaryscheduleroperator.v1.2.0
        version: 1.2.0
      - name: secondaryscheduleroperator.v1.1.4
        version: 1.1.4
      - name: secondaryscheduleroperator.v1.1.3
        version: 1.1.3
      - name: secondaryscheduleroperator.v1.1.2
        version: 1.1.2
      - name: secondaryscheduleroperator.v1.1.1
        version: 1.1.1
      - name: secondaryscheduleroperator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: openshift-secondary-scheduler-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: dns-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: dns-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSHealthCheckProbe",
                "metadata": {
                  "name": "$NAME"
                },
                "spec": {
                  "additionalHeadersRef": {
                    "name": "headers"
                  },
                  "address": "192.168.0.16",
                  "allowInsecureCertificate": true,
                  "failureThreshold": 5,
                  "hostname": "test.com",
                  "interval": "60s",
                  "path": "/healthz",
                  "port": 443,
                  "protocol": "HTTPS"
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSRecord",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "dns-operator",
                    "app.kubernetes.io/instance": "dnsrecord-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "dnsrecord",
                    "app.kubernetes.io/part-of": "dns-operator"
                  },
                  "name": "dnsrecord-sample"
                },
                "spec": {
                  "endpoints": [
                    {
                      "dnsName": "dnsrecord-simple.example.com",
                      "recordTTL": 60,
                      "recordType": "A",
                      "targets": [
                        "52.215.108.61",
                        "52.30.101.221"
                      ]
                    }
                  ],
                  "providerRef": {
                    "name": "dns-provider-creds"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
          createdAt: "2025-04-07T08:13:58Z"
          description: Configures how north-south traffic from outside the network
            is balanced and reaches Gateways
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Connectivity Link"]'
          operators.operatorframework.io/builder: operator-sdk-v1.33.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kuadrant/dns-operator
          support: kuadrant
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DNSHealthCheckProbe is the Schema for the dnshealthcheckprobes
              API
            displayName: DNSHealthCheckProbe
            kind: DNSHealthCheckProbe
            name: dnshealthcheckprobes.kuadrant.io
            version: v1alpha1
          - description: DNSRecord is the Schema for the dnsrecords API
            displayName: DNSRecord
            kind: DNSRecord
            name: dnsrecords.kuadrant.io
            version: v1alpha1
        description: Configures how north-south traffic from outside the network is
          balanced and reaches Gateways - Part of Red Hat Connectivity Link
        displayName: Red Hat - DNS Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dns
        - kuadrant
        links:
        - name: GitHub
          url: https://github.com/kuadrant/dns-operator
        - name: Documentation
          url: https://github.com/Kuadrant/dns-operator/tree/main/docs
        maintainers:
        - email: mnairn@redhat.com
          name: Michael Nairn
        - email: pbrookes@redhat.com
          name: Phil Brookes
        - email: cbrookes@redhat.com
          name: Craig Brookes
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/rhcl-1/dns-operator-bundle@sha256:44d82d5f8434fb6e88d2da406e87df0bf4274661d9d77554b80e09d19fce03de
        - registry.redhat.io/rhcl-1/dns-rhel9-operator@sha256:8ad3876e36ba2f76355499d98a97672f54f26754fc9d219d9cb1c5b9f9874430
        version: 1.1.0
      entries:
      - name: dns-operator.v1.1.0
        version: 1.1.0
      - name: dns-operator.v1.0.2
        version: 1.0.2
      name: stable
    defaultChannel: stable
    packageName: dns-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: topology-aware-lifecycle-manager
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: "4.18"
    - currentCSV: topology-aware-lifecycle-manager.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ran.openshift.io/v1alpha1",
                "kind": "ClusterGroupUpgrade",
                "metadata": {
                  "name": "ClusterGroupUpgrade-sample"
                },
                "spec": {
                  "preCaching": true
                }
              }
            ]
          capabilities: Basic Install
          containerImage: registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
          description: The topology aware lifecycle manager is an operator that facilitates
            platform upgrades of group of clusters
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-operators
          operatorframework.io/suggested-namespace-template: |-
            {
              "apiVersion": "v1",
              "kind": "Namespace",
              "metadata": {
                "name": "openshift-operators",
                "annotations": {
                  "workload.openshift.io/allowed": "management"
                }
              }
            }
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: Red Hat
          repository: https://github.com/openshift-kni/topology-aware-lifecycle-manager
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterGroupUpgrade is the Schema for the ClusterGroupUpgrades
              API
            displayName: Cluster Group Upgrade
            kind: ClusterGroupUpgrade
            name: clustergroupupgrades.ran.openshift.io
            version: v1alpha1
          - description: ImageBasedGroupUpgrade is the schema for upgrading a group
              of clusters using IBU
            displayName: Image-Based Group Upgrade
            kind: ImageBasedGroupUpgrade
            name: imagebasedgroupupgrades.lcm.openshift.io
            version: v1alpha1
          - description: PreCachingConfig is the Schema for the precachingconfigs
              API
            displayName: Pre-caching Config
            kind: PreCachingConfig
            name: precachingconfigs.ran.openshift.io
            version: v1alpha1
        description: |-
          Topology Aware Lifecycle Manager is an operator that facilitates
          platform and operator upgrades of group of clusters
        displayName: Topology Aware Lifecycle Manager
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - acm lifecyclemanagement upgrades cluster
        links:
        - name: Cluster Group Upgrades Operator
          url: https://topology-aware-lifecycle-manager.domain
        maturity: alpha
        minKubeVersion: 1.31.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-rhel9-operator@sha256:d3d09f6c87d2e7b7fe979650df2d15c61027ad5253148f076bd2d9190655bc69
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-operator-bundle@sha256:4ef2260dd4af7b52354bac62f032aa3d48411e8c47c123d6ee7f2df5b8462350
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:5903b64e4c53728ef99d2dcea5dceff80413745bb23f1afaa5e3cdb2128e91ae
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-aztp-rhel9@sha256:4cf6b3c4b1398353fd5c6ed685e004f6a39fb0cd8162e5ed0e07c2a6720383b7
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-precache-rhel9@sha256:4f62d666a72fd569b55017313d809d271b6e263917c9a52a588fe2e0c8ec981c
        - registry.redhat.io/openshift4/topology-aware-lifecycle-manager-recovery-rhel9@sha256:9aa51c20003da785eb639e0c0634a80d323337267845141090d8f8665d279b01
        version: 4.18.1
      entries:
      - name: topology-aware-lifecycle-manager.v4.18.1
        version: 4.18.1
      - name: topology-aware-lifecycle-manager.v4.18.0
        version: 4.18.0
      name: stable
    defaultChannel: stable
    packageName: topology-aware-lifecycle-manager
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Aqua Security, Inc.
      provider-url: ""
    name: aqua
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: aqua-operator.v2022.4.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaCsp",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "namespace": "aqua",
                    "version": "2022.4",
                    "requirements": true
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "dbDiskSize": 10,
                    "databaseSecret": {
                      "key": "db-password",
                      "name": "aqua-database-password"
                    }
                  },
                  "database": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "gateway": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "server": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "route": true,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaDatabase",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "database",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "diskSize": 10,
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "runAsNonRoot": false,
                  "gateway": {
                    "host": "aqua-gateway",
                    "port": 8443
                  },
                  "token": "<<your-token>>"
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaKubeEnforcer",
                "metadata": {
                  "name": "aqua"
                },
                "spec": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "config": {
                    "gateway_address": "aqua-gateway.aqua:8443",
                    "cluster_name": "aqua-secure",
                    "imagePullSecret": "aqua-registry",
                    "kubebench_image": "docker.io/aquasec/kube-bench:v0.7.1"
                  },
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "<<KUBE_ENFORCER_TAG>>",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  }
                }
              },
              {
              "apiVersion": "operator.aquasec.com/v1alpha1",
              "kind": "AquaLightning",
              "metadata": {
                "name": "aqua"
              },
              "spec": {
                "global": {
                  "gateway_address": "aqua-gateway.aqua:8443",
                  "cluster_name": "Default-cluster-name"
                },
                "common": {
                  "imagePullSecret": "aqua-registry"
                },
                "kubeEnforcer": {
                  "infra": {
                    "version": "2022.4",
                    "serviceAccount": "aqua-kube-enforcer-sa"
                  },
                  "token": "<<KUBE_ENFORCER_GROUP_TOKEN>>",
                  "allowAnyVersion": null,
                  "deploy": {
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "repository": "kube-enforcer",
                      "pullPolicy": "Always"
                    }
                  },
                  "starboard": {
                    "infra": {
                      "serviceAccount": "starboard-operator"
                    },
                    "config": {
                      "imagePullSecret": "starboard-registry"
                    },
                    "deploy": {
                      "replicas": 1
                    }
                  },
                  "env": []
                },
                "enforcer": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "deploy": {
                    "image": {
                      "repository": "enforcer",
                      "registry": "registry.aquasec.com",
                      "tag": "2022.4",
                      "pullPolicy": "IfNotPresent"
                    }
                  },
                  "token": "<<ENFORCER_GROUP_TOKEN>>",
                  "env": [],
                  "secret": {
                    "name": null,
                    "key": null
                  },
                  "runAsNonRoot": null,
                  "rhcosVersion": null
                }
              }
            },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaGateway",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "ClusterIP",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "gateway",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaScanner",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "imagePullSecret": "aqua-registry"
                  },
                  "deploy": {
                    "replicas": 1,
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "scanner",
                      "tag": "<<IMAGE TAG>>"
                    }
                  },
                  "runAsNonRoot": false,
                  "login": {
                    "username": "<<YOUR AQUA USER NAME>>",
                    "password": "<<YOUR AQUA USER PASSWORD>>",
                    "host": "http://aqua-server:8080",
                    "token": ""
                  }
                }
              },
              {
                "apiVersion": "operator.aquasec.com/v1alpha1",
                "kind": "AquaServer",
                "metadata": {
                  "name": "aqua",
                  "namespace": "aqua"
                },
                "spec": {
                  "infra": {
                    "serviceAccount": "aqua-sa",
                    "version": "2022.4"
                  },
                  "common": {
                    "databaseSecret": {
                      "name": "<<EXTERNAL DB PASSWORD SECRET NAME>>",
                      "key": "<<EXTERNAL DB PASSWORD SECRET KEY>>"
                    },
                    "imagePullSecret": "aqua-registry",
                    "splitDB": false
                  },
                  "externalDb": {
                    "host": "<<EXTERNAL DB IP OR DNS>>",
                    "port": "<<EXTERNAL DB PORT>>",
                    "username": "<<EXTERNAL DB USERNAME>>",
                    "password": "<<EXTERNAL DB PASSWORD (if secret does not exist)>>"
                  },
                  "deploy": {
                    "replicas": 1,
                    "service": "LoadBalancer",
                    "image": {
                      "registry": "registry.aquasec.com",
                      "repository": "console",
                      "tag": "<<IMAGE TAG>>",
                      "pullPolicy": "Always"
                    }
                  },
                  "runAsNonRoot": false
                }
              },
              {
                  "apiVersion": "operator.aquasec.com/v1alpha1",
                  "kind": "AquaCloudConnector",
                  "metadata": {
                    "name": "aqua"
                  },
                  "spec": {
                    "infra": {
                      "serviceAccount": "aqua-sa",
                      "version": "2022.4"
                    },
                    "common": {
                      "imagePullSecret": "aqua-registry"
                    },
                    "deploy": {
                      "image": {
                        "repository": "aqua-cloud-connector",
                        "registry": "registry.aquasec.com",
                        "tag": "<<IMAGE TAG>>"
                      },
                      "replicas": 1
                    },
                    "runAsNonRoot": false,
                    "gateway": {
                      "host": "aqua-gateway",
                      "port": "8443"
                    },
                    "login": {
                      "host": "aqua-gateway:8443",
                      "password": "<<YOUR AQUA USER PASSWORD>>",
                      "token": "<<YOUR AQUA CLOUD CONNECTOR TOKEN>>",
                      "username": "<<YOUR AQUA USER NAME>>"
                    },
                    "tunnels": [
                      {
                        "host": "<account>.azurecr.io",
                        "port": "443"
                      },
                      {
                        "host": "gke.gcr.io",
                        "port": "443"
                      },
                      {
                        "host": "<account>.jfrog.io",
                        "port": "443"
                      },
                      {
                        "host": "1.1.1.1",
                        "port": "443"
                      },
                      {
                        "region": "us-east1",
                        "type": "ecr"
                      }
                    ]
                  }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Security
          certified: "false"
          containerImage: docker.io/aquasec/aqua-operator:2022.4.18
          createdAt: "2020-03-25T08:00:00Z"
          description: The Aqua Security Operator runs within a Openshift cluster
            and provides a means to deploy and manage Aqua Security cluster and components.
          operators.operatorframework.io/internal-objects: '["aquastarboards.aquasecurity.github.io","configauditreports.aquasecurity.github.io",
            "clusterconfigauditreports.aquasecurity.github.io"]'
          repository: https://github.com/aquasecurity/aqua-operator
          support: Aqua Security, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Aqua Security CSP Deployment with Aqua Operator
            displayName: AquaCsp
            kind: AquaCsp
            name: aquacsps.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Lightning Enforcers Deployment with Aqua Operator
            displayName: AquaLightningEnforcers
            kind: AquaLightning
            name: aqualightnings.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Database Deployment with Aqua Operator
            displayName: AquaDatabase
            kind: AquaDatabase
            name: aquadatabases.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Enforcer Deployment with Aqua Operator
            displayName: AquaEnforcer
            kind: AquaEnforcer
            name: aquaenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security KubeEnforcer Deployment with Aqua Operator
            displayName: AquaKubeEnforcer
            kind: AquaKubeEnforcer
            name: aquakubeenforcers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Gateway Deployment with Aqua Operator
            displayName: AquaGateway
            kind: AquaGateway
            name: aquagateways.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Scanner Deployment with Aqua Operator
            displayName: AquaScanner
            kind: AquaScanner
            name: aquascanners.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Security Server Deployment with Aqua Operator
            displayName: AquaServer
            kind: AquaServer
            name: aquaservers.operator.aquasec.com
            version: v1alpha1
          - description: Aqua Starboard for kube-enforcer.
            displayName: AquaStarboard
            kind: AquaStarboard
            name: aquastarboards.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ClusterConfigAuditReport
            kind: ClusterConfigAuditReport
            name: clusterconfigauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: For Aqua Starboard.
            displayName: ConfigAuditReport
            kind: ConfigAuditReport
            name: configauditreports.aquasecurity.github.io
            version: v1alpha1
          - description: Aqua Security Cloud Connector Deployment with Aqua Operator
            displayName: AquaCloudConnector
            kind: AquaCloudConnector
            name: aquacloudconnector.operator.aquasec.com
            version: v1alpha1
        description: |-
          The Aqua Security Operator runs within an OpenShift cluster and provides a means to deploy and manage the Aqua Security cluster and components :
          * Server (aka “console”)
          * Database (for production environments we recommend to use an external database and not the Aqua default database)
          * Gateway
          * Enforcer (aka “agent”)
          * KubeEnforcer
          * Scanner
          * CSP (package that contains the Server, Database, and Gateway)
          * LightningEnforcer (package that contains Enforcer and Kube Enforcer)

          Use the Aqua-Operator to
          * Deploy Aqua Security components on OpenShift
          * Scale up Aqua Security components with extra replicas
          * Assign metadata tags to Aqua Security components

          ## Before You Begin Using the Operator CRDs

          Obtain access to the Aqua registry - https://www.aquasec.com/about-us/contact-us/


          ## Instructions and exmaples

          Follow instructions and exmaples
          [here](https://github.com/aquasecurity/aqua-operator/blob/2022.4.0/docs/DeployOpenShiftOperator.md)
        displayName: Aqua Security Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - aqua-security
        - scanning
        - security
        - runtime-security
        links:
        - name: Aqua Security
          url: https://www.aquasec.com/
        - name: Aqua Operator Github
          url: https://github.com/aquasecurity/aqua-operator
        maintainers:
        - email: support@aquasec.com
          name: Aqua Support, Aqua Security
        maturity: alpha
        provider:
          name: Aqua Security, Inc.
        relatedImages:
        - docker.io/aquasec/aqua-operator:2022.4.18
        - quay.io/community-operator-pipeline-prod/aqua@sha256:ca2b998fb0fa4e80d6f4c3be57041e8e55e43cf0b5f1fee579733fda9ea6a107
        version: 2022.4.18
      entries:
      - name: aqua-operator.v2022.4.18
        version: 2022.4.18
      - name: aqua-operator.v2022.4.17
        version: 2022.4.17
      name: 2022.4.0
    defaultChannel: 2022.4.0
    packageName: aqua
    provider:
      name: Aqua Security, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ses-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ses-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
          createdAt: "2025-09-19T19:37:25Z"
          description: AWS SES controller is a service controller for managing SES
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ConfigurationSet represents the state of an AWS ses ConfigurationSet
              resource.
            displayName: ConfigurationSet
            kind: ConfigurationSet
            name: configurationsets.ses.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Simple Email Service (SES) resources in AWS from within your Kubernetes cluster.

          **About Amazon SES**

          Amazon Simple Email Service (SES) is an email platform that provides an easy, cost-effective way for you to send and receive email using your own email addresses and domains.
          For example, you can send marketing emails such as special offers, transactional emails such as order confirmations, and other types of correspondence such as newsletters. When you use Amazon SES to receive mail, you can develop software solutions such as email autoresponders, email unsubscribe systems, and applications that generate customer support tickets from incoming emails.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon SES
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ses
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon SES Developer Resources
          url: https://aws.amazon.com/ses/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ses maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ses-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ses-controller@sha256:0cdbea8336c2825bf1bc8dbd3f5235ff41d97fbed30106e7775c6e993f92d8b6
        version: 1.1.1
      entries:
      - name: ack-ses-controller.v1.1.1
        version: 1.1.1
      - name: ack-ses-controller.v1.1.0
        version: 1.1.0
      - name: ack-ses-controller.v1.0.13
        version: 1.0.13
      - name: ack-ses-controller.v1.0.12
        version: 1.0.12
      - name: ack-ses-controller.v1.0.11
        version: 1.0.11
      - name: ack-ses-controller.v1.0.10
        version: 1.0.10
      - name: ack-ses-controller.v1.0.9
        version: 1.0.9
      - name: ack-ses-controller.v1.0.8
        version: 1.0.8
      - name: ack-ses-controller.v1.0.7
        version: 1.0.7
      - name: ack-ses-controller.v1.0.6
        version: 1.0.6
      - name: ack-ses-controller.v1.0.5
        version: 1.0.5
      - name: ack-ses-controller.v1.0.3
        version: 1.0.3
      - name: ack-ses-controller.v1.0.2
        version: 1.0.2
      - name: ack-ses-controller.v1.0.1
        version: 1.0.1
      - name: ack-ses-controller.v1.0.0
        version: 1.0.0
      - name: ack-ses-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-ses-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: observability-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: candidate
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      - name: observability-operator.v0.0.22
        version: 0.0.22
      name: development
    - currentCSV: observability-operator.v0.0.28
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "MonitoringStack",
                "metadata": {
                  "labels": {
                    "mso": "example"
                  },
                  "name": "sample-monitoring-stack"
                },
                "spec": {
                  "logLevel": "debug",
                  "resourceSelector": {
                    "matchLabels": {
                      "app": "demo"
                    }
                  },
                  "retention": "1d"
                }
              },
              {
                "apiVersion": "monitoring.rhobs/v1alpha1",
                "kind": "ThanosQuerier",
                "metadata": {
                  "name": "example-thanos"
                },
                "spec": {
                  "selector": {
                    "matchLabels": {
                      "mso": "example"
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          containerImage: quay.io/rhobs/observability-operator:0.0.28
          createdAt: "2023-11-09T07:43:16Z"
          description: A Go based Kubernetes operator to setup and manage highly available
            Monitoring Stack using Prometheus, Alertmanager and Thanos Querier.
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/internal-objects: |-
            [
              "prometheuses.monitoring.rhobs", "alertmanagers.monitoring.rhobs",
              "thanosrulers.monitoring.rhobs"
            ]
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/rhobs/observability-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AlertmanagerConfig configures the Prometheus Alertmanager,
              specifying how alerts should be grouped, inhibited and notified to external
              systems
            displayName: AlertmanagerConfig
            kind: AlertmanagerConfig
            name: alertmanagerconfigs.monitoring.rhobs
            version: v1alpha1
          - description: Alertmanager describes an Alertmanager cluster
            displayName: Alertmanager
            kind: Alertmanager
            name: alertmanagers.monitoring.rhobs
            version: v1
          - description: MonitoringStack is the Schema for the monitoringstacks API
            displayName: MonitoringStack
            kind: MonitoringStack
            name: monitoringstacks.monitoring.rhobs
            version: v1alpha1
          - description: PodMonitor defines monitoring for a set of pods
            displayName: PodMonitor
            kind: PodMonitor
            name: podmonitors.monitoring.rhobs
            version: v1
          - description: Probe defines monitoring for a set of static targets or ingresses
            displayName: Probe
            kind: Probe
            name: probes.monitoring.rhobs
            version: v1
          - description: PrometheusAgent defines a Prometheus agent deployment
            displayName: PrometheusAgent
            kind: PrometheusAgent
            name: prometheusagents.monitoring.rhobs
            version: v1alpha1
          - description: Prometheus defines a Prometheus deployment
            displayName: Prometheus
            kind: Prometheus
            name: prometheuses.monitoring.rhobs
            version: v1
          - description: PrometheusRule defines recording and alerting rules for a
              Prometheus instance
            displayName: PrometheusRule
            kind: PrometheusRule
            name: prometheusrules.monitoring.rhobs
            version: v1
          - description: ScrapeConfig defines a namespaced Prometheus scrape_config
              to be aggregated across multiple namespaces into the Prometheus configuration
            displayName: ScrapeConfig
            kind: ScrapeConfig
            name: scrapeconfigs.monitoring.rhobs
            version: v1alpha1
          - description: ServiceMonitor defines monitoring for a set of services
            displayName: ServiceMonitor
            kind: ServiceMonitor
            name: servicemonitors.monitoring.rhobs
            version: v1
          - description: ThanosQuerier outlines the Thanos querier components, managed
              by this stack
            displayName: ThanosQuerier
            kind: ThanosQuerier
            name: thanosqueriers.monitoring.rhobs
            version: v1alpha1
          - description: ThanosRuler defines a ThanosRuler deployment
            displayName: ThanosRuler
            kind: ThanosRuler
            name: thanosrulers.monitoring.rhobs
            version: v1
        description: |2+

          Observability Operator is a Go based Kubernetes operator to setup and manage highly available Monitoring stack using Prometheus, Alertmanager and Thanos Querier.
          ### Supported Features
          - Setup multiple Highly Available Monitoring stack using Prometheus, Alertmanager and Thanos Querier
          - Customizable configuration for managing Prometheus deployments
          - Customizable configuration for managing Alertmanager deployments
          - Customizable configuration for managing Thanos Querier deployments
          ### Documentation
          - **[Documentation](https://github.com/rhobs/observability-operator/tree/main/docs)**
          - **[Developer Docs](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md)**
          ### Contributing
          **[See Contribution Guidelines](https://github.com/rhobs/observability-operator/blob/main/docs/developer.md#contributions)** to contribute to the project by
          - Raising Bugs and Feature Requests
          - Fixing Issues by opening a Pull Request
          - Improving Documentation
          ###  License
          Licensed under the [Apache License, Version 2.0](http://www.apache.org/licenses/LICENSE-2.0)

        displayName: Observability Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - observability
        - monitoring
        - prometheus
        - thanos
        links:
        - name: Observability Operator
          url: https://rhobs-handbook.netlify.app/products/observability-operator
        - name: GitHub
          url: https://github.com/rhobs/observability-operator
        maintainers:
        - email: jfajersk@redhat.com
          name: Jan Fajerski
        - email: sthaha@redhat.com
          name: Sunil Thaha
        - email: spasquie@redhat.com
          name: Simon Pasquier
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - quay.io/openshift-community-operators/observability-operator@sha256:182b95a93db9b95606405c24686cbb96054efb0c2236822569f30b6c27992534
        - quay.io/rhobs/obo-admission-webhook:v0.69.0-rhobs1
        - quay.io/rhobs/obo-prometheus-operator:v0.69.0-rhobs1
        - quay.io/rhobs/observability-operator:0.0.28
        version: 0.0.28
      entries:
      - name: observability-operator.v0.0.28
        version: 0.0.28
      - name: observability-operator.v0.0.25
        version: 0.0.25
      name: stable
    defaultChannel: stable
    packageName: observability-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: rhpam-kogito-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: rhpam-kogito-operator.7.13.5-4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoBuild",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "gitSource": {
                    "contextDir": "ruleunit-quarkus-example",
                    "uri": "https://github.com/kiegroup/kogito-examples"
                  },
                  "type": "RemoteSource"
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoInfra",
                "metadata": {
                  "name": "kogitoinfra-sample"
                },
                "spec": {
                  "resource": {
                    "apiVersion": "kafka.strimzi.io/v1beta2",
                    "kind": "Kafka",
                    "name": "kogito-kafka"
                  }
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoRuntime",
                "metadata": {
                  "name": "ruleunit-quarkus-example"
                },
                "spec": {
                  "replicas": 1
                }
              },
              {
                "apiVersion": "rhpam.kiegroup.org/v1",
                "kind": "KogitoSupportingService",
                "metadata": {
                  "name": "kogito-jobs-service"
                },
                "spec": {
                  "replicas": 1,
                  "serviceType": "JobsService"
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
          description: RHPAM Kogito Operator for deployment and management of Kogito
            services.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: '["Red Hat Process Automation
            Manager", "Red Hat Decision Manager"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kiegroup/rhpam-kogito-operator
          support: Red Hat
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KogitoBuild handles how to build a custom Kogito service
              in a Kubernetes/OpenShift cluster.
            displayName: Kogito Build
            kind: KogitoBuild
            name: kogitobuilds.rhpam.kiegroup.org
            version: v1
          - description: "KogitoInfra is the resource to bind a Custom Resource (CR)
              not managed by Kogito Operator to a given deployed Kogito service. \n
              It holds the reference of a CR managed by another operator such as Strimzi.
              For example: one can create a Kafka CR via Strimzi and link this resource
              using KogitoInfra to a given Kogito service (custom or supporting, such
              as Data Index). \n Please refer to the Kogito Operator documentation
              (https://docs.jboss.org/kogito/release/latest/html_single/) for more
              information."
            displayName: Kogito Infra
            kind: KogitoInfra
            name: kogitoinfras.rhpam.kiegroup.org
            version: v1
          - description: KogitoRuntime is a custom Kogito service.
            displayName: Kogito Runtime
            kind: KogitoRuntime
            name: kogitoruntimes.rhpam.kiegroup.org
            version: v1
          - description: KogitoSupportingService deploys the Supporting service in
              the given namespace.
            displayName: Kogito Supporting Service
            kind: KogitoSupportingService
            name: kogitosupportingservices.rhpam.kiegroup.org
            version: v1
        description: |-
          RHPAM Kogito Operator is designed for deployment and management of Kogito services from source (only on OpenShift) and every piece of infrastructure that the services might need:

          * Builds a RHPAM Kogito service from a remote git repository and creates a custom image for the given runtime: Red Hat build of Quarkus or Springboot. [See more](https://docs.jboss.org/kogito/release/latest/html_single/#con-kogito-operator-deployment-options_kogito-deploying-on-openshift).
          * Deploy RHPAM Kogito service from user provided image.
          * Setup Grafana Dashboard when Grafana Operator is available in cluster.
          * Create Prometheus monitor object when Prometheus operator is available in cluster.

          **IMPORTANT!** RHPAM Kogito Operator may need Grafana and Prometheus Operator to configure monitoring in cases if you do not provide these services beforehand. If you need these features to be automatically provisioned for you, bear in mind that those operators must be installed in the same namespace as the RHPAM Kogito Operator.
        displayName: RHPAM Kogito Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloud
        - kogito
        - quarkus
        - springboot
        - operator
        - native
        - automation
        links:
        - name: Product Page
          url: https://access.redhat.com/products/red-hat-process-automation-manager
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/red_hat_process_automation_manager/7.13/html/getting_started_with_red_hat_build_of_kogito_in_red_hat_process_automation_manager/assembly-deploying-kogito-microservices-on-openshift
        - name: Source
          url: https://github.com/kiegroup/kogito-operator/
        maintainers:
        - email: bsig-cloud@redhat.com
          name: Red Hat
        maturity: beta
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:9492c67eaa7ab37df6f5fa623c4aac955b7d7ffaab9ae2e57dea745e82907e40
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator-bundle@sha256:0192000a2994a62a904ce4aabfc1ff04caf2fa9d779d27c016922b9121877fac
        - registry.redhat.io/rhpam-7/rhpam-kogito-rhel8-operator@sha256:19497d4aff0ea3ea114dccb0d0e265458bb14ba6d325353eb82b554547dbd2e4
        version: 7.13.5-4
      entries:
      - name: rhpam-kogito-operator.7.13.5-4
        version: 7.13.5-4
      - name: rhpam-kogito-operator.7.13.5-3
        version: 7.13.5-3
      - name: rhpam-kogito-operator.7.13.5-2
        version: 7.13.5-2
      - name: rhpam-kogito-operator.7.13.5-1
        version: 7.13.5-1
      - name: rhpam-kogito-operator.7.13.4-3
        version: 7.13.4-3
      - name: rhpam-kogito-operator.7.13.4-2
        version: 7.13.4-2
      - name: rhpam-kogito-operator.7.13.4-1
        version: 7.13.4-1
      - name: rhpam-kogito-operator.7.13.3-1
        version: 7.13.3-1
      - name: rhpam-kogito-operator.7.13.2-3
        version: 7.13.2-3
      - name: rhpam-kogito-operator.7.13.2-2
        version: 7.13.2-2
      - name: rhpam-kogito-operator.7.13.2-1
        version: 7.13.2-1
      - name: rhpam-kogito-operator.7.13.1-2
        version: 7.13.1-2
      - name: rhpam-kogito-operator.7.13.1-1
        version: 7.13.1-1
      - name: rhpam-kogito-operator.7.13.0-1
        version: 7.13.0-1
      - name: rhpam-kogito-operator.v7.12.0-4
        version: 7.12.0-4
      - name: rhpam-kogito-operator.v7.12.0-3
        version: 7.12.0-3
      - name: rhpam-kogito-operator.v7.12.0-2
        version: 7.12.0-2
      name: 7.x
    defaultChannel: 7.x
    packageName: rhpam-kogito-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: SynaXG Engineering Team
      provider-url: ""
    name: synaxg-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: synaxg-operator.v1.2.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Reset",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3reset-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3reset",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3reset-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753",
                  "sg3ResetList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0"
                        }
                      ]
                    }
                  ]
                }
              },
              {
                "apiVersion": "card.synaxg.app.com/v1",
                "kind": "SG3Update",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "synaxg-operator",
                    "app.kubernetes.io/instance": "sg3update-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "sg3update",
                    "app.kubernetes.io/part-of": "synaxg-operator"
                  },
                  "name": "sg3update-sample"
                },
                "spec": {
                  "image": "registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1",
                  "sg3UpdateList": [
                    {
                      "nodeName": "test",
                      "pciList": [
                        {
                          "enable": "false",
                          "pciAddr": "000d:01:00.0",
                          "sdkImageAddr": "quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9"
                        }
                      ]
                    }
                  ]
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
          createdAt: "2025-07-01T03:42:07Z"
          description: SynaXG Operator will work with the SuperMicro Server + SG3
            card  and Openshift Container Platform.The Marvell kernel module “octeon_ep”
            and “octeon_ep_vf” are necessary to the Operator. The Operator can send
            command of “reset” and “update” to SG3 board via the VF channel.As a result,
            users can realize the process of “reset card” and “update firmware” on
            the web GUI, and monitor the executing result.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: free text
          operators.operatorframework.io/builder: operator-sdk-v1.31.0-ocp
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          support: SynaXG
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SG3Reset is the Schema for the sg3resets API
            displayName: SG3 Reset
            kind: SG3Reset
            name: sg3resets.card.synaxg.app.com
            version: v1
          - description: SG3Update is the Schema for the sg3updates API
            displayName: SG3 Update
            kind: SG3Update
            name: sg3updates.card.synaxg.app.com
            version: v1
        description: SynaXG Operator will work with the SuperMicro Server + SG3 card  and
          Openshift Container Platform.The Marvell kernel module “octeon_ep” and “octeon_ep_vf”
          are necessary to the Operator. The Operator can send command of “reset”
          and “update” to SG3 board via the VF channel.As a result, users can realize
          the process of “reset card” and “update firmware” on the web GUI, and monitor
          the executing result.
        displayName: SynaXG vDU card 102X Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - vDU
        - 102X
        - SynaXG
        links:
        - name: Synaxg Operator
          url: https://synaxg-operator.domain
        maintainers:
        - email: notice@synaxg.com
          name: Mantosh Malhotra
        maturity: alpha
        provider:
          name: SynaXG Engineering Team
        relatedImages:
        - registry.connect.redhat.com/synaxg/synaxg-operator@sha256:ff3d9ab5506c4487e5b6697ffe960584aadfe9f16408e7976d345c3d93375171
        - registry.connect.redhat.com/synaxg/synaxg-reset-utility@sha256:4050d5731a21bbec6e26165b8dc2b7632577ffa376e7eb056aa3245cf9574753
        - registry.connect.redhat.com/synaxg/synaxg-upgrade-utility@sha256:7ffec3249edd2467f2414c2ce4232e8e436278fd1b43a51b66f54b6ea10a60d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:38ee3484ba60bb474e7231e9d0718157386716cea7ffd6da6f32ddd551a5243d
        - quay.io/synaxgcom/sdk-img@sha256:1ed57013f278e5a858f71f343649652dbd52b6f3cf451374e45b6ac427fa76f9
        - registry.connect.redhat.com/synaxg/synaxg-operator-utility@sha256:3b22a8022a97cb1ffdb47ffcc9d9d562601f48669aa14fd58cd8ba6128b39bd7
        version: 1.2.0
      entries:
      - name: synaxg-operator.v1.2.0
        version: 1.2.0
      name: alpha
    defaultChannel: alpha
    packageName: synaxg-operator
    provider:
      name: SynaXG Engineering Team
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Anchore Inc.
      provider-url: ""
    name: anchore-engine
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: anchore-engine.v1.0.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.anchore.io/v1alpha1",
                "kind": "AnchoreEngine",
                "metadata": {
                  "name": "anchoreengine-sample"
                },
                "spec": {
                  "anchoreGlobal": {
                    "defaultAdminEmail": "example@email.com"
                  },
                  "postgresql": {
                    "extraEnv": [
                      {
                        "name": "POSTGRESQL_USER",
                        "value": "anchoreengine"
                      },
                      {
                        "name": "POSTGRESQL_PASSWORD",
                        "value": "anchore-postgres,123"
                      },
                      {
                        "name": "POSTGRESQL_DATABASE",
                        "value": "anchore"
                      },
                      {
                        "name": "PGUSER",
                        "value": "postgres"
                      }
                    ],
                    "postgresDatabase": "anchore",
                    "postgresPassword": "anchore-postgres,123",
                    "postgresUser": "anchoreengine"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "true"
          containerImage: registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
          createdAt: "2021-10-06T23:53:51Z"
          description: Anchore Engine - container image scanning service for policy-based
            security, best-practice and compliance enforcement.
          operators.operatorframework.io/builder: operator-sdk-v1.13.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/anchore/engine-operator
          support: anchore.slack.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Deploys Anchore Engine
            displayName: Anchore Engine Operator
            kind: AnchoreEngine
            name: anchoreengines.charts.anchore.io
            version: v1alpha1
        description: "Anchore Engine is an open source software system that provides
          a centralized service for analyzing container images, \n scanning for security
          vulnerabilities, and enforcing deployment policies. Anchore Engine is provided
          as a Docker \n container image that can be run standalone or with an orchestration
          platform such as Kubernetes. Anchore Engine allows \n users to perform detailed
          analysis on their container images, run queries, produce reports, and define
          policies that \n can be used in CI/CD pipelines. Users can extend Anchore
          Engine with plugins that add new queries, image analysis, and \n policies.
          Anchore Engine can be accessed directly through a RESTful API or via the
          Anchore CLI.\n\n The Anchore Engine Operator is based on the official stable
          [Helm Chart](https://github.com/anchore/anchore-charts/tree/master/stable/anchore-engine)."
        displayName: Anchore Engine Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - analysis
        - docker
        - anchore
        - anchore-engine
        - image
        - security
        - vulnerability
        - scanner
        links:
        - name: Anchore Inc.
          url: https://www.anchore.com
        - name: Anchore Engine
          url: https://github.com/anchore/anchore-engine
        maintainers:
        - email: bradyt@anchore.com
          name: Brady Todhunter
        - email: zach@anchore.com
          name: Zach Hill
        maturity: alpha
        provider:
          name: Anchore Inc.
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-96@sha256:e8177c5de05ccdd6d12e360b65c63c889b2de725d573ffdedd8914230b118639
        - registry.connect.redhat.com/anchore/engine-operator-bundle@sha256:fbbe7e6c1d75c4de2f47e2c825c930568e85f1134545e9d890a0c9f3d9187a4d
        - registry.connect.redhat.com/anchore/engine-operator@sha256:baf0f5acd319bc4d0132c88922ee3f357fdcdc278d931f243d0cbec3f6d416c6
        - registry.connect.redhat.com/anchore/engine0@sha256:68e09d5ff695b46f24de3d7610f88dfc14b77b38758bbd68d03350dbae3deb0f
        version: 1.0.1
      entries:
      - name: anchore-engine.v1.0.1
        version: 1.0.1
      name: alpha
    defaultChannel: alpha
    packageName: anchore-engine
    provider:
      name: Anchore Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-airflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: airflow-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
          description: Stackable Operator for Apache Airflow
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/airflow-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for AirflowClusterSpec via `CustomResource`
            displayName: airflowclusters.airflow.stackable.tech
            kind: AirflowCluster
            name: airflowclusters.airflow.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Airflow
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - airflow
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/airflow-operator@sha256:8450e6398c0898efac66ddf57929a714619d2bd45396369b73bcaeeebf61eaa0
        - registry.connect.redhat.com/stackable/stackable-apache-airflow-operator@sha256:ffee816bc94bea75ed3e393c6e2f4378ed77df1e8988ada9583b429d3e8328b8
        version: 25.7.0
      entries:
      - name: airflow-operator.v25.7.0
        version: 25.7.0
      - name: airflow-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-airflow-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibmcloud-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibmcloud-operator.v1.1.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Service","metadata":{"annotations":{"ibmcloud.ibm.com/self-healing":"enabled"},"name":"mytranslator"},"spec":{"plan":"lite","serviceClass":"language-translator"}},{"apiVersion":"ibmcloud.ibm.com/v1","kind":"Binding","metadata":{"name":"binding-translator"},"spec":{"parameters":[{"name":"serviceid_crn","value":"crn:v1:bluemix:public:iam-identity::a/ACCOUNT_ID::serviceid:SERVICE_ID"}],"secretName":"translator-secret","serviceName":"mytranslator"}}]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: cloudoperators/ibmcloud-operator:1.1.0
          createdAt: "2022-03-17T04:17:40Z"
          description: The IBM Cloud Operator provides a Kubernetes CRD-Based API
            to manage the lifecycle of IBM public cloud services.
          repository: https://github.com/IBM/cloud-operators
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Binding is an instance of a service binding resource on IBM
              Cloud. A Binding creates a secret with the service instance credentials.
            displayName: Binding
            kind: Binding
            name: bindings.ibmcloud.ibm.com
            version: v1
          - description: Service is an instance of a Service resource on IBM Cloud.
            displayName: Service
            kind: Service
            name: services.ibmcloud.ibm.com
            version: v1
        description: |-
          With the IBM Cloud Operator, you can provision and bind [IBM public cloud services](https://cloud.ibm.com/catalog#services) to your Kubernetes cluster in a Kubernetes-native way. The IBM Cloud Operator is based on the [Kubernetes custom resource definition (CRD) API](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) so that your applications can create, update, and delete IBM Cloud services from within the cluster by calling Kubnernetes APIs, instead of needing to use several IBM Cloud APIs in addition to configuring your app for Kubernetes.


          ## Features

          * **Service provisioning**: Create any service with any plan that is available in the [IBM Cloud catalog](https://cloud.ibm.com/catalog#services).

          * **Bindings management**: Automatically create Kubernetes secrets in your Kubernetes cluster with the credentials to bind to any provisioned service to the cluster.


          ## Upgrading the operator


          To upgrade an OperatorHub installation, use the OpenShift web console to perform the upgrade. If upgrading from the `alpha` channel subscription to `stable`, follow the additional upgrade instructions below.

          ### Upgrading to version v0.3 or v1.0
          **IMPORTANT NOTICE:** v0.1 and v0.2 used a different naming scheme for secrets and configmaps. Before you update the IBM Cloud Operator, create secret and configmap resources with these names and copy the contents of your previous resources to the new resources. Then, the upgraded operator recognizes and continues to update the resources.

          Existing `Service` and `Binding` resources **do not** need modification for the upgrade.

          | Previous names (v0.1 and v0.2)         | **Current names (v0.3 and v1.0)**             | Description                                                                           |
          |:---------------------------------------|:--------------------------------------------|:--------------------------------------------------------------------------------------|
          | secret-ibm-cloud-operator              | **ibmcloud-operator-secret**                | Secret with the API key, scoped to the namespace.                                     |
          | config-ibm-cloud-operator              | **ibmcloud-operator-defaults**              | ConfigMap with the default values for new resources.                                  |
          | ibm-cloud-operator                     | **ibmcloud-operator-config**                | ConfigMap with the management namespace configuration.                                |
          | ${namespace}-secret-ibm-cloud-operator | **${namespace}-ibmcloud-operator-secret**   | Management namespace Secret with the API key for ${namespace}.                        |
          | ${namespace}-config-ibm-cloud-operator | **${namespace}-ibmcloud-operator-defaults** | Management namespace ConfigMap with default values for new resources in ${namespace}. |

          **Tip:** Forgot to update the secret and configmap names before upgrading? The operator will not take action on Service and Binding resources until the new secrets and configmaps have been created. Creating these after the upgrade will also work.


          ### OperatorHub stable channel
          Continuous updates and bug fixes are provided in the latest stable channel. Subscribing to the `stable` channel in OperatorHub means you automatically get the latest updates without breaking backward compatibility.

          ### Changelog

          For a list of recent changes, see the [releases page](https://github.com/IBM/cloud-operators/releases).


          ## Prerequisites

          1.  Have an [IBM Cloud account](https://cloud.ibm.com/registration).
          2.  Have a cluster that runs Kubernetes version 1.11 or later (OpenShift 3.11 or later).
          3.  Install the required command line tools.
              *   [IBM Cloud CLI](https://cloud.ibm.com/docs/cli?topic=cloud-cli-getting-started) (`ibmcloud`)
              *   [Kubernetes CLI](https://kubernetes.io/docs/tasks/tools/install-kubectl/) (`kubectl`)
          4.  Log in to your IBM Cloud account from the CLI.

              ```bash
              ibmcloud login
              ```

          5.  Target the appropriate resource group (`-g`) and default region (`-r`) for provisioning services:

              ```bash
              ibmcloud target -g default -r us-south
              ```

          6.  If you also want to provision Cloud Foundry services, target the appropriate `org` and `space` (`--cf`).

              ```bash
              ibmcloud target --cf
              ```

          7.  Set the Kubernetes context of your CLI to your cluster so that you can run `kubectl` commands. For example, if your cluster runs OpenShift, use the `oc login` command.


          ## Setting up the operator

          You can use an installation script to set up the IBM Cloud Operator.
          By default, the installation script stores an API key in a Kubernetes secret in your cluster that can be accessed by the IBM Cloud Operator.
          Next, the script sets default values that are used to provision IBM Cloud services, like the resource group and region to provision the services in. Later, you can override any default value in the `Service` custom resource. Finally, the script deploys the operator in your cluster.

          To use your own API key, set the `IBMCLOUD_API_KEY` environment variable to the key before running the installation script:

              export IBMCLOUD_API_KEY="CUSTOM_API_KEY"

          If installed with Operator Hub, the operator will run in the `openshift-operators` namespace. Otherwise, it will run in the `ibmcloud-operator-system` namespace.

          Prefer to create the secrets and defaults yourself? See the [IBM Cloud Operator installation guide](https://github.com/IBM/cloud-operators/blob/master/docs/install.md).

          ### Using a service ID

          By default, the installation script creates an IBM Cloud API key that impersonates your user credentials, to use to set up the IBM Cloud Operator. However, you might want to create a service ID in IBM Cloud Identity and Access Managment (IAM). By using a service ID, you can control access for the IBM Cloud Operator without having the permissions tied to a particular user, such as if that user leaves the company. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-serviceids).

          1.  Create a service ID in IBM Cloud IAM. If possible, do not use spaces in the names for your IAM credentials. When you use the operator binding feature, any spaces are replaced with underscores.

              ```bash
              ibmcloud iam service-id-create serviceid-ico -d service-ID-for-ibm-cloud-operator
              ```

          2.  Assign the service ID access to the required permissions to work with the IBM Cloud services that you want the operator to manage. The required permissions vary with each IBM Cloud service. You can also scope an access policy to different regions or resource groups. For example, the following command grants the service ID the **Administrator** platform role in the default resource group in the US South (Dallas) region. For more information, see the [IBM Cloud docs](https://cloud.ibm.com/docs/account?topic=account-userroles).

              ```bash
              ibmcloud iam service-policy-create serviceid-ico --roles Administrator --resource-group-name default --region us-south
              ```

          3.  Create an API key for the service ID.

              ```bash
              ibmcloud iam service-api-key-create apikey-ico serviceid-ico -d api-key-for-ibm-cloud-operator
              ```

          4.  Set the API key of the service ID as your CLI environment variable. Now, when you run the installation script, the script uses the service ID's API key. The following command is an example for macOS.

              ```bash
              export IBMCLOUD_API_KEY=<apikey-ico-value>
              ```

          5.  Confirm that the API key environment variable is set in your CLI.

              ```bash
              echo $IBMCLOUD_API_KEY
              ```

          6.  Follow the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to the IBM Cloud account that owns the service ID.


          ### Installing the operator for OpenShift clusters

          Before you begin, complete the [prerequisite steps](https://github.com/IBM/cloud-operators/blob/master/README.md#prerequisites) to log in to IBM Cloud and your cluster, and optionally set up a [service ID API key](https://github.com/IBM/cloud-operators/blob/master/README.md#using-a-service-id).

          To configure the latest release for OpenShift before installing via the OperatorHub, run the following script:

          *   **Latest release**:

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash
              ```

          *   **Specific release**: Replace `-v 0.0.0` with the specific version that you want to install.

              ```bash
              curl -sL https://raw.githubusercontent.com/IBM/cloud-operators/master/hack/configure-operator.sh | bash -s -- -v 0.0.0 store-creds
              ```


          ## Using the IBM Cloud Operator

          To use the IBM Cloud Operator, create a service instance and then bind the service to your cluster. For more information, see the [sample configuration files](https://github.com/IBM/cloud-operators/blob/master/config/samples), [user guide](https://github.com/IBM/cloud-operators/blob/master/docs/user-guide.md), and [reference documentation](https://github.com/IBM/cloud-operators/blob/master/README.md#reference).

          #### Step 1: Creating a service instance

          1.  To create an instance of an IBM public cloud service, first create a `Service` custom resource file. For more options, see the [Service properties](https://github.com/IBM/cloud-operators/blob/master/README.md#service-properties) reference doc.
              *   `<SERVICE_CLASS>` is the IBM Cloud service that you want to create. To list IBM Cloud services, run `ibmcloud catalog service-marketplace` and use the **Name** value from the output.
              *   `<PLAN>` is the plan for the IBM Cloud service that you want to create, such as `free` or `standard`. To list available plans, run `ibmcloud catalog service <SERVICE_CLASS> | grep plan`.

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Service
              metadata:
                  name: myservice
              spec:
                  plan: <PLAN>
                  serviceClass: <SERVICE_CLASS>
              ```

          2.  Create the service instance in your cluster.

              ```bash
              kubectl apply -f filepath/myservice.yaml
              ```

          3.  Check that your service status is **Online** in your cluster.

              ```bash
              kubectl get services.ibmcloud
              NAME           STATUS   AGE
              myservice      Online   12s
              ```

          4.  Verify that your service instance is created in IBM Cloud.

              ```bash
              ibmcloud resource service-instances | grep myservice
              ```

          #### Step 2: Binding the service instance

          1.  To bind your service to the cluster so that your apps can use the service, create a `Binding` custom resource, where the `serviceName` field is the name of the `Service` custom resource that you previously created. For more options, see [Binding properties](https://github.com/IBM/cloud-operators/blob/master/README.md#binding-properties).

              ```yaml
              apiVersion: ibmcloud.ibm.com/v1
              kind: Binding
              metadata:
                  name: mybinding
              spec:
                  serviceName: myservice
              ```

          2.  Create the binding in your cluster.

              ```bash
              kubectl apply -f filepath/mybinding.yaml
              ```

          3.  Check that your service status is **Online**.

              ```bash
              kubectl get bindings.ibmcloud
              NAME         STATUS   AGE
              mybinding    Online   25s
              ```

          4.  Check that a secret of the same name as your binding is created. The secret contains the service credentials that apps in your cluster can use to access the service.

              ```bash
              kubectl get secrets
              NAME         TYPE      DATA   AGE
              mybinding    Opaque    6      102s
              ```
        displayName: IBM Cloud Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: Source on GitHub
          url: https://github.com/ibm/cloud-operators
        maintainers:
        - email: johnstarich@johnstarich.com
          name: John Starich
        - email: mvaziri@us.ibm.com
          name: vazirim
        - email: greg.lanthier@ca.ibm.com
          name: Greg Lanthier
        - email: halacs@users.noreply.github.com
          name: Halacs
        - email: artberger@users.noreply.github.com
          name: Art
        maturity: alpha
        minKubeVersion: 1.16.0
        provider:
          name: IBM
        relatedImages:
        - cloudoperators/ibmcloud-operator:1.1.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.5.0
        - quay.io/openshift-community-operators/ibmcloud-operator@sha256:b22e89daf6b7f0fd55ba6db7eb7d7a48cbaef2844dfe20e558d7b821482e9480
        version: 1.1.0
      entries:
      - name: ibmcloud-operator.v1.1.0
        version: 1.1.0
      name: stable
    defaultChannel: stable
    packageName: ibmcloud-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-cloudfront-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-cloudfront-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "CachePolicy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Distribution",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "cloudfront.services.k8s.aws/v1alpha1",
                "kind": "Function",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
          createdAt: "2025-09-19T21:12:14Z"
          description: AWS CloudFront controller is a service controller for managing
            CloudFront resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CachePolicy represents the state of an AWS cloudfront CachePolicy
              resource.
            displayName: CachePolicy
            kind: CachePolicy
            name: cachepolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Distribution represents the state of an AWS cloudfront Distribution
              resource.
            displayName: Distribution
            kind: Distribution
            name: distributions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: Function represents the state of an AWS cloudfront Function
              resource.
            displayName: Function
            kind: Function
            name: functions.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginAccessControl represents the state of an AWS cloudfront
              OriginAccessControl resource.
            displayName: OriginAccessControl
            kind: OriginAccessControl
            name: originaccesscontrols.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: OriginRequestPolicy represents the state of an AWS cloudfront
              OriginRequestPolicy resource.
            displayName: OriginRequestPolicy
            kind: OriginRequestPolicy
            name: originrequestpolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: ResponseHeadersPolicy represents the state of an AWS cloudfront
              ResponseHeadersPolicy resource.
            displayName: ResponseHeadersPolicy
            kind: ResponseHeadersPolicy
            name: responseheaderspolicies.cloudfront.services.k8s.aws
            version: v1alpha1
          - description: VPCOrigin represents the state of an AWS cloudfront VPCOrigin
              resource.
            displayName: VPCOrigin
            kind: VPCOrigin
            name: vpcorigins.cloudfront.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon CloudFront resources in AWS from within your Kubernetes cluster.

          **About Amazon CloudFront**

          Amazon CloudFront is a web service that speeds up distribution of your static and dynamic web content, such as .html, .css, .js, and image files, to your users. CloudFront delivers your content through a worldwide network of data centers called edge locations.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon CloudFront
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - cloudfront
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon CloudFront Developer Resources
          url: https://aws.amazon.com/cloudfront/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: cloudfront maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/cloudfront-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-cloudfront-controller@sha256:235ad7b4ad07644d9a638909c681620b24a85148a8655b7c903faf6aa42c2242
        version: 1.2.1
      entries:
      - name: ack-cloudfront-controller.v1.2.1
        version: 1.2.1
      - name: ack-cloudfront-controller.v1.2.0
        version: 1.2.0
      - name: ack-cloudfront-controller.v1.1.6
        version: 1.1.6
      - name: ack-cloudfront-controller.v1.1.4
        version: 1.1.4
      - name: ack-cloudfront-controller.v1.1.3
        version: 1.1.3
      - name: ack-cloudfront-controller.v1.1.2
        version: 1.1.2
      - name: ack-cloudfront-controller.v1.1.1
        version: 1.1.1
      - name: ack-cloudfront-controller.v1.1.0
        version: 1.1.0
      - name: ack-cloudfront-controller.v1.0.10
        version: 1.0.10
      - name: ack-cloudfront-controller.v1.0.9
        version: 1.0.9
      - name: ack-cloudfront-controller.v1.0.8
        version: 1.0.8
      - name: ack-cloudfront-controller.v1.0.7
        version: 1.0.7
      - name: ack-cloudfront-controller.v1.0.6
        version: 1.0.6
      - name: ack-cloudfront-controller.v1.0.2
        version: 1.0.2
      - name: ack-cloudfront-controller.v1.0.1
        version: 1.0.1
      - name: ack-cloudfront-controller.v1.0.0
        version: 1.0.0
      - name: ack-cloudfront-controller.v0.0.15
        version: 0.0.15
      - name: ack-cloudfront-controller.v0.0.14
        version: 0.0.14
      - name: ack-cloudfront-controller.v0.0.13
        version: 0.0.13
      - name: ack-cloudfront-controller.v0.0.12
        version: 0.0.12
      - name: ack-cloudfront-controller.v0.0.11
        version: 0.0.11
      - name: ack-cloudfront-controller.v0.0.10
        version: 0.0.10
      - name: ack-cloudfront-controller.v0.0.9
        version: 0.0.9
      - name: ack-cloudfront-controller.v0.0.8
        version: 0.0.8
      - name: ack-cloudfront-controller.v0.0.6
        version: 0.0.6
      - name: ack-cloudfront-controller.v0.0.5
        version: 0.0.5
      - name: ack-cloudfront-controller.v0.0.4
        version: 0.0.4
      - name: ack-cloudfront-controller.v0.0.3
        version: 0.0.3
      - name: ack-cloudfront-controller.v0.0.2
        version: 0.0.2
      - name: ack-cloudfront-controller.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: ack-cloudfront-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Opstree Solutions
      provider-url: ""
    name: mongodb-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: mongodb-operator.v0.3.0
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDB","metadata":{"name":"mongodb"},"spec":{"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}},{"apiVersion":"opstreelabs.in/v1alpha1","kind":"MongoDBCluster","metadata":{"name":"mongodb"},"spec":{"clusterSize":3,"kubernetesConfig":{"image":"quay.io/opstree/mongo:v5.0.6","imagePullPolicy":"IfNotPresent","securityContext":{"fsGroup":1001}},"storage":{"accessModes":["ReadWriteOnce"],"storageSize":"1Gi","storageClass":"gp2"},"mongoDBSecurity":{"mongoDBAdminUser":"admin","secretRef":{"name":"mongodb-secret","key":"password"}}}}]'
          capabilities: Seamless Upgrades
          categories: Database,  OpenShift Optional, Storage
          certified: "false"
          containerImage: quay.io/opstree/mongodb-operator:v0.3.0
          createdAt: 05-04-2022
          description: A golang based operator to create and manage MongoDB standalone
            setup and cluster setup
          k8sMaxVersion: ""
          k8sMinVersion: 1.17.0
          repository: https://github.com/ot-container-kit/mongodb-operator
          support: ""
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Mongo DB Cluster
            displayName: MongoDBCluster
            kind: MongoDBCluster
            name: mongodbclusters.opstreelabs.in
            version: v1alpha1
          - description: Mongo DB
            displayName: MongoDB
            kind: MongoDB
            name: mongodbs.opstreelabs.in
            version: v1alpha1
        description: |+
          MongoDB Operator is an operator created in Golang to create, update, and manage MongoDB standalone, replicated, and arbiter replicated setup on Kubernetes and Openshift clusters. This operator is capable of doing the setup for MongoDB with all the required best practices.

          ### Supported Features

          * MongoDB replicated cluster setup
          * MongoDB standalone setup
          * MongoDB replicated cluster failover and recovery
          * Monitoring support with MongoDB Exporter
          * Password based authentication for MongoDB
          * Kubernetes's resources for MongoDB standalone and cluster
          * Customizable configuration changes in MongoDB
          * Insightful Grafana dashboards

          ### Documentation

          **[Documentation](https://ot-mongodb-operator.netlify.app/)**
          ### Getting Help

          This project is managed by OpsTree Solutions. If you have any queries or suggestions, mail us at opensource@opstree.com.
          ### Contributing
          You can contribute to this project by:-

          * Raising Bugs and Feature Requests
          * Fixing issue by opening a Pull Request
          * Improving Documentation

          ### License

          MongoDB  Operator is licensed under [Apache License, Version 2.0](https://github.com/OT-CONTAINER-KIT/mongodb-operator/blob/master/LICENSE)


        displayName: MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - Kubernetes
        - Database
        - MongoDB
        - Opstree
        - Operatork
        - Cloud Native
        links:
        - name: Blog
          url: https://blog.opstree.com
        - name: GitHub
          url: https://github.com/opstree
        - name: Opensource
          url: https://opstree.github.io
        - name: Documentation
          url: https://ot-mongodb-operator.netlify.app/
        - name: Buildpiper
          url: https://buildpiper.io
        maintainers:
        - email: abhishek.dubey@opstree.com
          name: Abhishek Dubey
        - email: sandeep@opstree.com
          name: Sandeep Rawat
        maturity: alpha
        minKubeVersion: 1.17.0
        provider:
          name: Opstree Solutions
        relatedImages:
        - quay.io/openshift-community-operators/mongodb-operator@sha256:1c06d3e936348ceeed773eb45144c9dd702e50ba293610421a93e3b3942ac9ef
        - quay.io/opstree/mongodb-operator:v0.3.0
        version: 0.3.0
      entries:
      - name: mongodb-operator.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: mongodb-operator
    provider:
      name: Opstree Solutions
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: raftt.io
      provider-url: https:///www.raftt.io
    name: raftt-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: raftt-operator.v1.25.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "raftt.io/v1beta1",
                "kind": "Raftt",
                "metadata": {
                  "name": "raftt",
                  "namespace": "raftt"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Developer Tools
          containerImage: quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
          createdAt: "2025-04-29T11:31:02Z"
          description: This operator installs and configures a Raftt instance
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/raftt-operator/support?utm_source=openshift_console
          operatorframework.io/suggested-namespace: raftt
          operators.openshift.io/valid-subscription: Raftt commercial license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          support: Raftt systems
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Raftt represents an installation of Raftt cluster controller
              & supporting services
            displayName: Raftt
            kind: Raftt
            name: raftts.raftt.io
            version: v1beta1
        description: |
          Raftt allows fast code iterations and debugging over complex Kubernetes environments.
          Raftt provides a localhost development experience for applications running on Kubernetes.
          You can use Raftt to orchestrate the environments for you, so you get the optimal development experience -
          iterations that take under a second, OOTB debugging of your code as it runs in the cluster, easy introspection, and more.
        displayName: Raftt
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - dx
        - development
        - debugging
        links:
        - name: Raftt Operator
          url: https://raftt-operator.domain
        - name: Documentation
          url: https://docs.raftt.io
        maintainers:
        - email: ops@raftt.io
          name: Raftt
        maturity: alpha
        provider:
          name: raftt.io
          url: https:///www.raftt.io
        relatedImages:
        - quay.io/raftt/raftt-operator@sha256:73edc6d8416ee3815d7e59bb26c33e3d35d97cf1ceadd8fa457041be13e175ca
        - registry.connect.redhat.com/raftt/operator-bundle@sha256:7108686c168c35acd996e61a981bdd885b454837cf6823a48a28cdeeebbb0af9
        version: 1.25.2
      entries:
      - name: raftt-operator.v1.25.2
        version: 1.25.2
      name: alpha
    defaultChannel: alpha
    packageName: raftt-operator
    provider:
      name: raftt.io
      url: https:///www.raftt.io
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Ab Initio
      provider-url: ""
    name: abinitio-runtime-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: abinitio-runtime-operator.v4.3.3-1
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/About-running-jobs-in-Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-FOUR-Reference/Co-Operating-System-runtime-operator-API-reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/latest/Co-Operating-System/Co-Operating-System-Guide-to-Cloud-and-Container-Computing/PART-TWO-Running-jobs-in-Kubernetes/Configuring-and-running-Co-Operating-System-jobs-in-Kubernetes/Example-manifests-for-CoOperatingSystemRuntime-jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.27.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:665efa8cc2ba97ad31f219161cdd53e15c1aef16b6c05ac60506be9bbc9ffc9a
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:d0bd55bad99bfd1ce09da5ee885a38229829570ebf80e1acc4341dff17a54e64
        version: 4.3.3-1
      entries:
      - name: abinitio-runtime-operator.v4.3.3-1
        version: 4.3.3-1
      - name: abinitio-runtime-operator.v4.3.1-0
        version: 4.3.1-0
      name: release-4.3
    - currentCSV: abinitio-runtime-operator.v4.2.3-0
      currentCSVDesc:
        annotations:
          capabilities: Basic Install
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: abinitio-system
          operators.operatorframework.io/builder: operator-sdk-v1.8.0-ocp
          operators.operatorframework.io/internal-objects: '["cooperatingsystemruntimepoolmetrics.cloud.abinitio.com",
            "cooperatingsystemruntimeprofiles.cloud.abinitio.com"]'
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Defines a *runtime claim*, an interface for reserving a runtime
              object (CoOperatingSystemRuntime) in a runtime pool (CoOperatingSystemRuntimePool)
              in order to launch a job.
            displayName: Co>Operating System Runtime Claim
            kind: CoOperatingSystemRuntimeClaim
            name: cooperatingsystemruntimeclaims.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Pool Metric
            kind: CoOperatingSystemRuntimePoolMetric
            name: cooperatingsystemruntimepoolmetrics.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime pool*, a set of templated runtime objects
              (CoOperatingSystemRuntime) that are available on demand for running
              jobs. A runtime pool's capacity is configurable and can be scheduled
              in response to usage patterns.
            displayName: Co>Operating System Runtime Pool
            kind: CoOperatingSystemRuntimePool
            name: cooperatingsystemruntimepools.cloud.abinitio.com
            version: v1
          - description: For internal use only.
            displayName: Co>Operating System Runtime Profile
            kind: CoOperatingSystemRuntimeProfile
            name: cooperatingsystemruntimeprofiles.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime object*, a set of coordinated resources
              used to run a data processing job in parallel across a launcher pod
              and a dynamically scaled set of worker pods. Such a job leverages the
              full power of the Co>Operating System for processing complex enterprise
              data.
            displayName: Co>Operating System Runtime
            kind: CoOperatingSystemRuntime
            name: cooperatingsystemruntimes.cloud.abinitio.com
            version: v1
          - description: Defines a *runtime template*, a reusable specification (CoOperatingSystemRuntimeSpec)
              that can be referenced and overridden by runtime objects as needed,
              making it possible to specify jobs concisely. The runtime objects in
              a runtime pool (CoOperatingSystemRuntimePool) are created from a runtime
              template.
            displayName: Co>Operating System Runtime Template
            kind: CoOperatingSystemRuntimeTemplate
            name: cooperatingsystemruntimetemplates.cloud.abinitio.com
            version: v1
        description: |-
          The Co>Operating System runtime operator brings the power of Ab Initio's Co>Operating
          System to Kubernetes by defining a set of custom API resources for provisioning and
          running data processing jobs.

          The operator is installed along with the following custom resource definitions (CRDs):

          - CoOperatingSystemRuntime
          - CoOperatingSystemRuntimeClaim
          - CoOperatingSystemRuntimePool
          - CoOperatingSystemRuntimeTemplate

          In combination, these resources give you significant flexibility in scheduling, running,
          and monitoring your jobs, which can run in parallel across a dynamically scaled set of
          containers. For reference documentation for these resources, see
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference)
          in the Ab Initio Help.

          The following topics describe the basic requirements and procedures for creating custom
          resources for running jobs after you have installed the runtime operator.

          ## Requirements for using custom resources

          The runtime operator is fully configured when you install it in your OpenShift cluster
          through OperatorHub. Importantly, the operator is installed in unrestricted mode,
          which means it can manage the CoOperatingSystemRuntime family of CRDs in any namespace
          in your cluster.

          For creating custom resources in order to run jobs using the operator, you must meet
          the requirements and configure the settings described in the following sections.

          ### Container images for job agents

          To run jobs of kind CoOperatingSystemRuntime, you must have a way to build the following
          OCI container images:

          - A *processing* image with the Co>Operating System server software installed in it.
          - A *sandbox workspace* image with the graphs and other files needed to run your jobs.
          This image is based on the Co>Operating System image and functions as the job's launcher
          agent, which coordinates with a set of worker agents (processing containers) in order to
          run a job in parallel.

          For building and managing these container images, you need the following:

          - A Co>Operating System installation package.
          - Access to a technical repository containing the projects that make up your application.
          - A container-building tool.
          - Access to a private container registry for hosting the images you build. You need pull
          and push access to this registry from your container build context and pull access from
          your OpenShift cluster (for example, via a Kubernetes Secret).

          > WARNING! Never push container images with Ab Initio software to a public registry. Push
          them to a private container registry only.

          The Co>Operating System includes reference Dockerfiles and example scripts for building
          container images. Typically, you build these images as part of your CI/CD pipeline. For
          more information, contact your Ab Initio administrator or Ab Initio Support.

          ### Ab Initio key servers

          The key servers that you use to enable jobs must be reachable from your cluster. You
          specify these key servers via the `AB_HOSTNAME_KEYSERVER_URLS` configuration variable in
          the pod specifications for launcher and worker agents in the CoOperatingSystemRuntime
          resource.

          To discuss the software activation key requirements for jobs that run in Kubernetes,
          contact your key server administrator or Ab Initio Support.

          ### Namespace and role considerations

          The runtime operator is installed with the cluster-wide RBAC roles and role bindings that
          it needs in order to manage custom resources and run jobs.

          You typically use a separate namespace or namespaces for creating and managing custom
          resources to run jobs. Each of these namespaces must be equipped with a service
          account that can list pod limits and get and patch CoOperatingSystemRuntime resources.

          For each namespace in which you intend to create custom resources, create the needed RBAC
          role and role binding by applying the following Role and RoleBinding manifests:

          ```yaml
          apiVersion: rbac.authorization.k8s.io/v1
          kind: Role
          metadata:
            name: cooperatingsystemruntime-role
          rules:
          - apiGroups:
            - ""
            resources:
            - limitranges
            verbs:
            - list
          - apiGroups:
            - cloud.abinitio.com
            resources:
            - cooperatingsystemruntimes
            verbs:
            - get
            - patch
          ---
          apiVersion: rbac.authorization.k8s.io/v1
          kind: RoleBinding
          metadata:
            name: run-cooperatingsystemruntime-rolebinding
          subjects:
          - kind: ServiceAccount
            name: default
          roleRef:
            kind: Role
            name: cooperatingsystemruntime-role
            apiGroup: rbac.authorization.k8s.io
          ```

          One way to do this is to save these manifests to a file and specify a namespace when
          using `oc` to apply them to the cluster. In the following example, the manifests are
          saved to a file named `runtime-rbac.yaml` and the role and role binding are to be created
          for the `default` service account in the `hello-world` namespace:

          ```
          oc -n hello-world apply -f runtime-rbac.yaml
          ```

          To bind the role to a service account other than `default`, simply update the value of
          `subjects[0].name` in the provided RoleBinding manifest.

          ### Dynamic volume provisioning

          The launcher and worker template sections of the CoOperatingSystemRuntime resource
          require you to specify persistent volume claims (PVCs) for job recovery data. To ensure
          that these PVCs can be bound to approriate storage volumes, dynamic volume provisioning
          must be enabled in your cluster.

          Work with your cluster administrator to identify the storage needs of your jobs and to
          ensure that dynamic volume provisioning is enabled so that PVCs can be matched to
          persistent volumes of the right storage class.

          For more information about dynamic storage provisioning, see the
          [OpenShift documentation](https://docs.openshift.com/container-platform/latest/storage/dynamic-provisioning.html).

          ## Quick start and examples

          For an overview of the key concepts associated with running jobs in Kubernetes, see
          ["About running jobs in Kubernetes"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/About_running_jobs_in_Kubernetes)
          in the Ab Initio Help.

          As described in this topic, you might begin by creating a CoOperatingSystemRuntime
          manifest that defines the configuration and compute resources needed for a standalone test
          job. You could then use the specification for this job to create a CoOperatingSystemRuntimeTemplate
          resource, which would collect common settings and make it possible to concisely specify
          CoOperatingSystemRuntime jobs.

          You could also use a runtime pool (of kind CoOperatingSystemRuntime) to extend these basic
          resources: such a pool consists of an actively running group of templated runtime objects whose
          pods can be used to run jobs on demand. Finally, for reserving resources in a pool, you could
          create a CoOperatingSystemRuntimeClaim resource.

          For more information about these custom resources, refer to the
          ["Co>Operating System runtime operator API reference"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Co_Operating_System_runtime_operator_API_reference).

          For a set of integrated examples (including complete YAML manifests) that show you how to
          create and use the custom resources, see ["Example manifests for running jobs"](https://docs.abinitio.com/r/422/Co_Operating_System/Co_Operating_System_Guide_to_Cloud_and_Container_Computing/Configuring_and_running_Co_Operating_System_jobs_in_Kubernetes/Creating_custom_resources_for_running_jobs/Example_manifests_for_running_jobs).
        displayName: Co>Operating System Runtime Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - abinitio
        - runtime
        - operator
        links:
        - name: Ab Initio
          url: https://www.abinitio.com/
        maintainers:
        - email: support@abinitio.com
          name: Ab Initio Support
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Ab Initio
        relatedImages:
        - registry.connect.redhat.com/abinitio/cooperating-system-runtime-operator@sha256:ee30210a153e8b3ba09080737216d881cc6a07e1a888e9684be0d17fd4ff791d
        - registry.connect.redhat.com/abinitio/runtime-operator-bundle@sha256:b9ffc18b2d88dfab92583af35b27e0ae36bceac7778cb448ab2ea2eae2ab3e94
        version: 4.2.3-0
      entries:
      - name: abinitio-runtime-operator.v4.2.3-0
        version: 4.2.3-0
      name: stable
    defaultChannel: release-4.3
    packageName: abinitio-runtime-operator
    provider:
      name: Ab Initio
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Jooho Lee
      provider-url: https://github.com/jooho/nfs-provisioner-operator
    name: nfs-provisioner-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: nfs-provisioner-operator.v0.0.8
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "cache.jhouse.com/v1alpha1",
                "kind": "NFSProvisioner",
                "metadata": {
                  "name": "nfsprovisioner-sample"
                },
                "spec": {
                  "nfsImageConfiguration": {
                    "image": "k8s.gcr.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439",
                    "imagePullPolicy": "IfNotPresent"
                  },
                  "scForNFS": "nfs",
                  "scForNFSPvc": "local-sc",
                  "storageSize": "1G"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Storage
          certified: "false"
          containerImage: quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
          createdAt: "2025-08-08T20:23:50Z"
          description: Create and manage NFS Server and Provisioner
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/jooho/nfs-provisioner-operator
          support: NFS
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NFSProvisioner is the Schema for the nfsprovisioners API
            displayName: NFS Provisioner App
            kind: NFSProvisioner
            name: nfsprovisioners.cache.jhouse.com
            version: v1alpha1
        description: "This operator deploy NFS server with local storage and also
          provide provisioner for storageClass.\n### Core Capabilities\n* **NFS Server:**
          Deployed\n* **NFS Provisioner:** Help customers to create PV using StorageClass\n*
          **StorageClass:** Dynamically create PV for requested PVC\n### NFS Provisioner
          Operator Features\n* NFS Server can use localStorage PVC or HostPath on
          the node\n### Must-Gather \n> isv-cli must-gather --image=quay.io/jooholee/nfs-provisioner-operator-must-gather:0.0.1\n"
        displayName: NFS Provisioner Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - nfs
        - storage
        links:
        - name: Nfs Provisioner Operator
          url: https://github.com/jooho/nfs-provisioner-operator
        maintainers:
        - email: ljhiyh@gmail.com
          name: Jooho
        maturity: alpha
        provider:
          name: Jooho Lee
          url: https://github.com/jooho/nfs-provisioner-operator
        relatedImages:
        - quay.io/jooholee/nfs-provisioner-operator@sha256:d9c013967421ec72644a588a155975bf856a856f1ef38ba71dfea306fdb47acd
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b
        - quay.io/community-operator-pipeline-prod/nfs-provisioner-operator@sha256:d7ef0fa3b5a1a6094b6fd60c6197a52cb14766865b3549806de100bb3522821a
        version: 0.0.8
      entries:
      - name: nfs-provisioner-operator.v0.0.8
        version: 0.0.8
      - name: nfs-provisioner-operator.v0.0.7
        version: 0.0.7
      - name: nfs-provisioner-operator.v0.0.6
        version: 0.0.6
      - name: nfs-provisioner-operator.v0.0.5
        version: 0.0.5
      - name: nfs-provisioner-operator.v0.0.4
        version: 0.0.4
      name: alpha
    defaultChannel: alpha
    packageName: nfs-provisioner-operator
    provider:
      name: Jooho Lee
      url: https://github.com/jooho/nfs-provisioner-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-kafka-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-kafka-controller.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kafka.services.k8s.aws/v1alpha1",
                "kind": "Cluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
          createdAt: "2025-09-20T21:41:07Z"
          description: AWS Kafka controller is a service controller for managing Kafka
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Cluster represents the state of an AWS kafka Cluster resource.
            displayName: Cluster
            kind: Cluster
            name: clusters.kafka.services.k8s.aws
            version: v1alpha1
          - description: Configuration represents the state of an AWS kafka Configuration
              resource.
            displayName: Configuration
            kind: Configuration
            name: configurations.kafka.services.k8s.aws
            version: v1alpha1
          - description: ServerlessCluster represents the state of an AWS kafka ServerlessCluster
              resource.
            displayName: ServerlessCluster
            kind: ServerlessCluster
            name: serverlessclusters.kafka.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon Kafka resources in AWS from within your Kubernetes cluster.

          **About Amazon Kafka**

          Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed AWS streaming data service makes it easy for you to build and run applications that use Apache Kafka to process streaming data without needing to become an expert in operating Apache Kafka clusters. Amazon MSK manages the provisioning, configuration, and maintenance of Apache Kafka clusters and Apache ZooKeeper nodes for you. Amazon MSK also shows key Apache Kafka performance metrics in the AWS Console.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.
        displayName: AWS Controllers for Kubernetes - Amazon Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Kafka Developer Resources
          url: https://aws.amazon.com/msk/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: kafka maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/kafka-controller:1.2.1
        - quay.io/community-operator-pipeline-prod/ack-kafka-controller@sha256:1a32352f30d0835299e14e8d0f0bac5585396dd6a10a63fe633e3d81c4d71b50
        version: 1.2.1
      entries:
      - name: ack-kafka-controller.v1.2.1
        version: 1.2.1
      - name: ack-kafka-controller.v1.2.0
        version: 1.2.0
      - name: ack-kafka-controller.v1.1.3
        version: 1.1.3
      - name: ack-kafka-controller.v1.1.2
        version: 1.1.2
      - name: ack-kafka-controller.v1.1.1
        version: 1.1.1
      - name: ack-kafka-controller.v1.1.0
        version: 1.1.0
      - name: ack-kafka-controller.v1.0.14
        version: 1.0.14
      - name: ack-kafka-controller.v1.0.13
        version: 1.0.13
      - name: ack-kafka-controller.v1.0.12
        version: 1.0.12
      - name: ack-kafka-controller.v1.0.11
        version: 1.0.11
      - name: ack-kafka-controller.v1.0.10
        version: 1.0.10
      - name: ack-kafka-controller.v1.0.8
        version: 1.0.8
      - name: ack-kafka-controller.v1.0.7
        version: 1.0.7
      - name: ack-kafka-controller.v1.0.6
        version: 1.0.6
      - name: ack-kafka-controller.v1.0.5
        version: 1.0.5
      - name: ack-kafka-controller.v1.0.4
        version: 1.0.4
      - name: ack-kafka-controller.v1.0.3
        version: 1.0.3
      - name: ack-kafka-controller.v1.0.2
        version: 1.0.2
      - name: ack-kafka-controller.v1.0.1
        version: 1.0.1
      - name: ack-kafka-controller.v1.0.0
        version: 1.0.0
      - name: ack-kafka-controller.v0.0.14
        version: 0.0.14
      - name: ack-kafka-controller.v0.0.12
        version: 0.0.12
      - name: ack-kafka-controller.v0.0.11
        version: 0.0.11
      - name: ack-kafka-controller.v0.0.10
        version: 0.0.10
      - name: ack-kafka-controller.v0.0.9
        version: 0.0.9
      - name: ack-kafka-controller.v0.0.8
        version: 0.0.8
      - name: ack-kafka-controller.v0.0.7
        version: 0.0.7
      - name: ack-kafka-controller.v0.0.6
        version: 0.0.6
      - name: ack-kafka-controller.v0.0.5
        version: 0.0.5
      - name: ack-kafka-controller.v0.0.4
        version: 0.0.4
      - name: ack-kafka-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-kafka-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Windup
      provider-url: ""
    name: windup-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: windup-operator.v0.0.25
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
             {
               "apiVersion": "windup.jboss.org/v1",
               "kind": "Windup",
               "metadata": {
                 "name": "windup",
                 "labels": {
                   "application": "windup"
                 }
               },
               "spec": {}
             }
            ]
          capabilities: Basic Install
          categories: Modernization & Migration
          certified: "false"
          containerImage: quay.io/windupeng/windup-operator-native:0.0.25
          createdAt: 01/22/24
          description: Windup is an analysis tool that supports the modernization
            and migration of Java applications.
          imagePullPolicy: Always
          operatorframework.io/suggested-namespace: windup
          repository: https://github.com/windup/windup-operator/
          support: https://issues.redhat.com/projects/WINDUP
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: "Create a Windup Web Console application.  \n\nTo access
              the Windup Web Console application from the \"Developer\" perspective,
              please go to Topology and click on the \"Open URL\" icon on the web-console
              pod.  \n\nAlternatively from the \"Administrator\" perspective, please
              go to Networking->Routes and click on the Location hyperlink.  \n\nKnown
              issue  \nIf you want to customize the Windup Web Console instance's
              parameters and you can not see any in the `Form view`, please switch
              to the `YAML view` and change them as needed.\n"
            displayName: Windup
            kind: Windup
            name: windups.windup.jboss.org
            version: v1
        description: |
          Windup is a web console application that supports large-scale Java application modernization and migration projects across a broad range of transformations and use cases.

          It analyzes application code, supports effort estimation, accelerates code migration, and enables users to move applications to containers.

          For more information please refer to the [https://windup.github.io/](https://windup.github.io/) page.
        displayName: Windup Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - migration
        - modernization
        - windup
        links:
        - name: Website
          url: https://windup.github.io
        - name: GitHub
          url: https://github.com/windup/windup-operator
        maintainers:
        - email: migrate@redhat.com
          name: Windup
        maturity: alpha
        provider:
          name: Windup
        relatedImages:
        - quay.io/community-operator-pipeline-prod/windup-operator@sha256:4e2bafba54eb01b0ae77555d2ff3da2c34286c7c173756dd20a90de28be36535
        - quay.io/windupeng/windup-operator-native:0.0.25
        version: 0.0.25
      entries:
      - name: windup-operator.v0.0.25
        version: 0.0.25
      - name: windup-operator.v0.0.24
        version: 0.0.24
      - name: windup-operator.v0.0.23
        version: 0.0.23
      - name: windup-operator.v0.0.22
        version: 0.0.22
      - name: windup-operator.v0.0.21
        version: 0.0.21
      - name: windup-operator.v0.0.20
        version: 0.0.20
      - name: windup-operator.v0.0.19
        version: 0.0.19
      - name: windup-operator.v0.0.18
        version: 0.0.18
      - name: windup-operator.v0.0.17
        version: 0.0.17
      - name: windup-operator.v0.0.16
        version: 0.0.16
      - name: windup-operator.v0.0.15
        version: 0.0.15
      - name: windup-operator.v0.0.14
        version: 0.0.14
      - name: windup-operator.v0.0.13
        version: 0.0.13
      - name: windup-operator.v0.0.12
        version: 0.0.12
      - name: windup-operator.v0.0.11
        version: 0.0.11
      name: alpha
    defaultChannel: alpha
    packageName: windup-operator
    provider:
      name: Windup
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Broadcom
      provider-url: https://www.broadcom.com/
    name: layer7-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: layer7-operator.v1.2.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Gateway",
                "metadata": {
                  "name": "ssg"
                },
                "spec": {
                  "app": {
                    "image": "docker.io/caapim/gateway:11.1.2",
                    "management": {
                      "cluster": {
                        "hostname": "gateway.brcmlabs.com",
                        "password": "7layer"
                      },
                      "password": "7layer",
                      "username": "admin"
                    },
                    "replicas": 1,
                    "service": {
                      "ports": [
                        {
                          "name": "https",
                          "port": 8443,
                          "protocol": "TCP",
                          "targetPort": 8443
                        },
                        {
                          "name": "management",
                          "port": 9443,
                          "protocol": "TCP",
                          "targetPort": 9443
                        }
                      ],
                      "type": "LoadBalancer"
                    }
                  },
                  "license": {
                    "accept": false,
                    "secretName": "gateway-license"
                  },
                  "version": "11.1.2"
                },
                "status": {}
              },
              {
                "apiVersion": "security.brcmlabs.com/v1",
                "kind": "Repository",
                "metadata": {
                  "name": "my-repository"
                },
                "spec": {
                  "auth": {},
                  "branch": "main",
                  "enabled": true,
                  "endpoint": "https://github.com/\u003cusername\u003e/\u003crepository\u003e",
                  "type": "git"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Api",
                "metadata": {
                  "name": "l7api-sample"
                },
                "spec": {
                  "deploymentTags": [
                    "ssg",
                    "ssg1"
                  ],
                  "graphmanBundle": "jsongz",
                  "portalPublished": false,
                  "serviceUrl": "/v1/api"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7Portal",
                "metadata": {
                  "name": "dev-portal"
                },
                "spec": {
                  "auth": {},
                  "deploymentTags": [
                    "ssg"
                  ],
                  "enabled": true,
                  "endpoint": "dev-ssg.brcmlabs.com",
                  "enrollmentBundle": "",
                  "portalTenant": "portal"
                }
              },
              {
                "apiVersion": "security.brcmlabs.com/v1alpha1",
                "kind": "L7StateStore",
                "metadata": {
                  "name": "state-store-example"
                },
                "spec": {
                  "redis": {
                    "database": 0,
                    "groupName": "l7gw",
                    "masterPassword": "7layer",
                    "standalone": {
                      "host": "redis-standalone",
                      "port": 6379
                    },
                    "storeId": "test",
                    "type": "standalone"
                  },
                  "type": "redis"
                }
              }
            ]
          capabilities: Basic Install
          categories: Security
          certified: "false"
          containerImage: docker.io/caapim/layer7-operator:v1.2.1
          createdAt: "2025-07-04T17:22:57Z"
          description: The Layer7 Operator covers all aspects of deploying, maintaining
            and upgrading Layer7 API Gateways in Kubernetes.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: layer7-operator-system
          operators.openshift.io/valid-subscription: Layer7 (Broadcom) API Gateway
            license
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/caapim/layer7-operator
          support: Broadcom Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Gateway is the Schema for the Gateway Custom Resource
            displayName: Gateway
            kind: Gateway
            name: gateways.security.brcmlabs.com
            version: v1
          - description: L7Api is the Schema for the l7apis API
            displayName: L7 Api
            kind: L7Api
            name: l7apis.security.brcmlabs.com
            version: v1alpha1
          - description: L7Portal is the Schema for the l7portals API
            displayName: L7 Portal
            kind: L7Portal
            name: l7portals.security.brcmlabs.com
            version: v1alpha1
          - description: L7StateStore is the Schema for the l7statestores API
            displayName: L7 State Store
            kind: L7StateStore
            name: l7statestores.security.brcmlabs.com
            version: v1alpha1
          - description: Repository is the Schema for the repositories API
            displayName: Repository
            kind: Repository
            name: repositories.security.brcmlabs.com
            version: v1
        description: |
          The [Layer7 Operator](https://github.com/CAAPIM/layer7-operator) covers all aspects of deploying, maintaining and upgrading Layer7 API Gateways in Kubernetes. The Layer7 Operator is exclusive to Kubernetes and utilizes the Container Gateway with specific emphasis on running in Ephemeral mode (without an external MySQL policy database).

          ## Getting Started
          This [Getting Started](https://github.com/CAAPIM/layer7-operator/wiki/Getting-Started) guide covers a simple introduction to the Layer7 Operator. You can find more documentation [here](https://github.com/CAAPIM/layer7-operator/wiki).
        displayName: Layer7 Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - layer7-operator
        links:
        - name: Layer7 Operator
          url: https://github.com/caapim/layer7-operator
        maintainers:
        - email: gary.vermeulen@broadcom.com
          name: Gary Vermeulen
        maturity: alpha
        provider:
          name: Broadcom
          url: https://www.broadcom.com/
        relatedImages:
        - docker.io/caapim/gateway:11.1.2
        - docker.io/caapim/graphman-static-init:1.0.3
        - docker.io/caapim/layer7-operator:v1.2.1
        - quay.io/community-operator-pipeline-prod/layer7-operator:1.2.1
        version: 1.2.1
      entries:
      - name: layer7-operator.v1.2.1
        version: 1.2.1
      - name: layer7-operator.v1.2.0
        version: 1.2.0
      name: preview
    defaultChannel: preview
    packageName: layer7-operator
    provider:
      name: Broadcom
      url: https://www.broadcom.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Percona
      provider-url: ""
    name: percona-server-mongodb-operator-certified-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: percona-server-mongodb-operator-certified-rhmp.v1.20.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDB",
                "metadata": {
                  "name": "my-cluster-name",
                  "finalizers": [
                    "percona.com/delete-psmdb-pods-in-order"
                  ]
                },
                "spec": {
                  "crVersion": "1.20.1",
                  "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a",
                  "imagePullPolicy": "Always",
                  "initImage": "registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46",
                  "updateStrategy": "SmartUpdate",
                  "upgradeOptions": {
                    "versionServiceEndpoint": "https://check.percona.com",
                    "apply": "disabled",
                    "schedule": "0 2 * * *",
                    "setFCV": false
                  },
                  "secrets": {
                    "users": "my-cluster-name-secrets",
                    "encryptionKey": "my-cluster-name-mongodb-encryption-key"
                  },
                  "pmm": {
                    "enabled": false,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9",
                    "serverHost": "monitoring-service"
                  },
                  "replsets": [
                    {
                      "name": "rs0",
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      },
                      "nonvoting": {
                        "enabled": false,
                        "size": 3,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "podDisruptionBudget": {
                          "maxUnavailable": 1
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        },
                        "volumeSpec": {
                          "persistentVolumeClaim": {
                            "resources": {
                              "requests": {
                                "storage": "3Gi"
                              }
                            }
                          }
                        }
                      },
                      "arbiter": {
                        "enabled": false,
                        "size": 1,
                        "affinity": {
                          "antiAffinityTopologyKey": "kubernetes.io/hostname"
                        },
                        "resources": {
                          "limits": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          },
                          "requests": {
                            "cpu": "300m",
                            "memory": "0.5G"
                          }
                        }
                      }
                    }
                  ],
                  "sharding": {
                    "enabled": true,
                    "configsvrReplSet": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "expose": {
                        "enabled": false,
                        "type": "ClusterIP"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "volumeSpec": {
                        "persistentVolumeClaim": {
                          "resources": {
                            "requests": {
                              "storage": "3Gi"
                            }
                          }
                        }
                      }
                    },
                    "mongos": {
                      "size": 3,
                      "affinity": {
                        "antiAffinityTopologyKey": "kubernetes.io/hostname"
                      },
                      "podDisruptionBudget": {
                        "maxUnavailable": 1
                      },
                      "resources": {
                        "limits": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        },
                        "requests": {
                          "cpu": "300m",
                          "memory": "0.5G"
                        }
                      },
                      "expose": {
                        "type": "ClusterIP"
                      }
                    }
                  },
                  "backup": {
                    "enabled": true,
                    "image": "registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d",
                    "serviceAccountName": "percona-server-mongodb-operator",
                    "pitr": {
                      "enabled": false,
                      "oplogOnly": false,
                      "compressionType": "gzip",
                      "compressionLevel": 6
                    }
                  }
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBBackup",
                "metadata": {
                  "name": "backup1"
                },
                "spec": {
                  "clusterName": "my-cluster-name",
                  "storageName": "s3-us-west"
                }
              },
              {
                "apiVersion": "psmdb.percona.com/v1",
                "kind": "PerconaServerMongoDBRestore",
                "metadata": {
                  "name": "restore1"
                },
                "spec": {
                  "backupName": "backup1",
                  "clusterName": "my-cluster-name"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Database
          certified: "true"
          containerImage: registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
          createdAt: "2025-06-05T11:15:53.428Z"
          description: Percona Distribution for MongoDB Operator automates the creation,
            modification, or deletion of items in your Percona Server for MongoDB
            environment
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/percona-server-mongodb-operator-certified-rhmp/support?utm_source=openshift_console
          repository: https://github.com/percona/percona-server-mongodb-operator
          support: Percona
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Instance of a Percona Server for MongoDB replica set
            displayName: PerconaServerMongoDB
            kind: PerconaServerMongoDB
            name: perconaservermongodbs.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Backup
            displayName: PerconaServerMongoDBBackup
            kind: PerconaServerMongoDBBackup
            name: perconaservermongodbbackups.psmdb.percona.com
            version: v1
          - description: Instance of a Percona Server for MongoDB Restore
            displayName: PerconaServerMongoDBRestore
            kind: PerconaServerMongoDBRestore
            name: perconaservermongodbrestores.psmdb.percona.com
            version: v1
        description: |2+

          ## Percona is Cloud Native
          The Percona Distribution for MongoDB Kubernetes Operator automates the creation, modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance modification, or deletion of items in your Percona Server for MongoDB environment. The Operator contains the necessary Kubernetes settings to maintain a consistent Percona Server for MongoDB instance.

          Consult the [documentation](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/index.html) on the Percona Kubernetes Operator for Percona Server for MongoDB for complete details on capabilities and options.

          ### Supported Features

          * **Scale Your Cluster** - change the `size` parameter to [add or remove members](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/scaling.html) of the replica set. Three is the minimum recommended size for a functioning replica set.

          * **Add Monitoring** - [Percona Monitoring and Management (PMM) can be easily deployed](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/monitoring.html) to monitor your Percona Server for MongoDB replica set(s). The recommended installation process uses Helm, the package manager for Kubernetes.

          * **Automate Your Backups** - [configure automatic backups](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/backups.html) to run on a scheduled basis or run an on-demand backup at any time. Backups are performed using Percona Backup for MongoDB (PBM) and can be stored on local PVs or in any S3-compatible cloud storage provider. physical Backups - [configure physical backups](https://docs.percona.com/percona-operator-for-mongodb/backups.html#physical)
          * **Automated volume expansion** - Kubernetes supports the Persistent Volume expansion as a stable feature since v1.24. Using it with the Operator previously involved manual operations. Now this is automated, and users can resize their PVCs by just changing the value of the resources.requests.storage option in the PerconaServerMongoDB custom resource. This feature is in a technical preview stage and is not recommended for production environments.

          ### Common Configurations

          * **Set Member as Arbiter** - [Set up a replica set which contains an arbiter](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/arbiter.html), which participates in elections but does not store any data. This reduces storage costs while helping maintain replica set integrity.

          * **Expose Members Outside K8S** - [by appropriately configuring the ServiceType](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/expose.html) you can expose replica set members outside of Kubernetes or provide statically assigned IP addresses.

          * **Utilize Local Storage Options** - [with support for Local Storage you can mount existing data directories](https://www.percona.com/doc/kubernetes-operator-for-psmongodb/storage.html) into your replica set managed by Kubernetes or utilize high performance hardware for local storage rather than network storage for your database.

          ### Before You Start

          Add the PSMDB user `Secret` to Kubernetes. User information must be placed in the data section of the `secrets.yaml`
          file with Base64-encoded logins and passwords for the user accounts.

          Below is a sample `secrets.yaml` file for the correct formatting.

          ```
          apiVersion: v1
          kind: Secret
          metadata:
            name: my-cluster-name-secrets
          type: Opaque
          data:
            MONGODB_BACKUP_USER: YmFja3Vw
            MONGODB_BACKUP_PASSWORD: YmFja3VwMTIzNDU2
            MONGODB_DATABASE_ADMIN_USER: ZGF0YWJhc2VBZG1pbgo=
            MONGODB_DATABASE_ADMIN_PASSWORD: ZGF0YWJhc2VBZG1pbjEyMzQ1Ngo=
            MONGODB_CLUSTER_ADMIN_USER: Y2x1c3RlckFkbWlu
            MONGODB_CLUSTER_ADMIN_PASSWORD: Y2x1c3RlckFkbWluMTIzNDU2
            MONGODB_CLUSTER_MONITOR_USER: Y2x1c3Rlck1vbml0b3I=
            MONGODB_CLUSTER_MONITOR_PASSWORD: Y2x1c3Rlck1vbml0b3IxMjM0NTY=
            MONGODB_USER_ADMIN_USER: dXNlckFkbWlu
            MONGODB_USER_ADMIN_PASSWORD: dXNlckFkbWluMTIzNDU2
            PMM_SERVER_API_KEY: dXNlckFkbWluMTIzNDU2
          ```
          ### Release Highlights
          * This release of Percona Operator for MongoDB fixes the failing backup that was caused by the Operator sending multiple requests to PBM. The issue was fixed by bypassing the cache for the backup controller and enabling direct communication with the API server for sending backup requests.

        displayName: Percona Distribution for MongoDB Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - mongodb
        - percona
        - database
        - replicaset
        - shard
        links:
        - name: Percona
          url: https://www.percona.com/
        - name: Percona Kubernetes Operators Landing Page
          url: https://www.percona.com/software/percona-kubernetes-operators
        - name: Documentation
          url: https://percona.github.io/percona-server-mongodb-operator/
        - name: Github
          url: https://github.com/percona/percona-server-mongodb-operator
        maintainers:
        - email: info@percona.com
          name: Percona
        maturity: stable
        provider:
          name: Percona
        relatedImages:
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-bundle-rhmp@sha256:875d81ba1b91a72d9bb4f1d86e2285992370a04ace3831b3c914127ab55a57e7
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:3c03fb18a5d73876576b80aca2b61e7c84ca052839ccb467d14e15f6d1f9d834
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:5871c3e16337c14e1636f71eca929a8386f3226bfbea027b372da47b8b29862d
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a34fe515df61e2da6347624b2029af6519bb4a1fa062cba3b5f0bcae2d34a86a
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a4f6c588c00a93502e5e13f1c849fb6ca946f57dd6ad64fe18cb96312ad753d9
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator-containers@sha256:a57926b53eef862b70a53818a7ccb0e9756f878f30d4128b608090b109bdc585
        - registry.connect.redhat.com/percona/percona-server-mongodb-operator@sha256:b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46
        version: 1.20.1
      entries:
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.1
        version: 1.20.1
      - name: percona-server-mongodb-operator-certified-rhmp.v1.20.0
        version: 1.20.0
      name: stable
    defaultChannel: stable
    packageName: percona-server-mongodb-operator-certified-rhmp
    provider:
      name: Percona
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Habana Labs Ltd.
      provider-url: https://habana.ai
    name: habana-ai-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: habana-ai-operator.v1.20.0-543-5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.20.0-543"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.20.0-4"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.20.0-543-2"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.20.0-543"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.20.0-543"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.20.0-543"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.20.0-543-4"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
          createdAt: "2025-02-26T09:03:44Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.20.0-543-5'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:317765eac4e0332a0c4f6bead5563b343a8c22e6e0eaa62276ff66be08608c30
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:d14c457106b6133fc41ef727ac40a6bf2f0e5c97f21659969d1daebdc1310b43
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:a673900f08c38c9bb40c90b3cddea580afcc6a99195d816cd0b5cae9831ca5c9
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:3ea80aa8994e8b19fbbdcd18ba2c35800774da3ca4e27a2514d0f5f9676786fa
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:cba03259b33c4cb9bcd22622ab0dda962846ea0a80c727779365cf55129ad308
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:957702b58ee66d7339a20b2a5ece783606d6ccebde627c58951d5ecd6bb1b13f
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c5beac6209d0671d3a49b63f1cf95401055b042430eafaefc3859b5c633f0e84
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:9ab54a84a00e8afc078eec2a7d643f09b1bd674e0e230733fbce7108e392e009
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:a657269d0212d952f5107514db63c3ea2b6c86ffed0a44f0f33eca96a242cf35
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:68d0f44170cbf3ea15b793e0503fb07ce7e32488011fc85ccb6c9234b3b9302f
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:d5b37eafe48446e1ffa492e195f4519b77d4a11f6750e5c71f27c1536bdfd2cc
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c662de244c5900c89b918b16982298676d9a6d553360a5716174b28d4d64a635
        version: 1.20.0-543-5
      entries:
      - name: habana-ai-operator.v1.20.0-543-5
        version: 1.20.0-543-5
      name: "1.20"
    - currentCSV: habana-ai-operator.v1.21.5-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.21.5-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.21.5-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.21.5-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.21.5-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
          createdAt: "2025-09-17T11:55:32Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.21.5-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:bd57e7059467f0ca1667d19a70bd9a6ad96b83bbcb53c09fdad5a262fee76af9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:d7f64d15c379524597aa2b2c5b4586c890442fa32356692013cf665b01b0f092
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:256cd8bebec3ed835e187e925a14a6b605ce9b3a17fe7d8886e0b32cde163d90
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:c84e997c7d8c22c4d4e902a29e45e5364f0ad306f17d03a7d8ddb4391142e167
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:bc3749ff153fff71e252820cbf22fcd0753b195b2204f9731e42af29fb7ddaa0
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:c7147359dcd6a2e5adee95259d37743ff302b1deaeb13b184044f4029d33f44a
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:a0b45ccfb6af54b1043ccc2336ad3d7990cc447af5ba3621f3eb1953cd40ffc5
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:a216ec6a6e7df50e067c557e9ed5871a3308b2ba42a4e41c9642f0c2c4503d88
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:c95d3ffd6c91be3418415740258d33a8dbcedad6fe7d6bfe68d4dd122027447b
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:1ec1264dcc4ee289ada0747083ad9b072bad2e33544a3a879383694c2974a885
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:0741da99140d137474b9e7edba921672eab8ff6277e1dd8bf3a4179e87f01dd4
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:9692d079ef18a364ed235c187482f329b3a5764a773e363acb1238044ad4a884
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:1e488c8ad31784dab0fe70144f5146d5809e2fc7c86424d1b512ab887787e28f
        version: 1.21.5-6
      entries:
      - name: habana-ai-operator.v1.21.5-6
        version: 1.21.5-6
      name: "1.21"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: "1.22"
    - currentCSV: habana-ai-operator.v1.22.1-6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "habanalabs.habana.ai/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "habana-ai"
                },
                "spec": {
                  "bmc_monitoring": {
                    "image": {
                      "repository": "vault.habana.ai/habana-bmc-exporter/bmc-exporter",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "250m",
                        "memory": "250Mi"
                      },
                      "requests": {
                        "cpu": "150m",
                        "memory": "100Mi"
                      }
                    }
                  },
                  "device_plugin": {
                    "image": {
                      "repository": "vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin",
                      "tag": "1.22.1-6"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "20m",
                        "memory": "64Mi"
                      },
                      "requests": {
                        "cpu": "10m",
                        "memory": "32Mi"
                      }
                    }
                  },
                  "driver": {
                    "driver_loader": {
                      "images": {
                        "rhel_8.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.2": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.4": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "rhel_9.6": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "tencentos_3.1": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_22.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                          "tag": "1.22.1-6"
                        },
                        "ubuntu_24.04": {
                          "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer",
                          "tag": "1.22.1-6"
                        }
                      },
                      "mlnx_ofed_repo_path": "artifactory/gaudi-installer/deps",
                      "mlnx_ofed_version": "mlnx-ofed-5.8-2.0.3.0-rhel8.4-x86_64.tar.gz",
                      "repo_path": "artifactory/gaudi-installer/repos",
                      "repo_server": "vault.habana.ai",
                      "resources": {
                        "limits": {
                          "cpu": "4000m",
                          "memory": "16Gi"
                        },
                        "requests": {
                          "cpu": "2000m",
                          "memory": "8Gi"
                        }
                      }
                    },
                    "driver_runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "feature_discovery": {
                    "nfd_plugin": false,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  },
                  "image_registry": "vault.habana.ai",
                  "metric_exporter": {
                    "interval": 20,
                    "port": 41611,
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/gaudi-metric-exporter/metric-exporter",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "150m",
                          "memory": "120Mi"
                        },
                        "requests": {
                          "cpu": "100m",
                          "memory": "100Mi"
                        }
                      }
                    }
                  },
                  "runtime": {
                    "configuration": {
                      "container_engine": "crio"
                    },
                    "runner": {
                      "image": {
                        "repository": "vault.habana.ai/habana-ai-operator/habana-container-runtime",
                        "tag": "1.22.1-6"
                      },
                      "resources": {
                        "limits": {
                          "cpu": "20m",
                          "memory": "64Mi"
                        },
                        "requests": {
                          "cpu": "10m",
                          "memory": "32Mi"
                        }
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
          createdAt: "2025-09-17T10:15:21Z"
          description: Manages Intel Gaudi AI accelerators within a Kubernetes cluster
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <1.22.1-6'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: habana-ai-operator
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: http://github.com/HabanaAI/habana-ai-operator
          support: Habana Labs Ltd.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy is the Schema for the clusterpolicies API.
            displayName: Cluster Policy
            kind: ClusterPolicy
            name: clusterpolicies.habanalabs.habana.ai
            version: v1
        description: |
          Kubernetes provides access to accelerators such as Intel Gaudi AI accelerators and other devices through the [Device Plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The Intel Gaudi uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all Intel Gaudi software components needed to provision and monitor the Intel Gaudi AI accelerators. These components include:
          * The Intel Gaudi kernel modules to enable the hardware
          * The Kubernetes DevicePlugin to allow the Intel Gaudi AI accelerator allocation via the kubelet
          ### Documentation
          Visit the [Intel Gaudi developer site](https://developer.habana.ai/) for more information.
          To get started with using Intel Gaudi with OpenShift, see the instructions in the [Intel Gaudi Documentation](https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html).
        displayName: Intel Gaudi Base Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - habanalabs
        - hardware
        - driver
        links:
        - name: Intel Gaudi Documentation
          url: https://docs.habana.ai/en/latest/Orchestration/HabanaAI_Operator/index.html
        maintainers:
        - email: erik.berlin@intel.com
          name: Erik Berlin
        - email: tal.david@intel.com
          name: Tal David
        - email: lukasz.wencel@intel.com
          name: Lukasz Wencel
        maturity: stable
        provider:
          name: Habana Labs Ltd.
          url: https://habana.ai
        relatedImages:
        - vault.habana.ai/docker-k8s-device-plugin/docker-k8s-device-plugin@sha256:c9ff6f3fe24ef87c7bfc71bdb1cd7900941df52f9ad1293badd7e14783ce7ef6
        - vault.habana.ai/habana-ai-operator/driver/rhel8.6/driver-installer@sha256:e34a6820a58e7e8090b521781c76acc3867c94f8291634de7b138f928e0bdb74
        - vault.habana.ai/habana-ai-operator/driver/tencentos3.1/driver-installer@sha256:38c21eae6249e5571921998660418e59870bc0880e7b8c9b95a4a8db30ef2e7d
        - vault.habana.ai/habana-bmc-exporter/bmc-exporter@sha256:1925586fda34c6fa444a59b1fa7be20b3cf16e845bf56566ad2a41c276b0cfe9
        - vault.habana.ai/habana-ai-operator/habana-container-runtime@sha256:8f413edfff1978b35ba3998ed5f1c6293e6764b1d2cd9faa08f0bd81dc6e3e10
        - vault.habana.ai/habana-ai-operator/habanalabs-feature-discovery@sha256:1e15f3de21c2b1e20f54d1f8a973290c6d940624506880c5bc3eb880011bcc02
        - vault.habana.ai/habana-ai-operator/driver/rhel9.6/driver-installer@sha256:2cf151c2a518e3e1c7c09efb8301e469d266d1f030fff2e6180b086cc2e04ee7
        - vault.habana.ai/habana-ai-operator/intel-gaudi-kubernetes-operator@sha256:131c807c327605610959cd9edcb92d3fbd0578e3e7b8e87faafdaf4637be0e63
        - quay.io/brancz/kube-rbac-proxy@sha256:c2ef3c3fcb4ffcfab30088b19c8ef18931aac8cf072c75ef72c654457e1d632b
        - vault.habana.ai/habana-ai-operator/driver/ubuntu22.04/driver-installer@sha256:7140877b23291a5db3e5e28def3c9d5e7cb7be62f046e2fb7ac45eff369cea84
        - registry.connect.redhat.com/ospid-627904f807c06bb948c71e73/habana-ai@sha256:2083f9ad23e59b90db35046760488dd55741f71c543ce6a78c96439f84ca550b
        - vault.habana.ai/habana-ai-operator/driver/rhel9.2/driver-installer@sha256:704b4bebdfe6b85693df4b213f67862859a554c21ce4e614736172f64b541ea7
        - vault.habana.ai/habana-ai-operator/driver/rhel9.4/driver-installer@sha256:392fcd2a4730299e6e3bab9f7e96fc3b710de8adb374a2c64b7316e097ed05f9
        - vault.habana.ai/habana-ai-operator/driver/ubuntu24.04/driver-installer@sha256:d1974d72774c626e77707e8f44bcc631ac142bfe2a5395714969a964c963e36e
        - vault.habana.ai/gaudi-metric-exporter/metric-exporter@sha256:53c1844b5e8eb08d0cc29865a732100208c5dfc0fe55df8b32e589e319bcab6c
        version: 1.22.1-6
      entries:
      - name: habana-ai-operator.v1.22.1-6
        version: 1.22.1-6
      name: stable
    defaultChannel: stable
    packageName: habana-ai-operator
    provider:
      name: Habana Labs Ltd.
      url: https://habana.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      pod-security.kubernetes.io/audit: privileged
      pod-security.kubernetes.io/enforce: privileged
      pod-security.kubernetes.io/warn: privileged
      provider: NVIDIA Corporation
      provider-url: ""
    name: gpu-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: stable
    - currentCSV: gpu-operator-certified.v1.10.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                  },
                  "driver": {
                    "enabled": true,
                    "use_ocp_driver_toolkit": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "operator": {
                    "defaultRuntime": "crio",
                    "deployGFD": true,
                    "initContainer": {
                    }
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
          createdAt: Mon Apr 18 10:06:40 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.10.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://coreos.com/blog/introducing-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:67c48a8370c34f68b46eb57339e87d01c04198368a0423553fab7565c862c4df
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:9abb063e6b74085728fead5ed5bbe3bd7b3f57de60e87f6b964bba2d46f1bc67
        - nvcr.io/nvidia/gpu-operator@sha256:c7f9074c1a7f58947c807f23f2eece3a8b04e11175127919156f8e864821d45a
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:60b1d25ebb3aada94985e359bb059d4491bf0148ac75d51b68a461670659e429
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:e8f06ce53415e8572f633b2169289d973993cb1e49d4dcc686ac134194d88f33
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:24d804e8f005d7aeca8343aa13e5f92295d8642a4c47cb24e3ac86a22543bc37
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:5b16056257acc51b517d9cdb1da3218693cefc214af93789e6e214fd2b4cacf1
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:ac13256c295126b854a07421c45de0721e7a57974b38b3e66eb1fdc3a4e29f2e
        - nvcr.io/nvidia/k8s-device-plugin@sha256:aa95c16106280e36a5e32d2fe4c66e8b70f5a114860c6f4ed5b1a4085c63601b
        - nvcr.io/nvidia/cuda@sha256:7f0f7dc8c993a86865d4d9f9b0e36d113ea15603d5d5279d2879fa2bc21c8236
        - nvcr.io/nvidia/driver@sha256:f99b66855e92451a59a722355dd52bf97eb985cc8a647172273de8178aeaa93c
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:c2e3d0fe41a0d227dbb70caec03c780cc76317515e5ab3875f31d50c63f41c66
        version: 1.10.1
      entries:
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.10
    - currentCSV: gpu-operator-certified.v1.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    },
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "dcgmExporter": {
                    "config": {
                      "name": ""
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                  },
                  "devicePlugin": {
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": true
                  },
                  "vgpuDeviceManager": {
                    "enabled": true,
                    "config": {
                      "name": "vgpu-devices-config",
                      "default": "default"
                      }
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
          createdAt: Thu Aug  4 11:57:56 PDT 2022
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=v1.9.0 <v1.11.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:f48e8b430248a9cadf716a9211bb0050d4a663d1f1386ea107e72917f82e5b5f
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3804e2b03c44d25dd410e3882b7fb6d1fde41f55110d7702af47136b042a63ff
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:8490d555131a57a11311561667ea9ec8cf554a399d2492a6f9e6677c01ec4b9a
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:39127d04c9faff3f22068dcf463ed48d02be684cd18e3f14db843bddbdd783a7
        - nvcr.io/nvidia/driver@sha256:b8799e7397fae5a27446737443091eed05b918816f53c400a48be71125ee22e8
        - nvcr.io/nvidia/cuda@sha256:4400b56f81095b24fc9dde93ad7ad7d0ca6a7e9736c4c7f1d4e2bcbf0afbe45c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7a0ccbcbe8d379bfc441e06504eb45d0d433a37e2a9e6dde83d13347296ebe81
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:b90a519e9f30fbc013e1806bc081268724e9e80b4035c67c167ab3b43a37704a
        - nvcr.io/nvidia/gpu-operator@sha256:53180f81fb7e1e4bcd49a556bc6285c2262abbab8083b29d2688b02d4200dd2d
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:16ee5b1a8d9217f761c40a06c7ae598cbc893fbc35d99ff9b84fa886ea6b9b37
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:3128cd81b56975643c1c88c6ceff6d72e918a67ec2d6a5a606b6e04753c5b0ce
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:10d825ccbfcf87f8c74f841fa89ae345f2d381fb47a5b2b31f61b72ebbbb0915
        - nvcr.io/nvidia/driver@sha256:6e2902d73c54cf57bec94edf166d206cfeb0b948b5aab70800ccbe025064b3a1
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:e88e610d09f1a7528ab32cb54259aee340eba2b2bf66b92a28157626daadc82d
        - nvcr.io/nvidia/driver@sha256:cf8ce04907d0e8decf90a8dccfa6134af9c39509d2f8a21c355ed71223acacdb
        - nvcr.io/nvidia/driver@sha256:8c9fabc1cd773e9526eb7e8526084f065f6cca9d0fd8dc2a672a0bb717f1cc60
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:2d6e86db0e2b0db64c00566938e1851e25d4bb1019b1e7b696e2b199cba33397
        version: 1.11.1
      entries:
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v1.11
    - currentCSV: gpu-operator-certified.v22.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
          createdAt: Mon Jan 30 15:43:24 PST 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <22.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:808529c14a8873324096fbd1622fb895c4dc7f60c1cf9e3ded646f9a81d7b496
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:50c2a2b998e467c82716efa79fe136aa6f7ea95fd23576cf384d251bb9628640
        - nvcr.io/nvidia/k8s-device-plugin@sha256:9c17d3a907eb77eb8f7b4f3faf52d8352e4252af92003f828083f80d629bd2c3
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:bec9f026d9b3d9404c78d6091817a359015c6a7aa411735b34138c1518853b5d
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:9a00cdfdddb73327ef8e8e0fa60e50926a388d380c551bba6a3d3012be40401d
        - nvcr.io/nvidia/driver@sha256:f2ca4612f3418e0e3963be8c3771ffe463cdc1fc1b810437a3aad9d76bb6499d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:00f1476548fbed9ee01961443a73bf65396c2e8bb2b84426f949dd56cb4d14cd
        - nvcr.io/nvidia/driver@sha256:f250d1e4989fd4821e81bad9227eeeaa531ca8810f3a40e996f1ca2af7945f96
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:9177a0ae30798b42d0387f6a20cd3ce3cd1799a91b7866bf812368764b05b1af
        - nvcr.io/nvidia/gpu-operator@sha256:54d5846c0291db621931a8544c9cb6437512470a93278ca9130ba73f2e2f0055
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:64d757b4c80b910e64647a84a1d592fab2ea3313ff6dce30c25c3a08e180bd74
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/driver@sha256:268c26781c46b36fe691638e807eb5c3a2cd077029ddb095ac75264adac04925
        - nvcr.io/nvidia/driver@sha256:c5a62bc2a6cce339798e4b7cf9cec47e289c2d30b40f451ba8c9e958a68e3e55
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:efb88937f73434994d1bbadc87b492a1df047aa9f8d6e9f5ec3b09536e6e7691
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:c3cf59dd5d6160eba5d816ade2e81b35ebb10f4884df67971f6ace36f8e6efc1
        - nvcr.io/nvidia/cuda@sha256:5f2a2d8977f2c59abe88394f05cc3c044194554f90994d8554b0e1117ab5900d
        - nvcr.io/nvidia/driver@sha256:b860ef04cc6cd8c481810fa19bef7221071f5a3862640b2ad6c2704728dc77c8
        version: 22.9.2
      entries:
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v22.9
    - currentCSV: gpu-operator-certified.v23.3.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "true"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
          createdAt: Mon May 22 15:22:15 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.3.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:3df118f2330e95db3c76173cca35e542b2da120ec0bb8f6aff83c4c96c0c2daf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80871275fbcf888d4683c58638c7a75b3c46091040cafc6210a8c93513adb146
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:708291528f26c70c561745967a1200edcb2f984687f749a49a2860fee580423f
        - nvcr.io/nvidia/driver@sha256:8ff0ee88fe6c4ff031ef642593f93625bfd9a48fe7be8a79a5c2736117e36941
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:21dfc9c56b5f8bce73e60361d6e83759c3fa14dc6afc2d5ebdf1b891a936daf6
        - nvcr.io/nvidia/cuda@sha256:62d845b4bd6de4be65f0b6daaae6831c29d4529955f99db5d26f5f7179106285
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:0d47dad29d2ef445b301c5c64717758eed43a606345b79f97bce2e64b40a91a8
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:5489125ceae5864280e4d6a9ab52ab0f650b3179349a7298c4a204feb60b661a
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:cab21c93987a5c884075efe0fb4a8abaa1997e1696cbc773ba69889f42f8329b
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:f10170b076badda8f669af8b0f7b47eb5470abaf35a67dda9c4e182d66138177
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:84ce86490d0d313ed6517f2ac3a271e1179d7478d86c772da3846727d7feddc3
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:03e43d6b1460cadb33509d5ee07a5e27603ab89d7ec35f67d8a51bb78745b463
        - nvcr.io/nvidia/driver@sha256:73b8928a1b1fb650d2ca8d6451a7d7c551fd0509f7511c2fd5bfdcef83ae4f5d
        - nvcr.io/nvidia/gpu-operator@sha256:08ca33f491195b7ba867be1e6642b5ff8aff6a4b0b7cfc0ec87619252ccfddc7
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:8f03f74076b5aa6672a5d75a398fa7515fcde57fdb1cf8079e015b3366efd3b5
        - nvcr.io/nvidia/driver@sha256:917f92c589c3124dba54f66f89c44b93b0beaf418f01d0da6a546e1670fdf3a9
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:2d7e32e3d30c2415b4eb0b48ff4ce5a4ccabaf69ede0486305ed51d26cab7713
        version: 23.3.2
      entries:
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.3
    - currentCSV: gpu-operator-certified.v23.6.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": false,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
          createdAt: Thu Aug 31 12:22:44 PDT 2023
          description: Automate the management and monitoring of NVIDIA GPUs.
          olm.skipRange: '>=1.9.0 <23.6.1'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/driver@sha256:b4517810f2884f345b4f2323e83e41d28f54cdc6e2caf0690bb7ed4f3947dbcc
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:0379c59b43e11861837787a0da22e70b528718373bb10de9e8b7752849bd4e7a
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:b30f5c64791dc5406b57eaf51ac40195dc1051a18cd1f38fd0a43c65f1d6dba9
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:9484110986c80ab83bc404066ca4b7be115124ec04ca16bce775403e92bfd890
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:d1c636cc17155c4764d331e050f9bf3ec9f39f6ec6ec14e972179b185ea9f5b8
        - nvcr.io/nvidia/gpu-operator@sha256:d2e0f8bd86e8e9f2ff3514e113657658584ca4a2d0b82dac2e47364cc0b30648
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:60086c33226bbe7f04b611c20d308fed5825226f8a0a5635bc58d8ed55cc08a5
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:bf06c8ad380a36e5014607bd3ab05b80e8e5e7e77d23d0dd89fe53a5564152fa
        - nvcr.io/nvidia/cuda@sha256:64c0ff5fd4a27d4b1d46378634f7a7398c75f87c18c70c988dde18910d461560
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:bb845160b32fd12eb3fae3e830d2e6a7780bc7405e0d8c5b816242d48be9daa8
        - nvcr.io/nvidia/k8s-device-plugin@sha256:f1617c3fe54fd9f6d4a00358e5da1a9fa9e8963a4ef6b3ed06f47d4971ddff4b
        - nvcr.io/nvidia/driver@sha256:3382e254056f28831767bc6729bc2594353a5ff2a28fe9f2d94396c597bb23d8
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:a229beba8540549c234a046e975bf4bf8c50d521dfe551af56f6e042b25ec017
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:00a91252be8e62da6f28cdf1ec3721f2e945ae296938b0b925ef9db2bf6dcc8b
        - nvcr.io/nvidia/driver@sha256:bbdc4415fe2e71eeea1b8a6e50ee562e11704dabb46db1db4aea807102886a35
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:878a6162a4fa0b61dae02c7cfc13ba4d28bb06f92fd62d0c521633ab507d675a
        version: 23.6.1
      entries:
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.6
    - currentCSV: gpu-operator-certified.v23.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
          createdAt: Wed Mar 6 16:43:36 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <23.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:714547d54e5fe4191019a1e5f1daffc7fab7481b619b79c378541dafc76c9e5d
        - nvcr.io/nvidia/driver@sha256:7481a3697783dcdca9ae78e7b548a6900e86ea33ab49ec14f0ba55db2fdb1a2e
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:5339c41013727fa5dd3ae5bc1d51eb90e02e5baa85d4039b1efdde30baeae1ae
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:6a05d6a1923fda756aed0dddf7ed23a83c30cf1e6c519fc39dd70c0309ec8257
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:9aefef081c3ab1123556374d2b15d0429f3990af2fbaccc3c9827801e1042703
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:925f4a47710e4318ed457930f5406174c1f6d28b1bf6b1bc310687fec0fde712
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:a67d8e92861a2dce5649105c07561e4422e9fe4ba81a6525dc0d70a7ef85f9c0
        - nvcr.io/nvidia/k8s-device-plugin@sha256:bb81ae8d034d1e1c6a342091a5f1dda5884dd43ab47e96624487a133d1edd4b3
        - nvcr.io/nvidia/driver@sha256:abda1ac56371d55917b96ff330109980f468e133c9d5705da0ef87429f14ccd7
        - nvcr.io/nvidia/gpu-feature-discovery@sha256:8d1ec78f2b1ddb7f0c47453d0427231190747bda411733a7dd0c8f5196f09e9c
        - nvcr.io/nvidia/gpu-operator@sha256:3988b3b570cf65b0d338cb29bd24b0b9bf1885e1563fbb398ef23cba0b59e6bb
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:59a3875e7a37eb370385e654184efa3a1b193c9ea352165818496b19cbe14aa4
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:27c44f4720a4abf780217bd5e7903e4a008ebdbcf71238c4f106a0c22654776c
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:011fb450af3fa2e8fe5d28d590e4c653631447bc23d149591ced3d89089c4f2c
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:387021553dfb16aab633228d42f63f04fa932b4f46add07527f296dfe97e5148
        version: 23.9.2
      entries:
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v23.9
    - currentCSV: gpu-operator-certified.v24.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
          createdAt: Tue Apr 30 14:18:26 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.3.0'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cuda@sha256:ae0623ec8634b6c88f815b88037763def160cbbac15013b77ddef257fc276c9a
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:2cc2119d20515f8ca7a61a0f3932578d69a45bcdea49e6c320a89c56f105e7d9
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:2edc1d4ed555830e70010c82558936198f5faa86fc29ecf5698219145102cfcc
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:1ae0b923bc57f47bab046b50c50110f6914bbaffbfef704df34b3fe332db2e31
        - nvcr.io/nvidia/k8s-device-plugin@sha256:80674b19898ecf9ed6949e39674da769d6feeeb01bea54b914ef9ff502834f49
        - nvcr.io/nvidia/driver@sha256:56c79482582cdfbc58d3134e8672637c5bf05f328880f76898f526143d04c6af
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:1532fc1351ca44f863feaaf54ba8810be7a62396f98dd292ccaf975fde31ce5b
        - nvcr.io/nvidia/driver@sha256:a836ccbe21da735aee1c39b81060ed5e2fdb4ffa339874baaf4634f1e9259f74
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:2442fd2ec0ee29746abbced727c53741f017e8f3f615321c1ec42c1c3ffa0b12
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:f0e4e14a3081417be8e6a5d855fb0cb69e1b63bc54a74f17e6d7084abe275588
        - nvcr.io/nvidia/gpu-operator@sha256:041e75a3df84039c2dbbd4b9d67763bd212138822dbb6dbc0008858c1c6eff8d
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:0c77725de8c42d248ed825453efd2e005f4900f4be384fd23084f6b721ddd0e0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:ecb867c5787e44f1889b257e6c03b07748689a7b293c8f1affb1d4cb449b235c
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:d8774d6afd37110199512636172823ae8749e5ea2e13760f57c255a74f47018c
        - nvcr.io/nvidia/driver@sha256:96f25c67e5b1072d5981080e12d65ec37eb9ef2fc0494499416aa801b0a34da3
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7798eb9b9424e09959fc808596f87bdb68ac5740174123c6646dee2166d3e7d2
        version: 24.3.0
      entries:
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.3
    - currentCSV: gpu-operator-certified.v24.6.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
          createdAt: Wed Sep 25 08:25:16 PDT 2024
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.6.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:15dab1273345df4a5844c4c761d064dbc4b592101251dc39174e597137123027
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:fa72cc3e0b09510d71a097b92ee101e057f1c11e1cb363172bd26b2d94272417
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:96380b95396b7f29d2ed2ec8ef1ad5a8bcf9f55051db723295dc0a20db845331
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:ae63fac9a4057a7646f0cf0ee0566e8928529adde05c4c0a017cda0599e381b2
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:740abc3ff657545c10effd5354f09af525200ed9a1b7623f0c2e8c7bd9e4a4e2
        - nvcr.io/nvidia/driver@sha256:a6d12fb5753f267dda25dfd38910f972bc632c006a24107fa50e20bba3642d7c
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7ad2c9f71fe06f9f7745ac8635f46740fbdff4f11edd468addfab81afcdfa534
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:969147c01d63be5d1fe458f32f1cc0c7408cf3062531db91408e2fc57b4d8a67
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:21f4c8b88716e8e6f732f9fb4c2efaef937c227491a8631c5e55036f80f39a4d
        - nvcr.io/nvidia/driver@sha256:858de27c152669f5a3cf4287406405b16dd5bb70c0373324eb735511997bb415
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:781fb47e264d9e0fbc8da5bd046e5e678316c866bc36ddd4b56d4eb0de682d5b
        - nvcr.io/nvidia/gpu-operator@sha256:8e0969cffc030a89c4acd68e64d41dd54e3bce8a794106b178d4dbd636a07f1c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:7bcc188703f2fac630f7a4ff8960e6733ac3a29adf6a946533b796d9a27b8acf
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:33de74eb590f071403c17b6c210c02963245851971168bc0c07c06c100a9f376
        - nvcr.io/nvidia/cuda@sha256:de5b598bca89850567c4c104411d66bb52f47c9179199e6a3be6829b7ac586a2
        - nvcr.io/nvidia/driver@sha256:07e11f85d54d49ec9648fb06e148b8d832ee1f9c3549a915eee853c9ef2949c2
        version: 24.6.2
      entries:
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.6
    - currentCSV: gpu-operator-certified.v24.9.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "useOpenKernelModules": false,
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": [
                        {
                          "name": "WITH_WORKLOAD",
                          "value": "false"
                        }
                      ]
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
          createdAt: Mon Jan 27 16:10:52 PST 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <24.9.2'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:cf39d78ced7fb5727a9668ee2cd44b14bb7a23a95b83d5464b7d755740e02121
        - nvcr.io/nvidia/gpu-operator@sha256:92f5262bb4934e0ea29eb10ec0ad4fc293164f2f2f94d48a935922791abbfdac
        - nvcr.io/nvidia/driver@sha256:026f8f2d29b7058ecaaa5b98666e28e1b3646b73e7d7cda3dc8026d47a929152
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:857f7669ae5ce2a43a3b5691fa970085098321ef8e90a896fa8dda3dee5b1d2b
        - nvcr.io/nvidia/cuda@sha256:b3b340847db9abd0c39703278f5638d3e65e1369e9a526a19b3138d5a6563527
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:7edd7a0413dcb39b6e3bcefaf06812f3293c8e480ca10783e821a561ed686200
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:4ffa1cd2a6497eb647a89ed259dcfb007554737b9d80f69bc173a2c3cd72a1da
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:e8398b2451996e5c64fc3855e7a3c93b667a85f69c0e259049d3b0012c21545d
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:34eab99d7992a57c35803ec5e0afd538d16e997341e951fef4c4019afb08793e
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:40830d3561c14743f484d45b498141f9e86b1308e16fae3978110783927264ab
        - nvcr.io/nvidia/driver@sha256:38b66a8d44cab9e2c62da9e101f32cd9dbcb5e02d8e57b47671284d374ca3695
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b072c5793be65eee556eaff1b9cbbd115a1ef29982be95b2959adfcb4bc72382
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:d934033f9ff89087d555850eb1da150eaaa27b11e5ad21db1b874b8bc0b8d35b
        - nvcr.io/nvidia/driver@sha256:d55b57938866e538acc3a71ca32f8cf87e71c591abd4a34695ee428e7ec2fa73
        - nvcr.io/nvidia/driver@sha256:6fe74322562c726c8fade184d8c45ebae3da7b1ea0a21f0ff9dc42c66c65e692
        - nvcr.io/nvidia/k8s-device-plugin@sha256:7089559ce6153018806857f5049085bae15b3bf6f1c8bd19d8b12f707d087dea
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:b897e25e03133513f9afdb5abe38cbf0f2ac0556c27518d9d80e792eb07ecf28
        version: 24.9.2
      entries:
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v24.9
    - currentCSV: gpu-operator-certified.v25.3.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nvidia.com/v1",
                "kind": "ClusterPolicy",
                "metadata": {
                  "name": "gpu-cluster-policy"
                },
                "spec": {
                  "operator": {
                    "defaultRuntime": "crio",
                    "use_ocp_driver_toolkit": true,
                    "initContainer": {
                    }
                  },
                  "sandboxWorkloads": {
                    "enabled": false,
                    "defaultWorkload": "container"
                  },
                  "driver": {
                    "enabled": true,
                    "useNvidiaDriverCRD": false,
                    "kernelModuleType": "auto",
                    "upgradePolicy": {
                      "autoUpgrade": true,
                      "drain": {
                        "deleteEmptyDir": false,
                        "enable": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "maxParallelUpgrades": 1,
                      "maxUnavailable": "25%",
                      "podDeletion": {
                        "deleteEmptyDir": false,
                        "force": false,
                        "timeoutSeconds": 300
                      },
                      "waitForCompletion": {
                        "timeoutSeconds": 0
                      }
                    },
                    "repoConfig": {
                      "configMapName": ""
                    },
                    "certConfig": {
                      "name": ""
                    },
                    "licensingConfig": {
                      "nlsEnabled": true,
                      "configMapName": ""
                    },
                    "virtualTopology": {
                      "config": ""
                    },
                    "kernelModuleConfig": {
                      "name": ""
                    }
                  },
                  "dcgmExporter": {
                    "enabled": true,
                    "config": {
                      "name": ""
                    },
                    "serviceMonitor": {
                      "enabled": true
                    }
                  },
                  "dcgm": {
                    "enabled": true
                  },
                  "daemonsets": {
                    "updateStrategy": "RollingUpdate",
                    "rollingUpdate": {
                      "maxUnavailable": "1"
                    }
                  },
                  "devicePlugin": {
                    "enabled": true,
                    "config": {
                      "name": "",
                      "default": ""
                    },
                    "mps": {
                      "root": "/run/nvidia/mps"
                    }
                  },
                  "gfd": {
                    "enabled": true
                  },
                  "migManager": {
                    "enabled": true
                  },
                  "nodeStatusExporter": {
                    "enabled": true
                  },
                  "mig": {
                    "strategy": "single"
                  },
                  "toolkit": {
                    "enabled": true
                  },
                  "validator": {
                    "plugin": {
                      "env": []
                    }
                  },
                  "vgpuManager": {
                    "enabled": false
                  },
                  "vgpuDeviceManager": {
                    "enabled": true
                  },
                  "sandboxDevicePlugin": {
                    "enabled": true
                  },
                  "vfioManager": {
                    "enabled": true
                  },
                  "gds": {
                    "enabled": false
                  },
                  "gdrcopy": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "nvidia.com/v1alpha1",
                "kind": "NVIDIADriver",
                "metadata": {
                  "name": "gpu-driver"
                },
                "spec": {
                  "driverType": "gpu",
                  "repository": "nvcr.io/nvidia",
                  "image": "driver",
                  "version": "sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735",
                  "nodeSelector": {},
                  "manager": {},
                  "repoConfig": {
                    "name": ""
                  },
                  "certConfig": {
                    "name": ""
                  },
                  "licensingConfig": {
                    "nlsEnabled": true,
                    "name": ""
                  },
                  "virtualTopologyConfig": {
                    "name": ""
                  },
                  "kernelModuleConfig": {
                    "name": ""
                  }
                }
              }
            ]
          capabilities: Deep Insights
          categories: AI/Machine Learning, OpenShift Optional
          certified: "true"
          containerImage: nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
          createdAt: Thu Sep 18 18:11:36 PDT 2025
          description: Automate the management and monitoring of NVIDIA GPUs.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.9.0 <25.3.4'
          operatorframework.io/suggested-namespace: nvidia-gpu-operator
          operators.operatorframework.io/builder: operator-sdk-v1.4.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          provider: NVIDIA
          repository: http://github.com/NVIDIA/gpu-operator
          support: NVIDIA
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NVIDIADriver allows you to deploy the NVIDIA driver
            displayName: NVIDIADriver
            kind: NVIDIADriver
            name: nvidiadrivers.nvidia.com
            version: v1alpha1
          - description: ClusterPolicy allows you to configure the GPU Operator
            displayName: ClusterPolicy
            kind: ClusterPolicy
            name: clusterpolicies.nvidia.com
            version: v1
        description: |
          Kubernetes provides access to special hardware resources such as NVIDIA GPUs, NICs, Infiniband adapters and other devices through the [device plugin framework](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/). However, configuring and managing nodes with these hardware resources requires configuration of multiple software components such as drivers, container runtimes or other libraries which are difficult and prone to errors.
          The NVIDIA GPU Operator uses the [operator framework](https://cloud.redhat.com/blog/introducing-the-operator-framework) within Kubernetes to automate the management of all NVIDIA software components needed to provision and monitor GPUs. These components include the NVIDIA drivers (to enable CUDA), Kubernetes device plugin for GPUs, the NVIDIA Container Runtime, automatic node labelling and NVIDIA DCGM exporter.
          Visit the official site of the [GPU Operator](https://github.com/NVIDIA/gpu-operator) for more information. For getting started with using the GPU Operator with OpenShift, see the instructions [here](https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/openshift/contents.html).
        displayName: NVIDIA GPU Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - gpu
        - cuda
        - compute
        - operator
        - deep learning
        - monitoring
        - tesla
        maintainers:
        - email: operator_feedback@nvidia.com
          name: NVIDIA
        maturity: stable
        provider:
          name: NVIDIA Corporation
        relatedImages:
        - nvcr.io/nvidia/gpu-operator@sha256:2e86875d61c030a9bd348a18bde8c4caa3d0c430a1aac642086b9cb1b3e3b015
        - nvcr.io/nvidia/cloud-native/dcgm@sha256:66c3d68ca2b35d47b5562300a10ff8711b953ba3a15df7dbc409d81613ebe5d0
        - nvcr.io/nvidia/driver@sha256:b6250c461c6226d5fb1383e3e8331e7e0bf07f9ed88a36988bbe5f6cc754d735
        - nvcr.io/nvidia/cloud-native/k8s-mig-manager@sha256:3191cc0684e92d655e72fe9faf3da95fe6b8a6c2eb4ad396f75ba9b98c021aaf
        - nvcr.io/nvidia/k8s-device-plugin@sha256:3c54348fe5a57e5700e7d8068e7531d2ef2d5f3ccb70c8f6bac0953432527abd
        - nvcr.io/nvidia/cloud-native/gpu-operator-validator@sha256:619e7bf48168d76a7e087c3bf6190fda9b10449c9839309fa4f9ed5b8a9e8804
        - nvcr.io/nvidia/driver@sha256:44422f10e899d3419fbd611a02fc1c6127e071a2067bd36e3debb043bf251de0
        - nvcr.io/nvidia/k8s/dcgm-exporter@sha256:6bb786c41fa7b1a6238c023566fdeaeeea3ba842a8daec015a7bd85a8037b249
        - registry.connect.redhat.com/nvidia/gpu-operator-bundle@sha256:cbd83ff041c3e6e9a5d76dc2cfa1894893c5e2f92583b16df9c418232cbce78c
        - nvcr.io/nvidia/k8s/container-toolkit@sha256:51c8f71d3b3c08ae4eb4853697e3f8e6f11e435e666e08210178e6a1faf8028f
        - nvcr.io/nvidia/cloud-native/vgpu-device-manager@sha256:c22dc67465c33d2f17932931229e61a6e7d9e32c5ed2ecefe8d10e873ffedfeb
        - nvcr.io/nvidia/cloud-native/gdrdrv@sha256:5c4e61f7ba83d7a64ff2523d447c209ce5bde1ddc79acaf1f32f19620b4912d6
        - nvcr.io/nvidia/kubevirt-gpu-device-plugin@sha256:119de9a331a47203858b99901f44d0c4a8052961b4e60327f4b100d0ab8c9df0
        - nvcr.io/nvidia/cloud-native/k8s-driver-manager@sha256:b689dfb7806ae51b0072620b47c26114f05477e8ad9725892ee10366e5d93d8d
        - nvcr.io/nvidia/driver@sha256:43aa99352e1dc47e9218f36340ae6328a6282ed1e7e9a61b4204475e980230ef
        - nvcr.io/nvidia/cuda@sha256:d19fe621624c4eb6ac931b8558daa3ecc0c3f07f1e2a52e0267e083d22dceade
        version: 25.3.4
      entries:
      - name: gpu-operator-certified.v25.3.4
        version: 25.3.4
      - name: gpu-operator-certified.v25.3.3
        version: 25.3.3
      - name: gpu-operator-certified.v25.3.2
        version: 25.3.2
      - name: gpu-operator-certified.v25.3.1
        version: 25.3.1
      - name: gpu-operator-certified.v25.3.0
        version: 25.3.0
      - name: gpu-operator-certified.v24.9.2
        version: 24.9.2
      - name: gpu-operator-certified.v24.9.1
        version: 24.9.1
      - name: gpu-operator-certified.v24.9.0
        version: 24.9.0
      - name: gpu-operator-certified.v24.6.2
        version: 24.6.2
      - name: gpu-operator-certified.v24.6.1
        version: 24.6.1
      - name: gpu-operator-certified.v24.6.0
        version: 24.6.0
      - name: gpu-operator-certified.v24.3.0
        version: 24.3.0
      - name: gpu-operator-certified.v23.9.2
        version: 23.9.2
      - name: gpu-operator-certified.v23.9.1
        version: 23.9.1
      - name: gpu-operator-certified.v23.9.0
        version: 23.9.0
      - name: gpu-operator-certified.v23.6.1
        version: 23.6.1
      - name: gpu-operator-certified.v23.6.0
        version: 23.6.0
      - name: gpu-operator-certified.v23.3.2
        version: 23.3.2
      - name: gpu-operator-certified.v23.3.1
        version: 23.3.1
      - name: gpu-operator-certified.v23.3.0
        version: 23.3.0
      - name: gpu-operator-certified.v22.9.2
        version: 22.9.2
      - name: gpu-operator-certified.v22.9.1
        version: 22.9.1
      - name: gpu-operator-certified.v22.9.0
        version: 22.9.0
      - name: gpu-operator-certified.v1.11.1
        version: 1.11.1
      - name: gpu-operator-certified.v1.11.0
        version: 1.11.0
      - name: gpu-operator-certified.v1.10.1
        version: 1.10.1
      - name: gpu-operator-certified.v1.10.0
        version: 1.10.0
      name: v25.3
    defaultChannel: v25.3
    packageName: gpu-operator-certified
    provider:
      name: NVIDIA Corporation
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-sfn-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-sfn-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "Activity",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "sfn.services.k8s.aws/v1alpha1",
                "kind": "StateMachine",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
          createdAt: "2025-09-19T18:41:33Z"
          description: AWS Step Functions controller is a service controller for managing
            Step Functions resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Activity represents the state of an AWS sfn Activity resource.
            displayName: Activity
            kind: Activity
            name: activities.sfn.services.k8s.aws
            version: v1alpha1
          - description: StateMachine represents the state of an AWS sfn StateMachine
              resource.
            displayName: StateMachine
            kind: StateMachine
            name: statemachines.sfn.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Step Functions resources in AWS from within your Kubernetes cluster.

          **About Amazon Step FUnctions**

          Step Functions is a serverless orchestration service that lets you combine AWS Lambda functions and other AWS services to build business-critical applications. Through Step Functions' graphical console, you see your application’s workflow as a series of event-driven steps.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon Step Functions
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sfn
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon Step Functions Developer Resources
          url: https://aws.amazon.com/step-functions/resources
        maintainers:
        - email: ack-maintainers@amazon.com
          name: step functions maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/sfn-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-sfn-controller@sha256:01036536877e5b236dc83774eb3b6fe9493eaa593cf5d1cb5b103798d0727bf8
        version: 1.1.1
      entries:
      - name: ack-sfn-controller.v1.1.1
        version: 1.1.1
      - name: ack-sfn-controller.v1.1.0
        version: 1.1.0
      - name: ack-sfn-controller.v1.0.29
        version: 1.0.29
      - name: ack-sfn-controller.v1.0.28
        version: 1.0.28
      - name: ack-sfn-controller.v1.0.27
        version: 1.0.27
      - name: ack-sfn-controller.v1.0.26
        version: 1.0.26
      - name: ack-sfn-controller.v1.0.25
        version: 1.0.25
      - name: ack-sfn-controller.v1.0.24
        version: 1.0.24
      - name: ack-sfn-controller.v1.0.23
        version: 1.0.23
      - name: ack-sfn-controller.v1.0.22
        version: 1.0.22
      - name: ack-sfn-controller.v1.0.21
        version: 1.0.21
      - name: ack-sfn-controller.v1.0.20
        version: 1.0.20
      - name: ack-sfn-controller.v1.0.19
        version: 1.0.19
      - name: ack-sfn-controller.v1.0.18
        version: 1.0.18
      - name: ack-sfn-controller.v1.0.17
        version: 1.0.17
      - name: ack-sfn-controller.v1.0.16
        version: 1.0.16
      - name: ack-sfn-controller.v1.0.15
        version: 1.0.15
      - name: ack-sfn-controller.v1.0.14
        version: 1.0.14
      - name: ack-sfn-controller.v1.0.13
        version: 1.0.13
      - name: ack-sfn-controller.v1.0.11
        version: 1.0.11
      - name: ack-sfn-controller.v1.0.10
        version: 1.0.10
      - name: ack-sfn-controller.v1.0.9
        version: 1.0.9
      - name: ack-sfn-controller.v1.0.8
        version: 1.0.8
      - name: ack-sfn-controller.v1.0.6
        version: 1.0.6
      - name: ack-sfn-controller.v1.0.5
        version: 1.0.5
      - name: ack-sfn-controller.v1.0.2
        version: 1.0.2
      - name: ack-sfn-controller.v1.0.1
        version: 1.0.1
      - name: ack-sfn-controller.v1.0.0
        version: 1.0.0
      - name: ack-sfn-controller.v0.1.3
        version: 0.1.3
      - name: ack-sfn-controller.v0.1.2
        version: 0.1.2
      - name: ack-sfn-controller.v0.1.1
        version: 0.1.1
      - name: ack-sfn-controller.v0.1.0
        version: 0.1.0
      - name: ack-sfn-controller.v0.0.19
        version: 0.0.19
      - name: ack-sfn-controller.v0.0.18
        version: 0.0.18
      - name: ack-sfn-controller.v0.0.17
        version: 0.0.17
      - name: ack-sfn-controller.v0.0.16
        version: 0.0.16
      - name: ack-sfn-controller.v0.0.15
        version: 0.0.15
      - name: ack-sfn-controller.v0.0.14
        version: 0.0.14
      - name: ack-sfn-controller.v0.0.13
        version: 0.0.13
      - name: ack-sfn-controller.v0.0.12
        version: 0.0.12
      - name: ack-sfn-controller.v0.0.11
        version: 0.0.11
      - name: ack-sfn-controller.v0.0.10
        version: 0.0.10
      - name: ack-sfn-controller.v0.0.9
        version: 0.0.9
      - name: ack-sfn-controller.v0.0.8
        version: 0.0.8
      - name: ack-sfn-controller.v0.0.7
        version: 0.0.7
      - name: ack-sfn-controller.v0.0.6
        version: 0.0.6
      - name: ack-sfn-controller.v0.0.5
        version: 0.0.5
      name: alpha
    defaultChannel: alpha
    packageName: ack-sfn-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app.kubernetes.io/version: 5.8.3
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Crunchy Data
      provider-url: https://www.crunchydata.com/
    name: postgresql
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: postgresoperator.v5.8.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "CrunchyBridgeCluster",
                "metadata": {
                  "name": "example-crunchybridgecluster"
                },
                "spec": {
                  "clusterName": "example-bridge-cluster",
                  "isHa": false,
                  "majorVersion": 16,
                  "plan": "standard-4",
                  "provider": "aws",
                  "region": "us-west-2",
                  "secret": "crunchy-bridge-api-key",
                  "storage": "10Gi"
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGAdmin",
                "metadata": {
                  "name": "example-pgadmin",
                  "namespace": "openshift-operators"
                },
                "spec": {
                  "dataVolumeClaimSpec": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "resources": {
                      "requests": {
                        "storage": "1Gi"
                      }
                    }
                  },
                  "serverGroups": [
                    {
                      "name": "Crunchy Postgres for Kubernetes",
                      "postgresClusterSelector": {}
                    }
                  ]
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PGUpgrade",
                "metadata": {
                  "name": "example-upgrade"
                },
                "spec": {
                  "fromPostgresVersion": 16,
                  "postgresClusterName": "example",
                  "toPostgresVersion": 17
                }
              },
              {
                "apiVersion": "postgres-operator.crunchydata.com/v1beta1",
                "kind": "PostgresCluster",
                "metadata": {
                  "name": "example"
                },
                "spec": {
                  "backups": {
                    "pgbackrest": {
                      "repos": [
                        {
                          "name": "repo1",
                          "volume": {
                            "volumeClaimSpec": {
                              "accessModes": [
                                "ReadWriteOnce"
                              ],
                              "resources": {
                                "requests": {
                                  "storage": "1Gi"
                                }
                              }
                            }
                          }
                        }
                      ]
                    }
                  },
                  "instances": [
                    {
                      "dataVolumeClaimSpec": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "resources": {
                          "requests": {
                            "storage": "1Gi"
                          }
                        }
                      },
                      "replicas": 1
                    }
                  ],
                  "postgresVersion": 16
                }
              }
            ]
          capabilities: Auto Pilot
          categories: Database
          containerImage: registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
          createdAt: 2019-12-31 19:40Z
          description: Production Postgres Made Easy
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          repository: https://github.com/CrunchyData/postgres-operator
          support: crunchydata.com
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CrunchyBridgeCluster is the Schema for the crunchybridgeclusters
              API
            displayName: Crunchy Bridge Cluster
            kind: CrunchyBridgeCluster
            name: crunchybridgeclusters.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGAdmin is the Schema for the PGAdmin API
            displayName: PGAdmin
            kind: PGAdmin
            name: pgadmins.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PGUpgrade is the Schema for the pgupgrades API
            displayName: PGUpgrade
            kind: PGUpgrade
            name: pgupgrades.postgres-operator.crunchydata.com
            version: v1beta1
          - description: PostgresCluster is the Schema for the postgresclusters API
            displayName: Postgres Cluster
            kind: PostgresCluster
            name: postgresclusters.postgres-operator.crunchydata.com
            version: v1beta1
        description: |-
          [Crunchy Postgres for Kubernetes](https://www.crunchydata.com/products/crunchy-postgresql-for-kubernetes), is the leading Kubernetes native
          Postgres solution. Built on PGO, the Postgres Operator from Crunchy Data, Crunchy Postgres for Kubernetes gives you a declarative Postgres
          solution that automatically manages your PostgreSQL clusters.

          Designed for your GitOps workflows, it is [easy to get started](https://access.crunchydata.com/documentation/postgres-operator/latest/quickstart)
          with Crunchy Postgres for Kubernetes. Within a few moments, you can have a production grade Postgres cluster complete with high availability, disaster
          recovery, and monitoring, all over secure TLS communications. Even better, Crunchy Postgres for Kubernetes lets you easily customize your Postgres
          cluster to tailor it to your workload!

          With conveniences like cloning Postgres clusters to using rolling updates to roll out disruptive changes with minimal downtime, Crunchy Postgres
          for Kubernetes is ready to support your Postgres data at every stage of your release pipeline. Built for resiliency and uptime, Crunchy Postgres
          for Kubernetes will keep your Postgres cluster in a desired state so you do not need to worry about it.

          Crunchy Postgres for Kubernetes is developed with many years of production experience in automating Postgres management on Kubernetes, providing
          a seamless cloud native Postgres solution to keep your data always available.

          Crunchy Postgres for Kubernetes is made available to users without an active Crunchy Data subscription in connection with Crunchy Data's
          [Developer Program](https://www.crunchydata.com/developers/terms-of-use).
          For more information, please contact us at [info@crunchydata.com](mailto:info@crunchydata.com).

          - **PostgreSQL Cluster Provisioning**: [Create, Scale, & Delete PostgreSQL clusters with ease][provisioning],
            while fully customizing your Pods and PostgreSQL configuration!
          - **High-Availability**: Safe, automated failover backed by a [distributed consensus based high-availability solution][high-availability].
            Uses [Pod Anti-Affinity][k8s-anti-affinity] to help resiliency; you can configure how aggressive this can be!
            Failed primaries automatically heal, allowing for faster recovery time. You can even create regularly scheduled
            backups as well and set your backup retention policy
          - **Disaster Recovery**: [Backups][backups] and [restores][disaster-recovery] leverage the open source [pgBackRest][] utility and
            [includes support for full, incremental, and differential backups as well as efficient delta restores][backups].
            Set how long you want your backups retained for. Works great with very large databases!
          - **Monitoring**: [Track the health of your PostgreSQL clusters][monitoring] using the open source [pgMonitor][] library and OpenTelemetry.
          - **Clone**: [Create new clusters from your existing clusters or backups][clone] with efficient data cloning.
          - **TLS**: All connections are over [TLS][tls]. You can also [bring your own TLS infrastructure][tls] if you do not want to use the provided defaults.
          - **Connection Pooling**: Advanced [connection pooling][pool] support using [pgBouncer][].
          - **Affinity and Tolerations**: Have your PostgreSQL clusters deployed to [Kubernetes Nodes][k8s-nodes] of your preference.
            Set your [pod anti-affinity][k8s-anti-affinity], node affinity, Pod tolerations and more rules to customize your deployment topology!
          - **PostgreSQL Major Version Upgrades**: Perform a [PostgreSQL major version upgrade][major-version-upgrade] declaratively.
          - **Database Administration**: Easily deploy [pgAdmin4][pgadmin] to administer your PostgresClusters' databases.
            The automatic discovery of PostgresClusters ensures that you are able to seamlessly access any databases within your environment from the pgAdmin4 GUI.
          - **Full Customizability**: Crunchy PostgreSQL for Kubernetes makes it easy to get your own PostgreSQL-as-a-Service up and running
            and fully customize your deployments, including:
              - Choose the resources for your Postgres cluster: [container resources and storage size][resize-cluster]. [Resize at any time][resize-cluster] with minimal disruption.
              - Use your own container image repository, including support `imagePullSecrets` and private repositories
              - [Customize your PostgreSQL configuration][customize-cluster]

          and much more!

          [backups]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery
          [clone]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/backups-disaster-recovery/disaster-recovery
          [customize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster
          [disaster-recovery]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/backups-disaster-recovery/disaster-recovery
          [high-availability]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/high-availability
          [major-version-upgrade]: https://access.crunchydata.com/documentation/postgres-operator/v5/guides/major-postgres-version-upgrade/
          [monitoring]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/day-two/monitoring
          [pool]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [provisioning]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/create-cluster
          [resize-cluster]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/cluster-management/resize-cluster
          [tls]: https://access.crunchydata.com/documentation/postgres-operator/latest/tutorials/day-two/customize-cluster#customize-tls

          [k8s-anti-affinity]: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#inter-pod-affinity-and-anti-affinity
          [k8s-nodes]: https://kubernetes.io/docs/concepts/architecture/nodes/

          [pgAdmin]: https://www.pgadmin.org/
          [pgBackRest]: https://www.pgbackrest.org
          [pgBouncer]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials/basic-setup/connection-pooling
          [pgMonitor]: https://github.com/CrunchyData/pgmonitor

          ## Post-Installation

          ### Tutorial

          Want to [learn more about the PostgreSQL Operator][tutorial]? Browse through the [tutorial][] to learn more about what you can do,
          [join the Discord server][discord] for community support, or check out the [PGO GitHub repo][ghrepo] to learn more
          about the open source Postgres Operator project that powers Crunchy Postgres for Kubernetes.

          [tutorial]: https://access.crunchydata.com/documentation/postgres-operator/v5/tutorials
          [discord]: https://discord.gg/a7vWKG8Ec9
          [ghrepo]: https://github.com/CrunchyData/postgres-operator
        displayName: Crunchy Postgres for Kubernetes
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - postgres
        - postgresql
        - database
        - sql
        - operator
        - crunchy data
        links:
        - name: Crunchy Data
          url: https://www.crunchydata.com/
        - name: Documentation
          url: https://access.crunchydata.com/documentation/postgres-operator/v5/
        maintainers:
        - email: info@crunchydata.com
          name: Crunchy Data
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Crunchy Data
          url: https://www.crunchydata.com/
        relatedImages:
        - registry.developers.crunchydata.com/crunchydata/postgres-operator:ubi9-5.8.3-0
        - quay.io/community-operator-pipeline-prod/postgresql@sha256:f9b7137598dcaf0774de71f4633c7e3b0359b0b40f313ad9c91dc8eca58592eb
        version: 5.8.3
      entries:
      - name: postgresoperator.v5.8.3
        version: 5.8.3
      - name: postgresoperator.v5.8.2
        version: 5.8.2
      - name: postgresoperator.v5.8.1
        version: 5.8.1
      - name: postgresoperator.v5.8.0
        version: 5.8.0
      - name: postgresoperator.v5.7.4
        version: 5.7.4
      - name: postgresoperator.v5.7.3
        version: 5.7.3
      - name: postgresoperator.v5.7.2
        version: 5.7.2
      - name: postgresoperator.v5.7.1
        version: 5.7.1
      - name: postgresoperator.v5.7.0
        version: 5.7.0
      - name: postgresoperator.v5.6.1
        version: 5.6.1
      - name: postgresoperator.v5.6.0
        version: 5.6.0
      - name: postgresoperator.v5.5.2
        version: 5.5.2
      - name: postgresoperator.v5.5.1
        version: 5.5.1
      - name: postgresoperator.v5.5.0
        version: 5.5.0
      - name: postgresoperator.v5.4.3
        version: 5.4.3
      - name: postgresoperator.v5.4.2
        version: 5.4.2
      - name: postgresoperator.v5.4.1
        version: 5.4.1
      - name: postgresoperator.v5.3.0
        version: 5.3.0
      - name: postgresoperator.v5.2.0
        version: 5.2.0
      - name: postgresoperator.v5.1.3
        version: 5.1.3
      - name: postgresoperator.v5.1.2
        version: 5.1.2
      - name: postgresoperator.v5.1.1
        version: 5.1.1
      - name: postgresoperator.v5.1.0
        version: 5.1.0
      name: v5
    defaultChannel: v5
    packageName: postgresql
    provider:
      name: Crunchy Data
      url: https://www.crunchydata.com/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      full_version: 4.18.11-1
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat, Inc.
      provider-url: ""
    name: odr-hub-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: odr-hub-operator.v4.17.13-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-09T14:20:05Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.17.13-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:2c36dace20878ffea08b9f6a95de8c7cc452daa69fe74f025fc777277726312e
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:36393d8945938325e4f330eee801ba28ab9b7604f58ea927004e17bb514f328f
        version: 4.17.13-rhodf
      entries:
      - name: odr-hub-operator.v4.17.13-rhodf
        version: 4.17.13-rhodf
      - name: odr-hub-operator.v4.17.12-rhodf
        version: 4.17.12-rhodf
      - name: odr-hub-operator.v4.17.11-rhodf
        version: 4.17.11-rhodf
      - name: odr-hub-operator.v4.17.10-rhodf
        version: 4.17.10-rhodf
      - name: odr-hub-operator.v4.17.9-rhodf
        version: 4.17.9-rhodf
      - name: odr-hub-operator.v4.17.8-rhodf
        version: 4.17.8-rhodf
      - name: odr-hub-operator.v4.17.7-rhodf
        version: 4.17.7-rhodf
      - name: odr-hub-operator.v4.17.6-rhodf
        version: 4.17.6-rhodf
      - name: odr-hub-operator.v4.17.5-rhodf
        version: 4.17.5-rhodf
      - name: odr-hub-operator.v4.17.4-rhodf
        version: 4.17.4-rhodf
      - name: odr-hub-operator.v4.17.3-rhodf
        version: 4.17.3-rhodf
      - name: odr-hub-operator.v4.17.2-rhodf
        version: 4.17.2-rhodf
      - name: odr-hub-operator.v4.17.1-rhodf
        version: 4.17.1-rhodf
      - name: odr-hub-operator.v4.17.0-rhodf
        version: 4.17.0-rhodf
      name: stable-4.17
    - currentCSV: odr-hub-operator.v4.18.11-rhodf
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east"
                },
                "spec": {
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRCluster",
                "metadata": {
                  "name": "east1"
                },
                "spec": {
                  "cidrs": [
                    "192.168.39.45/32"
                  ],
                  "clusterFence": "ManuallyFenced",
                  "region": "east",
                  "s3ProfileName": "s3-profile-of-east1"
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPlacementControl",
                "metadata": {
                  "name": "drplacementcontrol-sample",
                  "namespace": "application-namespace"
                },
                "spec": {
                  "drPolicyRef": {
                    "name": "drpolicy-sample"
                  },
                  "kubeObjectProtection": {
                    "captureInterval": "1m",
                    "captureOrder": [
                      {
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "name": "config"
                      },
                      {
                        "excludedResources": [
                          "ConfigMap",
                          "Secret"
                        ],
                        "includeClusterResources": true,
                        "name": "everything"
                      }
                    ],
                    "recoverOrder": [
                      {
                        "backupName": "config",
                        "includedResources": [
                          "ConfigMap",
                          "Secret"
                        ]
                      },
                      {
                        "backupName": "everything",
                        "includedResources": [
                          "*"
                        ]
                      }
                    ]
                  },
                  "placementRef": {
                    "kind": "PlacementRule",
                    "name": "application-placement-rule"
                  },
                  "preferredCluster": "east",
                  "pvcSelector": {
                    "matchLabels": {
                      "any-pvc-label": "value"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east",
                    "west"
                  ],
                  "replicationClassSelector": {
                    "matchLabels": {
                      "class": "ramen-replication-class"
                    }
                  },
                  "schedulingInterval": "1h",
                  "volumeSnapshotClassSelector": {
                    "matchLabels": {
                      "class": "ramen-snapshot-class"
                    }
                  }
                }
              },
              {
                "apiVersion": "ramendr.openshift.io/v1alpha1",
                "kind": "DRPolicy",
                "metadata": {
                  "name": "metrodr-drpolicy-sample"
                },
                "spec": {
                  "drClusters": [
                    "east1",
                    "east2"
                  ],
                  "replicationClassSelector": {},
                  "schedulingInterval": {},
                  "volumeSnapshotClassSelector": {}
                }
              }
            ]
          capabilities: Seamless Upgrades
          createdAt: "2025-09-11T15:27:38Z"
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.2.0 <4.18.11-rhodf'
          operatorframework.io/suggested-namespace: openshift-dr-system
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: '["OpenShift Platform Plus","OpenShift
            Data Foundation Essentials","OpenShift Data Foundation Advanced"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: DRCluster is the Schema for the drclusters API
            displayName: DRCluster
            kind: DRCluster
            name: drclusters.ramendr.openshift.io
            version: v1alpha1
          - description: DRPlacementControl is the Schema for the drplacementcontrols
              API
            displayName: DRPlacement Control
            kind: DRPlacementControl
            name: drplacementcontrols.ramendr.openshift.io
            version: v1alpha1
          - description: DRPolicy is the Schema for the drpolicies API
            displayName: DRPolicy
            kind: DRPolicy
            name: drpolicies.ramendr.openshift.io
            version: v1alpha1
        description: OpenShift DR Hub is a disaster-recovery orchestrator for stateful
          applications. It operates from an Advanced Cluster Management (ACM) hub
          cluster to orchestrate the recovery of application state, and scheduling
          of ACM PlacementRule for disaster recovery operations.
        displayName: Openshift DR Hub Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - Storage
        - Integration & Delivery
        - OpenShift Optional
        links:
        - name: Source Code
          url: https://github.com/red-hat-storage/ramen
        maintainers:
        - email: ocs-support@redhat.com
          name: Red Hat Support
        maturity: alpha
        provider:
          name: Red Hat, Inc.
        relatedImages:
        - registry.redhat.io/odf4/odr-hub-operator-bundle@sha256:bb352ffcfe7e79981f96fc5b5a4f2f557fc7fe39de27c36f195349bf4ad6248b
        - registry.redhat.io/odf4/odr-rhel9-operator@sha256:c4451e0287045b8fecfdc44757675257f5dbbab50b97a5ad4e4e009b23a30cf5
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:d27c5b239add0c308b0278af62c22d8545bc09fc32282b1ec0350e68654ad36b
        version: 4.18.11-rhodf
      entries:
      - name: odr-hub-operator.v4.18.11-rhodf
        version: 4.18.11-rhodf
      - name: odr-hub-operator.v4.18.10-rhodf
        version: 4.18.10-rhodf
      - name: odr-hub-operator.v4.18.9-rhodf
        version: 4.18.9-rhodf
      - name: odr-hub-operator.v4.18.8-rhodf
        version: 4.18.8-rhodf
      - name: odr-hub-operator.v4.18.7-rhodf
        version: 4.18.7-rhodf
      - name: odr-hub-operator.v4.18.6-rhodf
        version: 4.18.6-rhodf
      - name: odr-hub-operator.v4.18.5-rhodf
        version: 4.18.5-rhodf
      - name: odr-hub-operator.v4.18.4-rhodf
        version: 4.18.4-rhodf
      - name: odr-hub-operator.v4.18.3-rhodf
        version: 4.18.3-rhodf
      - name: odr-hub-operator.v4.18.2-rhodf
        version: 4.18.2-rhodf
      - name: odr-hub-operator.v4.18.1-rhodf
        version: 4.18.1-rhodf
      - name: odr-hub-operator.v4.18.0-rhodf
        version: 4.18.0-rhodf
      name: stable-4.18
    defaultChannel: stable-4.18
    packageName: odr-hub-operator
    provider:
      name: Red Hat, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Kong Inc.
      provider-url: https://konghq.com
    name: kong-gateway-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kong-gateway-operator.v0.0.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "ControlPlane",
                "metadata": {
                  "name": "controlplane-example"
                },
                "spec": {
                  "dataplane": "dataplane-example",
                  "gatewayClass": "kong"
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "DataPlane",
                "metadata": {
                  "name": "dataplane-example"
                },
                "spec": {
                  "env": [
                    {
                      "name": "KONG_LOG_LEVEL",
                      "value": "debug"
                    }
                  ]
                }
              },
              {
                "apiVersion": "gateway-operator.konghq.com/v1alpha1",
                "kind": "GatewayConfiguration",
                "metadata": {
                  "name": "kong",
                  "namespace": "default"
                },
                "spec": {
                  "controlPlaneDeploymentOptions": {
                    "env": [
                      {
                        "name": "CONTROLLER_LOG_LEVEL",
                        "value": "debug"
                      }
                    ]
                  },
                  "dataPlaneDeploymentOptions": {
                    "containerImage": "kong",
                    "version": "2.8"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking,Integration & Delivery
          operatorframework.io/suggested-namespace: kong-system
          operators.operatorframework.io/builder: operator-sdk-v1.23.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ControlPlane is the Schema for the controlplanes API
            displayName: Control Plane
            kind: ControlPlane
            name: controlplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: DataPlane is the Schema for the dataplanes API
            displayName: Data Plane
            kind: DataPlane
            name: dataplanes.gateway-operator.konghq.com
            version: v1alpha1
          - description: GatewayConfiguration is the Schema for the gatewayconfigurations
              API
            displayName: Gateway Configuration
            kind: GatewayConfiguration
            name: gatewayconfigurations.gateway-operator.konghq.com
            version: v1alpha1
        description: Kong Operator for Kubernetes. Technical preview.
        displayName: Kong
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - networking
        - gateway
        - ingress
        - kong
        - kubernetes
        - go
        links:
        - name: Github repository for the operator
          url: https://github.com/kong/gateway-operator
        maintainers:
        - email: support@konghq.com
          name: Kong Inc.
        maturity: alpha
        minKubeVersion: 1.22.0
        provider:
          name: Kong Inc.
          url: https://konghq.com
        relatedImages:
        - ghcr.io/kong/gateway-operator@sha256:76b227715a3e1915a28f8aaa1e2ff8495349d29d0d61e7b85a72d74b03fbcc7f
        - registry.connect.redhat.com/kong/gateway-operator-bundle@sha256:91b3d73bc75eb170035110e5c63747e4918f738841aa97b6df27d22e37b4879b
        - registry.connect.redhat.com/kong/kong-ingress-controller@sha256:f691db070e7c604da9e163800903d0942ec6fda879eb9dcd5d8e90618ae787da
        - registry.connect.redhat.com/kong/kong@sha256:54e7c7e57bf93810a7cc36327023b8699f754c4c40c5c289548d54170caa5db2
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:db06cc4c084dd0253134f156dddaaf53ef1c3fb3cc809e5d81711baa4029ea4c
        version: 0.0.2
      entries:
      - name: kong-gateway-operator.v0.0.2
        version: 0.0.2
      - name: kong-gateway-operator.v0.0.1
        version: 0.0.1
      name: alpha
    defaultChannel: alpha
    packageName: kong-gateway-operator
    provider:
      name: Kong Inc.
      url: https://konghq.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-hive-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: hive-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
          description: Stackable Operator for Apache Hive
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/hive-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for HiveClusterSpec via `CustomResource`
            displayName: hiveclusters.hive.stackable.tech
            kind: HiveCluster
            name: hiveclusters.hive.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Hive
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - hive
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/hive-operator@sha256:95ddee51958b27776e759184cb05ebabc814cfc1e762340175cce8757e4e9828
        - registry.connect.redhat.com/stackable/stackable-apache-hive-operator@sha256:49575f91d7c39ccbbecf2d66ea5a2e6a0ab40267aefb489330b7806ab6c9fb50
        version: 25.7.0
      entries:
      - name: hive-operator.v25.7.0
        version: 25.7.0
      - name: hive-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-hive-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: JFrog
      provider-url: ""
    name: openshiftartifactoryha-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: artifactory-ha-operator.v1.3.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "nginx": {
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
          createdAt: 2024-03-23 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/valid-subscription: Not required
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.
          If you are using operator version < 1.2.0 there is a issue with upgrade where artifactory version is not updated if you update operator version. As workaround you can update the Custom Resource OpenshiftArtifactoryHa which has hardcored image value.After upgrading to version 1.2.0 you can remove the below parameters from your CR so that future upgrade works without workaround. ``` artifactory-ha.artifactory.image.registry artifactory-ha.artifactory.image.repository artifactory-ha.artifactory.image.tag ```
          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: alpha
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:d7bd0e90aff6e8fc9fbc4119e5f1ec89eba93487b84356e3af95208c1250fe44
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:de69bb16052d5ac6fed9882eb65d49446b6fa1cc5ba830dd2a39e882712b1347
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:0986ca71419ae628892db11e7ca7da5ac1c60219c144ce9aeb2de6e683a45262
        - registry.redhat.io/rhel8/nginx-120@sha256:87f658744af695273c434fac8246b716728b691a17d5d0c5b89247362e61e45c
        version: 1.3.3
      entries:
      - name: artifactory-ha-operator.v1.3.3
        version: 1.3.3
      name: alpha
    - currentCSV: artifactory-ha-operator.v1.1.18
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.helm.k8s.io/v1",
                "kind": "OpenshiftArtifactoryHa",
                "metadata": {
                  "name": "openshiftartifactoryha"
                },
                "spec": {
                  "artifactory-ha": {
                    "artifactory": {
                      "image": {
                        "registry": "registry.connect.redhat.com",
                        "repository": "jfrog/artifactory-pro",
                        "tag": "7.27.3-1"
                      },
                      "joinKey": "OVERRIDE",
                      "masterKey": "OVERRIDE",
                      "uid": "1000721030",
                      "node": {
                        "replicaCount": 2,
                        "waitForPrimaryStartup": {
                          "enabled": false
                        }
                      }
                    },
                    "databaseUpgradeReady": true,
                    "database": {
                      "driver": "OVERRIDE",
                      "password": "OVERRIDE",
                      "type": "OVERRIDE",
                      "url": "OVERRIDE",
                      "user": "OVERRIDE"
                    },
                    "initContainerImage": "registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8",
                    "nginx": {
                      "uid": "1000720104",
                      "gid": "1000720107",
                      "http": {
                        "externalPort": 80,
                        "internalPort": 8080
                      },
                      "https": {
                        "externalPort": 443,
                        "internalPort": 8443
                      },
                      "image": {
                        "registry": "registry.redhat.io",
                        "repository": "rhel8/nginx-116",
                        "tag": "latest"
                      },
                      "service": {
                        "ssloffload": false
                      },
                      "tlsSecretName": "OVERRIDE"
                    },
                    "postgresql": {
                      "enabled": false
                    },
                    "waitForDatabase": true
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools,Integration & Delivery
          certified: "true"
          containerImage: registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
          createdAt: 2020-03-25 00:00:00+00:00
          description: JFrog Artifactory Enterprise deploys Artifactory in a high
            availability environment across multiple pods
          operators.operatorframework.io/builder: operator-sdk-v1.3.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/jfrog/JFrog-Cloud-Installers/tree/openshift4/Openshift4
          support: JFrog
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents Artifactory HA Instances
            displayName: Artifactory HA
            kind: OpenshiftArtifactoryHa
            name: openshiftartifactoryhas.charts.helm.k8s.io
            version: v1
        description: |2+
           ## Breaking change
          Please update to version 1.1.10 of the operator.
          Version 1.1.8 and 1.1.9 have issues. Please skip these versions.

          ## Overview
          Openshift Operator to deploy JFrog Artifactory Enterprise into your Openshift cluster.
          ## Security Context Constraints
          To deploy this helm chart you will need to be a cluster admin w/ access to the anyuid scc and add the operator service account to the anyuid scc.
          ```
          oc adm policy add-scc-to-user anyuid -z artifactory-ha-operator -n openshift-operators
          ```
          Add the service account for the Artifactory chart to deploy successfully:
          ``` oc adm policy add-scc-to-user anyuid -z openshiftartifactoryha-artifactory-ha -n artifactory ```
          ### Usage

          An external DB is required. The operator will not deploy a DB but will require you to specify the configuration values to connect to it.

          Search for JFrog and click JFrog Artifactory Enterprise Operator to install.

          Go to the Installed Operators.

          Wait for the JFrog Artifactory Enterprise Operator to complete the installation.

          Open the Operator and click on the provided API: Artifactory HA.

          Click Create New Instance and provide the following parameters for your DB configuration:

          ```
          DATABASE_TYPE
          DATABASE_DRIVER
          DATABASE_URL
          DATABASE_USER
          DATABASE_PASSWORD
          ```
          Master key and Join key must be supplied. To generate a new key for each run the command below:

          ```
          # Create a key
          export JOIN_KEY=$(openssl rand -hex 32)
          echo ${JOIN_KEY}
          ```

          To use TLS you will need to first create a k8s tls secret to store your .crt and .key file into.
          Then supply the value of this k8s secret into the TLS_SECRET field.
          ``` oc create secret tls my_tls_secret --cert=tls.crt --key=tls.key --namespace=my_namespace ```
          Click Create for Artifactory Enterprise to deploy into OpenShift and connect to it on the external IP exposed by the load balancer.
          ### Create a route
          To expose Artifactory from Openshift we recommend you create a new route in Openshift.
          You can either use the oc command line tool or the Openshift web console to generate a new route.
          Depending upon where you terminate TLS you will need to either specify pass through or edge.
          Command Line (Edge):
          ``` oc create route edge --service=openshiftartifactory-ha --cert=tls.crt --key=tls.key --ca-cert=ca.crt --hostname=www.example.com ```
          Or for more information visit the official Openshift documentation on routes here:
          https://docs.openshift.com/container-platform/4.6/networking/routes/route-configuration.html

        displayName: JFrog Artifactory Enterprise Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - DevOps
        - CI/CD
        - Developers
        - Software
        - Productivity
        - Artifact Repository
        - Repository Manager
        - Docker
        - Maven
        - Git
        - Helm
        - npm
        - go
        - golang
        - kubernetes
        - k8s
        - rpm
        - yum
        links:
        - name: JFrog
          url: https://www.jfrog.com
        - name: JFrog Artifact Repository
          url: https://jfrog.com/artifactory/
        - name: Artifactory Video
          url: https://www.youtube.com/watch?v=r2_A5CPo43U
        maintainers:
        - email: partner-support@jfrog.com
          name: JFrog
        maturity: stable
        provider:
          name: JFrog
        relatedImages:
        - registry.connect.redhat.com/jfrog/artifactory-operator-bundle@sha256:fc374491270b9566b0a5c80f3dd665b2227b1e8fb224eada7e5c98c82926ae08
        - registry.connect.redhat.com/jfrog/artifactory-operator@sha256:e292553e1b6e792ff19276c7f6565d7ed61dd95289eb176066329954bc93e371
        - registry.connect.redhat.com/jfrog/artifactory-pro@sha256:7289dde95d1e4d8e161a28cffd9c4a0cc96f92f7536d069ab63ec304a92254fd
        - registry.connect.redhat.com/jfrog/init@sha256:164b41720a37f75c6129c8a28d05cc56f821022313d44ff918cce4a2146d93e8
        - registry.redhat.io/rhel8/nginx-116@sha256:d71d368d6094a5777000a8d99164959ccb97fca355e855e24bd92c13a8df04f6
        version: 1.1.18
      entries:
      - name: artifactory-ha-operator.v1.1.18
        version: 1.1.18
      name: stable
    defaultChannel: alpha
    packageName: openshiftartifactoryha-operator
    provider:
      name: JFrog
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-ecrpublic-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-ecrpublic-controller.v1.1.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
          createdAt: "2025-09-20T01:21:52Z"
          description: AWS ECR PUBLIC controller is a service controller for managing
            ECR PUBLIC resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Repository represents the state of an AWS ecrpublic Repository
              resource.
            displayName: Repository
            kind: Repository
            name: repositories.ecrpublic.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon ECR PUBLIC resources in AWS from within your Kubernetes cluster.

          **About Amazon ECR PUBLIC**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon ECR PUBLIC
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - ecrpublic
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon ECR PUBLIC Developer Resources
          url: https://aws.amazon.com/ECR PUBLIC/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: ecrpublic maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/ecrpublic-controller:1.1.1
        - quay.io/community-operator-pipeline-prod/ack-ecrpublic-controller@sha256:6acdf23661d603339e998307e9d5130b21c45665f4b0c35019ea80aac3d7343e
        version: 1.1.1
      entries:
      - name: ack-ecrpublic-controller.v1.1.1
        version: 1.1.1
      - name: ack-ecrpublic-controller.v1.1.0
        version: 1.1.0
      - name: ack-ecrpublic-controller.v1.0.14
        version: 1.0.14
      - name: ack-ecrpublic-controller.v1.0.12
        version: 1.0.12
      name: alpha
    defaultChannel: alpha
    packageName: ack-ecrpublic-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: redhat-oadp-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: oadp-operator.v1.0.17
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "restic": {
                      "enable": true
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ResticRepository",
                "metadata": {
                  "name": "resticrepository",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: OpenShift Optional
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
          createdAt: "2024-09-09T20:32:09Z"
          description: |-
            OADP (OpenShift API for Data Protection) operator sets up and installs
            Velero on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.0.17'
          operatorframework.io/cluster-monitoring: "true"
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.21.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (Such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: A restic repository is an indicator of a connection from
              the restic server to the backupstoragelocation.
            displayName: ResticRepository
            kind: ResticRepository
            name: resticrepositories.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: OADP (OpenShift API for Data Protection) operator sets up and
          installs Data Protection Applications on the OpenShift platform.
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - velero
        - openshift
        - oadp
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: dymurray@redhat.com
          name: Dylan Murray
        - email: spampatt@redhat.com
          name: Shubham Dilip Pampattiwar
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-rhel8-operator@sha256:d5486149c26eb76d4b518230236409521e6b37d59ad1097762bae2a32b276e2e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel8@sha256:6e80289c93ce0ef26c776beca59eec792832dc7f2f57cb76a7321eabc12405c8
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:8b8d39f64cef8a1ed37a9ee44ab3bd5b98575ff32fd3962f94cf7a9c79e623ae
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel8@sha256:6173ab8b48c178ea8bad0d6b98e73801aa28135fcb7455146449a66543891e26
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel8@sha256:63ef9626a82b5d591abaff2e9fe76f52e2bac16b910da40af5e92920694a2a67
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel8@sha256:c88a193654c191820bee307d12ae5e721ef82740ec912de9f710481b386edde5
        - registry.redhat.io/oadp/oadp-mustgather-rhel8@sha256:f53f05a4f69a7584abb8f5767e7b6cae31546b465bbbf1894deaf54dc2b0108e
        - registry.redhat.io/oadp/oadp-velero-plugin-for-csi-rhel8@sha256:c3cbe6cf4158aec38c67582fb79358315d196de76e0da611b4b8d6183babeb58
        - registry.redhat.io/oadp/oadp-velero-rhel8@sha256:33cefb8803cd212932fe2bbb6052e53562d04084403f0166c8e077957b92c58d
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel8@sha256:ad206c2f78dbcdec43755eca3cb92229ee77e737e741dc276efec23f6379ecd9
        - registry.redhat.io/oadp/oadp-registry-rhel8@sha256:e90e8df1726698a79ec689358d323eea7a26a72290660db62443e3f0cd696a4e
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel8@sha256:66dfedd62d474aaed03b1fee142b91358271591b651a3766d8e0542d08da914f
        version: 1.0.17
      entries:
      - name: oadp-operator.v1.0.17
        version: 1.0.17
      - name: oadp-operator.v1.0.16
        version: 1.0.16
      - name: oadp-operator.v1.0.15
        version: 1.0.15
      name: stable-1.0
    - currentCSV: oadp-operator.v1.4.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "oadp.openshift.io/v1alpha1",
                "kind": "DataProtectionApplication",
                "metadata": {
                  "name": "velero-sample"
                },
                "spec": {
                  "backupLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-east-1"
                        },
                        "credential": {
                          "key": "cloud",
                          "name": "cloud-credentials"
                        },
                        "default": true,
                        "objectStorage": {
                          "bucket": "my-bucket-name",
                          "prefix": "velero"
                        },
                        "provider": "aws"
                      }
                    }
                  ],
                  "configuration": {
                    "nodeAgent": {
                      "enable": true,
                      "uploaderType": "restic"
                    },
                    "velero": {
                      "defaultPlugins": [
                        "openshift",
                        "aws",
                        "kubevirt"
                      ]
                    }
                  },
                  "snapshotLocations": [
                    {
                      "velero": {
                        "config": {
                          "profile": "default",
                          "region": "us-west-2"
                        },
                        "provider": "aws"
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Backup",
                "metadata": {
                  "name": "backup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "BackupStorageLocation",
                "metadata": {
                  "name": "backupstoragelocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DeleteBackupRequest",
                "metadata": {
                  "name": "deletebackuprequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "DownloadRequest",
                "metadata": {
                  "name": "downloadrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeBackup",
                "metadata": {
                  "name": "podvolumebackup",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "PodVolumeRestore",
                "metadata": {
                  "name": "podvolumerestore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Restore",
                "metadata": {
                  "name": "restore",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "Schedule",
                "metadata": {
                  "name": "schedule",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "ServerStatusRequest",
                "metadata": {
                  "name": "serverstatusrequest",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              },
              {
                "apiVersion": "velero.io/v1",
                "kind": "VolumeSnapshotLocation",
                "metadata": {
                  "name": "volumesnapshotlocation",
                  "namespace": "openshift-adp"
                },
                "spec": {}
              }
            ]
          capabilities: Seamless Upgrades
          categories: |
            Cloud Provider,Developer Tools,Modernization & Migration,OpenShift Optional,Storage
          certified: "false"
          containerImage: registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
          createdAt: "2025-07-10T04:59:11Z"
          description: OADP (OpenShift API for Data Protection) operator sets up and
            installs Data Protection Applications on the OpenShift platform.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "true"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=0.0.0 <1.4.5'
          operatorframework.io/suggested-namespace: openshift-adp
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
          operators.openshift.io/valid-subscription: '["OpenShift Kubernetes Engine",
            "OpenShift Container Platform", "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.34.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/openshift/oadp-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A backup repository is an indicator of a connection from
              the restic/kopia server to the backupstoragelocation.
            displayName: BackupRepository
            kind: BackupRepository
            name: backuprepositories.velero.io
            version: v1
          - description: Backup is a Velero resource that represents the capture of
              Kubernetes cluster state at a point in time (API objects and associated
              volume state).
            displayName: Backup
            kind: Backup
            name: backups.velero.io
            version: v1
          - description: BackupStorageLocation represents an object storage location
              (such as Amazon S3 Bucket) where Velero stores backup objects.
            displayName: BackupStorageLocation
            kind: BackupStorageLocation
            name: backupstoragelocations.velero.io
            version: v1
          - description: The CloudStorage API automates the creation of a bucket for
              object storage.
            displayName: CloudStorage
            kind: CloudStorage
            name: cloudstorages.oadp.openshift.io
            version: v1alpha1
          - description: DataDownload represents a data download of a volume snapshot.
              There is one DataDownload created per volume to be restored.
            displayName: DataDownload
            kind: DataDownload
            name: datadownloads.velero.io
            version: v2alpha1
          - description: DataProtectionApplication represents configuration to install
              a data protection application to safely backup and restore, perform
              disaster recovery and migrate Kubernetes cluster resources and persistent
              volumes.
            displayName: DataProtectionApplication
            kind: DataProtectionApplication
            name: dataprotectionapplications.oadp.openshift.io
            version: v1alpha1
          - description: DataUpload acts as the protocol between data mover plugins
              and data mover controller for the datamover backup operation
            displayName: DataUpload
            kind: DataUpload
            name: datauploads.velero.io
            version: v2alpha1
          - description: DeleteBackupRequest is a request to delete one or more backups.
            displayName: DeleteBackupRequest
            kind: DeleteBackupRequest
            name: deletebackuprequests.velero.io
            version: v1
          - description: DownloadRequest is a request to download an artifact from
              object storage such as a backup log file.
            displayName: DownloadRequest
            kind: DownloadRequest
            name: downloadrequests.velero.io
            version: v1
          - description: A velero pod volume backup is a restic backup of persistent
              volumes attached to a running pod.
            displayName: PodVolumeBackup
            kind: PodVolumeBackup
            name: podvolumebackups.velero.io
            version: v1
          - description: A velero pod volume restore is a restic restore of persistent
              volumes attached to a running pod.
            displayName: PodVolumeRestore
            kind: PodVolumeRestore
            name: podvolumerestores.velero.io
            version: v1
          - description: Restore is a Velero resource that represents the application
              of resources from a Velero backup to a target Kubernetes cluster.
            displayName: Restore
            kind: Restore
            name: restores.velero.io
            version: v1
          - description: Schedule is a Velero resource that represents a pre-scheduled
              or periodic Backup that should be run.
            displayName: Schedule
            kind: Schedule
            name: schedules.velero.io
            version: v1
          - description: ServerStatusRequest is a request to access current status
              information about the Velero server.
            displayName: ServerStatusRequest
            kind: ServerStatusRequest
            name: serverstatusrequests.velero.io
            version: v1
          - description: VolumeSnapshotLocation is a location where Velero stores
              volume snapshots.
            displayName: VolumeSnapshotLocation
            kind: VolumeSnapshotLocation
            name: volumesnapshotlocations.velero.io
            version: v1
        description: |
          **OpenShift API for Data Protection (OADP)** operator sets up and installs
          Velero on the OpenShift platform, allowing users to backup and restore
          applications.

          Backup and restore Kubernetes resources and internal images, at the
          granularity of a namespace, using a version of Velero appropriate for
          the installed version of OADP.

          OADP backs up Kubernetes objects and internal images by saving them as
          an archive file on object storage. OADP backs up persistent volumes
          (PVs) by creating snapshots with the native cloud snapshot API or with
          the Container Storage Interface (CSI). For cloud providers that do not
          support snapshots, OADP backs up resources and PV data with Restic or Kopia.

          * [Installing OADP for application backup and restore][install-doc]

          * [Installing OADP on a ROSA cluster and using STS, please follow the Getting Started Steps 1-3 in order to obtain the role ARN needed for using the standardized STS configuration flow via OLM][rosa-sts-doc]

          * [Frequently Asked Questions][faq-doc]

          [install-doc]: https://docs.openshift.com/container-platform/latest/backup_and_restore/application_backup_and_restore/installing/about-installing-oadp.html
          [rosa-sts-doc]: https://mobb.ninja/docs/misc/oadp/rosa-sts/
          [faq-doc]: https://access.redhat.com/articles/5456281
        displayName: OADP Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - disaster backup
        - disaster recovery
        - disaster restore
        - disaster
        - backup
        - restore
        - backup automation
        - recovery
        - data protection
        - data management
        - data
        - protection
        - management
        - application resilience
        - resilience
        - velero
        - openshift
        - oadp
        - cloud-native
        - replication
        - kopia
        - restic
        - network file system
        - nfs
        - simple cloud storage
        - s3
        - minio
        - cloud storage
        - data foundation
        - ceph
        - csi
        - container storage interface
        links:
        - name: OADP Operator
          url: https://github.com/openshift/oadp-operator
        maintainers:
        - email: oadp-maintainers@redhat.com
          name: OADP Operator maintainers
        maturity: stable
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/oadp/oadp-mustgather-rhel9@sha256:0cee6fecf052bae3ede6bb6f032b25c0c7cd63ced1e26328f2b02144f24fbb12
        - registry.redhat.io/oadp/oadp-operator-bundle@sha256:7317ed83242340597b5cdde8ee314606ddd563d0067ee85c12eb21037e2feb99
        - registry.redhat.io/oadp/oadp-velero-plugin-for-gcp-rhel9@sha256:00800ead11ea308794b1536d964f8a043e13ede9657cabb23afa42cba7cf2ce6
        - registry.redhat.io/oadp/oadp-velero-plugin-for-microsoft-azure-rhel9@sha256:9f05592bf0e5b76322ca582bdede387f9f4aabbbdc436549156c05d652c951d0
        - registry.redhat.io/oadp/oadp-velero-plugin-rhel9@sha256:3ee9d3e799cedc28c400d364b3a8790c7e88acd5f603b09ff512a22d91e265b3
        - registry.redhat.io/oadp/oadp-velero-restic-restore-helper-rhel9@sha256:e0647defc7c428f9a04a5f6d4532dc86e57b92f31b4f907253ae14af05c5993f
        - registry.redhat.io/oadp/oadp-velero-rhel9@sha256:17160f99940aa355dc91ab3d16ba6bf91faade6423434008cf8f6e5023b1e413
        - registry.redhat.io/oadp/oadp-velero-plugin-for-aws-rhel9@sha256:79d39186cae5696949813a0ea6c46eeededad29e6bde171d7a9b0011c993bb26
        - registry.redhat.io/oadp/oadp-rhel9-operator@sha256:ee010fb34448b772690a8a98808aa25aa1578a81061618f948b1d16b5d63d333
        - registry.redhat.io/oadp/oadp-velero-plugin-for-legacy-aws-rhel9@sha256:ab5c8d33684a361fa3bb10a006ef489be92cc955ad3d173bbfa77949dcc42aa8
        - registry.redhat.io/oadp/oadp-kubevirt-velero-plugin-rhel9@sha256:c3b32cb9f5b82021c08cd70467e68aad66746ef211d73612ecfe0af625706267
        version: 1.4.5
      entries:
      - name: oadp-operator.v1.4.5
        version: 1.4.5
      - name: oadp-operator.v1.4.4
        version: 1.4.4
      - name: oadp-operator.v1.4.3
        version: 1.4.3
      - name: oadp-operator.v1.4.2
        version: 1.4.2
      - name: oadp-operator.v1.4.1
        version: 1.4.1
      - name: oadp-operator.v1.4.0
        version: 1.4.0
      name: stable-1.4
    defaultChannel: stable-1.4
    packageName: redhat-oadp-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: serverless-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable
    - currentCSV: serverless-operator.v1.29.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-07-28T05:42:31Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.29.0 <1.29.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:a59ba5aeb8e66db3d8ba9074b8e553603c59e6b0e35b57b82c4ce3de589b94cc
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:6342358e6aed2339333ebf37fc223a41a0a3e199a0e4a840f3f66e5accf8f233
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:4408d1fcf34c0ffe9327d4c9c0bc772afa969d831bc5de1097164d10b4e8d900
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:3b7449287adf98f0cf0a1dc74e03b69218be3bcf5538327bc34ef776cc1c69c4
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:8273842d8e1714fa03c7e0809924b1ffc5ef282f79b9606bd8ac88c03e7a16bd
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:b780634fffe2257eb845f8c1d25498d9c654d0327c1f8841b40d90c176a2bfe0
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:3d7801f7ffe6420595779673a170235921a9090333f8ca51d5b7f664e19c82b2
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:61fb98802490323bb03ff4a7060334d8fbda9e9f95cd6366b29b1b6494e4ef8b
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:c4cff5ad5b0ac79e71a16b88640a312ab320817041317f50caf575f7e12a864e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:03a3d6f0314acf4da4f3d3d85f1e273a3ef25828d28f5b22f394441d55a645f3
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:216f58dc8c8166c0c69fc0312b667daad14cf39b7aeb7bc0fe1e313376f1a5fd
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:de0634d14962bdb8701a8c29e80b7b70684824462bc5203e5fa7f139839cb73b
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:0229a0a4f026f04f8fa0a1cd8e7afea10117fe3f12e21986ec6089769c8afcaf
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:16b39335e97bcc7c970a8ba385c61d8333e9912faaee813ccf8520d434b7bb7a
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:7dd11d385b3d5a178566d901650ecd45e4af395faec2141d57b98af8fc1e5151
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:3aab995872336e6ccb590ef6488c097cff38a91348055031145427a00aa6b732
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:fd86ea2a34de43014dea3ba2b8d705a3d94347d3a24a370a8bfd032ed239dc56
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:a10e76aa15707e708f82c8d9e8b5179bf67780d99f3c3e2e99b84ae2a0aa4ecb
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:fe916630003194dbcd3894dd880cad8b15a041a5e534cd8f76fde6296db00797
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:04866e16881f64ff2b5edd379eab0731c9d66a47c25ebfe79c8d8ad33d8116e6
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:a4943c0563c45ceccfda0a01e26be3a01c1437c89d927714c577bf6060495958
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:f37f494956adf0892ffca34d2a4c9b839ce28dc9f2779ca32e8801fb4f0e9105
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ad399867cbb712111db6affaf5c9e3573699ed0eb0def8ce6057de02054ced4f
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c12c471902b8e2596588d9e553ad59b501d979dd04f301d45f7d45e794737a8c
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:e5db11a0adb889f907cad797dad9cf2de09c433e5522fa333193cb35575802d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:b520305189c7567b2f8ac867b20468259fb22cd3edc5b4576af2bfcdb0c11855
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:594c3ac24bf59802e2d3fa127e4051277bc07b6444f934f535fddeb668ccc144
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:10788ca67809e5a2d76d6d72df7dd64ea4407773999707a7681e0cf1b5c4695f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:56e48afb1dcbf4e266ace944105f3c45042550d8d4a0528c982225a793d26d56
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:397eb425f383c1273c7976bea69a603530676a00a61f461ceb5f48d67901c4cd
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:ab90ec9b3cd9203fdcdd7c43abf95f48c9d7479a9c7af479eb1e4b6599ba9642
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:a3c5d5fd8014c432830be912aedebbc58f9c60a8cabd87ae6c42c7b8d27db8f9
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:aea5f200a845e1d4dc718d76465fdbb1eeac2b20ecae224d5d20a98214ad0357
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:31d45f10a8519addec9aac115c5edc8a2e14f9621c875d95f123d1291776ef4b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c2cb80bce34a74afa76151d2b44dc25ece73fb274cdfc6fdf243b1b0cb7d989f
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:bd134e64c156dc5c81d14522c9c9efcc1f56b2e972ac8ecb1bb1b130c76adc60
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:557e73127326be051aa6257e7daeb9721e2af04d2f06cc26cfc7a1073666b989
        version: 1.29.1
      entries:
      - name: serverless-operator.v1.29.1
        version: 1.29.1
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.29
    - currentCSV: serverless-operator.v1.30.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2023-10-30T14:03:25Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.30.0 <1.30.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Represents an installation of a particular version of Knative
              Serving
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Eventing
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: Represents an installation of a particular version of Knative
              Kafka components
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:a0c5a91348ec71697747113990652feb024143fbbd0cdd67e8ae672730383a79
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:66149df8549380673c24e1487d0c2c8eb638b7e502455671ca64ed6f32ff82c3
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:124087e6402caa375b24b976c3b9fd7d53890e20ac3b0b57ba3ac646e352c7ce
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:3c8bf480dbb05e09e8d07cab12e4736643f61bf9b5d98efa6bc0557eb31887bc
        - registry.redhat.io/ubi8/nodejs-16@sha256:8abeddc6ccfffbe313f87e5e069e18cb7047f2c7880ad63cf15bcd0a79ba71ff
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:01c8fee92889a84a19136b0843e4fed5db31c1f2a6d997669f103373fd995e36
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:53d3620bb0490d83aaf6a356bb7d3b9e7147222d1a2017f8317f05fa90cd3de9
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:3508434d5ab6b38c8a7466256f1158ebcd75e8b751afcc5255da8c21566804f0
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:51ac9aced282f62c8fb2d3696da3696b8496d8f877e9846fe3a4a743ccd63bea
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c933641efc6bf919b1945be5a01dfb7774ae865975b1bb1f44653b3e4ca9f1f
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:ce4b36c90955a356a5774410356aafdac2b60795526a3488b0b76ac2ec1a018c
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:b6ef281fca2b80d379b1e28fe3f23119df8ff765fb2cca291d514053fb512da6
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:e9505aba7c44cffc3f261a5dc7617a0e54ed98e78f73ea3c9c7a508560b792ed
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:12bc62246142a605f4efed0f6bb8843aa2df7699cc1c5ab53debb94b0682e9d1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:e8c756b659f6187790bc1a5115e0f71791bff96630859a905bc5b75d14596e0e
        - registry.redhat.io/ubi8/python-39@sha256:0ff58549041d32b685cc850aaa85da32d1ef73f99cd4b6e0f6882dad5f0399bf
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:96d6f52711ba433ab2f273be8080840a88e626e989027480424d19c2fe1e9454
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:e97eb2bb7eb529cbfe67c14e823ffa1269d614e2d35723896e89efa12a499c6a
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:4800c85e8de031bf77f6fae0e4c8e6c1a9912ccdfb77799cc2ea574c37d870ac
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:32445e4a73bc7c1e8007775adeb147175d934b17e7d3d4dcf422104dbe80a52b
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:21a638b481639c3c58e8b026d875917c30a5874b08fa4ceac40d447076b2bfa4
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ef116868c9c60e5ee246060688da49832571ff6a7901cfe143a42d048743736e
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:85a369d5696771d0932b61760336e52ae3f7c66f9a3def5312aefcac03ba516e
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:9a28cb1d4c4f4dd7d494829eeeb7122db2174a27747f915244672c95dc8f9d12
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:1455606c8edf611d996a48e29ead53e0bb4078f55384b7689058bb25fc59fb2c
        - registry.redhat.io/rhel8/buildah@sha256:a1e5cc0fb334e333e5eab69689223e8bd1f0c060810d260603b26cf8c0da2023
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:c9353a022e6f883b7c0b2d87b1bc4653172feb13806ebf0aa075d6f9f43c264e
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:6057a2d18406e00b0b7b1731870f707844e39193d4cce6de50137b77f17d220b
        - registry.redhat.io/ubi8/openjdk-17@sha256:af305e087061cbaa4107ce0b304a1ef987f6b54bae3f75bfd529cfd084cba9b5
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:2d3e314a6584fc6974517c1a992fb59f5133f54c2f6cfaa42f579c0d332e0300
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:06e9e073f092dce67d4227e282fd8eea6dc9f38b90a8d444cf448d80d027b616
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:cd1abeb2067afa44a0f1eca301cce37307582d56c1147820ff86b9b64a40588b
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:69ea22c646c09208ca68345d4b3508453d8c49f7f7fd9a4ff4c4cb78be7bff43
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:96c135ced2e3bff434addb012368d96d0778ef1d294ba507ae7e624811e20d46
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:a9c95fc208cc03625cf49bcebdaac73da823d1fa3bfae4239d7db20d3a0b8723
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:2264d2cd4ee2a720667d625ff42b6f8be4f2cd65b57a91eb6044da086a110740
        - registry.redhat.io/ocp-tools-4-tech-preview/source-to-image-rhel8@sha256:98d8cb3a255641ca6a1bce854e5e2460c20de9fb9b28e3cc67eb459f122873dd
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:1f7e65009168383cdcb6b2caa22fe4db49e03c71574973a8086f6c332dc5e26e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:16939b802f4c16a3d9d941917d0363325f40cdef28f6e1a982bc95401117674c
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3fb6e7626d2b0c92c2a2fef35e9b8171382b4bdf54393f2ab1281ae7ab441761
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:92a39ca90984cebb1938447a31c9170b52994ae98f83caa860acdb2c3ebb75c7
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:a5154f9d13a1c373b8ba47e2bcfb9d4474fd03e08db4134c92aa4f049536f1a8
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:d03f2bf8f5e7cba47102ba26ea0149ae6719076c645d5e451058392d6044b738
        version: 1.30.2
      entries:
      - name: serverless-operator.v1.30.2
        version: 1.30.2
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.30
    - currentCSV: serverless-operator.v1.31.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-02-01T19:15:42Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          olm.skipRange: '>=1.31.0 <1.31.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.13/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.24.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:887480662b1bfe09626fbca8ec7c5da879d097c46dca11764ea98474ec78ab5d
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:0a2989f688b966277652957b312df736ada43fff78fd89b17f60b2137e38f115
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-rhel8@sha256:22ffe08d6a0d5461158613f2b2a2ac6936025357e282a09197a750c42b5e6922
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:3a4e7412322755b4c278540c965a403f9e178f138bd4ad7929c1c24a8036dd14
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:df1aee67a161fa19d7f49ba56d53e0682cc1996436e985a12a2068fa419b54ea
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:7941b421b3e18bda8c565305241621fcdf1212cbbe0dc02f38f4ab0565d2bfcb
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:bf693eaa845373e1a06041626a9314ec00519a6fd2437b12f7f4f3f104864039
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:ef7d28580e75d746449a90cc0e5343ef3f4ddd75bd9fb866d97836bfbc3b314d
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:63e08279d2852d494d4f8836c54e2ce942a342f452a23ef437df9640bb2f5bc9
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:a82288782b61ea46d068cc09ee10608d3201c136abfc767e4d37734775101390
        - registry.redhat.io/ubi8/python-39@sha256:4f35cbcc6f4e108c69f7d2b70cce5afb138741b0c961506376d524e001b922f4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:eacdad8db61b23634c3f89bffa3f1c30ac5c7bbb5e938c6208003ff00f9595f1
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:3dab790fa845ed7eda088e5abdc87799da1354884a53ff1ff728739b8e278418
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:565becc77dbe3bf03d89b1399d6e1853253276a71e80b6b14069285d03cb14b2
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:80361e052d366ffffe9a99893cd83965bd69a9ff81d8c9262976de145bbc6184
        - registry.redhat.io/rhel8/buildah@sha256:b48f410efa0ff8ab0db6ead420a5d8d866d64af846fece5efb185230d7ecf591
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:8353ccb4f0d70597d61e4fd1f799d4bb74e5bc038f6b9916059e712aa70b97c1
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3b01ebdb0704dc6a1a993ac076819489495ab3fe32e3c14609e42d1358856be2
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:af5c30fc9f631d9ed110641b815315d8b4aef539f687277a3821d9939ca5dd46
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:de62762efa0df95ed23df56f5520490d8a231e47ce13c433c512254ceff81844
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:2572d98363cca2dd94984d6cefa573c5054fc206ee70048f901ad4b846d8dd52
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:d2a34100f6309c3cbac7c9ccd452b7bdab2a533afc99463bfbf91e3ba2706d54
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:e49220eaf3143a8958f0cb1f22115b2e834fb7177755433bdc592b753a54a8ec
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:3be4fc173b489d0beb768739f60274e07d95ffee86d70a8423e1f59678c805ea
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:433821e6c7f43da5ca2e38c0550722feb179a0bab021decd9a97ad020cb78b76
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:486bf60d17a5d265ebd65d590a6ee1a7e13bf3953ea797b2f6c2b5910789858f
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:1ee675125f7141e91f4ee57ebce082967834bbdabe1a870c221f31aeb8750be4
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:dbd9c0cac86f79b8d91ab65ae7543854144781d4cb17ef16183931ec9bc234a7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:1fcecc11b55007b9c6969a24940a391968378e4102eebf7ceabbe7efda138575
        - registry.redhat.io/openshift-serverless-1-tech-preview/eventing-istio-controller-rhel8@sha256:5b38e5c71e422161e40f9c37dd2f46212cb80e8e37b00247767276b957080397
        - registry.redhat.io/ubi8/openjdk-17@sha256:a8165bc2cd5051a96d6937e25ed03155bbd4b731da6e58cebfe2ea83209c16d8
        - registry.redhat.io/openshift-serverless-1/serving-domain-mapping-webhook-rhel8@sha256:99f5dcb92a33c047f0b68236a3a42ffaec78037dd0a5eb79d1144d62d6bfcf43
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:3181ad775ead55f317f96466c2719b2b73c5b35d31933733da21273955d160c4
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:677be039353b2cee51d5faa5d5a2abd6a7bd46992ffe9521eb033eea9925c785
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:cda04e3b7c2c1f37164733ad162ea0c4e52f14fdc8eb8bdef8507f0ed5454640
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:9e120ecdbea54eecbc5e0412c1774f8a2ee70d7fca682ecd0757b8627f70680b
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b37dd2cf4b459fecb6d681cac36918e486c3bd22501df50b90e20b15fe164d56
        - registry.redhat.io/ubi8/nodejs-16@sha256:f90b28179d1d5de68d42ef1a9166d70a9dfea85a1c6a4bf12dce2c59eeb2fbb1
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:9221a833ecc66fe2354aee7aaec2a2470b21e2d89cb6ed287ece32edab2a2edb
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:b2480154951109a159acfb88c5472a312a515d6a6b060d832b3f7bfc16f90ac5
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:d8bdd96c6a9b31f48c961ca438c254511db785fd34bd10b98084b702baf0ebab
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:c078071cd51c7fd5584039b54a2004331c05a897ee8b0689bfcb9204a8ebae80
        version: 1.31.1
      entries:
      - name: serverless-operator.v1.31.1
        version: 1.31.1
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.31
    - currentCSV: serverless-operator.v1.32.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-08-01T10:56:58Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.32.0 <1.32.2'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "fips",
            "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:16c026f09b377c44253cdb902b9a5fa0ae766fd0185cca1863798a212e0daee1
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.14/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:14ca32e999767776c3e585ff5eb1aa568f7d76fd3e5ae24cae05ce7564678080
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:2f0f84767f7d76dbbf4387ca2e272766267a2c46b5bc29f5d01f05dfa39cb944
        - registry.redhat.io/ubi8/openjdk-21@sha256:bb95ea866269ac8652f666ee1587d71f4f43da33493c2be7f6a9548bcfd2407c
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:3be20d63d7e82f35ac88274532d6ab7839f9a7306ca272e7d532ac03215faea9
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:1befc9969edd2062b3df69f45aee9771b33a2659c3a14db0aaaf5c3df985af58
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:b008da825ee2429f04cf39258389740c68056f82cf4bde0cbd3db850b6256885
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:4c9faea6fd162db3340aee8c11af11a08c9d79bbed104967c25df26c4933c8ae
        - registry.redhat.io/rhel8/buildah@sha256:55f5c7d88c77e1ab5a2ab8e877172b2b66cb196e21f10dc45148470d0ee8bbae
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:7c8cefd5823ad953f112b905c651f12a97491aafc008b2ba0111f0c9e83e50ae
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:828d31bf8a441091b8aed03da213a6b4ffae260070eff58c3319156f9bf04f5a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:e70dd3a9e067075eb56e5f2772642545e308348cd49b89357c02de3979b1a5be
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:9d36cde68bd822016e967148eebd2ef3e7ffab9dcdd37d17f33017e9b703d9b2
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:25cbdf78b67e7f5a360313ccfbe3b6e2c98873b4e9ddee46c115a623e77c7abb
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:20566173a8e907980e603d82fae4d96fcdfd5c43d9004c2e506cf0e09a47b91e
        - registry.redhat.io/ubi8/openjdk-17@sha256:a166bd2fedb99cb62ebf4695ab46e19c7e451f887940cdf0b8245c81e403aea2
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:b0db9f0fc85fcdcd5da980633a3a44be2bbabae068f89d0965263de38789e812
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:e55c70591c0498b6b85f0d68e304e88b830ea0b35219ba8881143555757be4b7
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:22b4d47725ed8cdd663ee26a03f31198fff9e71dc703532e70c42bdfb4840811
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:46b8dc1c8ea6c40401d4d1a9868ab3edbe8b9e92364a43add6722489e2eefa04
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:2a30ae808d5c32d812406ff4ac9b472e82b6bb0555e62eee4c2c6a972f8750c2
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:0507ea8745267fdb53329a5274d59d23c387225d2594af51dd2603b4ad99db0f
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1a1de677297525c0c5d317fe8a7a42684fe3f1b422a7d6bea1f70f0924653b08
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:3ffcf746df914f0949477b46f90cb2992bb63f8198d73ffd54208e7a777a9cb1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:990962ad245a6d62cd11551bd4e1f8b14dc08abbebe47beddb2f4874f78d4f47
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:ab9a1b952e9603b4ecc5f197715a1d871abc25ebb0c5e6470d4f0b365ae94276
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:da5f813dea324ffe9679524ea7f3801536fa48f7dbbd8da56a8492b916b15fc3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:9b6524f9c0e7f78e6ffb372a8587eaf5f4f1281eaf64f93dd00b63be84a4c21e
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:5930caa25ea875d29dff0ec5e235eaa13cb6dc98980d8045335eba1a09ae761d
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:04e5b3882223d45a5f5383ce6481efb7b2cbedea57ebd894caafb1454ad8c09d
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:70c26c7ac48d1eaaf1f95ca5c5b8ba8ee6b047a3c9c4a2c2db9c49a506bb2205
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:e7e6cbb46720eeca3f0f4b9829dfeeb3db373a83c89cd808b667ded4d5e139a4
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:ab09a1710aa2331e24df9e7283f3909fdb104c1b314dea88e919d0a396f334d3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:a5a7b2be50f211d4d9d555b2a2e2302096c139ff5011d984cc386f1530007f16
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:0729de0f3f03ed220216c6cb718fc34beece0ceacd926b5e2480bdfcf44df5cd
        - registry.redhat.io/ubi8/nodejs-16@sha256:69439fe83fa957089cb6fd3c1042ab71d719f81dd457a5d1fbc56c1e3ea600da
        - registry.redhat.io/ubi8/python-39@sha256:8c96e8648cc6651d0cefbd20496a89f208621601f9664554d3fdeef8ced25453
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:42f31eb8f12dd36541c8c0b6572d942c62e2064d5418428482d676d117a6c426
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:0551f1657dc158bd41eb2338dd6514d55c39fc77300cdf1744d8f5ab14e9ed1b
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:8033b7c1e903ba313134645c626eb3bf7ab1a0d68dc2575bc8e26884ad4f805f
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:e1e2273fdf6384d95e1128e54c0794d7d620291a272bfb1f91e0c2723be7f1b7
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:4933b76d95d411d597d4dc460fc46c92875d28e107cda661946206a9d934b8b5
        version: 1.32.2
      entries:
      - name: serverless-operator.v1.32.2
        version: 1.32.2
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.32
    - currentCSV: serverless-operator.v1.33.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:38:34Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.33.0 <1.33.3'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless#install-serverless-operator).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/installing-serverless)
          - [Develop Serverless Applications](https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:5c7f29554d04eb9348932953344f2ea0806d14e7704355619381bf395d993f58
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:a0399a993155e755409355740b260d9b0019559a62be3b6253d8693a4fb59286
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:62bf4bee0ea17ecc60366ecc71431adc79d6f0a5b9f59489913b474128c4adb0
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:9b94875e0e5d7abdb3b9ba7b04e217a2a9b7b4410d38d1608c86d1c21f929cc1
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:9b850afa750a67120f68a49f4834ba7f105563df22cbcbc563317fe1545640d1
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:09060de091d617bd4606f7a289b4528265ae0b7a04737f383e79efd5b0d3eb79
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:f938cc534fc4b24bb9449f1e1e1e904dc588e6646c98474de4b8c81d083b4d96
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:56c6ac59a68976741d2ba676a53e59fc494d04dc44b513d72fdf8a9ad894a54d
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:ccbda564735555785815f27f8f25ad9c6c27dabe4e262586fc51609331692527
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:93b2f5211e79abc38bf92f0b8fb8adf046303efa2fe62c47a2fef9dfb5662f55
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:e1cb1c24333605e52364456f9bd8d86d787d1bae17ef75dd04be61257fdb2256
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:6dbfa2573a2c78e450d3c1c04c6e7cbc50d9272233e828fe9b42e575ecf386c0
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:42d16e4d9f14a427e7a5b7e3a4ea07c258e0f2c74984baeee3b4926a8c5e2640
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:a8821b59d495e24598b27c6510b32ef19de2bb04ea1e568fc5afb9135d772036
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:eb95deb2105f14c6a9f3bb2cdd7306b446109a9b142a7bfc726b6429f270d9e5
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:d8ae7bc888f11121fbbba65074389b879bc22453bbabc551d79d0cee68fc33a3
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:af66cc7dbed88907d96a390fe5c8807a2a60bbe3276b8cd13b4b061cbaf7b162
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:510e98837a549f489b2d74939f3b9c7ed90a8260ba914910f681d5c0dec7655c
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:879fca768b46676f1df89de1cdc4ded8dac310209c146c2f381ed1deb19b06be
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:ea9e4de6228d5c3ed494c2ce5317ec290543ad40bcf7c43c2a3d6565e5d94e5d
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:c3dee2598088bda36ef474f4748943a29169ed9589aeb0f77bc79a7f6bef9907
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:359f8b9fbb4c3c2f79e8c70f3b50c994693b930f6f078c34c87112ebb93037e3
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ace72435e3e63ab21dff2345f958cf6bd171537ef2846c5d5d91cf821877140d
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:672022b6f2a080483d9d6f6f4a5f42e52e522a5797105f3465731809c6cf4947
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:eab4aa93b9d175045b1f01bff09681107d839285e0d633cd5146d2a2efab373c
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:05e3dd6255294b691783523f7722cfaf13aca1a1e1dab344a25eeea72aaf9f9e
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:993597ff83a4d62a23f642fadc92f182944b953770716b73b7d19492b5a462c3
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:b2be693c71cfc4ce934d5b01efb863dc6e391f93a6b15fe7eeb15914063a8294
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:b12272e1a9cb4f4bb62c4badc62b0bab110abb99484612dc781c46bfeac13cc0
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:a9c84ac7e7c3fb1e5dedaf949d54a3aee8a62397d2bd6022094671de00b651da
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:dd1f7c850f7cf2a4d70403deedd8a57ecc6e0ee5455d68fdfc49ff175fe42dde
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:e9b38f1df225238836aa829f3a2016372859fb8709cdd2b30496498669a90be1
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:28ae309e4cf7feabaf78ac6067e1173f9f5b33a71c06d6c27cf25e01a0b1ed64
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:2bd241898bd54dcb248ae86470430042dc545d45e7519ed470a1a9c8c63bdc37
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:b871bd08b782c9168e9a54340efffc0b0effadde8079915b848e4eff72bb3e95
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:28c9d05a3ba73b76f293e8f4902b574deb44f77a2bae8cf92473b8a8b10cee92
        version: 1.33.3
      entries:
      - name: serverless-operator.v1.33.3
        version: 1.33.3
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.33
    - currentCSV: serverless-operator.v1.34.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2024-11-20T08:53:53Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.34.0 <1.34.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Build strategies including Source-to-Image (S2I) and Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.34/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://access.redhat.com/documentation/en-us/openshift_container_platform/4.15/html/serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:6d042c614431aa9a9e77cf8ba0064423e903f7c418e367ecc23bc02a24a796c0
        - registry.redhat.io/openshift-serverless-1/eventing-webhook-rhel8@sha256:30cb41f09067959cfe3792e070fca62820fcca4035683e5c43dd46d1cce97c01
        - registry.redhat.io/openshift-serverless-1/eventing-storage-version-migration-rhel8@sha256:2adb0124f04d6c4a552ade79dd1499c8ca997edb45855bf438ca8043b2709e19
        - registry.redhat.io/openshift-serverless-1/eventing-apiserver-receive-adapter-rhel8@sha256:84994eabeb7342a0771c7cb1bdd70e7ee3928dd45874eafa64cde5bb385fe964
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-dispatcher-rhel8@sha256:b801ed7b1cb32d4a1522a959e97f280f2eff6fc92e47fea2ea022f180d37831a
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-post-install-rhel8@sha256:76480201188dd142206040af845bd55b44155af92e0c5ee9bcdb8660bea0cdb8
        - registry.redhat.io/ubi8/openjdk-21@sha256:095175ef8573de8055e67bd5c8a21098448ac73db4e0d6ccd5269dd36b4d263c
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-controller-rhel8@sha256:86ffb8be1762782d643c147cefcfbc2b4723b0f97d017314a99bd2275d05a1cf
        - registry.redhat.io/openshift-serverless-1/eventing-mtping-rhel8@sha256:d5ad3c9db7ce8e4fc950261f7c388476214a03d92e0043c7bbc4f0add2b1e69e
        - registry.redhat.io/ubi8/python-39@sha256:ac12e4706e0fdc464e791a3f274474ac1bb14d5af8f217c440fbb0c76d6a5d00
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/eventing-istio-controller-rhel8@sha256:ef7f34a025e40f93fbd51f3a4596998a6dc508db824f137ffa1fbaf90e88f4a8
        - registry.redhat.io/openshift-serverless-1/svls-must-gather-rhel8@sha256:a4b3d75903de1929326041e9885213074473f07fda91b299548d45680c295c00
        - registry.redhat.io/openshift-serverless-1/kourier-control-rhel8@sha256:b9d85323e110d8eb6600223c38f9d4cbe89b50d59e03bf1e6b8808f7cef9a3a5
        - registry.redhat.io/openshift-serverless-1/serving-webhook-rhel8@sha256:4b6f161c14c68536ec2c7679d30681afa9a1fa2d512548f0f591791a10a8b894
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-dispatcher-rhel8@sha256:d40daf67b36387faa9f55e70e50d10d02accfd19470c0ac11b78bf969947fed1
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-receiver-rhel8@sha256:d8b94bbce98325fe8079c11c0b7a6c59bad8a7edd0532a7d27ef1632ff0de708
        - registry.redhat.io/openshift-serverless-1/serving-activator-rhel8@sha256:144883b09d8a999fd4d2a109d074e44d7d8b959987512379aa71c71e4060ca28
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:02b834fd74da71ec37f6a5c0d10aac9a679d1a0f4e510c4f77723ef2367e858a
        - registry.redhat.io/openshift-serverless-1/serving-queue-rhel8@sha256:a5950f81cbc64d1edb64ae7f7f120c3435d36afcd14ffc34ae522af297d87e1e
        - registry.redhat.io/openshift-serverless-1/kn-cli-artifacts-rhel8@sha256:c0376504edae12e7af6b6cce02ac53b6b374f5058ebefcbdbd32628b551bb0c4
        - registry.redhat.io/openshift-serverless-1/serving-controller-rhel8@sha256:946daeff59b41a6f8c4842f8c2668f97ff1b49e2a8afae5d64ba3baa916b1d95
        - registry.redhat.io/openshift-serverless-1/eventing-controller-rhel8@sha256:1c82d20ccf5eb5f8e7725839f1060f4efdfd1c1b4773a99bd905e690a6e36f38
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-rhel8@sha256:3d17f12cc9da578549254ba1b8934b5afc66e453b1d5d667bb90281e45b8795c
        - registry.redhat.io/openshift-serverless-1/serving-autoscaler-hpa-rhel8@sha256:a9e8583c19ca3e01593f8c0e54731cb73df9b46b2ed47a2b1c711828c2295f82
        - registry.redhat.io/openshift-serverless-1-tech-preview/backstage-plugins-eventmesh-rhel8@sha256:09dd94090eea9cac7a1935a6d671cb97fd9fd5714c715c9d2aed461e468ca54a
        - registry.redhat.io/openshift-serverless-1/ingress-rhel8-operator@sha256:b07010595201d7cfe0ec4f18e612d323eb38464c74eb187f275617a744d9df12
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:095629572a40469eb651c9f81d6c10ff46254eff8b9813cadce54d6f0075c2fb
        - registry.redhat.io/openshift-serverless-1/eventing-in-memory-channel-controller-rhel8@sha256:1252960c0c0a4dc7b39090cde0fd5742a19453b1a65a08b737bc41d044a0882a
        - registry.redhat.io/openshift-serverless-1-tech-preview/knative-client-plugin-event-sender-rhel8@sha256:f94f0e8d21a6b6341b1db00513a1a2ed5cd84fcd898a376ecde942e940a518e7
        - registry.redhat.io/openshift-serverless-1/eventing-mtchannel-broker-rhel8@sha256:586fc8ed88f91298dbff31bc2fc4c63e0d6d9356e3358b2f853f49511c3ff9bf
        - registry.redhat.io/source-to-image/source-to-image-rhel8@sha256:6a6025914296a62fdf2092c3a40011bd9b966a6806b094d51eec5e1bd5026ef4
        - registry.redhat.io/openshift-serverless-1/serverless-rhel8-operator@sha256:633ea30ea3b6d5152bbdb61acb4271e5905d50fbe3c731fb4c8d8640a57f1970
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:45bc50a993d04b775a153e985a43d587d5455883fc4c6f31b772408297cf9040
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-ingress-rhel8@sha256:0f126893074cd0398654c4018df6e882b7610e148c82a41ac814f4bf3830bf9d
        - registry.redhat.io/openshift-serverless-1/func-utils-rhel8@sha256:e520d1b6b87f85dfffab5e53cfb70562ab7edf450061c07609da39e3b2498fdd
        - registry.redhat.io/ubi8/nodejs-20-minimal@sha256:acee3b00ccf8ff49bcebcf10e028e34a84d45f28f855363bafb3b93c75e84f11
        - registry.redhat.io/openshift-serverless-1/serving-storage-version-migration-rhel8@sha256:512c4d93dc13cc53d923a6990770a7527d6fb657278e2c5d8f0ece3cd5020155
        - registry.redhat.io/openshift-serverless-1/eventing-kafka-broker-webhook-rhel8@sha256:0f8d006d0e7468783739942a1c82c7aec376c9948f23841132c8ddece8ecba06
        - registry.redhat.io/openshift-serverless-1/eventing-mtbroker-filter-rhel8@sha256:75228e70fb1383bbdbcc64d0cf24bf5d977c09112e3f32316824188264c3a2bc
        - registry.redhat.io/openshift-serverless-1/knative-rhel8-operator@sha256:01ea7a06332d60ac16ae0a27adcaaede3f680a98a22ff995259a6c42ac0e358d
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:76a47d100de1f2afa23803bec503d9440605db03c7179b231f963e5663a3d12a
        - registry.redhat.io/openshift-serverless-1/client-kn-rhel8@sha256:33d62d13e16f47db9951ac826671f39fdc048fe7c3fb019eeb4bf9bd0f814222
        version: 1.34.1
      entries:
      - name: serverless-operator.v1.34.1
        version: 1.34.1
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.34
    - currentCSV: serverless-operator.v1.35.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.35.0 <1.35.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/infrastructure-features: '["disconnected", "proxy-aware"]'
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.35/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:dfcd11686f161d44786a205d12f9d628d9e847d40ffdb628ec9aa4a0da0b86f0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:018459cbba2e3169bc4be6564cf3747c9cd0677445912abf237f8fe6d58e46ac
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:ffaad2214fa897e89e92e574eb5cd1c671edab7f6eab87cf2ffd4ebf162b3cf4
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:732827652b7a7a49181e86d561831001c28b72b7e12d5e3c1d71b2e496003229
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:b66bb9531f05cd46360f787955c849208496e3e45541cb50b23bfb6b68570aa8
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:d0650e4d7e72241e07a3796f20931adba231fbf5e401a45d794761f2c5afae48
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:9b176477be4c782ce60d81a72bacb566069a499579ce6df5a987a0841f05ad5c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:e743dd93b7ae855825716403979db3f34cd300caad130e7d3011172b1aa4eaf7
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:818dc7bd69b453ba5882b4ccb82841f2d48c7552368417bac20bef1df64f52ff
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:b5db9a51bbb46640b23712c61e9c5a1c7bbdbcd9a53b0b92fa36b80dcaffffc0
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:f23d3ab058667c6fe9ee44dc249aa7e8449c1bb9581380ddbc66624511b286d2
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:2980eca437e4c1cd22a2b6c2ccae9c4d6b55921b3d575fbd4b9241f1f0a8a3a4
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:0751b9ea7ca197480c4cbca17654d0653d1c0c760ca3c818c2a1018dc952cdc8
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:e7aa54986c3d388c1bb00828c9ccb294ca0b6c26376c7593ed6bfddda0ef4e5d
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:d00a0059f6134b2519f9b1e034b81a943cae7ce391109703f24db21e56ee923d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:cb2ebe98481808e788ca70e7f0a9b3402818b622158f327b99a1e87c23226d2d
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:86386b54cc10743023884193b7d6c7159b90052bc5e4a0608a6cdfd47865c966
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:477b1ac0db0abf1e16ae31bb22fdb73dd98f4dbd243fa0a7e6e94852b16a813c
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:56e7a0158a713c70765ead847732db4e5f8439aacf31e9a4a22672b583028064
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:6b842b4eddfc71a2c8f6c1fcebb8f23a05f50c95be1ba2d20a1be660d3014c8e
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:0e58edf5ee340fdfb282faeb6c9bee8cd6323e7f29a9b282dbb91c2622131a9c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:0fa6d15d38601f266347cdd098e80bec9893f7e6b06e84efc4a8c50b3ff3c6cc
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:80a155d2637f47e249fd69760b250c9b7a1eeb6781de9308b54162bb3e1bce71
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:86074674b7926ad9bfc95556e50820792fcbc5e11ab02879378625ffabdd1889
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:3130a5159adb9b4eaab75a0209d2ac639156d5e9f2346a8b2e6e0e22fad4b45f
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:acb28d84cf03b8551efa32646864005dc3bc9bb341a696544892dfd0c87fbda6
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:d68fae6fc07efa9da9af0532aadb49f897a15afcdf9a98dc7517ff5e17735454
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:b2de6b658a1595084a202203c90704c3f9de018a7a904ca276112ad606e374f1
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:ab0342b325ee67cc2a7d321aa1eda9c6797dc91fe2b6f26e3cbab53fadf564c7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:6722aecbbbefa80dfaf93143e033c562cea0982514e2eea27654dd4638e1d072
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:020264063c50f2a4590f3d7524b271f0f77350a1f8fa3c35ca6e28f0162aa9a5
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:ab8f85bb18311b0bfee00561ef6b5d9faf810358c954ac2c6f39cf2a7739e07c
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:f56d21089566330ecbd9aa8acf14da332d3db7360aaf6da00c9a19df83f7334c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:8a4e5e0934197c32e0dfa1efc922d750940587fb61e0df49d476a2d035c53413
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:6f4fa254d923f3260ab0f7d6a0b459554f83b1abfcb575a5ce2c0135fba32427
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:4ca1bfaf2d7e9477583cddfa40f0b2ce59c20ff9ab20022038774dc3cc49662b
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:91456907c61952017daf07e376b565472ef86d67a2925f60e189c639ceb68595
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:da8d975527ba100313b92e87d6d2690fddae9988aa8d82d728221295955b7224
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:2cbe84afebac23b5397d144b4d5660af33290df451141fa3625607b6033983e0
        version: 1.35.1
      entries:
      - name: serverless-operator.v1.35.1
        version: 1.35.1
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.35
    - currentCSV: serverless-operator.v1.36.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeServing",
                "metadata": {
                  "name": "knative-serving"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.knative.dev/v1beta1",
                "kind": "KnativeEventing",
                "metadata": {
                  "name": "knative-eventing"
                },
                "spec": {
                }
              },
              {
                "apiVersion": "operator.serverless.openshift.io/v1alpha1",
                "kind": "KnativeKafka",
                "metadata": {
                  "name": "knative-kafka"
                },
                "spec": {
                  "broker": {
                    "enabled": false,
                    "defaultConfig": {
                      "numPartitions": 10,
                      "replicationFactor": 3,
                      "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                    }
                  },
                  "source": {
                    "enabled": false
                  },
                  "sink": {
                    "enabled": false
                  },
                  "channel": {
                    "enabled": false,
                    "bootstrapServers": "REPLACE_WITH_COMMA_SEPARATED_KAFKA_BOOTSTRAP_SERVERS"
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Networking,Integration & Delivery,Cloud Provider,Developer Tools
          certified: "false"
          createdAt: "2020-04-20T17:00:00Z"
          description: Deploy and manage event-driven serverless applications and
            functions using Knative.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=1.36.0 <1.36.1'
          operatorframework.io/suggested-namespace: openshift-serverless
          operators.openshift.io/must-gather-image: registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          repository: https://github.com/openshift-knative/serverless-operator
          support: Red Hat
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A platform for streamlined application deployment, traffic-based
              auto-scaling from zero to N, and traffic-split rollouts
            displayName: Knative Serving
            kind: KnativeServing
            name: knativeservings.operator.knative.dev
            version: v1beta1
          - description: An event-driven application platform that leverages CloudEvents
              with a simple HTTP interface
            displayName: Knative Eventing
            kind: KnativeEventing
            name: knativeeventings.operator.knative.dev
            version: v1beta1
          - description: An extension to Knative Eventing, merging HTTP accessibility
              with Apache Kafka's proven efficiency and reliability
            displayName: Knative Kafka
            kind: KnativeKafka
            name: knativekafkas.operator.serverless.openshift.io
            version: v1alpha1
        description: |-
          The Red Hat OpenShift Serverless operator provides a collection of APIs that
          enables containers, microservices and functions to run "serverless".
          Serverless applications can scale up and down (to zero) on demand and be triggered by a
          number of event sources. OpenShift Serverless integrates with a number of
          platform services, such as Monitoring and it is based on the open
          source project Knative.

          # Prerequisites
          Knative Serving (and Knative Eventing respectively) can only be installed into the
          `knative-serving` (`knative-eventing`) namespace. These namespaces will be
          automatically created when installing the operator.

          The components provided with the OpenShift Serverless operator require minimum cluster sizes on
          OpenShift Container Platform. For more information, see the documentation on [Getting started
          with OpenShift Serverless](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications).

          # Supported Features
          - **Easy to get started:** Provides a simplified developer experience to deploy
            and run cloud native applications on Kubernetes, providing powerful
            abstractions.
          - **Immutable Revisions:** Deploy new features performing canary, A/B or
            blue-green testing with gradual traffic rollout following best practices.
          - **Use any programming language or runtime of choice:** From Java, Python, Go
            and JavaScript to Quarkus, SpringBoot or Node.js.
          - **Automatic scaling:** Removes the requirement to configure numbers of replicas
            or idling behavior. Applications automatically scale to zero when not in use,
            or scale up to meet demand, with built in reliability and fault tolerance.
          - **Event Driven Applications:** You can build loosely coupled, distributed applications
            that can be connected to a variety of either built in or third party event sources,
            powered by operators.
          - **Ready for the hybrid cloud:** Provides true, portable serverless functionality,
            that can run anywhere OpenShift Container Platform runs. You can leverage data
            locality and SaaS as you need it.

          # Components & APIs
          This operator provides the following components:

          ## Knative Serving
          Knative Serving builds on Kubernetes to support deploying and serving of applications and functions as serverless containers.
          Serving simplifies the application deployment, dynamically scales based on in incoming traffic and supports custom rollout strategies with traffic splitting.
          Other features include:
          - Simplified deployment of serverless containers
          - Traffic-based auto-scaling, including scale-to-zero
          - Routing and network programming
          - Point-in-time application snapshots and their configurations

          ## Knative Eventing
          Knative Eventing provides a platform that offers composable primitives to enable late-binding event sources and
          event consumers.
          Knative Eventing supports the following architectural cloud-native concepts:

          - Services are loosely coupled during development and deployed independently to production
          - A producer can generate events before a consumer is listening, and a consumer can express an interest in an event or class of events that are not yet being produced.
          - Services can be connected to create new applications without modifying producer or consumer, and with the ability to select a specific subset of events from a particular producer.

          ## Knative Functions
          Knative Functions allows developers to write functions that let them focus on business logic.
          These functions are deployed as Knative Services and take advantage of Knative Serving and Eventing.
          Knative Functions bring greater efficiency, more scalability and faster development to facilitate rapid go-to-market.

          Other features include:
          - Buildpacks
          - Multiple runtimes, including Node.js, Quarkus and Go
          - Local developer experience through the kn CLI
          - Project templates
          - Support for receiving CloudEvents and plain HTTP requests

          ## Knative CLI `kn`
          The Knative client `kn` allows you to create Knative resources from the command line or from within
          Shell scripts.
          With its extensive help pages and autocompletion support, it frees you from memorizing the detailed structure of the Knative resource schemas.

          `kn` offers you:
          - Full support for managing all features of Knative Serving: Services, Revisions and Routes
          - Support for managing Knative Eventing entities: Sources, Brokers, Triggers, Channels and Subscriptions
          - A kubectl-like plugin architecture to extend the built-in functionality
          - Easy integration of Knative into Tekton pipelines by using `kn` in a Tekton task
          - Create, build and deploy Knative Functions for multiple runtimes, including Node.js, Quarkus, and Go

          # Further Information
          For documentation on OpenShift Serverless, see:
          - [Installation
          Guide](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index)
          - [Develop Serverless Applications](https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/serving/getting-started-with-knative-serving#serverless-applications)
        displayName: Red Hat OpenShift Serverless
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - serverless
        - FaaS
        - microservices
        - scale to zero
        - knative
        - serving
        - eventing
        - kafka
        links:
        - name: Documentation
          url: https://docs.redhat.com/en/documentation/red_hat_openshift_serverless/1.36/html/installing_openshift_serverless/index
        - name: Source Repository
          url: https://github.com/openshift-knative/serverless-operator
        maintainers:
        - email: support@redhat.com
          name: Serverless Team
        maturity: stable
        minKubeVersion: 1.25.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-serverless-1/kn-serving-queue-rhel8@sha256:074ebfc4494581611647ba90049655a470ca43743610a27be77db66a8633bac6
        - registry.access.redhat.com/ubi8/nodejs-20-minimal@sha256:a2a7e399aaf09a48c28f40820da16709b62aee6f2bc703116b9345fab5830861
        - registry.redhat.io/openshift-serverless-1/kn-ekb-kafka-controller-rhel8@sha256:6b7316cce3432d391adc6ac38c92ff0ee86dd7ae0101957cc65cd338cc69e352
        - registry.redhat.io/openshift-serverless-1/kn-backstage-plugins-eventmesh-rhel8@sha256:e3ae1cc6012f19631c012a07b113fb015977d2ee4f9faab3207fbc2464c25aa7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-sink-rhel8@sha256:d9b944e64391a996104afbe8a21f66316a948e9fb7b91734e861847285be59c7
        - registry.access.redhat.com/ubi8/python-39@sha256:27e795fd6b1b77de70d1dc73a65e4c790650748a9cfda138fdbd194b3d6eea3d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-controller-rhel8@sha256:4aeb68a22f63300e54391ec6b4cda8c4e665d904b5dcdd3c40155040525742a0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-s3-source-rhel8@sha256:27e74287774cda208f667bc2a386debe6b6255e8991abfb37c959c8525f2af53
        - registry.redhat.io/openshift-service-mesh/proxyv2-rhel8@sha256:1c19cf86fd05940ef3d0d6e79e6f371db993153f70baedb9107bbe73f236dc24
        - registry.redhat.io/openshift-serverless-1/net-istio-webhook-rhel8@sha256:a6f565d73375e827d830dca85838f4c4cb7f08b0c078acb0c0d567b25f0d256a
        - registry.redhat.io/openshift-serverless-1/kn-ekb-receiver-rhel8@sha256:f4d86b196457cdeebdc380b6f1b707c10f8dee789c7779c90ee1332e511af8c4
        - registry.redhat.io/openshift-serverless-1/kn-eventing-ingress-rhel8@sha256:c7c2e738e0d9da1d2d1993f68938e72c4709b1f28a54fa92224ef6a0e456b0aa
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtchannel-broker-rhel8@sha256:bf82b37ab8fd29c7a7cd846008a16d14b4cfdc1c09df4f731d71be3957e90123
        - registry.redhat.io/openshift-serverless-1/kn-eventing-mtping-rhel8@sha256:631eb3a8dd9d83ceafa7d8a4da4b1e2d4f2740e6017c766828dcc7575cf1322f
        - registry.redhat.io/openshift-serverless-1/kn-serving-controller-rhel8@sha256:187fb9d64b53c55a029ffce3c0f9e7c3304683fa8a0e832cfb7134e8a3d0c237
        - registry.redhat.io/openshift-serverless-1/kn-serving-webhook-rhel8@sha256:4344716eb2940c6a3fa9a278250f936602bb1549fece6037912837719c178b24
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-timer-source-rhel8@sha256:8fed5ef9a1f39a55077bb5026c8854d0ee7fc8d8ff5da277ce9c1bfc2ab2474b
        - registry.redhat.io/openshift-serverless-1/kn-eventing-webhook-rhel8@sha256:429d5d9568cca31fbe83c663276a300286af4cdbf3d35685afdbe5f6f2f8c6e0
        - registry.redhat.io/openshift-serverless-1/kn-plugin-func-func-util-rhel8@sha256:82e69a1890042f5636ac7bb645b1aaeaba0a4894851eccc1fa1c63296918e31a
        - registry.redhat.io/rhel8/buildah@sha256:3d505d9c0f5d4cd5a4ec03b8d038656c6cdbdf5191e00ce6388f7e0e4d2f1b74
        - registry.redhat.io/openshift-serverless-1/serverless-kn-operator-rhel8@sha256:d6f338e08fa937c8a4ccfec692424de4acded565f7ca3971ea507dffeb351503
        - registry.redhat.io/openshift-serverless-1/net-kourier-kourier-rhel8@sha256:3b04e9d8cd4b5e43a07ece838d10478723a6c6529350f010550b4a0d1d0950d1
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:ed89e221d7a0a1dd82f2fa75e9e70bff3dddcb8979594396590bd3a15673a152
        - registry.redhat.io/openshift-serverless-1/kn-eventing-jobsink-rhel8@sha256:f2d112ba9fdfccea988589b99b463c8a7157475f7a3f03678e8999e65b678cc3
        - registry.redhat.io/openshift-serverless-1/kn-eventing-controller-rhel8@sha256:259e5dab50d36cdbfc117a42325128c4e16f34c048ba64b36d4a89132ddb4c1a
        - registry.redhat.io/openshift-serverless-1/serverless-openshift-kn-rhel8-operator@sha256:78fee4c596f2848487ca704da3fa1074e3c9c99b3f96f66833aeba63836d7b93
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-source-rhel8@sha256:88323da16360981f27b8e4686fbbb64053e4e0e3c79c65d695edb317b334e1e0
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-log-sink-rhel8@sha256:30b8045ef5f925a36c58941386f1899bec125d005a08aeeeba5edd49b0859149
        - registry.redhat.io/openshift-serverless-1/serverless-operator-bundle@sha256:01d632b69b4a4112426afc4184c72eae6d8816974088e5980c9c808f0782b85c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-istio-controller-rhel8@sha256:c39cb88afaddc295d93a3774b8a20d001b82392df540fb502ee02316d82460a9
        - registry.access.redhat.com/ubi8/openjdk-21@sha256:441897a1f691c7d4b3a67bb3e0fea83e18352214264cb383fd057bbbd5ed863c
        - registry.redhat.io/openshift-serverless-1/kn-client-kn-rhel8@sha256:ee6533b71cf280877c9d7489dfd821c9209c6802bed7fa3fdb4f8f00fbeacb09
        - registry.redhat.io/openshift-serverless-1/kn-serving-storage-version-migration-rhel8@sha256:f3d87a7fa5b5f2818effcac3304bd56eddcdb0db97a18ac0f35a06fc7e223e7d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-apiserver-receive-adapter-rhel8@sha256:6210a946f373b9c36c9b31293debe6fed090008b8234c0ba17d37fa81646172e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-filter-rhel8@sha256:77e4689b2ce7a713e905553e0e42b8e128045f40cb6c6d4ee082f126cbc46b00
        - registry.redhat.io/openshift-serverless-1/kn-ekb-dispatcher-rhel8@sha256:9b328fff0cc8545fb9bb1f24f11d78cfac3ee76dedc11fa4def4262d79a60e8c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-transform-jsonata-rhel8@sha256:3af4c48030263eb5769e5f36a6dd13fd390d8566efaab37532f39e1df1c95c67
        - registry.redhat.io/openshift-serverless-1/kn-eventing-migrate-rhel8@sha256:a1834b620f875681388f0ae224dd0771773a8b61b1d8e5214a942f232854b87b
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-hpa-rhel8@sha256:596a7272d6de714d7dbc6e00e9e50a2872846c269547465252d953172ef0b009
        - registry.redhat.io/openshift-serverless-1/serverless-must-gather-rhel8@sha256:36e589adaf5ff2856f5aada90b3a2be893d9ff89f92b1b2c42ee59523c9c9c8e
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sns-sink-rhel8@sha256:51b5e18e8d4c1ca81d7c5ddfea6352a3887f757acfcc53bd2f8bd3b01337ed42
        - registry.redhat.io/openshift-serverless-1/kn-client-cli-artifacts-rhel8@sha256:3fe326ad9a00b112fc888ffd345c8d5042f0a22cf32ea575931192c39ce1fece
        - registry.redhat.io/openshift-serverless-1/net-istio-controller-rhel8@sha256:dcfe5fbc57b9f8520adbff368c8bf29eb491c9c26d1a1737a1722cffea83eb1b
        - registry.redhat.io/openshift-serverless-1/kn-plugin-event-sender-rhel8@sha256:0c2cdf917bbad1f5bca8a2a2ef12a96cfcd0c77c190cc35687eb57169b54ab6f
        - registry.redhat.io/openshift-serverless-1/kn-serving-activator-rhel8@sha256:5352e4eb4e426a41521276decadf204b6d23b4762a298e4d3dc73969e3eddc68
        - registry.redhat.io/openshift-serverless-1/kn-ekb-webhook-kafka-rhel8@sha256:3d780ad1ab280b5092b90e277485d589da2a18e8b08090af7d95aecd17af545d
        - registry.redhat.io/openshift-serverless-1/kn-eventing-channel-dispatcher-rhel8@sha256:52ba6b1252b6f3d41e2a0b3b880cb768010ab9dd925ac34596f6adaedcbbe07c
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-sqs-sink-rhel8@sha256:afa634e029bcf79a972f49372498d44b45bcfbdb6eb28ecaf4f0b0b2a0a1824b
        - registry.redhat.io/openshift-serverless-1/kn-ekb-post-install-rhel8@sha256:4274004b6289fa9bea3fa6e56d64077a328a6b5e63a9f0e9c41f6d561f682bd7
        - registry.redhat.io/openshift-serverless-1/kn-eventing-integrations-aws-ddb-streams-source-rhel8@sha256:a903480159ebf1d8652eeaa702d9885fb76bb106c02d75cec5c82c31d77d6410
        - registry.redhat.io/openshift-serverless-1/serverless-ingress-rhel8@sha256:a47b6b5237a87ccd90e5a49a53d06485fd55c87eebf12d01c335b87fc3a8c494
        - registry.redhat.io/openshift-serverless-1/kn-serving-autoscaler-rhel8@sha256:ed60e889a677fb139f2eb4bd9ab4fda6ecc39fd050ca2b6cc1c19717e428535a
        version: 1.36.1
      entries:
      - name: serverless-operator.v1.36.1
        version: 1.36.1
      - name: serverless-operator.v1.36.0
        version: 1.36.0
      - name: serverless-operator.v1.35.0
        version: 1.35.0
      - name: serverless-operator.v1.34.0
        version: 1.34.0
      - name: serverless-operator.v1.33.0
        version: 1.33.0
      - name: serverless-operator.v1.32.0
        version: 1.32.0
      - name: serverless-operator.v1.31.0
        version: 1.31.0
      - name: serverless-operator.v1.30.0
        version: 1.30.0
      - name: serverless-operator.v1.29.0
        version: 1.29.0
      - name: serverless-operator.v1.28.0
        version: 1.28.0
      - name: serverless-operator.v1.27.1
        version: 1.27.1
      - name: serverless-operator.v1.27.0
        version: 1.27.0
      - name: serverless-operator.v1.26.0
        version: 1.26.0
      - name: serverless-operator.v1.25.0
        version: 1.25.0
      - name: serverless-operator.v1.24.0
        version: 1.24.0
      - name: serverless-operator.v1.23.0
        version: 1.23.0
      - name: serverless-operator.v1.22.0
        version: 1.22.0
      - name: serverless-operator.v1.21.0
        version: 1.21.0
      - name: serverless-operator.v1.20.0
        version: 1.20.0
      name: stable-1.36
    defaultChannel: stable
    packageName: serverless-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Keysight
      provider-url: keysight.com
    name: keysight-wap-core
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: keysight-wap-core.v0.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "wap.loadcore.keysight.com/v1alpha1",
                "kind": "Core",
                "metadata": {
                  "name": "core-sample"
                },
                "spec": {
                  "auth": {
                    "enabled": true,
                    "tokenEndpoint": "https://keycloak.kcos-sso.svc.cluster.local.:8443/auth/realms/keysight/protocol/openid-connect/token"
                  },
                  "clusterDomain": "cluster.local",
                  "configService": {
                    "port": 8080
                  },
                  "environment": {
                    "openshift": {
                      "enable": true
                    }
                  },
                  "helper": {
                    "port": 8080
                  },
                  "image": {
                    "pgImage": "postgresql-ubi-image:14.2.0-3747-32",
                    "pgRepository": "prod.harbor.keysight.digital/loadcore",
                    "repository": "prod.harbor.keysight.digital/loadcore",
                    "tagSuffix": "-ocp"
                  },
                  "licenseService": {
                    "port": 8080
                  },
                  "persistence": {
                    "storageClass": "managed-loadcore-storage"
                  },
                  "rbac": {
                    "serviceAccountName": "load-core-service-account"
                  },
                  "restStatsService": {
                    "port": 8080
                  },
                  "resultService": {
                    "port": 8080,
                    "storage": "64Gi"
                  },
                  "sessionManager": {
                    "port": 8080
                  },
                  "statsService": {
                    "port": 8080
                  },
                  "testResultsService": {
                    "port": 8080
                  }
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2023-09-19T12:34:00Z"
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: Core
            name: cores.wap.loadcore.keysight.com
            version: v1alpha1
        description: Keysight WAP Core Operator
        displayName: Keysight WAP Core Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - loadcore
        - keysight
        links:
        - name: Keysight Wap Core
          url: https://keysight-wap-core.domain
        maintainers:
        - email: vlad.butusina@keysight.com
          name: Vlad Butusina
        maturity: alpha
        provider:
          name: Keysight
          url: keysight.com
        relatedImages:
        - prod.harbor.keysight.digital/loadcore/keysight-wap-core@sha256:001b0d08cd84a9ebc13cd8fd35b68d431b96f84d71e933c1123ed6c0d244fd37
        - registry.connect.redhat.com/loadcore/keysight-wap-core-bundle@sha256:2f8295f8f56ef2c56e8215d4125cf08facd2f1d07fad0ef3a01eb0a2ef5f0555
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:0848d89e9f94be04f17beeee75807d7fd11102737e6483ee674f2cf3ae28f213
        version: 0.1.21
      entries:
      - name: keysight-wap-core.v0.1.21
        version: 0.1.21
      name: alpha
    defaultChannel: alpha
    packageName: keysight-wap-core
    provider:
      name: Keysight
      url: keysight.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: ibm-namespace-scope-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ibm-namespace-scope-operator.v4.2.14
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.ibm.com/v1",
                "kind": "NamespaceScope",
                "metadata": {
                  "name": "namespacescope"
                },
                "spec": {
                  "namespaceMembers": [
                    "ibm-common-services",
                    "default"
                  ],
                  "restartLabels": {
                    "intent": "projected"
                  }
                }
              }
            ]
          capabilities: Seamless Upgrades
          containerImage: registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
          createdAt: "2025-02-19T00:54:24Z"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: <4.2.14
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
          repository: https://github.com/IBM/ibm-namespace-scope-operator
          support: IBM
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NamespaceScope is the Schema for the namespacescopes API
            displayName: Namespace Scope
            kind: NamespaceScope
            name: namespacescopes.operator.ibm.com
            version: v1
        description: This operator automates the extension of operator watch and service
          account permission scope to other namespaces in an openshift cluster.
        displayName: IBM NamespaceScope Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - IBM
        - Cloud
        links:
        - name: IBM Namespace Scope Operator
          url: https://github.com/IBM/ibm-namespace-scope-operator
        maintainers:
        - email: support@ibm.com
          name: IBM Support
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: IBM
        relatedImages:
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator@sha256:0774b1d916ebc354050612474800dda3a501c66a7a5aab6d18698c8c4e81ac4e
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator-bundle@sha256:c743dd7de2edbd08c6c73e20cf15f3b147491ad77a81677339e88caffb93e84b
        - registry.marketplace.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:c2246cd957294ce09d20c10a357fbcff5e3d8396eba15feb7fe73c45dc4f7506
        - registry.connect.redhat.com/cpopen/ibm-namespace-scope-operator@sha256:623a93173e8f703849238806b40aac843cb2ed016c8763a4f8ea1a0b0990eeea
        - registry.marketplace.redhat.com/cpopen/cpfs/ibm-licensing-scanner@sha256:7a2cd72736bf9d30341763fd1652ff4b9ce3c785da5123645ee7664ee38c7283
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-bundle@sha256:b6cb9e5976b749dd7818ac19d1cfd5e24673f3c052a41d6e2fe1d2d7094694b4
        - registry.marketplace.redhat.com/cpopen/ibm-license-service-scanner-operator-catalog@sha256:c016904e8c5cfcf8270241ad5e28b52f528859984ed894cdd5e1283fa90cd6e8
        version: 4.2.14
      entries:
      - name: ibm-namespace-scope-operator.v4.2.14
        version: 4.2.14
      name: v4.2
    defaultChannel: v4.2
    packageName: ibm-namespace-scope-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Accuknox
      provider-url: https://accuknox.com
    name: kubearmor-operator-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kubearmor-operator.v1.0.0
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Security
          com.redhat.openshift.versions: v4.12
          createdAt: "2023-05-01T05:17:07Z"
          description: The KubeArmor Operator deloys and manages KubeArmor daemonset
            and its componenets within a cluster.
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.28.0+git
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/kubearmor/KubeArmor
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions: {}
        description: |-
          The KubeArmor Operator deploys and manages KubeArmor and its components:
          - KubeArmor Daemonset
          - KubeArmor Relay Server
          - KubeArmor Controller to manage CRDs (KubeArmorPolicy, KubeArmorHostPolicy)

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior (such as process
          execution, file access, and networking operations) of pods, containers, and nodes (VMs) at the system level.

          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace identities by leveraging eBPF.

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)
          website: [https://kubearmor.io/](https://kubearmor.io/)
        displayName: Kubearmor Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kubearmor-operator
        - runtime-security
        - security
        - cloud-native
        links:
        - name: Kubearmor
          url: https://kubearmor.io
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: alpha
        provider:
          name: Accuknox
          url: https://www.accuknox.com/
        relatedImages:
        - docker.io/kubearmor/kubearmor-relay-server@sha256:695f9ade3382bcf858077e2a7b9b01e0713071517603778d72f7aca5a6d71323
        - docker.io/kubearmor/kubearmor@sha256:52f7049ecdbfa779663d2cd44ddd7f36aecd8ccd2ca91e69709b628600029f13
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:411a47151609b7da29974ef3ae3b04f1bb1526f4328fd3e1d9b0165380018049
        - docker.io/kubearmor/kubearmor-controller@sha256:ccec0fb8bb1e91bd48631cc89baf752b842f19c33b50edab08f8b33e57af64d4
        - docker.io/kubearmor/kubearmor-init@sha256:44fa534587972c5f85e3e6f362dd8268e9088a33c5657a21b72a7e8563164fb5
        - docker.io/kubearmor/kubearmor-operator@sha256:8c2d56781a2adf78ee58f956e01417f38b6ad2f319feada6470e93f4f9bf0808
        version: 1.0.0
      entries:
      - name: kubearmor-operator.v1.0.0
        version: 1.0.0
      name: alpha
    - currentCSV: kubearmor-operator.v1.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "operator.kubearmor.com/v1",
                "kind": "KubeArmorConfig",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/name": "kubearmorconfig",
                    "app.kubernetes.io/part-of": "kubearmoroperator"
                  },
                    "name": "kubearmorconfig-default",
                    "namespace": "placeholder"
                },
                "spec": {
                  "adapters": {
                    "elasticsearch": {
                      "url": "",
                      "enabled": false,
                      "auth": {
                        "secretName": "elastic-secret",
                        "usernameKey": "username",
                        "passwordKey": "password",
                        "allowInsecureTLS": false,
                        "caCertSecretName": "",
                        "caCertKey": "ca.crt"
                      },
                      "alertsIndex": "kubearmor-alerts"
                    }
                  },
                  "recommendedPolicies": {
                      "enable": true,
                      "matchExpressions": [
                          {
                              "key": "namespace",
                              "operator": "NotIn",
                              "values": [
                                  "kube-system",
                                  "kubearmor",
                                  "openshift-infra",
                                  "openshift-ingress",
                                  "openshift-ingress-canary",
                                  "openshift-ingress-operator",
                                  "openshift-insights",
                                  "openshift-kni-infra",
                                  "openshift-kube-apiserver",
                                  "openshift-kube-apiserver-operator",
                                  "openshift-kube-controller-manager",
                                  "openshift-kube-controller-manager-operator",
                                  "openshift-kube-scheduler",
                                  "openshift-kube-scheduler-operator",
                                  "openshift-kube-storage-version-migrator",
                                  "openshift-kube-storage-version-migrator-operator",
                                  "openshift-logging",
                                  "openshift-machine-api",
                                  "openshift-machine-config-operator",
                                  "openshift-managed-node-metadata-operator",
                                  "openshift-managed-upgrade-operator",
                                  "openshift-marketplace",
                                  "openshift-monitoring",
                                  "openshift-multus",
                                  "openshift-must-gather-operator",
                                  "openshift-network-diagnostics",
                                  "openshift-network-node-identity",
                                  "openshift-network-operator",
                                  "openshift-node",
                                  "openshift-nutanix-infra",
                                  "openshift-oauth-apiserver",
                                  "openshift-observability-operator",
                                  "openshift-ocm-agent-operator",
                                  "openshift-openstack-infra",
                                  "openshift-operator-lifecycle-manager",
                                  "openshift-operators",
                                  "openshift-operators-redhat",
                                  "openshift-osd-metrics",
                                  "openshift-ovirt-infra",
                                  "openshift-ovn-kubernetes",
                                  "openshift-package-operator",
                                  "openshift-rbac-permissions",
                                  "openshift-route-controller-manager",
                                  "openshift-route-monitor-operator",
                                  "openshift-security",
                                  "openshift-service-ca",
                                  "openshift-service-ca-operator",
                                  "openshift-splunk-forwarder-operator",
                                  "openshift-sre-pruning",
                                  "openshift-user-workload-monitoring",
                                  "openshift-validation-webhook",
                                  "openshift-velero",
                                  "openshift-vsphere-infra",
                                  "openshift",
                                  "openshift-addon-operator",
                                  "openshift-apiserver",
                                  "openshift-apiserver-operator",
                                  "openshift-aqua",
                                  "openshift-authentication",
                                  "openshift-authentication-operator",
                                  "openshift-backplane",
                                  "openshift-backplane-cee",
                                  "openshift-backplane-csa",
                                  "openshift-backplane-cse",
                                  "openshift-backplane-csm",
                                  "openshift-backplane-managed-scripts",
                                  "openshift-backplane-mcs-tier-two",
                                  "openshift-backplane-mobb",
                                  "openshift-backplane-srep",
                                  "openshift-backplane-tam",
                                  "openshift-cloud-controller-manager",
                                  "openshift-cloud-controller-manager-operator",
                                  "openshift-cloud-credential-operator",
                                  "openshift-cloud-ingress-operator",
                                  "openshift-cloud-network-config-controller",
                                  "openshift-cloud-platform-infra",
                                  "openshift-cluster-csi-drivers",
                                  "openshift-cluster-machine-approver",
                                  "openshift-cluster-node-tuning-operator",
                                  "openshift-cluster-samples-operator",
                                  "openshift-cluster-storage-operator",
                                  "openshift-cluster-version",
                                  "openshift-codeready-workspaces",
                                  "openshift-config",
                                  "openshift-config-managed",
                                  "openshift-config-operator",
                                  "openshift-console",
                                  "openshift-console-operator",
                                  "openshift-console-user-settings",
                                  "openshift-controller-manager",
                                  "openshift-controller-manager-operator",
                                  "openshift-custom-domains-operator",
                                  "openshift-customer-monitoring",
                                  "openshift-deployment-validation-operator",
                                  "openshift-dns",
                                  "openshift-dns-operator",
                                  "openshift-etcd",
                                  "openshift-etcd-operator",
                                  "openshift-host-network",
                                  "openshift-image-registry"
                              ]
                          }
                      ]
                  },
                  "defaultCapabilitiesPosture": "audit",
                  "defaultFilePosture": "audit",
                  "defaultNetworkPosture": "audit",
                  "defaultVisibility": "process,network",
                  "enableStdOutLogs": false,
                  "enableStdOutAlerts": false,
                  "enableStdOutMsgs": false,
                  "seccompEnabled": false
                }
              }
            ]
          alm-examples-metadata: '{"kubearmorconfig-default":{"description":"Default
            KubeArmorConfig CR for OpenShift"}}'
          capabilities: Seamless Upgrades
          categories: Security, Monitoring
          certified: "true"
          com.redhat.openshift.versions: v4.9
          containerImage: docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
          createdAt: "2024-03-20T17:43:15Z"
          description: KubeArmor Operator
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: kubearmor
          operators.openshift.io/infrastructure-features: '["disconnected"]'
          operators.openshift.io/valid-subscription: ""
          operators.operatorframework.io/builder: operator-sdk-v1.31.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/
          support: Accuknox
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: KubeArmorConfig is the Schema for the kubearmorconfigs API
            kind: KubeArmorConfig
            name: kubearmorconfigs.operator.kubearmor.com
            version: v1
        description: |2-

          ## Prerequisites

          1. Create a new project for KubeArmor (Recommended)

             ```
             oc new-project kubearmor
             ```

           2. Create [KubeArmor SCC](https://github.com/kubearmor/KubeArmor/blob/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml)

             ```
             oc create -f https://raw.githubusercontent.com/kubearmor/KubeArmor/main/pkg/KubeArmorOperator/config/rbac/kubearmor-scc.yaml
             ```

             Note: please change the serviceaccount namespace if you are using some other namespace than the kubearmor.

          KubeArmor is a cloud-native runtime security enforcement system that restricts the behavior
          (such as process execution, file access, and networking operations) of pods, containers,
          and nodes (VMs) at the system level.
          KubeArmor leverages Linux security modules (LSMs) such as AppArmor, SELinux, or BPF-LSM to enforce the
          user-specified policies. KubeArmor generates rich alerts/telemetry events with container/pod/namespace
          identities by leveraging eBPF.

          ## Features and Benefits

          ### Harden Infrastructure

          safeguards vital paths like certificate bundles, enforces security rules based on MITRE, STIGs, and CIS guidelines,
          and restricts access to raw database tables, enhancing overall cluster security.

          ### Least Permissive Access

          controls sensitive asset access through process and network whitelisting,
          reducing security risks and maintaining system integrity.

          ### Application Behavior

          in-depth monitoring of process executions, file system access, service bindings,
          ingress and egress connections, and sensitive system call profiling

          ## Documentation

          github: [https://github.com/kubearmor/KubeArmor](https://github.com/kubearmor/KubeArmor)

          website: [https://kubearmor.io/](https://kubearmor.io/)

          ## Installing the KubeArmor Operator

          The KubeArmor Operator, available through Red Hat's Operator catalog, offers stable update channel.
          **Stable Channel**: This channel provides the latest stable version of KubeArmor along with its patches.

          The KubeArmor operator is responsible for deploying various components of the KubeArmor application, including the KubeArmor
          Daemonset, KubeArmor Controller, KubeArmor Relay Server, and KubeArmor Snitch. The role of KubeArmor Snitch is to gather information
          about the enabled Linux security modules on the host, which is utilized by KubeArmor to enforce runtime security policies.
          Additionally, KubeArmor Snitch collects details about the container runtime socket and storage volume cluster configurations.
          This deployment is executed as a short-term job and necessitates root user privileges. KubeArmor Snitch is certified by RedHat
          as a container with root privileges.

          KubeArmor Operator comes with a custom resource:

          - **KubeArmorConfig** - User will provide the initial configurations of the KubeArmor using KubeArmorConfig CR.

          ### KubeArmorConfig Customization Options:

          | Parameter                                 | Description                                          | Default                                 |
          | ---                                       | ---                                                  | ---                                     |
          | `alertThrottling`                         | Enable throttling for the KubeArmor generated alerts | False                                   |
          | `defaultFilePosture`                      | Default file posture for KubeArmor                   | audit                                   |
          | `defaultNetworkPosture`                   | Default network posture for KubeArmor                | audit                                   |
          | `defaultCapabilitiesPosture`              | Default capabilities posture for KubeArmor           | audit                                   |
          | `defaultVisibility`                       | Default visibility for KubeArmor                     | process,file,network                    |
          | `enableStdOutLogs` |  show stdout logs for relay server | false |
          | `enableStdOutAlerts` |  show stdout alerts for relay server | false |
          | `enableStdOutMsgs` |  show stdout messages for relay server | false |
          | `kubearmorImage.image`                    | Default image for KubeArmor Daemonset                | kubearmor/kubearmor-ubi:stable          |
          | `kubearmorImage.imagePullPolicy`          | Default imagePullPolicy for KubeArmor Daemonset      | Always                                  |
          | `kubearmorInitImage.image`                | Default image for KubeArmor Init Container           | kubearmor/kubearmor-init:stable         |
          | `kubearmorInitImage.imagePullPolicy`      | Default imagePullPolicy for KubeArmor Init Container | Always                                  |
          | `kubearmorRelayImage.image`               | Default image for KubeArmor Relay Server             | kubearmor/kubearmor-relay-server:latest |
          | `kubearmorRelayImage.imagePullPolicy`     | Default imagePullPolicy for KubeArmor Relay Server   | Always                                  |
          | `kubearmorControllerImage.image`          | Default image for KubeArmor Controller               | kubearmor/kubearmor-controller:latest   |
          | `kubearmorControllerImage.imagePullPolicy`| Default imagePullPolicy for KubeArmor Controller     | Always                                  |
          | `maxAlertPerSec`                          | Maximum no of alerts per second to trigger throttling| 10                                      |
          | `recommendedPolicies`                     | Recommended Policies Spec                            | -                                       |
          | `recommendedPolicies.matchExpressions`    | Target Namespaces for recommended policies           | -                                       |
          | `recommendedPolicies.excludePolicy`       | Exclude policies from recommended policies           | -                                       |
          | `seccompEnabled`                          | Deploy KubeArmor Daemon with default seccomp profile | True                                    |
          | `throttleSec`                             | Thottling duration in seconds                        | 30                                      |

          Users are not required to specify the images in the KubearmorConfig custom resource (CR). By default, the operator will utilize the images published with the operator release.

          For detailed information on Default Posture refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/default_posture.md)

          similarly for default visibility refer [this guide](https://github.com/kubearmor/KubeArmor/blob/main/getting-started/kubearmor_visibility.md)

          ## Support

          For Support please join KubeArmor Slack [Here](https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA)
        displayName: KubeArmor Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - monitoring
        - security
        - alerting
        - runtime-security
        - cloud native
        - container security
        links:
        - name: Website
          url: https://kubearmor.io
        - name: Github
          url: https://github.com/kubearmor/KubeArmor.git
        - name: Slack
          url: https://join.slack.com/t/kubearmor/shared_invite/zt-1ltmqdbc6-rSHw~LM6MesZZasmP2hAcA
        maintainers:
        - email: ramakant@accuknox.com
          name: Ramakant Sharma
        maturity: stable
        provider:
          name: Accuknox
          url: https://accuknox.com
        relatedImages:
        - docker.io/kubearmor/kubearmor-controller@sha256:2d369138cdbb31468c41e9ce70eda986a16bbeb9dd1192cd8a011c9df8a83857
        - docker.io/kubearmor/kubearmor-init@sha256:e30ee508aa95d273f74a543e1b412b98c1714d04f093b59991c0ec2abaa450be
        - docker.io/kubearmor/kubearmor-operator@sha256:d6c6b16e636474568835ded97cecb5ed17bc6004501227f6b6d63522ed57b01c
        - docker.io/kubearmor/kubearmor-relay-server@sha256:0a1a20ef02bd05eaa4a900948cadee6ae7650b0dde70abe1935608eaa5f931b7
        - docker.io/kubearmor/kubearmor-snitch@sha256:c3d0f45590f9a0ebb50254047f12a7468844bcae7961eccb11f5ef578832c0eb
        - docker.io/kubearmor/kubearmor-ubi@sha256:629e295a2e64953bf95f3f3e842c42ff665ed6e4db32db8e43dcde4befdd3fc2
        - registry.connect.redhat.com/kubearmor/kubearmor-operator@sha256:076c345de449353cc9837c603024ba9d790d26c5a8cebfc97db61ded1482ab78
        - registry.connect.redhat.com/sumologic/kube-rbac-proxy@sha256:1153a4592746b05e019bde4d818d176ff9350c013f84d49829032540de882841
        version: 1.5.2
      entries:
      - name: kubearmor-operator.v1.5.2
        version: 1.5.2
      name: stable
    defaultChannel: stable
    packageName: kubearmor-operator-certified
    provider:
      name: Accuknox
      url: https://accuknox.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Digital.ai
      provider-url: https://digital.ai
    name: digitalai-release-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: digitalai-release-operator.v25.3.0-beta.813
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "autoscaling": {
                    "enabled": false,
                    "maxReplicas": 5,
                    "metrics": [],
                    "minReplicas": 3
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "cd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.0-debian-12-r4"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.0-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/extra/lib",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.5.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824 '
          createdAt: "2025-09-02T05:48:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql.\n\n```yaml\napiVersion:
          xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n
          \ global:\n    security:\n      allowInsecureImages: true\n  k8sSetup:\n
          \   platform: Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula:
          true\n  keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\n#### Configuration Details\n\nThe sample configuration
          uses:\n\n- Autogenerated limited license\n- Simple administrator initial
          password\n- Default keystore setup\n- Only one replica is set for all pods\n-
          All Pods are missing resource limit setup\n- Default storage class and minimal
          size for persistent storage\n- Release pods are using persistence access
          mode `ReadWriteOnce`, for Release pods the requirement is to have `ReadWriteMany`\n-
          Embedded PostgreSQL for DB management\n\n#### Security Configuration Details\n\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:f481b2e364b9fe417a1cbbe9eff70180a3b4153a14712a8935527e25f919ee0d
        - docker.io/xebialabsunsupported/release-operator@sha256:52914f17dc55dfaef55aed382e7269408fb676a213d12117eafa906d81506824
        - docker.io/xebialabsunsupported/xl-release@sha256:0548a1ba63da5ae09c59b67a30240bf48877ab8950e09b0f0f6fd7c88ba34c13
        version: 25.3.0-beta.813
      entries:
      - name: digitalai-release-operator.v25.3.0-beta.813
        version: 25.3.0-beta.813
      - name: digitalai-release-operator.v25.3.0-beta.711
        version: 25.3.0-beta.711
      - name: digitalai-release-operator.v25.3.0-beta.609
        version: 25.3.0-beta.609
      - name: digitalai-release-operator.v25.3.0-beta.512
        version: 25.3.0-beta.512
      name: beta
    - currentCSV: digitalai-release-operator.v24.1.10
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "fsGroup": null,
                    "runAsGroup": null,
                    "runAsUser": null
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false,
                        "fsGroup": null,
                        "runAsGroup": null,
                        "runAsUser": null
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false,
                      "fsGroup": null,
                      "runAsGroup": null,
                      "runAsUser": null
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnami/kubectl",
                        "tag": "1.28.7-debian-12-r3"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "tag": "1.26.0-debian-12-r1"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "tag": "1.10.1-debian-12-r5"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "tag": "15.6.0-debian-12-r7"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "tag": "3.12.13-debian-12-r2"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "tag": "12-debian-12-r21"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnami/os-shell",
                      "tag": "12-debian-12-r16"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {
              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
          createdAt: "2024-10-16T04:25:19Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    runAsUser: null\n    runAsGroup: null\n
          \   fsGroup: null\n  containerSecurityContext:\n    runAsUser: null\n    runAsGroup:
          null\n  securityContextConstraints:\n    enabled: false\n  volumePermissions:\n
          \   enabled: false\n  route:\n    enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n        runAsUser: null\n        runAsGroup: null\n        fsGroup:
          null\n      containerSecurityContext:\n        enabled: false\n        runAsUser:
          null\n        runAsGroup: null\n      securityContextConstraints:\n        enabled:
          false\n    volumePermissions:\n      enabled: false\n  rabbitmq:\n    install:
          true\n    persistence:\n      size: 1Gi\n      storageClass: ''\n    replicaCount:
          1\n    podSecurityContext:\n      enabled: false\n      runAsUser: null\n
          \     runAsGroup: null\n      fsGroup: null\n    containerSecurityContext:\n
          \     enabled: false\n      runAsUser: null\n      runAsGroup: null\n    securityContextConstraints:\n
          \     enabled: false\n    volumePermissions:\n      enabled: false\n```\n\nIf
          SCC is required, it must be managed manually. The necessary permissions
          for accessing SecurityContextConstraints must be granted when SCC is enabled;
          otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:771a9a173e033a3ad8b46f5c00a7036eaa88c8d8d1fbd89217325168998113ea
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:a85d4f1c8a686abc0b9032dddda75991216d44dfade214625723d7009be314d3
        - docker.io/xebialabs/release-operator@sha256:47d6e31f66fce52c6566240b85fb028890409ad3addeea134120d1afa4814883
        - docker.io/xebialabs/xl-release@sha256:d1b65493ed524d37c60483f44b8bcc877ff398c81d36c7e767f2254dadec98cc
        version: 24.1.10
      entries:
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable
    - currentCSV: digitalai-release-operator.v24.1.21
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "securityContextConstraints": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "securityContextConstraints": {
                      "enabled": false
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "securityContextConstraints": {
                    "enabled": false
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "library/busybox",
                      "tag": "stable"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  jvm {\n      override-jvm-options = {{ .Values.jvm.overrideOptions }}\n      networkaddress.cache.ttl = {{ index .Values.jvm.options \"networkaddress.cache.ttl\" }}\n  }\n}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "image": {
                        "tag": "v0.14.7"
                      },
                      "ingressClass": "haproxy-dair",
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/bash\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/bash\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/bash\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "bitnamilegacy/kubectl",
                        "tag": "1.33.3-debian-12-r1"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-client/xl",
                        "kube",
                        "upgrade",
                        "--skip-context-check",
                        "--local-repo",
                        "/opt/xebialabs/xl-op-blueprints",
                        "--answers",
                        "/opt/xebialabs/xl-client/generated_answers.yaml"
                      ],
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    },
                    "overrideOptions": true
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.26.2-debian-12-r10"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "15.6.0-debian-12-r20"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "securityContextConstraints": {
                        "enabled": true
                      },
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.12.14-debian-12-r7"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "securityContextConstraints": {
                      "enabled": true
                    },
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r49"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "securityContextConstraints": {
                    "enabled": true
                  },
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "bitnamilegacy/os-shell",
                      "tag": "12-debian-12-r49"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  }
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator without SCC creation and with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c '
          createdAt: "2025-08-22T08:54:25Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration disables SCC creation
          and uses embedded Postgresql and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind:
          DigitalaiRelease\nmetadata:\n  name: dair-min\nspec:\n  k8sSetup:\n    platform:
          Openshift\n  auth:\n    adminPassword: 'admin'\n  licenseAcceptEula: true\n
          \ keystore:\n    passphrase: 'test1234'\n    keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  securityContextConstraints:\n    enabled: false\n
          \ volumePermissions:\n    enabled: false\n  route:\n    enabled: false\n
          \   annotations:\n      haproxy.router.openshift.io/cookie_name: JSESSIONID\n
          \     haproxy.router.openshift.io/disable_cookies: 'false'\n      haproxy.router.openshift.io/rewrite-target:
          /\n      haproxy.router.openshift.io/timeout: 120s\n    hostname: '<mandatory-release-hostname>'\n
          \   path: /\n    tls:\n      enabled: true\n      termination: edge\n  postgresql:\n
          \   install: true\n    primary:\n      persistence:\n        size: 1Gi\n
          \       storageClass: ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n
          \       enabled: false\n      containerSecurityContext:\n        enabled:
          false\n      securityContextConstraints:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   securityContextConstraints:\n      enabled: false\n    volumePermissions:\n
          \     enabled: false\n```\n\nIf SCC is required, it must be managed manually.
          The necessary permissions for accessing SecurityContextConstraints must
          be granted when SCC is enabled; otherwise, the following error will occur:\n\n```bash\nsecuritycontextconstraints.security.openshift.io
          \\\"dair-doc-digitalai-release-privileged\\\" is forbidden: User \\\"system:serviceaccount:openshift-operators:xlr-controller-manager\\\"
          cannot get resource \\\"securitycontextconstraints\\\" in API group \\\"security.openshift.io\\\"
          at the cluster scope\"\n```\n\n#### Configuration Details\n\nThe sample
          configuration uses:\n\n- Autogenerated limited license\n- Simple administrator
          initial password\n- Default keystore setup\n- Only one replica is set for
          all pods\n- All Pods are missing resource limit setup\n- Default storage
          class and minimal size for persistent storage\n- Release pods are using
          persistence access mode `ReadWriteOnce`, for Release pods the requirement
          is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB management\n- Embedded
          RabbitMQ for message queue management\n\n#### Security Configuration Details\n\n-
          `securityContextConstraints.enabled: false` - disables creation of SCCs;\n-
          `podSecurityContext/containerSecurityContext` - that disables the use of
          specific UIDs or GIDs, so the IDs can be assigned from the defined ranges
          (for example from restricted SCC);\n- `volumePermissions.enabled: false`
          - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:b71367012d1ed17d3967cd77fc28b38c69c76f75848fb20021dbb176dc382eed
        - docker.io/xebialabs/release-operator@sha256:aca76f0f5db170ddfc6ad692ea2874b9efb83fe732540946d88a0c078907ec8c
        - docker.io/xebialabs/xl-release@sha256:1cadd1d6ad7389ac9fe398ffba246b78238e54c180045b9dd44fdeefa3cf2af5
        version: 24.1.21
      entries:
      - name: digitalai-release-operator.v24.1.21
        version: 24.1.21
      - name: digitalai-release-operator.v24.1.20
        version: 24.1.20
      - name: digitalai-release-operator.v24.1.19
        version: 24.1.19
      - name: digitalai-release-operator.v24.1.18
        version: 24.1.18
      - name: digitalai-release-operator.v24.1.17
        version: 24.1.17
      - name: digitalai-release-operator.v24.1.14
        version: 24.1.14
      - name: digitalai-release-operator.v24.1.10
        version: 24.1.10
      name: stable-v24.1
    - currentCSV: digitalai-release-operator.v24.3.11
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "clusterMode": "full",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"xl-release_cluster\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n    }\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "health": {
                    "enabled": true,
                    "periodScans": 10,
                    "probeFailureThreshold": 12,
                    "probesLivenessTimeout": 60,
                    "probesReadinessTimeout": 60
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "enabled": false
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.12.4-debian-12-r0"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "16.3.0-debian-12-r23"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\n{{ if eq .Values.queueType \"quorum\" -}}\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n{{- end }}\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "3.13.7-debian-12-r5"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "queueType": "classic",
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {},
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 200,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803 '
          createdAt: "2025-09-26T18:24:33Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:32b442eb6344d3ab64d2885f53817e61f0e8b72b5c6e28845c8e056fc176b666
        - docker.io/xebialabs/release-operator@sha256:725fb5f7cfa4098c4110a40f54dc646f7663267bcba53603a7c1303c44051803
        - docker.io/xebialabs/xl-release@sha256:e9e87975888fcd4c8c164db90cb0a5c7abde88bf272c59d97e465af7b49bc6be
        version: 24.3.11
      entries:
      - name: digitalai-release-operator.v24.3.11
        version: 24.3.11
      - name: digitalai-release-operator.v24.3.10
        version: 24.3.10
      - name: digitalai-release-operator.v24.3.9
        version: 24.3.9
      - name: digitalai-release-operator.v24.3.8
        version: 24.3.8
      - name: digitalai-release-operator.v24.3.7
        version: 24.3.7
      - name: digitalai-release-operator.v24.3.6
        version: 24.3.6
      - name: digitalai-release-operator.v24.3.5
        version: 24.3.5
      name: stable-v24.3
    - currentCSV: digitalai-release-operator.v25.1.5
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-min"
                },
                "spec": {
                  "auth": {
                    "adminPassword": "admin"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "hooks": {
                    "getLicense": {
                      "enabled": true
                    }
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q==",
                    "passphrase": "test1234"
                  },
                  "licenseAcceptEula": true,
                  "persistence": {
                    "accessModes": [
                      "ReadWriteOnce"
                    ],
                    "size": "1Gi",
                    "storageClass": ""
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "1Gi",
                        "storageClass": ""
                      },
                      "podSecurityContext": {
                        "enabled": false
                      },
                      "resourcesPreset": "small"
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "1Gi",
                      "storageClass": ""
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": 1,
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": 1,
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "dair-placeholders"
                },
                "spec": {
                  "appHostname": "\u003cmandatory-release-hostname\u003e",
                  "appProtocol": "http",
                  "auth": {
                    "adminPassword": "\u003cmandatory-admin-password\u003e"
                  },
                  "containerSecurityContext": {
                    "enabled": false
                  },
                  "diagnosticMode": {
                    "enabled": false
                  },
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-db-host\u003e:5432/\u003cxlr-database-name\u003e",
                        "username": "\u003cxlr-username\u003e"
                      },
                      "report": {
                        "maxPoolSize": 10,
                        "password": "\u003cxlr-report-password\u003e",
                        "url": "jdbc:postgresql://\u003cxlr-report-db-host\u003e:5432/\u003cxlr-report-database-name\u003e",
                        "username": "\u003cxlr-report-username\u003e"
                      }
                    },
                    "mq": {
                      "enabled": false,
                      "password": "\u003cpassword\u003e",
                      "queueName": "\u003cqueue-name\u003e",
                      "queueType": "\u003cclassic-or-quorum\u003e",
                      "url": "\u003cqueue-url\u003e",
                      "username": "\u003cusername\u003e"
                    }
                  },
                  "global": {
                    "security": {
                      "allowInsecureImages": true
                    }
                  },
                  "haproxy-ingress": {
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "enabled": false
                    },
                    "getLicense": {
                      "enabled": true
                    },
                    "installReleaseRunner": {
                      "answersSecret": "remote-runner-answers",
                      "enabled": false,
                      "releaseName": "remote-runner"
                    }
                  },
                  "ingress": {
                    "enabled": false
                  },
                  "k8sSetup": {
                    "platform": "Openshift"
                  },
                  "keystore": {
                    "keystore": "\u003cmandatory-enter base64 encoded keystore\u003e",
                    "passphrase": "\u003cmandatory-keystore-passphrase\u003e"
                  },
                  "license": "\u003clicense\u003e",
                  "licenseAcceptEula": true,
                  "nginx-ingress-controller": {
                    "install": false
                  },
                  "oidc": {
                    "accessTokenUri": "",
                    "clientId": "",
                    "clientSecret": "",
                    "emailClaim": "",
                    "enabled": false,
                    "fullNameClaim": "",
                    "issuer": "",
                    "keyRetrievalUri": "",
                    "logoutUri": "",
                    "postLogoutRedirectUri": "",
                    "redirectUri": "",
                    "rolesClaim": "",
                    "scopes": "[\"openid\"]",
                    "userAuthorizationUri": "",
                    "userNameClaim": ""
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "size": "\u003crelease-storage-size\u003e",
                    "storageClass": "\u003cmandatory-release-storage-class-rwx\u003e"
                  },
                  "podSecurityContext": {
                    "enabled": false
                  },
                  "postgresql": {
                    "install": true,
                    "primary": {
                      "containerSecurityContext": {
                        "enabled": false
                      },
                      "persistence": {
                        "size": "\u003cpostgresql-storage-size\u003e",
                        "storageClass": "\u003cpostgresql-storage-class-rwo\u003e"
                      },
                      "podSecurityContext": {
                        "enabled": false
                      }
                    },
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "rabbitmq": {
                    "containerSecurityContext": {
                      "enabled": false
                    },
                    "install": true,
                    "persistence": {
                      "size": "\u003crabbitmq-storage-size\u003e",
                      "storageClass": "\u003crabbitmq-storage-class-rwo\u003e"
                    },
                    "podSecurityContext": {
                      "enabled": false
                    },
                    "replicaCount": "\u003crabbitmq-replica-count\u003e",
                    "volumePermissions": {
                      "enabled": false
                    }
                  },
                  "replicaCount": "\u003crelease-replica-count\u003e",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": true,
                    "hostname": "\u003cmandatory-release-hostname\u003e",
                    "path": "/",
                    "tls": {
                      "enabled": true,
                      "termination": "edge"
                    }
                  },
                  "ssl": {
                    "enabled": false,
                    "keystore": "",
                    "keystoreKeypassword": "",
                    "keystorePassword": "",
                    "keystoreType": ""
                  },
                  "volumePermissions": {
                    "enabled": false
                  }
                }
              },
              {
                "apiVersion": "xlr.digital.ai/v1alpha1",
                "kind": "DigitalaiRelease",
                "metadata": {
                  "name": "digitalairelease-sample"
                },
                "spec": {
                  "affinity": {},
                  "appContextRoot": "/",
                  "appHostname": "",
                  "appProtocol": "http",
                  "args": [
                    "--",
                    "/opt/xebialabs/xl-release-server/bin/run-in-container.sh"
                  ],
                  "auth": {
                    "adminPassword": null,
                    "sessionStorage": false
                  },
                  "busyBox": {
                    "image": {
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-client",
                      "tag": "{{ .Chart.AppVersion }}"
                    }
                  },
                  "cluster": {
                    "management": {
                      "ssl": {
                        "enabled": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    },
                    "manager": "xlr-legacy",
                    "mode": "full",
                    "name": "xl-release_cluster",
                    "node": {
                      "datacenter": "default"
                    },
                    "remote": {
                      "sslEngine": {
                        "enabled": false,
                        "enabledAlgorithms": [
                          "TLS_DHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256",
                          "TLS_DHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384",
                          "TLS_RSA_WITH_AES_128_CBC_SHA"
                        ],
                        "hostnameVerification": false,
                        "keyPassword": null,
                        "keystore": {},
                        "keystorePassword": null,
                        "keystoreType": "pkcs12",
                        "requireMutualAuthentication": false,
                        "truststore": {},
                        "truststorePassword": null,
                        "truststoreType": "pkcs12"
                      }
                    }
                  },
                  "clusterDomain": "cluster.local",
                  "command": [
                    "/opt/xebialabs/tini"
                  ],
                  "commonAnnotations": {},
                  "commonLabels": {},
                  "configuration": {
                    "default-conf_xl-release-conf-template": {
                      "content": "xl {\n  cluster {\n    # mode: \"default\", \"hot-standby\", \"full\"\n    mode = ${XL_CLUSTER_MODE}\n    name = \"${XLR_CLUSTER_NAME}\"\n    manager = \"${XLR_CLUSTER_MANAGER}\"\n    pekko {\n      loglevel = \"INFO\"\n      actor.debug.receive = off\n      remote {\n          log-received-messages = off\n          log-sent-messages = off\n      }\n      {{- if .Values.storage.reportingEngine.downingProviderClass }}\n      cluster {\n        downing-provider-class = {{.Values.storage.reportingEngine.downingProviderClass | quote}}\n      }\n      {{- end }}\n\n      {{- if .Values.cluster.remote.sslEngine.enabled }}\n      remote {\n        artery {  \n          ssl {  \n            config-ssl-engine {  \n              enabled-algorithms = [\n                {{- $last := sub (len .Values.cluster.remote.sslEngine.enabledAlgorithms) 1 }}\n                {{ range $index, $v := .Values.cluster.remote.sslEngine.enabledAlgorithms }}\n                {{ $v | quote }}{{ if ne $index $last }},{{ end }}\n                {{- end }}\n              ]  \n              key-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEY_PASSWORD}\"  \n              key-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.keystore \"default\" (printf \"cluster_remote_keystore.%s\" .Values.cluster.remote.sslEngine.keystoreType) \"context\" $) }}\"  \n              key-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_KEYSTORE_PASSWORD}\"  \n              trust-store = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.remote.sslEngine.truststore \"default\" (printf \"cluster_remote_truststore.%s\" .Values.cluster.remote.sslEngine.truststoreType) \"context\" $) }}\"  \n              trust-store-password = \"${?CLUSTER_REMOTE_SSL_ENGINE_TRUSTSTORE_PASSWORD}\"  \n              hostname-verification = {{ if .Values.cluster.remote.sslEngine.hostnameVerification }}on{{ else }}off{{ end }}\n              require-mutual-authentication = {{ if .Values.cluster.remote.sslEngine.requireMutualAuthentication }}on{{ else }}off{{ end }}  \n            }  \n          }  \n          transport = tls-tcp  \n        }  \n      }\n      {{- end }}\n    }\n    {{- if .Values.cluster.management.ssl.enabled }}\n    xlr-pekko-native {\n      pekko {\n        management {\n          http {  \n            route-providers-read-only = true  \n          }  \n          ssl {  \n            enabled = true\n            keyPassword = \"${?CLUSTER_MANAGEMENT_SSL_KEY_PASSWORD}\"  \n            keyStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.keystore \"default\" (printf \"cluster_management_keystore.%s\" .Values.cluster.management.ssl.keystoreType) \"context\" $) }}\"  \n            keyStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_KEYSTORE_PASSWORD}\"  \n            keystoreType = {{ .Values.cluster.management.ssl.keystoreType }}  \n            trustStore = \"/opt/xebialabs/xl-release-server/certs/{{ include \"secrets.key\" (dict \"secretRef\" .Values.cluster.management.ssl.truststore \"default\" (printf \"cluster_management_truststore.%s\" .Values.cluster.management.ssl.truststoreType) \"context\" $) }}\"  \n            trustStorePassword = \"${?CLUSTER_MANAGEMENT_SSL_TRUSTSTORE_PASSWORD}\"  \n            trustStoreType = {{ .Values.cluster.management.ssl.truststoreType }}  \n          }  \n        }  \n      }\n    }\n    {{- end }}\n  }\n\n  server {\n      http2 {\n          enabled = ${XLR_HTTP2_ENABLED}\n      }\n      ssl {\n          sniHostCheck = {{ .Values.ssl.sniHostCheck }}\n      }\n      session {\n          storage {\n              enabled = {{ include \"common.tplvalues.render\" ( dict \"value\" .Values.auth.sessionStorage \"context\" $ ) }}\n          }\n      }\n  }\n\n  license {\n    kind = ${XL_LICENSE_KIND}\n    product = \"xl-release\"\n  }\n\n  database {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_DB_URL}\"\n    db-username=${XL_DB_USERNAME}\n    max-pool-size=${XL_DB_MAX_POOL_SIZE}\n  }\n\n  # TODO Release does not support (H2) running in one schema.\n  reporting {\n    db-driver-classname=\"${XL_DB_DRIVER}\"\n    db-password=\"\"\"${XL_REPORT_DB_PASSWORD}\"\"\"\n    db-url=\"${XL_REPORT_DB_URL}\"\n    db-username=${XL_REPORT_DB_USERNAME}\n    max-pool-size=${XL_REPORT_DB_MAX_POOL_SIZE}\n    {{- if or .Values.storage.reportingEngine.s3.enabled .Values.storage.reportingEngine.cleanupInterval }}\n    engine {\n      {{- if .Values.storage.reportingEngine.cleanupInterval }}\n      cleanUpInterval = \"{{ .Values.storage.reportingEngine.cleanupInterval }}\"\n      {{- end }}\n      {{- if or .Values.storage.reportingEngine.s3.enabled }}\n      type = \"xlr-s3\"\n      s3 {\n        bucket = \"{{ .Values.storage.reportingEngine.s3.bucket | required \"storage.reportingEngine.s3.bucket is required\" }}\"\n        base-key = \"{{ .Values.storage.reportingEngine.s3.baseKey | required \"storage.reportingEngine.s3.baseKey is required\" }}\"\n        region = \"{{ .Values.storage.reportingEngine.s3.region | required \"storage.reportingEngine.s3.region is required\" }}\"\n        {{- if .Values.storage.reportingEngine.s3.endpointUri }}\n        endpoint-uri = \"{{ .Values.storage.reportingEngine.s3.endpointUri }}\"\n        {{- end }}\n      }\n      {{- end }}\n    }\n    {{- end }}\n  }\n\n  # Task queue\n  queue {\n    embedded=${ENABLE_EMBEDDED_QUEUE}\n    connector=\"${XLR_TASK_QUEUE_CONNECTOR_TYPE}\"\n    password=\"\"\"${XLR_TASK_QUEUE_PASSWORD}\"\"\"\n    queueName=\"${XLR_TASK_QUEUE_NAME}\"\n    url=\"${XLR_TASK_QUEUE_URL}\"\n    username=\"${XLR_TASK_QUEUE_USERNAME}\"\n    queueType=\"${XLR_TASK_QUEUE_TYPE}\"\n  }\n\n  metrics {\n    enabled = ${XL_METRICS_ENABLED}\n  }\n\n  {{- if .Values.oidc.enabled }}\n  security {\n    auth {\n      providers {\n        oidc {\n          clientId=\"\"\n          clientId=${?OIDC_CLIENT_ID}\n          clientSecret=\"\"\n          clientSecret=${?OIDC_CLIENT_SECRET}\n          {{- if .Values.oidc.clientAuthMethod }}\n          clientAuthMethod={{ .Values.oidc.clientAuthMethod | quote }}\n          {{- end }}\n          {{- if .Values.oidc.clientAuthJwt.enable }}\n          clientAuthJwt {\n              jwsAlg={{ default \"\" .Values.oidc.clientAuthJwt.jwsAlg | quote }}\n              tokenKeyId={{ default \"\" .Values.oidc.clientAuthJwt.tokenKeyId | quote }}\n              {{- if .Values.oidc.clientAuthJwt.keyStore.enable }}\n              keyStore {\n                  path={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.path | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEYSTORE_PASSWORD}\n                  type={{ default \"\" .Values.oidc.clientAuthJwt.keyStore.type | quote }}\n              }\n              {{- end }}\n              {{- if .Values.oidc.clientAuthJwt.key.enable }}\n              key {\n                  alias={{ default \"\" .Values.oidc.clientAuthJwt.key.alias | quote }}\n                  password=\"\"\n                  password=${?OIDC_CLIENT_AUTH_JWT_KEY_PASSWORD}\n              }\n              {{- end }}\n          }\n          {{- end }}\n          issuer={{ .Values.oidc.issuer | quote }}\n          keyRetrievalUri={{ default \"\" .Values.oidc.keyRetrievalUri | quote }}\n          accessTokenUri={{ default \"\" .Values.oidc.accessTokenUri | quote }}\n          userAuthorizationUri={{ default \"\" .Values.oidc.userAuthorizationUri | quote }}\n          logoutUri={{ default \"\" .Values.oidc.logoutUri | quote }}\n          redirectUri={{ .Values.oidc.redirectUri | quote }}\n          postLogoutRedirectUri={{ .Values.oidc.postLogoutRedirectUri | quote }}\n          userNameClaim={{ default \"\" .Values.oidc.userNameClaim | quote }}\n          fullNameClaim={{ default \"\" .Values.oidc.fullNameClaim | quote }}\n          emailClaim={{ default \"\" .Values.oidc.emailClaim | quote }}\n          {{- if .Values.oidc.externalIdClaim }}\n          externalIdClaim={{ .Values.oidc.externalIdClaim | quote }}\n          {{- end }}\n          rolesClaim={{ default \"\" .Values.oidc.rolesClaim | quote }}\n          {{- if .Values.oidc.scopes }}\n          scopes={{ .Values.oidc.scopes }}\n          {{- else }}\n          scopes=[\"openid\"]\n          {{- end }}\n          {{- if .Values.oidc.idTokenJWSAlg }}\n          idTokenJWSAlg={{ .Values.oidc.idTokenJWSAlg | quote }}\n          {{- end }}\n          {{- if .Values.oidc.accessToken.enable }}\n          access-token {\n              issuer={{ default \"\" .Values.oidc.accessToken.issuer | quote }}\n              audience={{ default \"\" .Values.oidc.accessToken.audience | quote }}\n              keyRetrievalUri={{ default \"\" .Values.oidc.accessToken.keyRetrievalUri | quote }}\n              jwsAlg={{ default \"\" .Values.oidc.accessToken.jwsAlg | quote }}\n              secretKey=\"\"\n              secretKey=${?OIDC_ACCESS_TOKEN_SECRET_KEY}\n              }\n          {{- end }}\n          {{- if .Values.oidc.proxyHost }}\n          proxyHost={{ .Values.oidc.proxyHost | quote }}\n          {{- end }}\n          {{- if .Values.oidc.proxyPort }}\n          proxyPort={{ .Values.oidc.proxyPort | quote }}\n          {{- end }}\n        }\n      }\n    }\n  }\n  {{- end }}\n  {{- if or .Values.storage.jobLog.s3.enabled .Values.storage.jobLog.cleanupInterval .Values.storage.jobLog.maxItemsPerCleanup }}\n  job-log {\n    {{- if .Values.storage.jobLog.cleanupInterval }}\n    cleanup-interval = \"{{ .Values.storage.jobLog.cleanupInterval }}\"\n    {{- end }}\n    {{- if .Values.storage.jobLog.maxItemsPerCleanup }}\n    max-items-per-cleanup = \"{{ .Values.storage.jobLog.maxItemsPerCleanup }}\"\n    {{- end }}\n    {{- if or .Values.storage.jobLog.s3.enabled }}\n    type = \"xlr-s3\"\n    s3 {\n      bucket = \"{{ .Values.storage.jobLog.s3.bucket | required \"storage.jobLog.s3.bucket is required\" }}\"\n      base-key = \"{{ .Values.storage.jobLog.s3.baseKey | required \"storage.jobLog.s3.baseKey is required\" }}\"\n      region = \"{{ .Values.storage.jobLog.s3.region | required \"storage.jobLog.s3.region is required\" }}\"\n      {{- if .Values.storage.jobLog.s3.endpointUri }}\n      endpoint-uri = \"{{ .Values.storage.jobLog.s3.endpointUri }}\"\n      {{- end }}\n    }\n    {{- end }}\n  }\n  {{- end }}\n}\n\nmanagement.endpoint.prometheus.enabled = {{ .Values.metrics.prometheus.enabled | default false }}\n\n{{ .Values.xlReleaseConfExtra }}\n",
                      "mode": 432,
                      "path": "default-conf/xl-release.conf.template"
                    },
                    "node-conf_xl-release-conf-template": {
                      "content": "# The ${APP_HOME}/node-conf/xl-release.conf file provides node specific properties, which are then merged with the \n# ${APP_HOME}/conf/xl-release.conf file by the xl-platform when the instance is started.\n# The ${APP_HOME}/node-conf takes precedance over ${APP_HOME}/conf/xl-release.conf and this is specified in ${APP_HOME}/conf/xlr-wrapper-linux.conf\nxl {\n  cluster {\n    node {\n      id = ${XL_NODE_NAME}\n      hostname = ${XL_NODE_NAME}\n      datacenter = ${XLR_CLUSTER_DATACENTER}\n      clusterPort = 5531\n    }\n  }\n}\n",
                      "mode": 432,
                      "path": "node-conf/xl-release.conf.template"
                    }
                  },
                  "containerPorts": {
                    "releaseHttp": 5516,
                    "releaseHttps": 5543
                  },
                  "containerSecurityContext": {
                    "allowPrivilegeEscalation": false,
                    "capabilities": {
                      "drop": [
                        "ALL"
                      ]
                    },
                    "enabled": true,
                    "readOnlyRootFilesystem": true,
                    "runAsNonRoot": true,
                    "seccompProfile": {
                      "type": "RuntimeDefault"
                    }
                  },
                  "debugMode": {
                    "enabled": false,
                    "remoteJvmParams": "{{- if .Values.debugMode.enabled }} -agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=localhost:8001{{- end }}"
                  },
                  "defaultInitContainers": {
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    }
                  },
                  "diagnosticMode": {
                    "args": [
                      "--",
                      "sleep",
                      "infinity"
                    ],
                    "command": [
                      "/opt/xebialabs/tini"
                    ],
                    "enabled": false
                  },
                  "dnsConfig": {},
                  "dnsPolicy": "",
                  "enableEmbeddedQueue": false,
                  "external": {
                    "db": {
                      "enabled": false,
                      "main": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      },
                      "report": {
                        "maxPoolSize": "",
                        "password": null,
                        "url": "",
                        "username": null
                      }
                    },
                    "mq": {
                      "connector": "",
                      "enabled": false,
                      "password": null,
                      "queueName": "",
                      "queueType": "",
                      "url": "",
                      "username": null
                    }
                  },
                  "extraConfiguration": {},
                  "extraContainerPorts": [],
                  "extraDeploy": [],
                  "extraEnvVars": [],
                  "extraEnvVarsCM": "",
                  "extraEnvVarsSecret": "",
                  "extraSecrets": {},
                  "extraSecretsPrependReleaseName": false,
                  "extraVolumeMounts": [],
                  "extraVolumes": [],
                  "forceRemoveMissingTypes": false,
                  "forceUpgrade": true,
                  "fullnameOverride": "",
                  "generateXlConfig": true,
                  "global": {
                    "imagePullSecrets": [],
                    "imageRegistry": "",
                    "postgresql": {
                      "service": {
                        "ports": {
                          "postgresql": ""
                        }
                      }
                    },
                    "security": {
                      "allowInsecureImages": true
                    },
                    "storageClass": ""
                  },
                  "haproxy-ingress": {
                    "controller": {
                      "extraVolumeMounts": [
                        {
                          "mountPath": "/etc/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/lib/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        },
                        {
                          "mountPath": "/var/run/haproxy",
                          "name": "empty-dir",
                          "subPath": "haproxy"
                        }
                      ],
                      "extraVolumes": [
                        {
                          "emptyDir": {},
                          "name": "empty-dir"
                        }
                      ],
                      "ingressClass": "haproxy-dair",
                      "legacySecurityContext": false,
                      "securityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "service": {
                        "type": "LoadBalancer"
                      }
                    },
                    "install": false
                  },
                  "hooks": {
                    "genSelfSigned": {
                      "configuration": {
                        "bin_gen-self-signed": {
                          "content": "#!/bin/sh\n\necho \"Generating release self-signed cert\"\nHOSTNAME=\"{{- include \"release.hostname\" . -}}\"\nSTOREPASS=\"{{- .Values.ssl.keystorePassword -}}\"\nKEYPASS=\"{{- .Values.ssl.keystoreKeypassword -}}\"\nKEYTYPE=\"{{- .Values.ssl.keystoreType -}}\"\nkeytool -genkey -keyalg RSA -alias dai-release -keystore conf/keystore.$KEYTYPE -validity 365 -keysize 2048 -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\" \\\n  -dname \"CN=$HOSTNAME,OU=,O=Digital.ai Release,L=,ST=,C=\" \\\n  -ext \"SAN=DNS:{{- include \"common.names.fullname\" . -}}.local\"\nkeytool -export -alias dai-release -keystore conf/keystore.$KEYTYPE -rfc -file conf/public.cert -storepass \"$STOREPASS\" -storetype \"$KEYTYPE\" -keypass \"$KEYPASS\"\n",
                          "mode": 493,
                          "path": "bin/gen-self-signed.sh"
                        },
                        "bin_install-self-signed": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.genSelfSigned.name \"context\" $ ) }}\"\nKEYSTORE_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/keystore.{{- .Values.ssl.keystoreType -}}\"\nCERT_FILE_PATH=\"/opt/xebialabs/xl-release-server/conf/public.cert\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$KEYSTORE_FILE_PATH \\\n    --from-file=$CERT_FILE_PATH \\\n    --from-literal=host={{- include \"release.hostname\" . -}} \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-self-signed.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "genCommand": [
                        "/opt/xebialabs/xl-release-server/bin/gen-self-signed.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-self-signed.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-self-signed"
                    },
                    "getLicense": {
                      "configuration": {
                        "bin_get-license": {
                          "content": "#!/bin/sh\n\necho \"Requesting unregistered license\"\nSERVER_PATH_PART=https://download.xebialabs.com\necho -e $(curl -X POST \"${SERVER_PATH_PART}/api/unregistered/xl-release\" | jq --raw-output .license) \u003e ${APP_HOME}/conf/xl-release-license.lic\nfile_size=$(stat -c%s \"${APP_HOME}/conf/xl-release-license.lic\")\nif [ \"$file_size\" -lt 10 ]; then\n  echo \"License file is NOT valid\"\n  exit 1\nfi\n",
                          "mode": 493,
                          "path": "bin/get-license.sh"
                        },
                        "bin_install-license": {
                          "content": "#!/bin/sh\n\nSECRET_NAME=\"{{ include \"common.tplvalues.render\" ( dict \"value\" $.Values.hooks.getLicense.name \"context\" $ ) }}\"\nFILE_PATH=\"/opt/xebialabs/xl-release-server/conf/xl-release-license.lic\"\n\nif kubectl get secret \"$SECRET_NAME\" \u003e /dev/null 2\u003e\u00261; then\n  echo \"Secret '$SECRET_NAME' exists skipping creation.\"\nelse\n  kubectl create secret generic $SECRET_NAME \\\n    --from-file=$FILE_PATH \\\n    --dry-run=client \\\n    -o yaml | kubectl apply -f -\nfi\n",
                          "mode": 493,
                          "path": "bin/install-license.sh"
                        }
                      },
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": true,
                      "getCommand": [
                        "/opt/xebialabs/xl-release-server/bin/get-license.sh"
                      ],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": [
                        "/opt/xebialabs/xl-release-server/bin/install-license.sh"
                      ],
                      "name": "{{ include \"release.name\" . }}-license"
                    },
                    "installReleaseRunner": {
                      "answersSecret": "",
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "capabilities": {
                          "drop": [
                            "ALL"
                          ]
                        },
                        "enabled": true,
                        "readOnlyRootFilesystem": true,
                        "runAsNonRoot": true,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "deletePolicy": "before-hook-creation,hook-succeeded",
                      "enabled": false,
                      "extraEnvVars": [],
                      "image": {
                        "pullPolicy": "IfNotPresent",
                        "pullSecrets": [],
                        "registry": "docker.io",
                        "repository": "xebialabsunsupported/xl-client",
                        "tag": "{{ .Chart.AppVersion }}"
                      },
                      "installCommand": "#!/bin/sh\n\ncd /tmp; \n/opt/xebialabs/xl-client/xl kube upgrade \\\n  --skip-context-check \\\n  --local-repo /opt/xebialabs/xl-op-blueprints \\\n  --answers /opt/xebialabs/xl-client/generated_answers.yaml\n",
                      "name": "{{ include \"release.name\" . }}-install-runner",
                      "releaseName": ""
                    }
                  },
                  "hostAliases": [],
                  "image": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": [],
                    "registry": "docker.io",
                    "repository": "xebialabsunsupported/xl-release",
                    "tag": "{{ .Chart.AppVersion }}"
                  },
                  "ingress": {
                    "annotations": null,
                    "enabled": false,
                    "extraHosts": [],
                    "extraPaths": [],
                    "extraRules": [],
                    "extraTls": [],
                    "hostname": "",
                    "ingressClassName": "",
                    "path": "/",
                    "pathType": "ImplementationSpecific",
                    "secrets": [],
                    "selfSigned": false,
                    "tls": false
                  },
                  "initContainers": [],
                  "jvm": {
                    "options": {
                      "networkaddress.cache.ttl": 30
                    }
                  },
                  "jvmArgs": "",
                  "k8sSetup": {
                    "platform": "PlainK8s",
                    "validateValues": true
                  },
                  "keystore": {
                    "keystore": null,
                    "passphrase": null
                  },
                  "kubeVersion": "",
                  "license": null,
                  "licenseAcceptEula": false,
                  "lifecycleHooks": {},
                  "livenessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/liveness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "logback": {
                    "globalLoggingLevel": "info",
                    "scanEnabled": true,
                    "scanPeriod": "30 seconds"
                  },
                  "metrics": {
                    "jmx": {
                      "enabled": false
                    },
                    "prometheus": {
                      "enabled": false,
                      "podAnnotations": {
                        "prometheus.io/port": "{{ .Values.service.ports.metrics }}",
                        "prometheus.io/scrape": "true"
                      },
                      "rule": {
                        "additionalLabels": {},
                        "enabled": false,
                        "namespace": "",
                        "rules": []
                      },
                      "serviceMonitor": {
                        "annotations": {},
                        "basicAuth": {
                          "password": {
                            "key": "releasePassword",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          },
                          "username": {
                            "key": "releaseUsername",
                            "name": "{{ template \"common.names.fullname\" . }}",
                            "optional": false
                          }
                        },
                        "enabled": false,
                        "honorLabels": false,
                        "interval": "30s",
                        "jobLabel": "",
                        "labels": {},
                        "metricRelabelings": [],
                        "namespace": "",
                        "params": {},
                        "path": "/s/actuator/prometheus",
                        "podTargetLabels": {},
                        "relabelings": [],
                        "scrapeTimeout": "",
                        "selector": {
                          "service-type": "normal"
                        },
                        "targetLabels": {}
                      }
                    }
                  },
                  "nameOverride": "",
                  "namespaceOverride": "",
                  "networkPolicy": {
                    "additionalRules": [],
                    "allowExternal": true,
                    "enabled": false
                  },
                  "nginx-ingress-controller": {
                    "defaultBackend": {
                      "image": {
                        "repository": "bitnamilegacy/nginx",
                        "tag": "1.29.1-debian-12-r0"
                      }
                    },
                    "extraArgs": {
                      "ingress-class": "nginx-dair"
                    },
                    "image": {
                      "repository": "bitnamilegacy/nginx-ingress-controller",
                      "tag": "1.13.1-debian-12-r1"
                    },
                    "ingressClassResource": {
                      "controllerClass": "k8s.io/ingress-nginx-dair",
                      "name": "nginx-dair"
                    },
                    "install": false,
                    "replicaCount": 1,
                    "service": {
                      "type": "LoadBalancer"
                    }
                  },
                  "nodeAffinityPreset": {
                    "key": "",
                    "type": "",
                    "values": []
                  },
                  "nodeSelector": {},
                  "oidc": {
                    "accessToken": {
                      "audience": null,
                      "enable": false,
                      "issuer": null,
                      "jwsAlg": null,
                      "keyRetrievalUri": null,
                      "secretKey": null
                    },
                    "accessTokenUri": null,
                    "clientAuthJwt": {
                      "enable": false,
                      "jwsAlg": null,
                      "key": {
                        "alias": null,
                        "enable": false,
                        "password": null
                      },
                      "keyStore": {
                        "enable": false,
                        "password": null,
                        "path": null,
                        "type": null
                      },
                      "tokenKeyId": null
                    },
                    "clientAuthMethod": null,
                    "clientId": null,
                    "clientSecret": null,
                    "emailClaim": null,
                    "enabled": false,
                    "external": false,
                    "externalIdClaim": null,
                    "fullNameClaim": null,
                    "idTokenJWSAlg": null,
                    "issuer": null,
                    "keyRetrievalUri": null,
                    "logoutUri": null,
                    "postLogoutRedirectUri": null,
                    "proxyHost": null,
                    "proxyPort": null,
                    "redirectUri": null,
                    "rolesClaim": null,
                    "scopes": [
                      "openid"
                    ],
                    "userAuthorizationUri": null,
                    "userNameClaim": null
                  },
                  "pdb": {
                    "create": false,
                    "maxUnavailable": "",
                    "minAvailable": 1
                  },
                  "persistence": {
                    "accessModes": [
                      "ReadWriteMany"
                    ],
                    "annotations": {
                      "helm.sh/resource-policy": "keep"
                    },
                    "emptyDirPaths": [
                      "/opt/xebialabs/xl-release-server/cache",
                      "/opt/xebialabs/xl-release-server/conf",
                      "/opt/xebialabs/xl-release-server/driver/jdbc",
                      "/opt/xebialabs/xl-release-server/driver/mq",
                      "/opt/xebialabs/xl-release-server/log",
                      "/opt/xebialabs/xl-release-server/node-conf",
                      "/opt/xebialabs/xl-release-server/plugins",
                      "/opt/xebialabs/xl-release-server/work",
                      "/tmp"
                    ],
                    "enabled": true,
                    "existingClaim": "",
                    "paths": [
                      "/opt/xebialabs/xl-release-server/reports"
                    ],
                    "selector": {},
                    "single": true,
                    "size": "8Gi",
                    "storageClass": ""
                  },
                  "podAffinityPreset": "",
                  "podAnnotations": {},
                  "podAntiAffinityPreset": "soft",
                  "podLabels": {},
                  "podManagementPolicy": "OrderedReady",
                  "podSecurityContext": {
                    "enabled": true,
                    "fsGroup": 10001,
                    "runAsUser": 10001
                  },
                  "postgresql": {
                    "auth": {
                      "enablePostgresUser": true,
                      "postgresPassword": "postgres",
                      "username": "postgres"
                    },
                    "image": {
                      "repository": "bitnamilegacy/postgresql",
                      "tag": "17.4.0-debian-12-r19"
                    },
                    "install": true,
                    "primary": {
                      "configuration": "port = '5432'\nwal_level = 'replica'\nmax_wal_size = '400MB'\nmax_wal_senders = '16'\nwal_keep_size = '128MB'\nhot_standby = 'on'\nfsync = 'on'\nshared_preload_libraries = 'pgaudit'\npgaudit.log_catalog = 'off'\nlog_connections = 'false'\nlog_disconnections = 'false'\nlog_hostname = 'false'\nclient_min_messages = 'error'\nlisten_addresses = '*'\nmax_connections = 300\nshared_buffers = 80MB",
                      "initdb": {
                        "scriptsSecret": "{{ include \"postgresql.v1.primary.fullname\" . }}-release"
                      },
                      "persistence": {
                        "accessModes": [
                          "ReadWriteOnce"
                        ],
                        "enabled": true,
                        "existingClaim": "",
                        "size": "8Gi",
                        "storageClass": ""
                      },
                      "resourcesPreset": "small",
                      "service": {
                        "ports": {
                          "postgresql": 5432
                        },
                        "type": "ClusterIP"
                      }
                    },
                    "serviceAccount": {
                      "create": true
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "priorityClassName": "",
                  "rabbitmq": {
                    "auth": {
                      "existingErlangSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "password": "guest",
                      "username": "guest"
                    },
                    "clustering": {
                      "forceBoot": true
                    },
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "capabilities": {
                        "drop": [
                          "ALL"
                        ]
                      },
                      "readOnlyRootFilesystem": true,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "extraConfiguration": "raft.wal_max_size_bytes = 1048576\nquorum_queue.continuous_membership_reconciliation.enabled = true\nquorum_queue.continuous_membership_reconciliation.target_group_size = 3\nquorum_queue.continuous_membership_reconciliation.auto_remove = true\nquorum_queue.continuous_membership_reconciliation.interval = 600000\n",
                    "extraPlugins": "rabbitmq_amqp1_0, rabbitmq_jms_topic_exchange",
                    "image": {
                      "repository": "bitnamilegacy/rabbitmq",
                      "tag": "4.1.3-debian-12-r1"
                    },
                    "install": true,
                    "loadDefinition": {
                      "enabled": true,
                      "existingSecret": "{{ include \"common.names.fullname\" . }}-release",
                      "file": "/app/release_load_definition.json"
                    },
                    "persistence": {
                      "accessModes": [
                        "ReadWriteOnce"
                      ],
                      "enabled": true,
                      "size": "8Gi",
                      "storageClass": ""
                    },
                    "replicaCount": 3,
                    "service": {
                      "type": "ClusterIP"
                    },
                    "volumePermissions": {
                      "containerSecurityContext": {
                        "allowPrivilegeEscalation": false,
                        "readOnlyRootFilesystem": true,
                        "runAsGroup": 0,
                        "runAsNonRoot": false,
                        "runAsUser": 0,
                        "seccompProfile": {
                          "type": "RuntimeDefault"
                        }
                      },
                      "enabled": true,
                      "image": {
                        "repository": "bitnamilegacy/os-shell",
                        "tag": "12-debian-12-r51"
                      }
                    }
                  },
                  "rbac": {
                    "create": true
                  },
                  "readinessProbe": {
                    "enabled": true,
                    "failureThreshold": 3,
                    "initialDelaySeconds": 10,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "replicaCount": 3,
                  "resources": {},
                  "resourcesPreset": "micro",
                  "route": {
                    "annotations": {
                      "haproxy.router.openshift.io/cookie_name": "JSESSIONID",
                      "haproxy.router.openshift.io/disable_cookies": "false",
                      "haproxy.router.openshift.io/rewrite-target": "/",
                      "haproxy.router.openshift.io/timeout": "120s"
                    },
                    "enabled": false,
                    "hostname": "",
                    "path": "/",
                    "tls": {
                      "caCertificate": "",
                      "certificate": "",
                      "destinationCACertificate": "",
                      "enabled": false,
                      "insecureEdgeTerminationPolicy": "",
                      "key": "",
                      "secretName": "",
                      "selfSigned": false,
                      "termination": "edge"
                    }
                  },
                  "schedulerName": "",
                  "service": {
                    "annotations": {},
                    "annotationsHeadless": {},
                    "clusterIP": "",
                    "externalIPs": [],
                    "externalTrafficPolicy": "Cluster",
                    "extraPorts": [],
                    "labels": {
                      "service-type": "normal"
                    },
                    "loadBalancerIP": "",
                    "loadBalancerSourceRanges": [],
                    "nodePorts": {
                      "releaseHttp": "",
                      "releaseHttps": ""
                    },
                    "portEnabled": true,
                    "portNames": {
                      "releaseHttp": "release-http",
                      "releaseHttps": "release-https"
                    },
                    "ports": {
                      "releaseHttp": 80,
                      "releaseHttps": 443
                    },
                    "sessionAffinity": "None",
                    "sessionAffinityConfig": {},
                    "type": "ClusterIP"
                  },
                  "serviceAccount": {
                    "annotations": {},
                    "automountServiceAccountToken": true,
                    "create": true,
                    "name": ""
                  },
                  "sidecars": [],
                  "ssl": {
                    "enabled": false,
                    "keystore": {
                      "valueFrom": {
                        "secretKeyRef": {
                          "key": "keystore.{{ .Values.ssl.keystoreType }}",
                          "name": "{{ include \"common.tplvalues.render\" ( dict \"value\" .Values.hooks.genSelfSigned.name \"context\" $ ) }}"
                        }
                      }
                    },
                    "keystoreKeypassword": null,
                    "keystorePassword": null,
                    "keystoreType": "pkcs12",
                    "sniHostCheck": true
                  },
                  "startupProbe": {
                    "enabled": true,
                    "failureThreshold": 12,
                    "initialDelaySeconds": 60,
                    "path": "s/actuator/health/readiness",
                    "periodSeconds": 10,
                    "successThreshold": 1,
                    "timeoutSeconds": 20
                  },
                  "statefulsetAnnotations": {},
                  "statefulsetLabels": {},
                  "storage": {
                    "jobLog": {
                      "cleanupInterval": null,
                      "maxItemsPerCleanup": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    },
                    "reportingEngine": {
                      "cleanupInterval": null,
                      "s3": {
                        "baseKey": null,
                        "bucket": null,
                        "enabled": false,
                        "endpointUri": null,
                        "region": null
                      }
                    }
                  },
                  "terminationGracePeriodSeconds": 450,
                  "tolerations": [],
                  "topologySpreadConstraints": [],
                  "truststore": {
                    "params": "{{- if .Values.truststore.truststore }} -Djavax.net.ssl.trustStore=$(TRUSTSTORE) -Djavax.net.ssl.trustStorePassword=$(TRUSTSTORE_PASSWORD) -Djavax.net.ssl.trustStoreType=$(TRUSTSTORE_TYPE){{- end }}",
                    "password": null,
                    "truststore": {},
                    "type": "pkcs12"
                  },
                  "updateStrategy": {
                    "type": "RollingUpdate"
                  },
                  "useIpAsHostname": false,
                  "volumePermissions": {
                    "containerSecurityContext": {
                      "allowPrivilegeEscalation": false,
                      "readOnlyRootFilesystem": true,
                      "runAsGroup": 0,
                      "runAsNonRoot": false,
                      "runAsUser": 0,
                      "seccompProfile": {
                        "type": "RuntimeDefault"
                      }
                    },
                    "enabled": false,
                    "image": {
                      "digest": "",
                      "pullPolicy": "IfNotPresent",
                      "pullSecrets": [],
                      "registry": "docker.io",
                      "repository": "xebialabsunsupported/xl-release",
                      "tag": "{{ .Chart.AppVersion }}"
                    },
                    "resources": {
                      "limits": {
                        "cpu": "150m",
                        "ephemeral-storage": "2Gi",
                        "memory": "192Mi"
                      },
                      "requests": {
                        "cpu": "100m",
                        "ephemeral-storage": "50Mi",
                        "memory": "128Mi"
                      }
                    },
                    "script": "#!/bin/bash\n\ndeclare -a paths=( {{ range $path := .Values.persistence.paths }} \"{{ $path }}\"{{ end }} )\nfor path in \"${paths[@]}\"; do\n  echo \"Changing ownership to {{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }} for ${path}\"\n  chown \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\" \"${path}\"\n  find \"${path}\" -mindepth 1 -maxdepth 1 -not -name \".snapshot\" -not -name \"lost+found\" | \\\n    xargs -r chown -R \"{{ .Values.containerSecurityContext.runAsUser }}:{{ .Values.podSecurityContext.fsGroup }}\"\ndone\n"
                  },
                  "xlReleaseConfExtra": ""
                }
              }
            ]
          alm-examples-metadata: |-
            {

              "dair-min": {
                "description": "Minimal configuration for the Digital.ai Release operator with embedded Postgresql and RabbitMQ and route setup"
              },
              "dair-placeholders": {
                "description": "Configuration for the Digital.ai Release operator with placeholders to set. Set at least mandatory placeholders, others that are not set delete"
              },
              "digitalairelease-sample": {
                "description": "Sample configuration with all default values"
              },
            }
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: 'docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4 '
          createdAt: "2025-09-26T17:28:37Z"
          description: Digital.ai Release empowers developer teams to seamlessly release
            and deploy software on a large scale across hybrid environments, all while
            ensuring that compliance and security requirements are effectively in
            place.
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          repository: https://github.com/digital-ai/release-helm-chart
          support: Digital.ai
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: DigitalaiRelease
            name: digitalaireleases.xlr.digital.ai
            version: v1alpha1
        description: "Digital.ai Release empowers developer teams to seamlessly release
          and deploy software on a large scale across hybrid environments, all while
          ensuring that compliance and security requirements are effectively in place.\n\n##
          About this Operator\n\nDigital.ai Releas operator is based on helm chart
          that is available from [here](https://github.com/digital-ai/release-helm-chart).\nThe
          operator includes subcharts which are dependencies for the Release containers
          to run successfully. \nNginx Ingress and Haproxy ingress are disabled and
          not used with installations on the Openshift (and the images are not used
          in the installation)\n\n### Subcharts include:\n\n- [PostgresSQL](https://github.com/bitnami/charts/tree/main/bitnami/postgresql)\n-
          [RabbitMQ](https://github.com/bitnami/charts/tree/main/bitnami/rabbitmq)\n-
          [Nginx Ingress controller](https://github.com/bitnami/charts/tree/main/bitnami/nginx-ingress-controller)\n-
          [Haproxy Ingress controller](https://github.com/jcmoraisjr/haproxy-ingress/)\n\n##
          Installation\n\nThe installation can be done using the sample configuration
          provided below. Please note that this is a minimal configuration and it's
          not recommended for production use.\nThis configuration uses embedded Postgresql
          and RabbitMQ.\n\n```yaml\napiVersion: xlr.digital.ai/v1alpha1\nkind: DigitalaiRelease\nmetadata:\n
          \ name: dair-min\nspec:\n  global:\n    security:\n      allowInsecureImages:
          true\n  k8sSetup:\n    platform: Openshift\n  auth:\n    adminPassword:
          'admin'\n  licenseAcceptEula: true\n  keystore:\n    passphrase: 'test1234'\n
          \   keystore: 'zs7OzgAAAAIAAAABAAAAAwAWZGVwbG95aXQtcGFzc3N3b3JkLWtleQAAAY66Cf9nrO0ABXNyADNjb20uc3VuLmNyeXB0by5wcm92aWRlci5TZWFsZWRPYmplY3RGb3JLZXlQcm90ZWN0b3LNV8pZ5zC7UwIAAHhyABlqYXZheC5jcnlwdG8uU2VhbGVkT2JqZWN0PjY9psO3VHACAARbAA1lbmNvZGVkUGFyYW1zdAACW0JbABBlbmNyeXB0ZWRDb250ZW50cQB+AAJMAAlwYXJhbXNBbGd0ABJMamF2YS9sYW5nL1N0cmluZztMAAdzZWFsQWxncQB+AAN4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAARMA8ECD0yEJAHFhBOAgMDDUB1cQB+AAUAAACQUhNP1jw1dwOxWArpm0JBAX40fr4fvvAmyGGrx7mzlCVjb4uOxi4IroUoxcbBx8cpjiS6QzCUg6chsCQ0IABB6s7Tow2VR8vRlxXyxpJXvlUwfH8hwJn/ZK8rQuDsCkzd2rC+lYR0pEO0lUP3/AQ7dHJ4TgloyZOKKbWRNJvE4K4EOafqASuNTkyrr0SPflRydAAWUEJFV2l0aE1ENUFuZFRyaXBsZURFU3QAFlBCRVdpdGhNRDVBbmRUcmlwbGVERVO3iVKkgrEJQ78wacyjhp3UZBSB/Q=='\n
          \ replicaCount: 1\n  hooks:\n    getLicense:\n      enabled: true\n  persistence:\n
          \   storageClass: ''\n    accessModes:\n      - ReadWriteOnce\n    size:
          1Gi\n  podSecurityContext:\n    enabled: false\n  containerSecurityContext:\n
          \   enabled: false\n  volumePermissions:\n    enabled: false\n  route:\n
          \   enabled: false\n    annotations:\n      haproxy.router.openshift.io/cookie_name:
          JSESSIONID\n      haproxy.router.openshift.io/disable_cookies: 'false'\n
          \     haproxy.router.openshift.io/rewrite-target: /\n      haproxy.router.openshift.io/timeout:
          120s\n    hostname: '<mandatory-release-hostname>'\n    path: /\n    tls:\n
          \     enabled: true\n      termination: edge\n  postgresql:\n    install:
          true\n    primary:\n      persistence:\n        size: 1Gi\n        storageClass:
          ''\n      resourcesPreset: \"small\"\n      podSecurityContext:\n        enabled:
          false\n      containerSecurityContext:\n        enabled: false\n    volumePermissions:\n
          \     enabled: false\n  rabbitmq:\n    install: true\n    persistence:\n
          \     size: 1Gi\n      storageClass: ''\n    replicaCount: 1\n    podSecurityContext:\n
          \     enabled: false\n    containerSecurityContext:\n      enabled: false\n
          \   volumePermissions:\n      enabled: false\n```\n\n#### Configuration
          Details\n\nThe sample configuration uses:\n\n- Autogenerated limited license\n-
          Simple administrator initial password\n- Default keystore setup\n- Only
          one replica is set for all pods\n- All Pods are missing resource limit setup\n-
          Default storage class and minimal size for persistent storage\n- Release
          pods are using persistence access mode `ReadWriteOnce`, for Release pods
          the requirement is to have `ReadWriteMany`\n- Embedded PostgreSQL for DB
          management\n- Embedded RabbitMQ for message queue management\n\n#### Security
          Configuration Details\n\n- `podSecurityContext/containerSecurityContext`
          - that disables the use of specific UIDs or GIDs, so the IDs can be assigned
          from the defined ranges (for example from restricted SCC);\n- `volumePermissions.enabled:
          false` - disables automatic corrections of the mounted folders.\n\n## Customize
          Your Configuration\n\nConfigurations can be further personalized. Here are
          some areas you might want to consider:\n\n- **External DB:** For avoiding
          uncertified container images required by the subcharts, refer to these [instructions
          for setting up an external DB](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-db-migration.html).\n-
          **External Message Queue:** To set up an external message queue, follow
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-external-mq-migration.html).\n-
          **Custom License:** To apply a custom license, refer to this [process](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-license-update.html).\n-
          **Custom Configuration:** For modifying the configuration files, use this
          [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-customize.html).\n-
          **Truststore Setup:** For modifying default truststore for Release, use
          this [guide](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-setup-truststore.html).\n-
          **Plugin Management:** Managing plugins can be done using the following
          methods:\n    - [Plugin Management using XL CLI](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-plugin-management.html)\n
          \   - [Managing Plugins Offline for Kubernetes Environment](https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-release-offline-plugin-management.html)"
        displayName: Digital.ai Release
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - digitalai-release-operator
        - digitalai
        - digital.ai
        - release
        - devops
        links:
        - name: Digital.ai Release
          url: https://digital.ai/products/release/
        - name: Digital.ai Release Documentation
          url: https://docs.digital.ai/bundle/devops-release-version-v.24.1/page/release/operator/xl-op-before-you-begin.html
        - name: Digital.ai Release Helm Chart
          url: https://github.com/digital-ai/release-helm-chart
        maintainers:
        - email: support@digital.ai
          name: support
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Digital.ai
          url: https://digital.ai
        relatedImages:
        - docker.io/xebialabs/release-operator@sha256:838443baf39514c5e0ac0a09cab411942a3cc4a13e261c8c60138788fe8d96e4
        - docker.io/xebialabs/xl-release@sha256:9b63789cdb87745832f1283d3f0e28ac5a4514d767895e0914a6be6672f2d14f
        - registry.connect.redhat.com/digital-ai/release-operator@sha256:268aa90a9f6eebf6cdeb32dc47335e75e7483dcce4ea3331b1b736c2fd79bea4
        version: 25.1.5
      entries:
      - name: digitalai-release-operator.v25.1.5
        version: 25.1.5
      - name: digitalai-release-operator.v25.1.4
        version: 25.1.4
      - name: digitalai-release-operator.v25.1.3
        version: 25.1.3
      - name: digitalai-release-operator.v25.1.2
        version: 25.1.2
      - name: digitalai-release-operator.v25.1.1
        version: 25.1.1
      - name: digitalai-release-operator.v25.1.0
        version: 25.1.0
      name: stable-v25.1
    defaultChannel: stable-v24.3
    packageName: digitalai-release-operator
    provider:
      name: Digital.ai
      url: https://digital.ai
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-iam-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-iam-controller.v1.5.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Role",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Group",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "OpenIDConnectProvider",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "Policy",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              },
              {
                "apiVersion": "iam.services.k8s.aws/v1alpha1",
                "kind": "User",
                "metadata": {
                  "name": "example"
                },
                "spec": {}
              }
            ]
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
          createdAt: "2025-09-19T19:13:34Z"
          description: AWS IAM controller is a service controller for managing IAM
            resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Group represents the state of an AWS iam Group resource.
            displayName: Group
            kind: Group
            name: groups.iam.services.k8s.aws
            version: v1alpha1
          - description: InstanceProfile represents the state of an AWS iam InstanceProfile
              resource.
            displayName: InstanceProfile
            kind: InstanceProfile
            name: instanceprofiles.iam.services.k8s.aws
            version: v1alpha1
          - description: OpenIDConnectProvider represents the state of an AWS iam
              OpenIDConnectProvider resource.
            displayName: OpenIDConnectProvider
            kind: OpenIDConnectProvider
            name: openidconnectproviders.iam.services.k8s.aws
            version: v1alpha1
          - description: Policy represents the state of an AWS iam Policy resource.
            displayName: Policy
            kind: Policy
            name: policies.iam.services.k8s.aws
            version: v1alpha1
          - description: Role represents the state of an AWS iam Role resource.
            displayName: Role
            kind: Role
            name: roles.iam.services.k8s.aws
            version: v1alpha1
          - description: ServiceLinkedRole represents the state of an AWS iam ServiceLinkedRole
              resource.
            displayName: ServiceLinkedRole
            kind: ServiceLinkedRole
            name: servicelinkedroles.iam.services.k8s.aws
            version: v1alpha1
          - description: User represents the state of an AWS iam User resource.
            displayName: User
            kind: User
            name: users.iam.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Identity and Access Management (IAM) resources in AWS from within your Kubernetes cluster.

          **About Amazon IAM**

          AWS Identity and Access Management (IAM) is a web service that helps you securely control access to AWS resources. You use IAM to control who is authenticated (signed in) and authorized (has permissions) to use resources.

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon IAM
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - iam
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon IAM Developer Resources
          url: https://aws.amazon.com/iam/resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: iam maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - quay.io/community-operator-pipeline-prod/ack-iam-controller@sha256:2d904b6462679d578505409d9bc585795983df7a4f0dd6668253c724aaa5063b
        - public.ecr.aws/aws-controllers-k8s/iam-controller:1.5.1
        version: 1.5.1
      entries:
      - name: ack-iam-controller.v1.5.1
        version: 1.5.1
      - name: ack-iam-controller.v1.5.0
        version: 1.5.0
      - name: ack-iam-controller.v1.4.4
        version: 1.4.4
      - name: ack-iam-controller.v1.4.3
        version: 1.4.3
      - name: ack-iam-controller.v1.4.2
        version: 1.4.2
      - name: ack-iam-controller.v1.4.1
        version: 1.4.1
      - name: ack-iam-controller.v1.4.0
        version: 1.4.0
      - name: ack-iam-controller.v1.3.22
        version: 1.3.22
      - name: ack-iam-controller.v1.3.21
        version: 1.3.21
      - name: ack-iam-controller.v1.3.20
        version: 1.3.20
      - name: ack-iam-controller.v1.3.19
        version: 1.3.19
      - name: ack-iam-controller.v1.3.18
        version: 1.3.18
      - name: ack-iam-controller.v1.3.17
        version: 1.3.17
      - name: ack-iam-controller.v1.3.15
        version: 1.3.15
      - name: ack-iam-controller.v1.3.14
        version: 1.3.14
      - name: ack-iam-controller.v1.3.13
        version: 1.3.13
      - name: ack-iam-controller.v1.3.12
        version: 1.3.12
      - name: ack-iam-controller.v1.3.11
        version: 1.3.11
      - name: ack-iam-controller.v1.3.10
        version: 1.3.10
      - name: ack-iam-controller.v1.3.9
        version: 1.3.9
      - name: ack-iam-controller.v1.3.8
        version: 1.3.8
      - name: ack-iam-controller.v1.3.6
        version: 1.3.6
      - name: ack-iam-controller.v1.3.2
        version: 1.3.2
      - name: ack-iam-controller.v1.3.1
        version: 1.3.1
      - name: ack-iam-controller.v1.3.0
        version: 1.3.0
      - name: ack-iam-controller.v1.2.6
        version: 1.2.6
      - name: ack-iam-controller.v1.2.5
        version: 1.2.5
      - name: ack-iam-controller.v1.2.3
        version: 1.2.3
      - name: ack-iam-controller.v1.2.2
        version: 1.2.2
      - name: ack-iam-controller.v1.2.1
        version: 1.2.1
      - name: ack-iam-controller.v1.2.0
        version: 1.2.0
      - name: ack-iam-controller.v1.1.3
        version: 1.1.3
      - name: ack-iam-controller.v1.1.2
        version: 1.1.2
      - name: ack-iam-controller.v1.1.1
        version: 1.1.1
      - name: ack-iam-controller.v1.1.0
        version: 1.1.0
      - name: ack-iam-controller.v1.0.0
        version: 1.0.0
      - name: ack-iam-controller.v0.1.2
        version: 0.1.2
      - name: ack-iam-controller.v0.1.1
        version: 0.1.1
      - name: ack-iam-controller.v0.1.0
        version: 0.1.0
      - name: ack-iam-controller.v0.0.22
        version: 0.0.22
      - name: ack-iam-controller.v0.0.21
        version: 0.0.21
      - name: ack-iam-controller.v0.0.20
        version: 0.0.20
      - name: ack-iam-controller.v0.0.19
        version: 0.0.19
      - name: ack-iam-controller.v0.0.18
        version: 0.0.18
      - name: ack-iam-controller.v0.0.17
        version: 0.0.17
      - name: ack-iam-controller.v0.0.16
        version: 0.0.16
      - name: ack-iam-controller.v0.0.15
        version: 0.0.15
      - name: ack-iam-controller.v0.0.14
        version: 0.0.14
      - name: ack-iam-controller.v0.0.13
        version: 0.0.13
      - name: ack-iam-controller.v0.0.12
        version: 0.0.12
      - name: ack-iam-controller.v0.0.11
        version: 0.0.11
      - name: ack-iam-controller.v0.0.10
        version: 0.0.10
      - name: ack-iam-controller.v0.0.9
        version: 0.0.9
      - name: ack-iam-controller.v0.0.8
        version: 0.0.8
      - name: ack-iam-controller.v0.0.7
        version: 0.0.7
      - name: ack-iam-controller.v0.0.6
        version: 0.0.6
      - name: ack-iam-controller.v0.0.5
        version: 0.0.5
      - name: ack-iam-controller.v0.0.4
        version: 0.0.4
      - name: ack-iam-controller.v0.0.3
        version: 0.0.3
      name: alpha
    defaultChannel: alpha
    packageName: ack-iam-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: KubeRocketCI
      provider-url: https://docs.kuberocketci.io/
    name: edp-keycloak-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: edp-keycloak-operator.v1.19.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "ssoAutoRedirectEnabled": false,
                  "ssoRealmEnabled": false,
                  "ssoRealmName": "openshift"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "instagram",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "hideOnLoginPage": "true",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "instagram",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "instagram",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "realmName": "realm-sample"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.19.0
          createdAt: "2023-11-15T16:04:45Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/epam/edp-keycloak-operator
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: alpha
        minKubeVersion: 1.20.0
        provider:
          name: EPAM Delivery Platform
          url: https://epam.github.io/edp-install/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.19.0
        - quay.io/openshift-community-operators/edp-keycloak-operator@sha256:82ff141bb86599f12d4a516fd1b6c5252f9ff24ef2547be6c58e2f3d88219914
        version: 1.19.0
      entries:
      - name: edp-keycloak-operator.v1.19.0
        version: 1.19.0
      - name: edp-keycloak-operator.v1.18.2
        version: 1.18.2
      - name: edp-keycloak-operator.v1.17.1
        version: 1.17.1
      - name: edp-keycloak-operator.v1.13.0
        version: 1.13.0
      name: alpha
    - currentCSV: edp-keycloak-operator.v1.29.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "Keycloak",
                "metadata": {
                  "name": "keycloak-sample"
                },
                "spec": {
                  "secret": "my-keycloak-secret",
                  "url": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakAuthFlow",
                "metadata": {
                  "name": "keycloakauthflow-sample"
                },
                "spec": {
                  "alias": "MyBrowser",
                  "authenticationExecutions": [
                    {
                      "authenticator": "auth-cookie",
                      "priority": 0,
                      "requirement": "ALTERNATIVE"
                    },
                    {
                      "authenticator": "identity-provider-redirector",
                      "authenticatorConfig": {
                        "alias": "my-alias",
                        "config": {
                          "defaultProvider": "my-alias"
                        }
                      },
                      "priority": 1,
                      "requirement": "REQUIRED"
                    }
                  ],
                  "builtIn": false,
                  "description": "browser with idp",
                  "providerId": "basic-flow",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "topLevel": true
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-authorization-sample"
                },
                "spec": {
                  "adminFineGrainedPermissionsEnabled": true,
                  "authorization": {
                    "permissions": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Resource permission",
                        "logic": "POSITIVE",
                        "name": "resource-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "resources": [
                          "resource1"
                        ],
                        "type": "resource"
                      },
                      {
                        "decisionStrategy": "CONSENSUS",
                        "description": "Scope permission",
                        "logic": "POSITIVE",
                        "name": "scope-permission",
                        "policies": [
                          "role-policy"
                        ],
                        "scopes": [
                          "scope1"
                        ],
                        "type": "scope"
                      }
                    ],
                    "policies": [
                      {
                        "decisionStrategy": "AFFIRMATIVE",
                        "description": "Role policy",
                        "logic": "POSITIVE",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "aggregatedPolicy": {
                          "policies": [
                            "policy1",
                            "policy2"
                          ]
                        },
                        "description": "Aggregate policy",
                        "name": "aggregate-policy",
                        "type": "aggregate"
                      },
                      {
                        "clientPolicy": {
                          "clients": [
                            "client1",
                            "client2"
                          ]
                        },
                        "description": "Client policy",
                        "name": "client-policy",
                        "type": "client"
                      },
                      {
                        "description": "Group policy",
                        "groupPolicy": {
                          "groups": [
                            {
                              "extendChildren": true,
                              "name": "group1"
                            }
                          ]
                        },
                        "name": "group-policy",
                        "type": "group"
                      },
                      {
                        "description": "Role policy",
                        "name": "role-policy",
                        "rolePolicy": {
                          "roles": [
                            {
                              "name": "developer",
                              "required": true
                            }
                          ]
                        },
                        "type": "role"
                      },
                      {
                        "description": "Time policy",
                        "name": "time-policy",
                        "timePolicy": {
                          "notBefore": "2021-01-01T00:00:00Z",
                          "notOnOrAfter": "2021-12-31T23:59:59Z"
                        },
                        "type": "time"
                      },
                      {
                        "description": "User policy",
                        "name": "user-policy",
                        "type": "user",
                        "userPolicy": {
                          "users": [
                            "user1",
                            "user2"
                          ]
                        }
                      }
                    ],
                    "resources": [
                      {
                        "displayName": "Resource 1",
                        "iconUri": "https://example.com/icon.png",
                        "name": "resource1",
                        "scopes": [
                          "scope1"
                        ],
                        "type": "test"
                      }
                    ],
                    "scopes": [
                      "scope1"
                    ]
                  },
                  "authorizationServicesEnabled": true,
                  "clientId": "authorization-sample",
                  "directAccess": true,
                  "permission": {
                    "scopePermissions": [
                      {
                        "name": "token-exchange",
                        "policies": [
                          "policy1"
                        ]
                      }
                    ]
                  },
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-authorization-sample:client-secret-key",
                  "serviceAccount": {
                    "enabled": true
                  },
                  "webUrl": "https://example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClient",
                "metadata": {
                  "name": "keycloakclient-sample"
                },
                "spec": {
                  "advancedProtocolMappers": true,
                  "authenticationFlowBindingOverrides": {
                    "browser": "browser",
                    "directGrant": "direct grant"
                  },
                  "clientId": "agocd",
                  "defaultClientScopes": [
                    "argocd_groups"
                  ],
                  "directAccess": true,
                  "public": false,
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "secret": "$client-secret-name:client-secret-key",
                  "webUrl": "https://argocd.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakClientScope",
                "metadata": {
                  "name": "keycloakclientscope-sample"
                },
                "spec": {
                  "description": "Group Membership",
                  "name": "groups",
                  "protocol": "openid-connect",
                  "protocolMappers": [
                    {
                      "config": {
                        "access.token.claim": "true",
                        "claim.name": "groups",
                        "full.path": "false",
                        "id.token.claim": "true",
                        "userinfo.token.claim": "true"
                      },
                      "name": "groups",
                      "protocol": "openid-connect",
                      "protocolMapper": "oidc-group-membership-mapper"
                    }
                  ],
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealm",
                "metadata": {
                  "name": "keycloakrealm-sample"
                },
                "spec": {
                  "id": "d1-id-kc-realm-name",
                  "keycloakRef": {
                    "kind": "Keycloak",
                    "name": "keycloak-sample"
                  },
                  "organizationsEnabled": true,
                  "passwordPolicy": [
                    {
                      "type": "forceExpiredPasswordChange",
                      "value": "365"
                    },
                    {
                      "type": "length",
                      "value": "8"
                    }
                  ],
                  "realmEventConfig": {
                    "adminEventsDetailsEnabled": false,
                    "adminEventsEnabled": true,
                    "enabledEventTypes": [
                      "UPDATE_CONSENT_ERROR",
                      "CLIENT_LOGIN"
                    ],
                    "eventsEnabled": true,
                    "eventsExpiration": 15000,
                    "eventsListeners": [
                      "jboss-logging"
                    ]
                  },
                  "realmName": "d2-id-kc-realm-name",
                  "userProfileConfig": {
                    "attributes": [
                      {
                        "annotations": {
                          "inputType": "text"
                        },
                        "displayName": "Test Attribute",
                        "group": "test-group",
                        "multivalued": true,
                        "name": "test-attribute",
                        "permissions": {
                          "edit": [
                            "admin"
                          ],
                          "view": [
                            "admin",
                            "user"
                          ]
                        },
                        "required": {
                          "roles": [
                            "admin"
                          ],
                          "scopes": [
                            "profile"
                          ]
                        },
                        "selector": {
                          "scopes": [
                            "profile"
                          ]
                        },
                        "validations": {
                          "email": {
                            "max-local-length": {
                              "intVal": 64
                            }
                          },
                          "local-date": {},
                          "multivalued": {
                            "max": {
                              "stringVal": "10"
                            },
                            "min": {
                              "stringVal": "1"
                            }
                          },
                          "options": {
                            "options": {
                              "sliceVal": [
                                "option1",
                                "option2"
                              ]
                            }
                          }
                        }
                      }
                    ],
                    "groups": [
                      {
                        "annotations": {
                          "groupAnnotation": "groupAnnotation"
                        },
                        "displayDescription": "Test Group",
                        "displayHeader": "Test Group",
                        "name": "test-group"
                      }
                    ],
                    "unmanagedAttributePolicy": "ENABLED"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmComponent",
                "metadata": {
                  "name": "keycloakrealmcomponent-sample"
                },
                "spec": {
                  "config": {
                    "allowPasswordAuthentication": [
                      "true"
                    ],
                    "cachePolicy": [
                      "EVICT_WEEKLY"
                    ],
                    "debug": [
                      "true"
                    ],
                    "editMode": [
                      "READ_ONLY"
                    ],
                    "enabled": [
                      "true"
                    ],
                    "evictionDay": [
                      "3"
                    ],
                    "evictionHour": [
                      "5"
                    ],
                    "evictionMinute": [
                      "7"
                    ],
                    "kerberosRealm": [
                      "test-realm"
                    ],
                    "keyTab": [
                      "test-key-tab"
                    ],
                    "priority": [
                      "0"
                    ],
                    "serverPrincipal": [
                      "srv-principal-test"
                    ],
                    "updateProfileFirstLogin": [
                      "true"
                    ]
                  },
                  "name": "cr-kerb-test",
                  "providerId": "kerberos",
                  "providerType": "org.keycloak.storage.UserStorageProvider",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmGroup",
                "metadata": {
                  "name": "keycloakrealmgroup-sample"
                },
                "spec": {
                  "name": "ArgoCDAdmins",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "test-org-idp",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org-idp",
                  "config": {
                    "clientId": "test-org-client-id",
                    "clientSecret": "test-org-client-secret",
                    "kc.org.broker.redirect.mode.email-matches": "true",
                    "kc.org.domain": "example.com"
                  },
                  "enabled": true,
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmIdentityProvider",
                "metadata": {
                  "name": "keycloakrealmidentityprovider-sample"
                },
                "spec": {
                  "alias": "github",
                  "authenticateByDefault": false,
                  "config": {
                    "clientId": "foo",
                    "clientSecret": "$secretName:secretKey",
                    "syncMode": "IMPORT",
                    "useJwksUrl": "true"
                  },
                  "enabled": true,
                  "firstBrokerLoginFlowAlias": "first broker login",
                  "mappers": [
                    {
                      "config": {
                        "role": "role-tr",
                        "syncMode": "INHERIT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "oidc-hardcoded-role-idp-mapper",
                      "name": "test3212"
                    },
                    {
                      "config": {
                        "attribute": "foo",
                        "attribute.value": "bar",
                        "syncMode": "IMPORT"
                      },
                      "identityProviderAlias": "github",
                      "identityProviderMapper": "hardcoded-attribute-idp-mapper",
                      "name": "test-33221"
                    }
                  ],
                  "providerId": "github",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRole",
                "metadata": {
                  "name": "keycloakrealmrole-sample"
                },
                "spec": {
                  "composite": true,
                  "description": "default developer role",
                  "name": "developer",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  }
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmRoleBatch",
                "metadata": {
                  "name": "keycloakrealmrolebatch-sample"
                },
                "spec": {
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "roles": [
                    {
                      "composite": true,
                      "description": "default developer role",
                      "isDefault": false,
                      "name": "developer"
                    },
                    {
                      "composite": true,
                      "description": "default administrator role",
                      "isDefault": false,
                      "name": "administrator"
                    }
                  ]
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1",
                "kind": "KeycloakRealmUser",
                "metadata": {
                  "name": "keycloakrealmuser-sample"
                },
                "spec": {
                  "attributes": {
                    "baz": "jazz",
                    "foo": "bar"
                  },
                  "email": "john.snow13@example.com",
                  "emailVerified": true,
                  "enabled": true,
                  "firstName": "John",
                  "keepResource": true,
                  "lastName": "Snow",
                  "password": "12345678",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "keycloakrealm-sample"
                  },
                  "requiredUserActions": [
                    "UPDATE_PASSWORD"
                  ],
                  "username": "john.snow13"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloak",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloak-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloak",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloak-sample"
                },
                "spec": {
                  "secret": "keycloak-access",
                  "url": "https://keycloak.example.com"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "ClusterKeycloakRealm",
                "metadata": {
                  "labels": {
                    "app.kubernetes.io/created-by": "edp-keycloak-operator",
                    "app.kubernetes.io/instance": "clusterkeycloakrealm-sample",
                    "app.kubernetes.io/managed-by": "kustomize",
                    "app.kubernetes.io/name": "clusterkeycloakrealm",
                    "app.kubernetes.io/part-of": "edp-keycloak-operator"
                  },
                  "name": "clusterkeycloakrealm-sample"
                },
                "spec": {
                  "authenticationFlows": {
                    "browserFlow": "browserFlow-sample"
                  },
                  "clusterKeycloakRef": "clusterkeycloak-sample",
                  "organizationsEnabled": false,
                  "realmName": "realm-sample"
                }
              },
              {
                "apiVersion": "v1.edp.epam.com/v1alpha1",
                "kind": "KeycloakOrganization",
                "metadata": {
                  "name": "test-keycloak-organization",
                  "namespace": "default"
                },
                "spec": {
                  "alias": "test-org",
                  "attributes": {
                    "department": [
                      "engineering",
                      "qa"
                    ],
                    "location": [
                      "us-east"
                    ]
                  },
                  "description": "Test organization",
                  "domains": [
                    "example.com",
                    "test.com"
                  ],
                  "identityProviders": [
                    {
                      "alias": "test-org-idp"
                    }
                  ],
                  "name": "Test Organization",
                  "realmRef": {
                    "kind": "KeycloakRealm",
                    "name": "test-org-realm"
                  },
                  "redirectUrl": "https://example.com/redirect"
                }
              }
            ]
          capabilities: Deep Insights
          categories: Security
          containerImage: docker.io/epamedp/keycloak-operator:1.29.0
          createdAt: "2025-09-26T12:16:55Z"
          description: An Operator for managing Keycloak
          operators.operatorframework.io/builder: operator-sdk-v1.41.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/epam/edp-keycloak-operator
          support: KubeRocketCI
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: ClusterKeycloakRealm is the Schema for the clusterkeycloakrealms
              API.
            displayName: Cluster Keycloak Realm
            kind: ClusterKeycloakRealm
            name: clusterkeycloakrealms.v1.edp.epam.com
            version: v1alpha1
          - description: ClusterKeycloak is the Schema for the clusterkeycloaks API.
            displayName: Cluster Keycloak
            kind: ClusterKeycloak
            name: clusterkeycloaks.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakAuthFlow is the Schema for the keycloak authentication
              flow API.
            displayName: Keycloak Auth Flow
            kind: KeycloakAuthFlow
            name: keycloakauthflows.v1.edp.epam.com
            version: v1
          - description: KeycloakClient is the Schema for the keycloak clients API.
            displayName: Keycloak Client
            kind: KeycloakClient
            name: keycloakclients.v1.edp.epam.com
            version: v1
          - description: KeycloakClientScope is the Schema for the keycloakclientscopes
              API.
            displayName: Keycloak Client Scope
            kind: KeycloakClientScope
            name: keycloakclientscopes.v1.edp.epam.com
            version: v1
          - description: KeycloakOrganization is the Schema for the organizations
              API.
            displayName: Keycloak Organization
            kind: KeycloakOrganization
            name: keycloakorganizations.v1.edp.epam.com
            version: v1alpha1
          - description: KeycloakRealmComponent is the Schema for the keycloak component
              API.
            displayName: Keycloak Realm Component
            kind: KeycloakRealmComponent
            name: keycloakrealmcomponents.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmGroup is the Schema for the keycloak group API.
            displayName: Keycloak Realm Group
            kind: KeycloakRealmGroup
            name: keycloakrealmgroups.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmIdentityProvider is the Schema for the keycloak
              realm identity provider API.
            displayName: Keycloak Realm Identity Provider
            kind: KeycloakRealmIdentityProvider
            name: keycloakrealmidentityproviders.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRoleBatch is the Schema for the keycloak roles
              API.
            displayName: Keycloak Realm Role Batch
            kind: KeycloakRealmRoleBatch
            name: keycloakrealmrolebatches.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmRole is the Schema for the keycloak group API.
            displayName: Keycloak Realm Role
            kind: KeycloakRealmRole
            name: keycloakrealmroles.v1.edp.epam.com
            version: v1
          - description: KeycloakRealm is the Schema for the keycloak realms API.
            displayName: Keycloak Realm
            kind: KeycloakRealm
            name: keycloakrealms.v1.edp.epam.com
            version: v1
          - description: KeycloakRealmUser is the Schema for the keycloak user API.
            displayName: Keycloak Realm User
            kind: KeycloakRealmUser
            name: keycloakrealmusers.v1.edp.epam.com
            version: v1
          - description: Keycloak is the Schema for the keycloaks API.
            displayName: Keycloak
            kind: Keycloak
            name: keycloaks.v1.edp.epam.com
            version: v1
        description: |
          Keycloak Operator is an operator that is responsible for establishing
          a connection to provided Keycloak Server, reconciling Keycloak entities (realms,
          roles, groups, users, etc) according to the created CRs.

          ## Quick Start

          1. Create a User in the Keycloak `Master` realm, and assign a `create-realm` role.

          2. Insert newly created user credentials into Kubernetes secret:

              ```yaml
              apiVersion: v1
              kind: Secret
              metadata:
                name:  keycloak-access
              type: Opaque
              data:
                username: dXNlcg==   # base64-encoded value of "user"
                password: cGFzcw==   # base64-encoded value of "pass"
              ```

          3. Create Custom Resource `kind: Keycloak` with Keycloak instance URL and secret created on the previous step:

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: Keycloak
              metadata:
                name: keycloak-sample
              spec:
                secret: keycloak-access             # Secret name
                url: https://keycloak.example.com   # Keycloak URL
              ```

              Wait for the `.status` field with  `status.connected: true`

          4. Create Keycloak realm and group using Custom Resources:

            ```yaml
            apiVersion: v1.edp.epam.com/v1
            kind: KeycloakRealm
            metadata:
              name: keycloakrealm-sample
            spec:
              realmName: realm-sample
              keycloakOwner: keycloak-sample   # the name of `kind: Keycloak`
              ```

              ```yaml
              apiVersion: v1.edp.epam.com/v1
              kind: KeycloakRealmGroup
              metadata:
                name: argocd-admins
              spec:
                name: ArgoCDAdmins
                realm: keycloakrealm-sample   # the name of `kind: KeycloakRealm`
              ```
        displayName: EDP Keycloak Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - authentication
        - authorization
        - edp
        - idp
        - keycloak
        - kuberocketci
        - oauth
        - oidc
        - operator
        - saml
        - sso
        links:
        - name: Edp Keycloak Operator
          url: https://github.com/epam/edp-keycloak-operator
        maintainers:
        - email: SupportEPMD-EDP@epam.com
          name: epmd-edp
        maturity: stable
        minKubeVersion: 1.20.0
        provider:
          name: KubeRocketCI
          url: https://docs.kuberocketci.io/
        relatedImages:
        - docker.io/epamedp/keycloak-operator:1.29.0
        - quay.io/community-operator-pipeline-prod/edp-keycloak-operator@sha256:10151387a79d0fccb47114b56525f28f507faee96e4fe29babd2ec46c3dd9343
        version: 1.29.0
      entries:
      - name: edp-keycloak-operator.v1.29.0
        version: 1.29.0
      name: stable
    defaultChannel: stable
    packageName: edp-keycloak-operator
    provider:
      name: KubeRocketCI
      url: https://docs.kuberocketci.io/
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operator-bamoe-businessautomation: "true"
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: IBM
      provider-url: ""
    name: bamoe-businessautomation-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: bamoe-businessautomation-operator.8.0.8-1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.kiegroup.org/v2","kind":"KieApp","metadata":{"name":"rhpam-trial"},"spec":{"environment":"rhpam-trial"}}]'
          capabilities: Seamless Upgrades
          categories: Integration & Delivery
          certified: "true"
          containerImage: registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
          createdAt: "2025-07-22 17:40:17"
          description: Deploys and manages IBM Business Automation Manager Open Editions
            environment.
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.openshift.io/infrastructure-features: '["Disconnected"]'
          operators.openshift.io/valid-subscription: '["IBM Business Automation Manager
            Open Edition"]'
          repository: https://github.com/kiegroup/kie-cloud-operator
          support: https://www.ibm.com/mysupport
          tectonic-visibility: ocs
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: A project prescription running an IBM BAMOE environment.
            displayName: KieApp
            kind: KieApp
            name: kieapps.app.kiegroup.org
            version: v2
        description: |-
          Deploys and manages IBM Business Automation Manager Open Editions environment.

          * **IBM Process Automation Manager** is a platform for developing containerized microservices and applications that automate business decisions and processes. It includes business process management (BPM), business rules management (BRM), and business resource optimization and complex event processing (CEP) technologies. It also includes a user experience platform to create engaging user interfaces for process and decision services with minimal coding.
        displayName: IBM Business Automation
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - kieapp
        - pam
        - decision
        - kie
        - cloud
        - bpm
        - process
        - automation
        - operator
        links:
        - name: Product Page
          url: https://ibm.com
        - name: Documentation
          url: https://www.ibm.com/docs/en/ibamoe
        maintainers:
        - email: mysphelp@us.ibm.com
          name: IBM
        maturity: 8.x-stable
        provider:
          name: IBM
        relatedImages:
        - registry.redhat.io/openshift4/ose-cli@sha256:b2b93def4eb5edeccfa2d8ba22e493646325842674dff0446b45677bf641a342
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:f714b9c1555e7f27a3c9cf2cf687a74188f36b1283d4e09cec4db3c084d9fe64
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:efa55e17190b0ebe1618452f786876ef354b8fbec25c316aed7ed5bb1ce88116
        - registry.redhat.io/ibm-bamoe/bamoe-operator-bundle@sha256:54df7eba0292459c0491ba39c15bfcf793e1c0aef042a4b3f6c587b0d45f77af
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:82d8ebde7eb3ecfbf72e5b71e7e392d3a3597fc084b6cf35e8fec23bfe41210d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:fb326be806b337b0ae1711f974d15e073aebcbad77aaa1123a63cf4f331da5f2
        - registry.redhat.io/jboss-datagrid-7/datagrid73-openshift@sha256:4056b1cf8ff5995f56eecfdc087d9ece56a216dba6bd73acc5693f0778e8d573
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:25025e489c15108edffe5270d257e1ed9be0a67aa067bcfc184647793aac2617
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:01bf97beffb3647968aae052cdd401d6d562c35f3e78ac483dbf55c44a71dc2f
        - registry.redhat.io/amq7/amq-broker@sha256:792c1cb19e62519cc47d834650fe5a6a6ca589a37d0b53d4f4549e80cabae3f9
        - registry.redhat.io/datagrid/datagrid-8-rhel8@sha256:98e9a6618558bc8a1b9cb91d1c9b3380968d335092f6e953d1abe6af96abf65c
        - registry.redhat.io/ibm-bamoe/bamoe-dashbuilder-rhel8@sha256:b3b7388b3b116ef37aabbb04b5d2c3e36e58d2fc95f1e59795b423010e071fc3
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:3fcd9ae0bc194c198cbb5ee94272e80f302658733e467766bd1789357e9fd015
        - registry.redhat.io/rhscl/mysql-80-rhel7@sha256:e4f535c78bc2fdb58914c52fb805185035ff31515452b4df714f9ab623ba383a
        - registry.redhat.io/rhscl/postgresql-13-rhel7@sha256:338dd09681dcf860e6a976a3bc48a1474c2376d698815e4eb611ae9c5aa68966
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:5ed8821328413ec11859ae317679f0692afff9945f134fc8909e07a3ac387794
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:a287a2fbaded371b8fcae5e4124b2d905bb228f94da0fc899ca2600b378dca0d
        - registry.redhat.io/openshift4/ose-oauth-proxy@sha256:457fc73423fa94591bc7e35729a25c58d3465583afccf342fb338b6e11188575
        - registry.redhat.io/ibm-bamoe/bamoe-smartrouter-rhel8@sha256:2f1a7519617b5c0d082f943de0db295370649d968d699d76cfd829f4c9826a0f
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-rhel8@sha256:c3553707936cb9eee8a6a622643f72c91e1b0c11add5ec9ad5865242a9e133b4
        - registry.redhat.io/ibm-bamoe/bamoe-rhel8-operator@sha256:0646da8707d10b3cc2e2344269abd8797db34127a3326bf76bbc1547aa4bbd73
        - registry.redhat.io/ibm-bamoe/bamoe-controller-rhel8@sha256:6cd0f07856d612dca46b0a01a93b796f7a0f112b98cedfd6754f9fd770ee843d
        - registry.redhat.io/ibm-bamoe/bamoe-businesscentral-monitoring-rhel8@sha256:ef1d77bb085172f927ee53d1674c85e2acaf1f580d7cad55bae8bf2c33c49c01
        - registry.redhat.io/ibm-bamoe/bamoe-kieserver-rhel8@sha256:c9f08ddc6111cdd74faa0db5721d5730fc7716fccf4dc2a658595f7e0c184c10
        - registry.redhat.io/ibm-bamoe/bamoe-process-migration-rhel8@sha256:3fd45f2dabca7220772d823275a5697b42dd35f53c27671b093117933ef2d4f6
        - registry.redhat.io/amq7/amq-broker-rhel8@sha256:ae79b5dab78c4f48993fd8999b09bdcaa918b8e30a9f750bf24d411710f71928
        - registry.redhat.io/rhscl/postgresql-12-rhel7@sha256:191237ccffc476e764714a69e35d9421dcee5ddd7184442d7a8306a06b97e6bb
        version: 8.0.8-1
      entries:
      - name: bamoe-businessautomation-operator.8.0.8-1
        version: 8.0.8-1
      - name: bamoe-businessautomation-operator.8.0.7-2
        version: 8.0.7-2
      - name: bamoe-businessautomation-operator.8.0.7-1
        version: 8.0.7-1
      - name: bamoe-businessautomation-operator.8.0.6-3
        version: 8.0.6-3
      - name: bamoe-businessautomation-operator.8.0.6-2
        version: 8.0.6-2
      - name: bamoe-businessautomation-operator.8.0.6-1
        version: 8.0.6-1
      - name: bamoe-businessautomation-operator.8.0.5-3
        version: 8.0.5-3
      - name: bamoe-businessautomation-operator.8.0.5-2
        version: 8.0.5-2
      - name: bamoe-businessautomation-operator.8.0.5-1
        version: 8.0.5-1
      - name: bamoe-businessautomation-operator.8.0.4-3
        version: 8.0.4-3
      - name: bamoe-businessautomation-operator.8.0.4-2
        version: 8.0.4-2
      - name: bamoe-businessautomation-operator.8.0.4-1
        version: 8.0.4-1
      - name: bamoe-businessautomation-operator.8.0.3-4
        version: 8.0.3-4
      - name: bamoe-businessautomation-operator.8.0.3-3
        version: 8.0.3-3
      - name: bamoe-businessautomation-operator.8.0.3-2
        version: 8.0.3-2
      - name: bamoe-businessautomation-operator.8.0.3-1
        version: 8.0.3-1
      - name: bamoe-businessautomation-operator.8.0.2-3
        version: 8.0.2-3
      - name: bamoe-businessautomation-operator.8.0.2-2
        version: 8.0.2-2
      - name: bamoe-businessautomation-operator.8.0.2-1
        version: 8.0.2-1
      - name: bamoe-businessautomation-operator.8.0.1-2
        version: 8.0.1-2
      - name: bamoe-businessautomation-operator.8.0.1-1
        version: 8.0.1-1
      - name: bamoe-businessautomation-operator.8.0.0-1
        version: 8.0.0-1
      name: 8.x-stable
    defaultChannel: 8.x-stable
    packageName: bamoe-businessautomation-operator
    provider:
      name: IBM
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: cincinnati-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: update-service-operator.v5.0.3
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "updateservice.operator.openshift.io/v1",
                "kind": "UpdateService",
                "metadata": {
                  "name": "sample"
                },
                "spec": {
                  "replicas": 1
                }
              }
            ]
          capabilities: Basic Install
          description: Creates and maintains an OpenShift Update Service instance
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "true"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: openshift-update-service
          operators.openshift.io/valid-subscription: '["OpenShift Container Platform",
            "OpenShift Platform Plus"]'
          operators.operatorframework.io/builder: operator-sdk-v1.9.0+git
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: UpdateService is the Schema for the updateservices API.
            displayName: Update Service
            kind: UpdateService
            name: updateservices.updateservice.operator.openshift.io
            version: v1
        description: |-
          # Use Case
          Running an Update Service instance in a cluster is appealing for offline OpenShift
          clusters or for admins that want to provide their own graph data instead of
          using [https://github.com/openshift/cincinnati-graph-data](https://github.com/openshift/cincinnati-graph-data/).

          # About Update Service
          Update Service uses **Cincinnati** which is an update protocol designed to facilitate
          automatic updates. It describes a particular method for representing transitions between
          releases of a project and allowing a client to perform automatic updates between these
          releases.

          ## Components
          A **Graph Builder** iterates over the release payloads hosted by the
          storage component and builds a DAG of the releases. It is responsible for
          verifying that the graph described by the releases is acyclic and connected.

          A **Policy Engine** is in charge of altering a client's view of the graph
          by applying a set of filters which are defined within the particular Policy
          Engine instance. Both the input to and the output from Policy Engines is a
          graph, allowing multiple Policy Engines to be chained together. The first
          Policy Engine in a chain will fetch its graph from the Graph Builder and the
          last Policy Engine in a chain will serve the modified graph to the client.

          An **Update Service client** is the end consumer of the release payloads. The
          client periodically queries the Policy Engine for updates and applys them if
          available.

          # Query OpenShift's Update Service Endpoint
          $ curl --silent --header 'Accept:application/json' 'https://api.openshift.com/api/upgrades_info/v1/graph?arch=amd64&channel=stable-4.2' | jq '. as $graph | $graph.nodes | map(.version == "4.2.13") | index(true) as $orig | $graph.edges | map(select(.[0] == $orig)[1]) | map($graph.nodes[.])'
        displayName: OpenShift Update Service
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - Update Service
        - UpdateService
        - update-service
        - OSUS
        - Cincinnati
        - version
        - upgrade
        links:
        - name: Update Service Operator
          url: https://github.com/openshift/cincinnati-operator
        - name: OpenShift Update Service
          url: https://github.com/openshift/cincinnati
        - name: OpenShift Graph Data
          url: https://github.com/openshift/cincinnati-graph-data
        maintainers:
        - email: aos-team-ota@redhat.com
          name: OpenShift Update Service maintainers
        maturity: alpha
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift-update-service/cincinnati-operator-bundle@sha256:50e9c88ca31182c54f9c4bc6682642a5115fa280044e6c125f99626c69d999ad
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8-operator@sha256:759f4de107a5b429933c5e44046ffb4ed75c674aea75440513a6df8fa93a83a3
        - registry.redhat.io/openshift-update-service/openshift-update-service-rhel8@sha256:a9468a662babb272917df92bc9c827535b4b37883edb440b6d2d7c32ffb2145e
        version: 5.0.3
      entries:
      - name: update-service-operator.v5.0.3
        version: 5.0.3
      - name: update-service-operator.v5.0.2
        version: 5.0.2
      - name: update-service-operator.v5.0.1
        version: 5.0.1
      - name: update-service-operator.v5.0.0
        version: 5.0.0
      - name: update-service-operator.v4.9.1
        version: 4.9.1
      - name: update-service-operator.v4.9.0
        version: 4.9.0
      name: v1
    defaultChannel: v1
    packageName: cincinnati-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: ""
    name: numaresources-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-operators
    catalogSourceDisplayName: Red Hat Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: numaresources-operator.v4.18.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesOperator",
                "metadata": {
                  "name": "numaresourcesoperator"
                },
                "spec": {
                  "nodeGroups": [
                    {
                      "machineConfigPoolSelector": {
                        "matchLabels": {
                          "pools.operator.machineconfiguration.openshift.io/worker": ""
                        }
                      }
                    }
                  ]
                }
              },
              {
                "apiVersion": "nodetopology.openshift.io/v1alpha1",
                "kind": "NUMAResourcesScheduler",
                "metadata": {
                  "name": "numaresourcesscheduler"
                },
                "spec": {
                  "imageSpec": "URL_OF_SCHEDULER_IMAGE_FROM_REDHAT_REGISTRY",
                  "logLevel": "Normal",
                  "schedulerName": "topo-aware-scheduler"
                }
              }
            ]
          capabilities: Basic Install
          createdAt: "2025-03-25T14:27:42Z"
          features.operators.openshift.io/cnf: "true"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "true"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=4.17.0 <4.18.1'
          operatorframework.io/cluster-monitoring: "true"
          operators.openshift.io/valid-subscription: "[\n\t\"OpenShift Kubernetes
            Engine\",\n\t\"OpenShift Container Platform\",\n\t\"OpenShift Platform
            Plus\"\n]"
          operators.operatorframework.io/builder: operator-sdk-v1.36.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesOperator is the Schema for the numaresourcesoperators
              API
            displayName: NUMA Resources Operator
            kind: NUMAResourcesOperator
            name: numaresourcesoperators.nodetopology.openshift.io
            version: v1alpha1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1
          - description: NUMAResourcesScheduler is the Schema for the numaresourcesschedulers
              API
            displayName: NUMA Aware Scheduler
            kind: NUMAResourcesScheduler
            name: numaresourcesschedulers.nodetopology.openshift.io
            version: v1alpha1
        description: NUMA resources exporter operator
        displayName: numaresources-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        keywords:
        - RTE
        - NUMA
        links:
        - name: Numaresources Operator
          url: https://github.com/openshift-kni/numaresources-operator
        maintainers:
        - email: fromani@redhat.com
          name: fromani
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Red Hat
        relatedImages:
        - registry.redhat.io/openshift4/numaresources-operator-bundle@sha256:a372f2b3b61c0012618837e349d389f9f300ccd2ae0fa005e9cd7d379b8569a1
        - registry.redhat.io/openshift4/numaresources-rhel9-operator@sha256:da231dc87dad63f4fe28e3b438a1ca9a96f04839389f7c3c19e8368bf4604283
        version: 4.18.1
      entries:
      - name: numaresources-operator.v4.18.1
        version: 4.18.1
      - name: numaresources-operator.v4.18.0
        version: 4.18.0
      name: "4.18"
    defaultChannel: "4.18"
    packageName: numaresources-operator
    provider:
      name: Red Hat
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Stackable GmbH
      provider-url: https://stackable.tech
    name: stackable-kafka-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      name: "25.7"
    - currentCSV: kafka-operator.v25.7.0
      currentCSVDesc:
        annotations:
          capabilities: Full Lifecycle
          categories: Storage
          containerImage: quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
          description: Stackable Operator for Apache Kafka
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "false"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: stackable-operators
          repository: https://github.com/stackabletech/kafka-operator
          support: Stackable GmbH
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Auto-generated derived type for KafkaClusterSpec via `CustomResource`
            displayName: kafkaclusters.kafka.stackable.tech
            kind: KafkaCluster
            name: kafkaclusters.kafka.stackable.tech
            version: v1alpha1
        description: |
          The Stackable Data Platform was designed with openness and flexibility in mind. It provides a curated selection of the best open source data apps like Apache Kafka, Apache Druid, Trino and Apache Spark. All data apps work together seamlessly, and can be added or removed in no time. Based on Kubernetes, it runs everywhere - on-prem or in the cloud.

          For more information please visit our [web page](https://stackable.tech/en/open-source-data-platform)
        displayName: Stackable Operator for Apache Kafka
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - kafka
        maintainers:
        - email: info@stackable.tech
          name: Stackable GmbH
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Stackable GmbH
          url: https://stackable.tech
        relatedImages:
        - quay.io/stackable/kafka-operator@sha256:82bb213628174f624c2975dc766a8ba08377e136998d3139c2fe94497337c551
        - registry.connect.redhat.com/stackable/stackable-apache-kafka-operator@sha256:bc6b416914d7dd066667809358d725ba1d245fe468e77e34fa459c10de2a09a6
        version: 25.7.0
      entries:
      - name: kafka-operator.v25.7.0
        version: 25.7.0
      - name: kafka-operator.v25.3.0
        version: 25.3.0
      name: stable
    defaultChannel: stable
    packageName: stackable-kafka-operator
    provider:
      name: Stackable GmbH
      url: https://stackable.tech
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: redhat.com
    name: special-resource-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: special-resource-operator.v4.9.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sro.openshift.io/v1beta1",
                "kind": "SpecialResource",
                "metadata": {
                  "name": "simple-kmod"
                },
                "spec": {
                  "chart": {
                    "name": "simple-kmod",
                    "repository": {
                      "name": "example",
                      "url": "file:///charts/example"
                    },
                    "version": "0.0.1"
                  },
                  "driverContainer": {
                    "source": {
                      "git": {
                        "ref": "master",
                        "uri": "https://github.com/openshift-psap/kvc-simple-kmod.git"
                      }
                    }
                  },
                  "namespace": "simple-kmod",
                  "set": {
                    "apiVersion": "sro.openshift.io/v1beta1",
                    "buildArgs": [
                      {
                        "name": "KMODVER",
                        "value": "SRO"
                      }
                    ],
                    "kind": "Values",
                    "kmodNames": [
                      "simple-kmod",
                      "simple-procfs-kmod"
                    ]
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning, OpenShift Optional, Storage
          containerImage: quay.io/openshift-psap/special-resource-operator:release-4.9
          description: The special resource operator is an orchestrator for resources
            in a cluster required to deploy special hardware and software. Examples
            include hardware accelerators or software defined storage, where extra
            management or kernel modules are needed.
          olm.skipRange: '>=4.6.0 <4.9.0'
          operators.operatorframework.io/builder: operator-sdk-v1.2.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v2
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SpecialResource is the Schema for the specialresources API
            displayName: Special Resource
            kind: SpecialResource
            name: specialresources.sro.openshift.io
            version: v1beta1
          required:
          - description: Discovers special resources and labels the nodes with the
              PCI vendor ID.
            displayName: Node Feature Discovery
            kind: NodeFeatureDiscovery
            name: nodefeaturediscoveries.nfd.openshift.io
            version: v1
        description: "## About the managed application \nThe SRO will deploy the complete
          stack for enabling the accelerator in a cluster. \nIt manages the drivers,
          device-plugin, node-exporter, grafana and feature discovery.\n## About this
          Operator\n README: [https://github.com/openshift-psap/special-resource-operator](https://github.com/openshift-psap/special-resource-operator)\n\n##
          Prerequisites for enabling this Operator\nThe SRO has a dependency on NFD
          (Node Feature Discovery)\n"
        displayName: Special Resource Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - GPI
        links:
        - name: README
          url: https://github.com/openshift-psap/special-resource-operator
        maintainers:
        - email: zkaiser@redhat.com
          name: Zvonko Kaiser
        - email: dagray@redhat.com
          name: David Gray
        maturity: alpha
        provider:
          name: Red Hat
          url: redhat.com
        relatedImages:
        - quay.io/openshift-community-operators/special-resource-operator@sha256:df87c400713311f2495d6d02769acf29e5251330194f992c11719063095fdaea
        - quay.io/openshift-psap/special-resource-operator:release-4.9
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy
        version: 4.9.0
      entries:
      - name: special-resource-operator.v4.9.0
        version: 4.9.0
      - name: special-resource-operator.v4.8.2
        version: 4.8.2
      - name: special-resource-operator.v4.8.1
        version: 4.8.1
      - name: special-resource-operator.v4.8.0
        version: 4.8.0
      name: alpha
    defaultChannel: alpha
    packageName: special-resource-operator
    provider:
      name: Red Hat
      url: redhat.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      app: kube-green
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: kube-green Community
      provider-url: https://kube-green.dev
    name: kube-green
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kube-green.v0.5.2
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kube-green.com/v1alpha1",
                "kind": "SleepInfo",
                "metadata": {
                  "labels": {
                    "app": "kube-green"
                  },
                  "name": "sleepinfo-sample"
                },
                "spec": {
                  "excludeRef": [
                    {
                      "apiVersion": "apps/v1",
                      "kind": "Deployment",
                      "name": "api-gateway"
                    }
                  ],
                  "sleepAt": "20:00",
                  "suspendCronJobs": true,
                  "timeZone": "Europe/Rome",
                  "wakeUpAt": "08:00",
                  "weekdays": "1-5"
                }
              }
            ]
          capabilities: Seamless Upgrades
          categories: Developer Tools
          containerImage: docker.io/kubegreen/kube-green:0.5.2
          createdAt: "2023-10-07T15:15:08Z"
          description: Suspend your pods when no-one's using them to save energy and
            restore when necessary
          operators.operatorframework.io/builder: operator-sdk-v1.29.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/kube-green/kube-green
          support: kube-green Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SleepInfo is the Schema for the sleepinfos API
            displayName: SleepInfo
            kind: SleepInfo
            name: sleepinfos.kube-green.com
            version: v1alpha1
        description: |
          ## About this Operator

          [kube-green](https://kube-green.dev) shuts down (some of) your resources when you don't need them, reducing the consumption of resources of the cluster.

          Install this operator and apply the SleepInfo CRD to stop your dev namespaces from running during weekends, nights or when it is not necessary.
          See the [docs](https://kube-green.dev/docs/getting-started/) for more information.
          This operator is responsible to suspend and restart pods at specific times, defined by the SleepInfo CRD.

          An example of SleepInfo:

          ```yaml
          apiVersion: kube-green.com/v1alpha1
          kind: SleepInfo
          metadata:
            name: working-hours
          spec:
            weekdays: "1-5"
            sleepAt: "20:00"
            wakeUpAt: "08:00"
            timeZone: "Europe/Rome"
          ```

          This configuration stop pods everyday from monday to friday at 20:00 and restore them at 8:00.
          In this way, your cluster nodes could downscale, producing less CO2 and saving money.
        displayName: kube-green
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - green
        - green software
        - resources
        - downscale
        - autoscale
        - climate change
        links:
        - name: Homepage
          url: https://kube-green.dev
        - name: Source Code
          url: https://github.com/kube-green/kube-green
        maintainers:
        - email: bianchidavide12+kubegreen@gmail.com
          name: Davide
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: kube-green Community
          url: https://kube-green.dev
        relatedImages:
        - docker.io/kubegreen/kube-green:0.5.2
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.14.1
        - quay.io/openshift-community-operators/kube-green@sha256:e0c9f5ecdef1f7172af098f20cce17afebbedf02c0e0bd516d4d6aae736d1b27
        version: 0.5.2
      entries:
      - name: kube-green.v0.5.2
        version: 0.5.2
      - name: kube-green.v0.5.1
        version: 0.5.1
      - name: kube-green.v0.5.0
        version: 0.5.0
      - name: kube-green.v0.4.1
        version: 0.4.1
      - name: kube-green.v0.4.0
        version: 0.4.0
      - name: kube-green.v0.3.1
        version: 0.3.1
      - name: kube-green.v0.3.0
        version: 0.3.0
      name: alpha
    defaultChannel: alpha
    packageName: kube-green
    provider:
      name: kube-green Community
      url: https://kube-green.dev
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Apache KIE
      provider-url: ""
    name: sonataflow-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: sonataflow-operator.v10.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "docker.io/apache",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.25.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache Software Foundation
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: Sonata Flow Build
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: Sonata Flow Cluster Platform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: Sonata Flow Platform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: Sonata Flow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://github.com/apache/incubator-kie-tools
        maintainers:
        - email: dev@kie.apache.org
          name: Apache Software Foundation
        maturity: alpha
        minKubeVersion: 1.23.0
        provider:
          name: Apache Software Foundation
        relatedImages:
        - docker.io/apache/incubator-kie-sonataflow-operator:10.0.0
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.0
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:8ad2282afef801c794e316c93a12834d688cfa44766132231f93013c3124944f
        version: 10.0.0
      entries:
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: alpha
    - currentCSV: sonataflow-operator.v10.1.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlow",
                "metadata": {
                  "annotations": {
                    "sonataflow.org/description": "Greeting example on k8s!",
                    "sonataflow.org/version": "0.0.1"
                  },
                  "name": "greeting"
                },
                "spec": {
                  "flow": {
                    "functions": [
                      {
                        "name": "greetFunction",
                        "operation": "sysout",
                        "type": "custom"
                      }
                    ],
                    "start": "ChooseOnLanguage",
                    "states": [
                      {
                        "dataConditions": [
                          {
                            "condition": "${ .language == \"English\" }",
                            "transition": "GreetInEnglish"
                          },
                          {
                            "condition": "${ .language == \"Spanish\" }",
                            "transition": "GreetInSpanish"
                          }
                        ],
                        "defaultCondition": "GreetInEnglish",
                        "name": "ChooseOnLanguage",
                        "type": "switch"
                      },
                      {
                        "data": {
                          "greeting": "Hello from JSON Workflow, "
                        },
                        "name": "GreetInEnglish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "data": {
                          "greeting": "Saludos desde JSON Workflow, "
                        },
                        "name": "GreetInSpanish",
                        "transition": "GreetPerson",
                        "type": "inject"
                      },
                      {
                        "actions": [
                          {
                            "functionRef": {
                              "arguments": {
                                "message": ".greeting+.name"
                              },
                              "refName": "greetFunction"
                            },
                            "name": "greetAction"
                          }
                        ],
                        "end": true,
                        "name": "GreetPerson",
                        "type": "operation"
                      }
                    ]
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowBuild",
                "metadata": {
                  "name": "greeting"
                },
                "spec": {
                  "timeout": "360s"
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowClusterPlatform",
                "metadata": {
                  "name": "sonataflow-clusterplatform"
                },
                "spec": {
                  "platformRef": {
                    "name": "sonataflow-platform",
                    "namespace": "sonataflow-operator-system"
                  }
                }
              },
              {
                "apiVersion": "sonataflow.org/v1alpha08",
                "kind": "SonataFlowPlatform",
                "metadata": {
                  "name": "sonataflow-platform"
                },
                "spec": {
                  "build": {
                    "config": {
                      "registry": {
                        "address": "host/namespace",
                        "secret": "regcred"
                      }
                    }
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Application Runtime
          containerImage: docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
          createdAt: "2025-09-11T14:44:56Z"
          description: SonataFlow Kubernetes Operator for deploying workflow applications
            based on the CNCF Serverless Workflow specification
          operators.operatorframework.io/builder: operator-sdk-v1.35.0
          operators.operatorframework.io/internal-objects: '["sonataflowbuilds.sonataflow.org"]'
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/apache/incubator-kie-tools/packages/sonataflow-operator
          support: Apache KIE
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: SonataFlowBuild is an internal custom resource to control
              workflow build instances in the target platform
            displayName: SonataFlowBuild
            kind: SonataFlowBuild
            name: sonataflowbuilds.sonataflow.org
            version: v1alpha08
          - description: SonataFlowClusterPlatform is the Schema for the sonataflowclusterplatforms
              API
            displayName: SonataFlowClusterPlatform
            kind: SonataFlowClusterPlatform
            name: sonataflowclusterplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlowPlatform is the descriptor for the workflow platform
              infrastructure.
            displayName: SonataFlowPlatform
            kind: SonataFlowPlatform
            name: sonataflowplatforms.sonataflow.org
            version: v1alpha08
          - description: SonataFlow is the descriptor representation for a workflow
              application based on the CNCF Serverless Workflow specification.
            displayName: SonataFlow
            kind: SonataFlow
            name: sonataflows.sonataflow.org
            version: v1alpha08
        description: |-
          SonataFlow Kubernetes Operator for deploying workflow applications
          based on the [CNCF Serverless Workflow specification](https://serverlessworkflow.io/):

          * Deploy workflow applications using the [dev profile](https://sonataflow.org/serverlessworkflow/latest/cloud/operator/developing-workflows.html), suited for the your development cycle
          * Build workflow applications based on the platform you're currently working on.
        displayName: SonataFlow Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - sonataflow
        - cncf
        - serverless
        - serverlessworkflow
        links:
        - name: Product Page
          url: https://sonataflow.org/serverlessworkflow/latest/index.html
        maintainers:
        - email: dev@kie.apache.org
          name: Apache KIE
        maturity: stable
        minKubeVersion: 1.23.0
        provider:
          name: Apache KIE
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.13.1
        - quay.io/community-operator-pipeline-prod/sonataflow-operator@sha256:d4ec9966370c3e770209042e437c53b0780401150f5385069b8476e26c343919
        - docker.io/apache/incubator-kie-sonataflow-operator:10.1.0
        version: 10.1.0
      entries:
      - name: sonataflow-operator.v10.1.0
        version: 10.1.0
      - name: sonataflow-operator.v10.0.0
        version: 10.0.0
      - name: sonataflow-operator.v1.40.1
        version: 1.40.1
      - name: sonataflow-operator.v1.40.0
        version: 1.40.0
      name: stable
    defaultChannel: stable
    packageName: sonataflow-operator
    provider:
      name: Apache KIE
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Joget, Inc
      provider-url: ""
    name: joget-dx-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: joget-dx-openshift-operator.v0.0.36
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"app.joget.com/v1alpha1","kind":"JogetDX","metadata":{"name":"example-joget"},"spec":{"size":1}}]'
          capabilities: Seamless Upgrades
          categories: Application Runtime,Developer Tools
          certified: "true"
          containerImage: registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
          createdAt: "2019-07-23 12:59:59"
          description: No-code/low-code application platform to visually build, run
            and maintain apps
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/support
          support: Joget, Inc
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Joget DX on JBoss EAP 7
            displayName: Joget DX on JBoss EAP 7
            kind: JogetDX
            name: jogetdx.app.joget.com
            version: v1alpha1
        description: "Joget DX is the next generation open source no-code / low-code
          application platform that combines the best of Rapid Application Development,
          Business Process Automation and Workflow Management. \nJoget DX empowers
          business users, non-coders or coders with a single platform to easily build,
          deliver, monitor and maintain enterprise applications.\n\nThis operator
          installs a Joget DX cluster running on JBoss EAP 7.\n\n### Features\n* Build
          full-fledged apps e.g. CRM, HR, Healthcare, etc\n* Drag and drop forms,
          lists, UI\n* Add workflow to automate processes\n* Extend via plugins\n*
          Apps are mobile optimized and cloud ready\n* Download ready-made apps from
          the Joget Marketplace\n\n### Before You Start\nDeploy a [MySQL](https://docs.openshift.com/online/pro/using_images/db_images/mysql.html)
          or [MariaDB](https://docs.openshift.com/online/pro/using_images/db_images/mariadb.html)
          database.\n\n### Post Deployment\nAccess the service URL and complete the
          one-time [Database Setup](https://dev.joget.org/community/display/DX7/Setting+Up+Database)\n\n###
          More Information\nMore information about Joget DX on JBoss EAP 7 is available
          in the [Joget Knowledge Base](https://dev.joget.org/community/display/DX7/Automated+Deployment+on+Red+Hat+OpenShift+with+the+Joget+Operator)\n"
        displayName: Joget DX Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - lowcode
        - appdev
        - workflow
        - automation
        links:
        - name: Joget
          url: https://www.joget.com
        maintainers:
        - email: info@joget.com
          name: Joget, Inc
        maturity: alpha
        provider:
          name: Joget, Inc
        relatedImages:
        - registry.connect.redhat.com/joget/joget-dx-operator@sha256:79856e88008da33fc0323e197819fec39fbd6b6bf2b03a0bf799b517bcf8e749
        - registry.connect.redhat.com/joget/joget-dx7-eap7@sha256:7245501f5fccf011d75c034b7d55542cfe3bd2c64cc7e48b3224a48b623af282
        - registry.connect.redhat.com/joget/joget-dx-operator-bundle@sha256:10c0eed1e8ca3154f4c815241f6275f40113ae2ba0abdc940ad9a04682edafef
        version: 0.0.36
      entries:
      - name: joget-dx-openshift-operator.v0.0.36
        version: 0.0.36
      - name: joget-dx-openshift-operator.v0.0.35
        version: 0.0.35
      - name: joget-dx-openshift-operator.v0.0.34
        version: 0.0.34
      - name: joget-dx-openshift-operator.v0.0.33
        version: 0.0.33
      - name: joget-dx-openshift-operator.v0.0.32
        version: 0.0.32
      - name: joget-dx-openshift-operator.v0.0.31
        version: 0.0.31
      - name: joget-dx-openshift-operator.v0.0.30
        version: 0.0.30
      - name: joget-dx-openshift-operator.v0.0.29
        version: 0.0.29
      - name: joget-dx-openshift-operator.v0.0.28
        version: 0.0.28
      - name: joget-dx-openshift-operator.v0.0.27
        version: 0.0.27
      - name: joget-dx-openshift-operator.v0.0.26
        version: 0.0.26
      - name: joget-dx-openshift-operator.v0.0.25
        version: 0.0.25
      - name: joget-dx-openshift-operator.v0.0.24
        version: 0.0.24
      - name: joget-dx-openshift-operator.v0.0.23
        version: 0.0.23
      - name: joget-dx-openshift-operator.v0.0.22
        version: 0.0.22
      - name: joget-dx-openshift-operator.v0.0.21
        version: 0.0.21
      - name: joget-dx-openshift-operator.v0.0.20
        version: 0.0.20
      name: alpha
    defaultChannel: alpha
    packageName: joget-dx-operator
    provider:
      name: Joget, Inc
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: directpv
      provider-url: ""
    name: minio-volume-manager-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: private-dpv-operator.v5.0.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "charts.quay.io/v1alpha1",
                "kind": "AistorVolumemanager",
                "metadata": {
                  "name": "aistorvolumemanager-sample"
                },
                "spec": {
                  "annotations": {},
                  "controller": {
                    "podAnnotations": {},
                    "podLabels": {},
                    "replicas": 3
                  },
                  "images": {
                    "pullPolicy": "IfNotPresent",
                    "pullSecrets": []
                  },
                  "labels": {},
                  "license": "",
                  "nodeserver": {
                    "nodeSelector": {},
                    "podAnnotations": {},
                    "podLabels": {},
                    "tolerations": []
                  }
                }
              }
            ]
          capabilities: Full Lifecycle
          categories: Storage,AI/Machine Learning
          createdAt: "2025-07-07T18:31:48Z"
          description: MinIO AIStor is the standard for building large scale AI data
            infrastructure.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "true"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operatorframework.io/suggested-namespace: aistor
          operators.openshift.io/valid-subscription: Contact MinIO for subscription
            information.
          operators.operatorframework.io.bundle.channel.default.v1: stable
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: helm.sdk.operatorframework.io/v1
          support: min.io
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - kind: AistorVolumemanager
            name: aistorvolumemanagers.charts.quay.io
            version: v1alpha1
        description: directpv
        displayName: directpv
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - directpv
        links:
        - name: Private Dpv Operator
          url: https://private-dpv-operator.domain
        maintainers:
        - email: cesarcelis@minio.io
          name: celis
        maturity: alpha
        provider:
          name: directpv
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy@sha256:d8cc6ffb98190e8dd403bfe67ddcb454e6127d32b87acc237b3e5240f70a20fb
        - quay.io/cniackz4/directpv-operator@sha256:abe7b5029a0956943b5af20cd66192a6fa816998523c05b671dc1280a0219683
        - registry.connect.redhat.com/minio/volume-manager-operator@sha256:e3a69a9ef99d98d1cd8eb43d3acca48a6806ccae5b6747fff0ce79712a7d363d
        version: 5.0.0
      entries:
      - name: private-dpv-operator.v5.0.0
        version: 5.0.0
      name: stable
    defaultChannel: stable
    packageName: minio-volume-manager-operator
    provider:
      name: directpv
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: certified-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/arch.ppc64le: supported
      operatorframework.io/arch.s390x: supported
      operatorframework.io/os.linux: supported
      provider: Turbonomic, Inc.
      provider-url: ""
    name: prometurbo-certified
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: certified-operators
    catalogSourceDisplayName: Certified Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: prometurbo-operator.v8.17.40-beta.1
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5-beta.1 <8.17.40-beta.1'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:a1605506884f6bce7471d79c6243728bce991ceb381d7fe0c51f1cdcbeb96e22
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:43bf1b9f06a7ace06e0b87921ba9b62f430ee3c8b2357c713054151e01f9cd02
        version: 8.17.40-beta.1
      entries:
      - name: prometurbo-operator.v8.17.40-beta.1
        version: 8.17.40-beta.1
      name: beta
    - currentCSV: prometurbo-operator.v8.17.30
      currentCSVDesc:
        annotations:
          alm-examples: '[{"apiVersion":"charts.helm.k8s.io/v1","kind":"Prometurbo","metadata":{"name":"prometurbo-release"},"spec":{"serverMeta":{"turboServer":"https://Turbo_server_URL"}}}]'
          capabilities: Basic Install
          categories: Monitoring
          certified: "false"
          createdAt: ""
          description: Turbonomic Workload Automation for Multicloud simultaneously
            optimizes performance, compliance, and cost in real-time. Workloads are
            precisely resourced, automatically, to perform while satisfying business
            constraints.
          features.operators.openshift.io/cnf: "false"
          features.operators.openshift.io/cni: "false"
          features.operators.openshift.io/csi: "false"
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          olm.skipRange: '>=8.7.5 <8.17.30'
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
          repository: https://github.com/turbonomic/prometurbo/tree/master/deploy/prometurbo-operator
          support: Turbonomic, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Turbonomic Workload Automation for Multicloud simultaneously
              optimizes performance, compliance, and cost in real-time. Workloads
              are precisely resourced, automatically, to perform while satisfying
              business constraints.
            displayName: Prometurbo Operator
            kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1
          - kind: Prometurbo
            name: prometurbos.charts.helm.k8s.io
            version: v1alpha1
        description: |-
          ### Application Resource Management for Kubernetes
          Turbonomic AI-powered Application Resource Management simultaneously optimizes performance, compliance, and cost in real time.
          Software manages the complete application stack, automatically. Applications are continually resourced to perform while satisfying business constraints.

          Turbonomic makes workloads smart — enabling them to self-manage and determines the specific actions that will drive continuous health:

          * Continuous placement for Pods (rescheduling)
          * Continuous scaling for applications and  the underlying cluster.

          It assures application performance by giving workloads the resources they need when they need them.

          ### How does it work?
          Turbonomic uses a container — PromeTurbo — that runs next to your Prometheus sever, which monitors your environment.
          PromeTurbo sends application and infrastructure monitoring data back to the Turbonomic analytics engine.
          Turbonomic determines the right actions that drive continuous health, including continuous placement for Pods and continuous scaling for applications and the underlying cluster.

          ### Get Started
          Please follow [this](https://www.ibm.com/docs/en/tarm/8.15.x?topic=configuration-prometheus) for an overview of Prometurbo and deployment instructions.
        displayName: Prometurbo Operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: false
          type: AllNamespaces
        links:
        - name: Turbonomic, Inc.
          url: https://www.turbonomic.com/
        - name: Prometurbo Operator
          url: https://github.com/IBM/turbonomic-container-platform/tree/main/prometurbo
        maintainers:
        - email: turbodeploy@turbonomic.com
          name: turbodeploy
        maturity: alpha
        provider:
          name: Turbonomic, Inc.
        relatedImages:
        - icr.io/cpopen/prometurbo-operator@sha256:5ebd9fb9277844967f55b985fc44069c1c0209360735c5321eea3aaef8628528
        - registry.connect.redhat.com/turbonomic/prometurbo-operator-bundle@sha256:e668de3c5a2eed2b2b89515860f735be8ac03c2646ab7b65ae3c1270262bc1ed
        version: 8.17.30
      entries:
      - name: prometurbo-operator.v8.17.30
        version: 8.17.30
      name: stable
    defaultChannel: stable
    packageName: prometurbo-certified
    provider:
      name: Turbonomic, Inc.
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/arch.arm64: supported
      operatorframework.io/os.linux: supported
      provider: Amazon, Inc.
      provider-url: https://aws.amazon.com
    name: ack-opensearchserverless-controller
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: ack-opensearchserverless-controller.v0.2.1
      currentCSVDesc:
        annotations:
          alm-examples: '[]'
          capabilities: Basic Install
          categories: Cloud Provider
          certified: "false"
          containerImage: public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
          createdAt: "2025-09-20T21:43:17Z"
          description: AWS OpenSearchServerless controller is a service controller
            for managing OpenSearchServerless resources in Kubernetes
          operatorframework.io/suggested-namespace: ack-system
          operators.operatorframework.io/builder: operator-sdk-v1.28.0
          operators.operatorframework.io/project_layout: unknown
          repository: https://github.com/aws-controllers-k8s
          support: Community
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Collection represents the state of an AWS opensearchserverless
              Collection resource.
            displayName: Collection
            kind: Collection
            name: collections.opensearchserverless.services.k8s.aws
            version: v1alpha1
          - description: SecurityPolicy represents the state of an AWS opensearchserverless
              SecurityPolicy resource.
            displayName: SecurityPolicy
            kind: SecurityPolicy
            name: securitypolicies.opensearchserverless.services.k8s.aws
            version: v1alpha1
        description: |-
          Manage Amazon OpenSearchServerless resources in AWS from within your Kubernetes cluster.

          **About Amazon OpenSearchServerless**

          {ADD YOUR DESCRIPTION HERE}

          **About the AWS Controllers for Kubernetes**

          This controller is a component of the [AWS Controller for Kubernetes](https://github.com/aws/aws-controllers-k8s) project. This project is currently in **developer preview**.

          **Pre-Installation Steps**

          Please follow the following link: [Red Hat OpenShift](https://aws-controllers-k8s.github.io/community/docs/user-docs/openshift/)
        displayName: AWS Controllers for Kubernetes - Amazon OpenSearchServerless
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: true
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - opensearchserverless
        - aws
        - amazon
        - ack
        links:
        - name: AWS Controllers for Kubernetes
          url: https://github.com/aws-controllers-k8s/community
        - name: Documentation
          url: https://aws-controllers-k8s.github.io/community/
        - name: Amazon OpenSearchServerless Developer Resources
          url: https://aws.amazon.com/OpenSearchServerless/developer-resources/
        maintainers:
        - email: ack-maintainers@amazon.com
          name: opensearchserverless maintainer team
        maturity: alpha
        provider:
          name: Amazon, Inc.
          url: https://aws.amazon.com
        relatedImages:
        - public.ecr.aws/aws-controllers-k8s/opensearchserverless-controller:0.2.1
        - quay.io/community-operator-pipeline-prod/ack-opensearchserverless-controller@sha256:e31d29d7c3389932e42c82645fd7b697aa16e03e20b17be0f355ef6b301d5ab8
        version: 0.2.1
      entries:
      - name: ack-opensearchserverless-controller.v0.2.1
        version: 0.2.1
      - name: ack-opensearchserverless-controller.v0.2.0
        version: 0.2.0
      - name: ack-opensearchserverless-controller.v0.1.2
        version: 0.1.2
      - name: ack-opensearchserverless-controller.v0.1.1
        version: 0.1.1
      - name: ack-opensearchserverless-controller.v0.1.0
        version: 0.1.0
      name: alpha
    defaultChannel: alpha
    packageName: ack-opensearchserverless-controller
    provider:
      name: Amazon, Inc.
      url: https://aws.amazon.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: redhat-marketplace
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Pachyderm, Inc.
      provider-url: https://pachyderm.com
    name: pachyderm-operator-rhmp
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: redhat-marketplace
    catalogSourceDisplayName: Red Hat Marketplace
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: pachyderm-operator.v0.1.4
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "Pachyderm",
                "metadata": {
                  "name": "pachyderm-sample"
                },
                "spec": {
                  "console": {
                    "disable": true
                  },
                  "pachd": {
                    "metrics": {
                      "disable": false
                    },
                    "storage": {
                      "amazon": {
                        "credentialSecretName": "pachyderm-aws-secret"
                      },
                      "backend": "AMAZON"
                    }
                  },
                  "version": "v2.1.6"
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermExport",
                "metadata": {
                  "name": "pachydermexport-sample"
                },
                "spec": {
                  "backup": {
                    "target": "pachyderm-sample"
                  }
                }
              },
              {
                "apiVersion": "aiml.pachyderm.com/v1beta1",
                "kind": "PachydermImport",
                "metadata": {
                  "name": "pachydermimport-sample"
                },
                "spec": {
                  "backup": "backups/pachyderm-backup-20220610.tar.gz",
                  "destination": {
                    "name": "pachyderm-restore",
                    "namespace": "pachyderm-workloads"
                  },
                  "storageSecret": "s3-credentials-secret"
                }
              }
            ]
          capabilities: Basic Install
          categories: AI/Machine Learning
          certified: "true"
          containerImage: registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
          createdAt: "2021-05-20T08:05:00Z"
          description: Pachyderm provides the data layer that allows data science
            teams to productionize and scale their machine learning lifecycle with
            data driven automation, petabyte scalability and end-to-end reproducibility.
          marketplace.openshift.io/remote-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/pricing?utm_source=openshift_console
          marketplace.openshift.io/support-workflow: https://marketplace.redhat.com/en-us/operators/pachyderm-operator-rhmp/support?utm_source=openshift_console
          operators.operatorframework.io/builder: operator-sdk-v1.22.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/pachyderm/openshift-operator
          support: Pachyderm, Inc.
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: PachydermExport is the Schema for the pachydermexports API
            displayName: Pachyderm Export
            kind: PachydermExport
            name: pachydermexports.aiml.pachyderm.com
            version: v1beta1
          - description: PachydermImport is the Schema for the pachydermimports API
            displayName: Pachyderm Import
            kind: PachydermImport
            name: pachydermimports.aiml.pachyderm.com
            version: v1beta1
          - description: Pachyderm is the Schema for the pachyderms API
            displayName: Pachyderm
            kind: Pachyderm
            name: pachyderms.aiml.pachyderm.com
            version: v1beta1
        description: "The Pachyderm operator aims to make deploying and managing Pachyderm
          instances on Openshift container platform easier.\n\nPachyderm is the data
          foundation for machine learning.  Pachyderm provides industry leading data
          versioning, pipelines and lineage that allow data science teams to automate
          the machine learning lifecycle and optimize their machine learning operations
          (MLOps). \n\n## Getting started\n\n1. Pachyderm uses an object store to
          store your data. Set the pachd storage backend at `pachyderm.spec.pachd.storage.backend`
          to `AMAZON`, `MICROSOFT`, `GOOGLE` or `MINIO` depending on your storage
          provider. Additionally, create a user with read/write access to the object
          store.\n\n2. Setup the user credentials in the section, `pachyderm.spec.pachd.storage`.
          Required fields depend on the backend type set in step 1 above.\n\n3. Finally,
          deploy the pachyderm cluster in its own namespace.\n\nFor more information
          on getting started, reference the documentation.\n"
        displayName: Pachyderm
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - pachyderm
        - versioning
        - machine learning
        - artificial intelligence
        links:
        - name: Documentation
          url: https://docs.pachyderm.com/latest/
        maintainers:
        - email: kyle@pachyderm.io
          name: Kyle White
        - email: eochieng@redhat.com
          name: Edmund Ochieng
        maturity: beta
        minKubeVersion: v1.19.0
        provider:
          name: Pachyderm, Inc.
          url: https://pachyderm.com
        relatedImages:
        - registry.redhat.io/rhel8/postgresql-13@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/init-utils@sha256:5c0258bfa0203079925648c0a7dda6d05a29b6f461854821a1eec541777ff359
        - registry.connect.redhat.com/pachyderm/pachd@sha256:116477fefdf413ea02c6ac3d013f1f91d3f6e4b16aed2408c9890832cdb41430
        - registry.connect.redhat.com/pachyderm/pachyderm-bundle-rhmp@sha256:d54904777c861e00858ddd8f1218cc39db47fbfb1225b8d35902ed4e13e357b1
        - registry.connect.redhat.com/pachyderm/etcd@sha256:7f330b51874379b1d4b71d94e37e0ecfa52b86d628296136970688cf204f53b5
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:4bb562f98edc2d94299f25080dc3dd8290e5f8db2710fe8f76a5cd71442bb4c7
        - registry.connect.redhat.com/pachyderm/worker@sha256:374f83d92899707bc87b0318fda6bffa78d2e9e523b827bb3d40693a2d1bdecd
        - registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:5e33f9d095952866b9743cc8268fb740cce6d93439f00ce333a2de1e5974837e
        - registry.connect.redhat.com/pachyderm/pachyderm-operator@sha256:ecfab15a6b7bcf7cbdc71dda8ba8f408ee1e6c8f2c240779e56d0f006b439ce7
        - registry.connect.redhat.com/pachyderm/pgbouncer@sha256:c3b6949227df24abb8934d111a34b44e87fb16194ac34a5d0fe284cff02e9ac0
        - registry.connect.redhat.com/pachyderm/backup-handler@sha256:e2ef6bd8191e8ad58c765c0db8f5bb21ea7123caed0c0f58fe681a024201306d
        version: 0.1.4
      entries:
      - name: pachyderm-operator.v0.1.4
        version: 0.1.4
      name: stable
    defaultChannel: stable
    packageName: pachyderm-operator-rhmp
    provider:
      name: Pachyderm, Inc.
      url: https://pachyderm.com
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Red Hat
      provider-url: https://github.com/Kuadrant/kuadrant-operator
    name: kuadrant-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: kuadrant-operator.v0.3.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "authScheme": {
                    "identity": [
                      {
                        "apiKey": {
                          "allNamespaces": true,
                          "selector": {
                            "matchLabels": {
                              "app": "toystore"
                            }
                          }
                        },
                        "credentials": {
                          "in": "authorization_header",
                          "keySelector": "APIKEY"
                        },
                        "name": "friends"
                      }
                    ],
                    "response": [
                      {
                        "json": {
                          "properties": [
                            {
                              "name": "userID",
                              "valueFrom": {
                                "authJSON": "auth.identity.metadata.annotations.secret\\.kuadrant\\.io/user-id"
                              }
                            }
                          ]
                        },
                        "name": "rate-limit-apikey",
                        "wrapper": "envoyDynamicMetadata",
                        "wrapperKey": "ext_auth_data"
                      }
                    ]
                  },
                  "rules": null,
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rateLimits": [
                    {
                      "configurations": [
                        {
                          "actions": [
                            {
                              "metadata": {
                                "default_value": "no-user",
                                "descriptor_key": "userID",
                                "metadata_key": {
                                  "key": "envoy.filters.http.ext_authz",
                                  "path": [
                                    {
                                      "segment": {
                                        "key": "ext_auth_data"
                                      }
                                    },
                                    {
                                      "segment": {
                                        "key": "userID"
                                      }
                                    }
                                  ]
                                }
                              }
                            }
                          ]
                        }
                      ],
                      "limits": [
                        {
                          "conditions": [],
                          "maxValue": 2,
                          "seconds": 10,
                          "variables": [
                            "userID"
                          ]
                        }
                      ]
                    }
                  ],
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.3.1
          createdAt: "2023-06-01T12:24:17Z"
          operators.operatorframework.io/builder: operator-sdk-v1.28.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: Enable AuthN and AuthZ based access control on workloads
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta1
          - description: Kuadrant is the Schema for the kuadrants API
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: Enable access control on workloads based on HTTP rate limiting
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta1
        description: A Kubernetes Operator to manage the lifecycle of the Kuadrant
          system
        displayName: Kuadrant Operator
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.8.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.2.0
        - quay.io/kuadrant/kuadrant-operator:v0.3.1
        - quay.io/openshift-community-operators/kuadrant-operator@sha256:398c9230878feb756ad393d7de5675a7370c883a18acf73483a660e9303d807c
        version: 0.3.1
      entries:
      - name: kuadrant-operator.v0.3.1
        version: 0.3.1
      - name: kuadrant-operator.v0.3.0
        version: 0.3.0
      - name: kuadrant-operator.v0.2.1
        version: 0.2.1
      - name: kuadrant-operator.v0.2.0
        version: 0.2.0
      name: alpha
    - currentCSV: kuadrant-operator.v0.11.1
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "DNSPolicy",
                "metadata": {
                  "name": "dnspolicy-sample"
                },
                "spec": {
                  "healthCheck": {
                    "endpoint": "/",
                    "protocol": "HTTP"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1alpha1",
                "kind": "TLSPolicy",
                "metadata": {
                  "name": "tlspolicy-sample"
                },
                "spec": {
                  "issuerRef": {
                    "group": "cert-manager.io",
                    "kind": "ClusterIssuer",
                    "name": "self-signed-ca"
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "Gateway",
                    "name": "example-gateway"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta1",
                "kind": "Kuadrant",
                "metadata": {
                  "name": "kuadrant-sample"
                },
                "spec": {}
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "AuthPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "rules": {
                    "authentication": {
                      "apikey": {
                        "apiKey": {
                          "selector": {}
                        }
                      },
                      "credentials": {
                        "authorizationHeader": {
                          "prefix": "APIKEY"
                        }
                      }
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              },
              {
                "apiVersion": "kuadrant.io/v1beta2",
                "kind": "RateLimitPolicy",
                "metadata": {
                  "name": "toystore"
                },
                "spec": {
                  "limits": {
                    "toys": {
                      "rates": [
                        {
                          "duration": 1,
                          "limit": 50,
                          "unit": "minute"
                        }
                      ]
                    }
                  },
                  "targetRef": {
                    "group": "gateway.networking.k8s.io",
                    "kind": "HTTPRoute",
                    "name": "toystore"
                  }
                }
              }
            ]
          capabilities: Basic Install
          categories: Integration & Delivery
          containerImage: quay.io/kuadrant/kuadrant-operator:v0.11.0
          createdAt: "2024-10-02T20:30:45Z"
          description: '[DEPRECATED] Follow https://docs.kuadrant.io instead'
          operators.operatorframework.io/builder: operator-sdk-v1.32.0
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: https://github.com/Kuadrant/kuadrant-operator
          support: kuadrant
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: AuthPolicy enables authentication and authorization for service
              workloads in a Gateway API network
            displayName: AuthPolicy
            kind: AuthPolicy
            name: authpolicies.kuadrant.io
            version: v1beta2
          - description: DNSPolicy configures how North-South based traffic should
              be balanced and reach the gateways
            displayName: DNSPolicy
            kind: DNSPolicy
            name: dnspolicies.kuadrant.io
            version: v1alpha1
          - description: Kuadrant configures installations of Kuadrant Service Protection
              components
            displayName: Kuadrant
            kind: Kuadrant
            name: kuadrants.kuadrant.io
            version: v1beta1
          - description: RateLimitPolicy enables rate limiting for service workloads
              in a Gateway API network
            displayName: RateLimitPolicy
            kind: RateLimitPolicy
            name: ratelimitpolicies.kuadrant.io
            version: v1beta2
          - description: TLSPolicy provides tls for gateway listeners by managing
              the lifecycle of tls certificates
            displayName: TLSPolicy
            kind: TLSPolicy
            name: tlspolicies.kuadrant.io
            version: v1alpha1
        description: |-
          ## Deprecated
          This Kuadrant community operator is deprecated, Follow https://docs.kuadrant.io instead.
        displayName: '[DEPRECATED] Kuadrant Operator'
        installModes:
        - supported: false
          type: OwnNamespace
        - supported: false
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - api
        - api-management
        - Kuadrant
        - kubernetes
        - openshift
        - cloud-service-protection
        - rate-limiting
        - authentication
        - authorization
        links:
        - name: Kuadrant Operator
          url: https://github.com/Kuadrant/kuadrant-operator
        - name: Kuadrant Docs
          url: https://kuadrant.io
        maintainers:
        - email: eastizle@redhat.com
          name: Eguzki Astiz Lezaun
        - email: mcassola@redhat.com
          name: Guilherme Cassolato
        - email: didier@redhat.com
          name: Didier Di Cesare
        maturity: alpha
        minKubeVersion: 1.19.0
        provider:
          name: Red Hat
          url: https://github.com/Kuadrant/kuadrant-operator
        relatedImages:
        - oci://quay.io/kuadrant/wasm-shim:v0.6.0
        - quay.io/community-operator-pipeline-prod/kuadrant-operator@sha256:f21456392af3b37ae48d77dc7f19fa23c6946269555bb1280978b29420bbf31e
        - quay.io/kuadrant/kuadrant-operator:v0.11.0
        version: 0.11.1
      entries:
      - name: kuadrant-operator.v0.11.1
        version: 0.11.1
      - name: kuadrant-operator.v0.11.0
        version: 0.11.0
      - name: kuadrant-operator.v0.10.0
        version: 0.10.0
      - name: kuadrant-operator.v0.7.2
        version: 0.7.2
      - name: kuadrant-operator.v0.4.1
        version: 0.4.1
      - name: kuadrant-operator.v0.4.0
        version: 0.4.0
      name: stable
    defaultChannel: stable
    packageName: kuadrant-operator
    provider:
      name: Red Hat
      url: https://github.com/Kuadrant/kuadrant-operator
- apiVersion: packages.operators.coreos.com/v1
  kind: PackageManifest
  metadata:
    creationTimestamp: "2025-02-23T05:23:24Z"
    labels:
      catalog: community-operators
      catalog-namespace: openshift-marketplace
      operatorframework.io/arch.amd64: supported
      operatorframework.io/os.linux: supported
      provider: Foundation Model Stack
      provider-url: ""
    name: multi-nic-cni-operator
    namespace: openshift-must-gather-tf84h
  spec: {}
  status:
    catalogSource: community-operators
    catalogSourceDisplayName: Community Operators
    catalogSourceNamespace: openshift-marketplace
    catalogSourcePublisher: Red Hat
    channels:
    - currentCSV: multi-nic-cni-operator.v1.3.0
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.3.0",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
          createdAt: "2024-04-17T04:51:40Z"
          description: Automate container multi-network configuration
          features.operators.openshift.io/disconnected: "true"
          features.operators.openshift.io/fips-compliant: "false"
          features.operators.openshift.io/proxy-aware: "false"
          features.operators.openshift.io/tls-profiles: "false"
          features.operators.openshift.io/token-auth-aws: "false"
          features.operators.openshift.io/token-auth-azure: "false"
          features.operators.openshift.io/token-auth-gcp: "false"
          operators.operatorframework.io/builder: operator-sdk-v1.34.1
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v3
          repository: http://github.com/foundation-model-stack/multi-nic-cni
          support: Foundation Model Stack Team
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        maturity: alpha
        minKubeVersion: 1.25.0
        provider:
          name: Foundation Model Stack
        relatedImages:
        - gcr.io/kubebuilder/kube-rbac-proxy:v0.8.0
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.3.0
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator@sha256:6b7a1e34c33b16c464745f70658a10d6ea3d96099c1e2ab008e96589ef73e7eb
        version: 1.3.0
      entries:
      - name: multi-nic-cni-operator.v1.3.0
        version: 1.3.0
      - name: multi-nic-cni-operator.v1.2.9
        version: 1.2.9
      - name: multi-nic-cni-operator.v1.2.8
        version: 1.2.8
      - name: multi-nic-cni-operator.v1.2.2
        version: 1.2.2
      - name: multi-nic-cni-operator.v1.2.1
        version: 1.2.1
      - name: multi-nic-cni-operator.v1.0.2
        version: 1.0.2
      name: alpha
    - currentCSV: multi-nic-cni-operator.v1.2.7
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.7",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
          createdAt: "2025-05-09T12:45:12Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.7
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.7
        version: 1.2.7
      entries:
      - name: multi-nic-cni-operator.v1.2.7
        version: 1.2.7
      - name: multi-nic-cni-operator.v1.2.3
        version: 1.2.3
      - name: multi-nic-cni-operator.v1.2.0
        version: 1.2.0
      - name: multi-nic-cni-operator.v1.1.0
        version: 1.1.0
      name: beta
    - currentCSV: multi-nic-cni-operator.v1.2.6
      currentCSVDesc:
        annotations:
          alm-examples: |-
            [
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "CIDR",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3"
                },
                "spec": {
                  "cidr": [
                    {
                      "hosts": [
                        {
                          "hostIP": "10.244.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                          "podCIDR": "192.168.0.0/26"
                        },
                        {
                          "hostIP": "10.244.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth1",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.0.64-26",
                          "podCIDR": "192.168.0.64/26"
                        }
                      ],
                      "interfaceIndex": 0,
                      "netAddress": "10.244.0.0/16",
                      "vlanCIDR": "192.168.0.0/18"
                    },
                    {
                      "hosts": [
                        {
                          "hostIP": "10.0.0.168",
                          "hostIndex": 0,
                          "hostName": "node-1",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.0-26",
                          "podCIDR": "192.168.64.0/26"
                        },
                        {
                          "hostIP": "10.0.0.167",
                          "hostIndex": 1,
                          "hostName": "node-2",
                          "interfaceName": "eth2",
                          "ippool": "multi-nic-cni-operator-ipvlanl3-192.168.64.64-26",
                          "podCIDR": "192.168.64.64/26"
                        }
                      ],
                      "interfaceIndex": 1,
                      "netAddress": "10.0.0.0/16",
                      "vlanCIDR": "192.168.64.0/18"
                    }
                  ],
                  "config": {
                    "hostBlock": 8,
                    "interfaceBlock": 2,
                    "masterNets": [],
                    "name": "multi-nic-cni-operator-ipvlanl3",
                    "subnet": "192.168.0.0/16",
                    "type": "ipvlan",
                    "vlanMode": "l3"
                  }
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nic-cni-operator-multi-nicd",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "Config",
                "metadata": {
                  "annotations": {
                    "note": "The resource with default settings will be auto-created by multi-nic-cni-operator-controller-manager.\nPlease modify if need.\n"
                  },
                  "name": "multi-nicd"
                },
                "spec": {
                  "addRoutePath": "/addl3",
                  "cniType": "multi-nic",
                  "contextTimeoutMinutes": 2,
                  "daemon": {
                    "env": [
                      {
                        "name": "DAEMON_PORT",
                        "value": "11000"
                      },
                      {
                        "name": "RT_TABLE_PATH",
                        "value": "/opt/rt_tables"
                      }
                    ],
                    "image": "ghcr.io/foundation-model-stack/multi-nic-cni-daemon:v1.2.6",
                    "mounts": [
                      {
                        "hostpath": "/var/lib/cni/bin",
                        "name": "cnibin",
                        "podpath": "/host/opt/cni/bin"
                      },
                      {
                        "hostpath": "/var/lib/kubelet/device-plugins",
                        "name": "device-plugin",
                        "podpath": "/var/lib/kubelet/device-plugins"
                      },
                      {
                        "hostpath": "/etc/iproute2/rt_tables",
                        "name": "rt-tables",
                        "podpath": "/opt/rt_tables"
                      }
                    ],
                    "port": 11000,
                    "resources": {
                      "requests": {
                        "cpu": "100m",
                        "memory": "50Mi"
                      }
                    },
                    "securityContext": {
                      "privileged": true
                    }
                  },
                  "deleteRoutePath": "/deletel3",
                  "getInterfacePath": "/interface",
                  "ipamType": "multi-nic-ipam",
                  "joinPath": "/join",
                  "logLevel": 4,
                  "longReconcileMinutes": 10,
                  "normalReconcileMinutes": 1,
                  "urgentReconcileSeconds": 5
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "multi-nic-cni-operator-highspeed",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "DeviceClass",
                "metadata": {
                  "name": "highspeed"
                },
                "spec": {
                  "ids": [
                    {
                      "products": [
                        "1019",
                        "101d"
                      ],
                      "vendor": "15b3"
                    },
                    {
                      "products": [
                        "efa0",
                        "efa1"
                      ],
                      "vendor": "1d0f"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-node-1",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "HostInterface",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "node-1"
                },
                "spec": {
                  "hostName": "node-1",
                  "interfaces": [
                    {
                      "hostIP": "10.0.0.168",
                      "interfaceName": "eth1",
                      "netAddress": "10.0.0.0/16",
                      "pciAddress": "0000:00:04.0",
                      "product": "1000",
                      "vendor": "1af4"
                    },
                    {
                      "hostIP": "10.1.0.168",
                      "interfaceName": "eth2",
                      "netAddress": "10.1.0.0/16",
                      "pciAddress": "0000:00:05.0",
                      "product": "1000",
                      "vendor": "1af4"
                    }
                  ]
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "multi-nic-cni-operator-ipvlanl3-192.168.0.0-26",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "IPPool",
                "metadata": {
                  "annotations": {
                    "note": "This resource will be auto-generated and controlled by multi-nic-cni-operator-controller-manager.\nDo not create/modify.\n"
                  },
                  "name": "ipvlanl3-192.168.0.0-26"
                },
                "spec": {
                  "allocations": [],
                  "excludes": [],
                  "hostName": "node-1",
                  "interfaceName": "eth1",
                  "netAttachDef": "multi-nic-cni-operator-ipvlanl3",
                  "podCIDR": "192.168.0.0/26",
                  "vlanCIDR": "192.168.0.0/18"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "multi-nic-cni-operator-ipvlanl3",
                  "namespace": "multi-nic-cni-operator-system"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              },
              {
                "apiVersion": "multinic.fms.io/v1",
                "kind": "MultiNicNetwork",
                "metadata": {
                  "name": "ipvlanl3"
                },
                "spec": {
                  "ipam": "{\n  \"type\": \"multi-nic-ipam\",\n  \"hostBlock\": 8, \n  \"interfaceBlock\": 2,\n  \"vlanMode\": \"l3\"\n}\n",
                  "multiNICIPAM": true,
                  "plugin": {
                    "args": {
                      "mode": "l3"
                    },
                    "cniVersion": "0.3.0",
                    "type": "ipvlan"
                  },
                  "subnet": "192.168.0.0/16"
                }
              }
            ]
          capabilities: Basic Install
          categories: Networking
          containerImage: ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
          createdAt: "2025-03-24T14:49:54Z"
          operators.operatorframework.io/builder: operator-sdk-v1.39.2
          operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
        apiservicedefinitions: {}
        customresourcedefinitions:
          owned:
          - description: CIDR is the Schema for the cidrs API
            displayName: CIDR
            kind: CIDR
            name: cidrs.multinic.fms.io
            version: v1
          - description: Config is the Schema for the configs API
            displayName: Config
            kind: Config
            name: configs.multinic.fms.io
            version: v1
          - description: DeviceClass is the Schema for the deviceclasses API
            displayName: Device Class
            kind: DeviceClass
            name: deviceclasses.multinic.fms.io
            version: v1
          - description: HostInterface is the Schema for the hostinterfaces API
            displayName: Host Interface
            kind: HostInterface
            name: hostinterfaces.multinic.fms.io
            version: v1
          - description: IPPool is the Schema for the ippools API
            displayName: IPPool
            kind: IPPool
            name: ippools.multinic.fms.io
            version: v1
          - description: MultiNicNetwork is the Schema for the multinicnetworks API
            displayName: Multi Nic Network
            kind: MultiNicNetwork
            name: multinicnetworks.multinic.fms.io
            version: v1
        description: "Multi-NIC CNI Operator helps to attaching secondary network
          interfaces that is linked to \ndifferent network interfaces on host (NIC)
          to pod provides benefits of network segmentation \nand top-up network bandwidth
          in the containerization system.\n\nMulti-NIC CNI is the CNI plugin operating
          on top of Multus CNI. \nHowever, instead of defining and handling each secondary
          network interface one by one, \nthis CNI automatically discovers all available
          secondary interfaces and handles them as a NIC pool. \nWith this manner,
          it can provide the following benefits.\n\ni) Common secondary network definition:
          User can manage only one network definition for \n   multiple secondary
          interfaces with a common CNI main plugin such as ipvlan, macvlan, and sr-iov.\n\nii)
          Common NAT-bypassing network solution: All secondary NICs on each host can
          be assigned \n    with non-conflict CIDR and non-conflict L3 routing configuration
          that can omit an overlay networking overhead. \n    Particularyly, the CNI
          is built-in with L3 IPVLAN solution composing of the following functionalities.\n\n
          \   1. Interface-host-devision CIDR Computation: compute allocating CIDR
          range for each host and each interface \n       from a single global subnet
          with the number of bits for hosts and for interface.\n    2. L3 Host Route
          Configuration: configure L3 routes (next hop via dev) in host route table
          according to the computed CIDR.\n    3. Distributed IP Allocation Management:
          manage IP allocation/deallocation distributedly via the communication \n
          \      between CNI program and daemon at each host.\n\niii) Policy-based
          secondary network attachment: Instead of statically set the desired \n     host's
          master interface name one by one, user can define a policy on attaching
          multiple secondary network \n     interfaces such as specifying only the
          number of desired interfaces, filtering only highspeed NICs.\n\n## Instructions\n\nPlease
          check the [User Guide](https://foundation-model-stack.github.io/multi-nic-cni/user_guide/)"
        displayName: multi-nic-cni-operator
        installModes:
        - supported: true
          type: OwnNamespace
        - supported: true
          type: SingleNamespace
        - supported: false
          type: MultiNamespace
        - supported: true
          type: AllNamespaces
        keywords:
        - multi-nic
        - cni
        - network
        links:
        - name: Multi-NIC CNI Operator
          url: https://github.com/foundation-model-stack/multi-nic-cni
        maintainers:
        - email: sunyanan.choochotkaew1@ibm.com
          name: Sunyanan Choochotkaew
        - email: chiba@jp.ibm.com
          name: Tatsuhiro Chiba
        - email: minzhang@redhat.com
          name: Min Zhang
        - email: ebendror@redhat.com
          name: Eran Ben Dror
        maturity: alpha
        provider:
          name: Foundation Model Stack
        relatedImages:
        - ghcr.io/foundation-model-stack/multi-nic-cni-controller:v1.2.6
        - quay.io/community-operator-pipeline-prod/multi-nic-cni-operator:1.2.6
        version: 1.2.6
      entries:
      - name: multi-nic-cni-operator.v1.2.6
        version: 1.2.6
      - name: multi-nic-cni-operator.v1.2.5
        version: 1.2.5
      - name: multi-nic-cni-operator.v1.2.4
        version: 1.2.4
      - name: multi-nic-cni-operator.v1.0.5
        version: 1.0.5
      - name: multi-nic-cni-operator.v1.0.4
        version: 1.0.4
      - name: multi-nic-cni-operator.v1.0.3
        version: 1.0.3
      name: stable
    defaultChannel: stable
    packageName: multi-nic-cni-operator
    provider:
      name: Foundation Model Stack
kind: List
metadata:
  resourceVersion: ""
